Before Imagen came out in May 2022, models like DALL·E, GLIDE, and Latent Diffusion had already made big progress in turning text into images. Imagen stood out because it used a large frozen T5 language model to understand text, which gave it much better results. It also found that making the text model bigger mattered more than making the image model bigger. Imagen produced very realistic 1024×1024 images and set a new quality benchmark. Google didn’t invent the idea but refined it in a smarter way, combining language and diffusion models more effectively.
No recent users
Guests online: 294
Loading server status...
Did Imagen really achieve something new, or did others already publish similar work, and Google is just taking credit for polishing it?
Original: English
Technology
2 weeks, 4 days ago by
ModernSlave
Edited 3 days, 17 hours ago
Log in to add a comment.
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Anonymous
- Mdi
- plxic
- ModernSlave
- klop
Comments
No comments yet.