* This blog post is a summary of this video.

Uncovering Del 3: How OpenAI's Latest AI Art Model Sets New Standards

Author: What's AI by Louis BouchardTime: 2023-12-28 19:55:00

Table of Contents

Introduction to DALL-E 3 Image Generation Model

Last year we were blown away by DALL-E 2, the first super impressive text-to-image model by Anthropic. But today, prepare to step into a world where art and technology merge like never before with its third version. Let's dive into DALL-E 3 with a brand new paper Anthropic just released and uncover the advancements that set it leagues ahead of DALL-E 2.

Overview of DALL-E 3

Trained on highly descriptive generated image captions, DALL-E 3 doesn't just follow prompts, it breathes life into them. The results are incredible and it not only understands prompts, but it is also able to understand the story behind your prompt. The progress since 2020 is just unbelievable.

Advancements Over Previous Versions

At the heart of DALL-E 3's progress is a robust image captioner. It's all about the image captions, so the text fed during training along with the image. It should be able to generate the new image captioner is the main factor why DALL-E 3 is so much better than DALL-E 2.

Robust Image Captioner Enables Descriptive Prompts

Previous models were initially trained in a self-supervised way with image text pairs scraped from the internet directly. Imagine an Instagram picture and its caption or hashtags. It's not always that informative or even linked. The authors of the post mainly describe the main subject in the picture, not the whole story behind it or its environment and text that appears in the image along with the main subject.

Limitations of Previous Training Methods

Likewise, they don't see where everything is placed in the image, which will be super useful information to ensure the accurate recreation of a similar image. Even worse, lots of captions are just jokes or non-related thoughts or poems shared along the images. At this point, training with such data is pretty much shooting yourself in the foot.

Generating Synthetic Captions

What if you would instead have the perfect captions - super detailed with all the spatial information needed to recreate it? That would be perfect. But how can we have this information from millions of pictures? We could hire hundreds or thousands of humans to describe the images accurately, or we could use another model to understand images and generate better captions first. Well, that's what they did first: craft a powerful image captioner model, then use it on your current large dataset of image caption pairs to improve them.

Evaluating DALL-E 3 Performance

Human Evaluations Prefer DALL-E 3 Images

In evaluations, DALL-E 3 outshines DALL-E 2 with human raters consistently preferring the images generated by the newer model. It's also much better quantitatively measured on different benchmarks.

Quantitative Benchmarks

Like the T2I comp bench evaluation benchmark created by Anthropic, which is a benchmark consisting of 6,000 compositional text prompts with several evaluation metrics specifically designed to evaluate compositional text-image generation models.

Current Limitations and Future Outlook

To recap, DALL-E 3 is a huge step forward in prompt following and has amazing qualitative results, but it still has its limitations. It struggles with image generation features like spatial awareness. It's just really hard to have descriptive enough descriptions with location information for all objects. Also, this third version is already much better at generating text on screen, something all previous models really struggle with, but it's still quite unreliable. We will have to wait for DALL-E 4 to have the proper text generated in images.

Conclusion

Another problem with DALL-E 3 comes from the image captioner model. They have reported that the captioner is prone to hallucinating important details about an image. It often likes to give more details than less, even if it's to create them from nothing. Thus called hallucination, I guess this is just a regular behavior of LLMs. Maybe because good human writers like to give details and a good story, and the model was trained on this style of writing anyway. There's no complete fix on this new model hallucination problem, which is why you should always be careful when using those language models or even image ones in this case.

FAQ

Q: How does Del 3 generate images?
A: Del 3 is trained on highly descriptive generated image captions that allow it to recreate detailed images that accurately reflect prompts. It uses an advanced image captioner to generate descriptive captions.

Q: What are the key improvements in Del 3?
A: The main improvements are using a robust image captioner model to generate descriptive training data and improved spatial/compositional understanding compared to previous versions.

Q: How well does Del 3 perform?
A: Human evaluations show users strongly prefer Del 3 over previous versions. It also achieves state-of-the-art results on computational benchmarks.

Q: What are the limitations of Del 3?
A: It still struggles with spatial awareness and reliably generating text in images. The image captioner can also sometimes 'hallucinate' non-existent details.