this post was submitted on 20 Sep 2024
18 points (90.9% liked)

Stable Diffusion

4308 readers
4 users here now

Discuss matters related to our favourite AI Art generation technology

Also see

Other communities

founded 1 year ago
MODERATORS
 

Abstract

In this work, we introduce OmniGen, a new diffusion model for unified image generation. Unlike popular diffusion models (e.g., Stable Diffusion), OmniGen no longer requires additional modules such as ControlNet or IP-Adapter to process diverse control conditions. OmniGenis characterized by the following features: 1) Unification: OmniGen not only demonstrates text-to-image generation capabilities but also inherently supports other downstream tasks, such as image editing, subject-driven generation, and visual-conditional generation. Additionally, OmniGen can handle classical computer vision tasks by transforming them into image generation tasks, such as edge detection and human pose recognition. 2) Simplicity: The architecture of OmniGen is highly simplified, eliminating the need for additional text encoders. Moreover, it is more user-friendly compared to existing diffusion models, enabling complex tasks to be accomplished through instructions without the need for extra preprocessing steps (e.g., human pose estimation), thereby significantly simplifying the workflow of image generation. 3) Knowledge Transfer: Through learning in a unified format, OmniGen effectively transfers knowledge across different tasks, manages unseen tasks and domains, and exhibits novel capabilities. We also explore the model's reasoning capabilities and potential applications of chain-of-thought mechanism. This work represents the first attempt at a general-purpose image generation model, and there remain several unresolved issues. We will open-source the related resources at this https URL to foster advancements in this field.

Paper: https://arxiv.org/abs/2409.11340

Code: https://github.com/VectorSpaceLab/OmniGen (coming soon)

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 9 points 1 month ago (1 children)

Only 3.8B parameters according to the paper, so it ought to be quite easy on the hardware as well if they do.

[–] [email protected] 10 points 1 month ago (2 children)

That's kind of unbelievable given what they say it can do.

[–] [email protected] 3 points 1 month ago (1 children)

Do they have a schedule for the release?

[–] [email protected] 4 points 1 month ago

Doesn't seem like it.

[–] [email protected] 3 points 1 month ago

I must agree. But since they claim “few shot” all along the paper and they publish the failure case ( which I rarely see) I tend to believe that this model can do everything they said BUT not consistently. But this is look like a very promising POC for this kind of models