Meta has just introduced an advanced generative AI model with a primary focus on generating images from text, aiming to achieve exceptional performance in this specific domain.
This unveiling comes at a time when AI-driven image generators have gained significant popularity and accessibility, with numerous prominent companies and emerging startups relying on these models for their daily operations.
As per media reports, Meta anticipates that their latest AI model will produce more coherent and visually appealing imagery while efficiently analyzing input prompts. Unlike the current widely used AI-based image generators, such as DALL-E2, Google's Imagen, and Stable Diffusion, which employ a diffusion process for art creation, Meta's model takes a different approach.
It gradually refines the image by removing noise, enabling it to learn and operate effectively based on the provided prompts.
On the other hand, the diffusion process proves to be resource-intensive, expensive, and time-consuming. In contrast, Meta's CM3leon model adopts an attention mechanism that prioritizes the input prompt, be it in the form of text or an image.
This innovation is expected to significantly enhance efficiency, reducing the need for extensive computational power and a large dataset compared to other models.
CM3leon, developed by Meta, undergoes training using a dataset comprising millions of licensed images, even amid legal challenges concerning information misuse the company has encountered.
Unlike conventional image generators, CM3leon exhibits noteworthy capabilities in handling complex objects and comprehending prompts, as demonstrated by some of the intricate designs it generates.