Get the latest tech news

Meta releases its first open AI model that can process images


An important step to catch up to ChatGPT and Gemini.

The new model, Llama 3.2, could allow developers to create more advanced AI applications, like augmented reality apps that provide real-time understanding of video, visual search engines that sort images based on content, or document analysis that summarizes long chunks of text for you. Developers will have to do little except add this “new multimodality and be able to show Llama images and have it communicate,” Ahmad Al-Dahle, vice president of generative AI at Meta, told The Verge. There’s still a place for the (slightly) older Llama 3.1, though: that model, released in July, included a version with 405 billion parameters, which will theoretically be more capable when it comes to generating text.

Get the Android app

Or read this on The Verge

Read more on:

Photo of Meta

Meta

Photo of images

images

Photo of open AI model

open AI model

Related news:

News photo

Meta’s going to put AI-generated images in your Facebook and Instagram feeds

News photo

Meta’s AI can now talk to you in the voices of Awkwafina, John Cena, and Judi Dench

News photo

Meta rolls out its first major vision models to rival Anthropic, OpenAI