Get the latest tech news

Apple Publishes Details About New 'MM1' AI Model


Apple researchers have developed a new method for training large language models (LLMs) that seamlessly integrates both text and visual information. ...

By utilizing a diverse dataset comprising image-caption pairs, interleaved image-text documents, and text-only data, Apple's claims that the MM1 model sets a new standard in AI's ability to perform tasks such as image captioning, visual question answering, and natural language inference with a high level of accuracy. Apple's research focuses on the combination of different types of training data and model architectures, which enables the AI to understand and generate language based on a mix of visual and linguistic cues. This version apparently exhibits remarkable capabilities for multi-step reasoning over multiple images using few-shot "chain-of-thought" prompting, a technique that allows the AI to perform complex, open-ended problem solving based on minimal examples.

Get the Android app

Or read this on Mac Rumors

Read more on:

Photo of Apple

Apple

Photo of Details

Details

Photo of model

model

Related news:

News photo

Apple is in talks to license Gemini AI for iPhones

News photo

What is DarwinAI, Apple’s newly acquired start-up?

News photo

Apple and Google are reportedly talking. Could Gemini come to iPhone?