Get the latest tech news
Apple Publishes Details About New 'MM1' AI Model
Apple researchers have developed a new method for training large language models (LLMs) that seamlessly integrates both text and visual information. ...
By utilizing a diverse dataset comprising image-caption pairs, interleaved image-text documents, and text-only data, Apple's claims that the MM1 model sets a new standard in AI's ability to perform tasks such as image captioning, visual question answering, and natural language inference with a high level of accuracy. Apple's research focuses on the combination of different types of training data and model architectures, which enables the AI to understand and generate language based on a mix of visual and linguistic cues. This version apparently exhibits remarkable capabilities for multi-step reasoning over multiple images using few-shot "chain-of-thought" prompting, a technique that allows the AI to perform complex, open-ended problem solving based on minimal examples.
Or read this on Mac Rumors