Get the latest tech news

OpenAI’s newest AI model can hold a humanlike conversation | GPT-4o can see, hear and speak with near-instant response times.


GPT-4o can see, hear and speak with near-instant response times. The multimodal model will roll out for free over the next few weeks.

Users can relay visuals — through their phone camera, by uploading documents, or by sharing their screen — all while conversing with the AI model as if they are in a video call. Mira Murati, chief technology officer at OpenAI, said during a livestream demonstration that making advanced AI tools available to users for free is a “very important” component of the company’s mission. An AI assistant that can reason in real time using vision, text and voice would enable the technology to perform a creative range of tasks — such as walking users through a math problem, translating languages during a conversation and reading human facial expressions.

Get the Android app

Or read this on r/tech

Read more on:

Photo of OpenAI

OpenAI

Photo of newest AI model

newest AI model

Photo of GPT-4o

GPT-4o

Related news:

News photo

OpenAI now has an AI model with vision, and everyone else should be scared

News photo

Google showcases a potential answer to GPT-4o ahead of I/O

News photo

After Microsoft invested $10 billion, OpenAI snubs Windows 11 as it releases ChatGPT app first on Mac. “We’re just prioritizing where our users are.”