Get the latest tech news

Mutlimodal neural networks converge to a shared statistical model of reality


We argue that representations in AI models, particularly deep networks, are converging. First, we survey many examples of convergence in the literature: over time and across multiple domains, the ways by which different neural networks represent data are becoming more aligned. Next, we demonstrate convergence across data modalities: as vision models and language models get larger, they measure distance between datapoints in a more and more alike way. We hypothesize that this convergence is driving toward a shared statistical model of reality, akin to Plato's concept of an ideal reality. We term such a representation the platonic representation and discuss several possible selective pressures toward it. Finally, we discuss the implications of these trends, their limitations, and counterexamples to our analysis.

View a PDF of the paper titled The Platonic Representation Hypothesis, by Minyoung Huh and 3 other authors View PDF Abstract:We argue that representations in AI models, particularly deep networks, are converging. First, we survey many examples of convergence in the literature: over time and across multiple domains, the ways by which different neural networks represent data are becoming more aligned.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Reality

Reality

Photo of Mutlimodal

Mutlimodal

Related news:

News photo

Tesla’s Lofty AI Dreams Already Baked Into Towering Valuation

News photo

The reality of Kotlin's LSP and Tooling situation

News photo

Walmart and Roblox Are Teaming Up To Make Virtual E-commerce a Reality