Get the latest tech news

Language models pack billions of concepts into 12k dimensions


In a recent 3Blue1Brown video series on transformer models, Grant Sanderson posed a fascinating question: How can a relatively modest embedding space of 12,288 dimensions (GPT-3) accommodate millions of distinct real-world concepts? The answer lies at the intersection of high-dimensional geometry and a remarkable mathematical result known as the

In a recent 3Blue1Brown video series on transformer models, Grant Sanderson posed a fascinating question: How can a relatively modest embedding space of 12,288 dimensions (GPT-3) accommodate millions of distinct real-world concepts? "Archery" shares some semantic space with "precision" and "sport""Fire" overlaps with both "heat" and "passion""Gelatinous" relates to physical properties and food textures"Southern-ness" encompasses culture, geography, and dialect"Basketball" connects to both athletics and geometry"Green" spans color perception and environmental consciousness"Altruistic" links moral philosophy with behavioral patterns Sphere Packings, Lattices and Groups by Conway and Sloane Database-friendly random projections: Johnson-Lindenstrauss with binary coins by Achlioptas Hadamard Matrices, Sequences, and Block Designs by Seberry and Yamada

Get the Android app

Or read this on Hacker News

Read more on:

Photo of billions

billions

Photo of Dimensions

Dimensions

Photo of language models

language models

Related news:

News photo

U.S. sanctions cyber scammers who stole billions from Americans

News photo

US Taxpayers Will Pay Billions in New Fossil Fuel Subsidies Thanks to the Big Beautiful Bill

News photo

This 2FA phishing scam pwned a developer - and endangered billions of npm downloads