Get the latest tech news

The Matrix: A Bayesian learning model for LLMs


In this paper, we introduce a Bayesian learning model to understand the behavior of Large Language Models (LLMs). We explore the optimization metric of LLMs, which is based on predicting the next token, and develop a novel model grounded in this principle. Our approach involves constructing an ideal generative text model represented by a multinomial transition probability matrix with a prior, and we examine how LLMs approximate this matrix. We discuss the continuity of the mapping between embeddings and multinomial distributions, and present the Dirichlet approximation theorem to approximate any prior. Additionally, we demonstrate how text generation by LLMs aligns with Bayesian learning principles and delve into the implications for in-context learning, specifically explaining why in-context learning emerges in larger models where prompts are considered as samples to be updated. Our findings indicate that the behavior of LLMs is consistent with Bayesian Learning, offering new insights into their functioning and potential applications.

View a PDF of the paper titled The Matrix: A Bayesian learning model for LLMs, by Siddhartha Dalal and Vishal Misra We explore the optimization metric of LLMs, which is based on predicting the next token, and develop a novel model grounded in this principle. Our findings indicate that the behavior of LLMs is consistent with Bayesian Learning, offering new insights into their functioning and potential applications.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Transformers

Transformers

Photo of Improve

Improve

Related news:

News photo

Block* and AgentFormer – Playing with blocks and Transformers (yay)

News photo

Searchformer: Beyond A* – Better planning with transformers via search dynamics

News photo

Transformers One turns Cybertron’s greatest warriors into bumbling youths in first trailer