Get the latest tech news

Modeling Others' Minds as Code


Accurate prediction of human behavior is essential for robust and safe human-AI collaboration. However, existing approaches for modeling people are often data-hungry and brittle because they either make unrealistic assumptions about rationality or are too computationally demanding to adapt rapidly. Our key insight is that many everyday social interactions may follow predictable patterns; efficient "scripts" that minimize cognitive load for actors and observers, e.g., "wait for the green light, then go." We propose modeling these routines as behavioral programs instantiated in computer code rather than policies conditioned on beliefs and desires. We introduce ROTE, a novel algorithm that leverages both large language models (LLMs) for synthesizing a hypothesis space of behavioral programs, and probabilistic inference for reasoning about uncertainty over that space. We test ROTE in a suite of gridworld tasks and a large-scale embodied household simulator. ROTE predicts human and AI behaviors from sparse observations, outperforming competitive baselines -- including behavior cloning and LLM-based methods -- by as much as 50% in terms of in-sample accuracy and out-of-sample generalization. By treating action understanding as a program synthesis problem, ROTE opens a path for AI systems to efficiently and effectively predict human behavior in the real-world.

None

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Code

Code

Photo of minds

minds

Photo of Modeling Others

Modeling Others

Related news:

News photo

Don't Force Your LLM to Write Terse [Q/Kdb] Code: An Information Theory Argument

News photo

Codev lets enterprises avoid vibe coding hangovers with a team of agents that generate and document code

News photo

New computer model helps reveal how the brain both adapts and misfires