Get the latest tech news
AI21 CEO says transformers not right for AI agents due to error perpetuation
Ori Goshen, CEO of enterprise AI company AI21, believes alternative model architectures will make AI agents work better.
Ari Goshen, AI21 CEO, said in an interview with VentureBeat that Transformers, the most popular model architecture, has limitations that would make a multi-agent ecosystem difficult. Goshen strongly supports alternative architectures like Mamba and AI21’s Jamba, mainly because he believes transformer models are too expensive and unwieldy to run. Instead of an attention mechanism that forms the backbone of transformer models, Mamba can prioritize different data and assign weights to inputs, optimize memory usage, and use a GPU’s processing power.
Or read this on Venture Beat