Get the latest tech news
Meta AI develops compact language model for mobile devices
Key innovations in Meta's MobileLLM include prioritizing model depth over width, implementing embedding sharing and grouped-query attention and utilizing a novel immediate block-wise weight-sharing technique.
Meta AI researchers have unveiled MobileLLM, a new approach to creating efficient language models designed for smartphones and other resource-constrained devices. As progress in very large language models shows signs of slowing, researchers are increasingly exploring the potential of more compact, specialized designs. The focus on efficiency and on-device deployment puts MobileLLM in a similar category to what some researchers call Small Language Models (SLMs), despite the “LLM” in its name.
Or read this on Venture Beat