Get the latest tech news

SmolLM2


New from [Loubna Ben Allal](https://loubnabnl.github.io/) and her research team at Hugging Face: > SmolLM2 is a family of compact language models available in three size: 135M, 360M, and 1.7B parameters. …

SmolLM2 is a family of compact language models available in three size: 135M, 360M, and 1.7B parameters. They are capable of solving a wide range of tasks while being lightweight enough to run on-device. It was trained on 11 trillion tokens using a diverse dataset combination: FineWeb-Edu, DCLM, The Stack, along with new mathematics and coding datasets that we curated and will release soon.

Get the Android app

Or read this on Hacker News