Get the latest tech news

Granite 4.1: IBM's 8B Model Matching 32B MoE


IBM just released Granite 4.1, a family of open source language models built specifically for enterprise use. Three sizes, Apache 2.0 licensed and trained on 15 trillion tokens with a level of pipeline obsession that's worth understanding. But there's one result in the benchmarks I keep coming back to. The 8B model. Dense architecture, no MoE tricks, no extended reasoning chains. It matches or beats Granite 4.0-H-Small across basically every benchmark they ran. That older model has 32B parameters with 9B active. This one has 8 billion. Full stop. That result is either very impressive or it means the old model was underbuilt. Probably both. Here's how they built it, what the numbers actually say, and whether any of it matters for your use case.

None

Get the Android app

Or read this on Hacker News

Read more on:

Photo of IBM

IBM

Photo of granite

granite

Photo of 32b

32b

Related news:

News photo

IBM launches Bob with multi-model routing and human checkpoints to turn AI coding into a secure production system

News photo

IBM to Add 750 Jobs in AI, Quantum Computing at Chicago Tech Hub

News photo

IBM's AI coding 'partner' Bob hits general availability