Get the latest tech news

Microsoft BitNet: inference framework for 1-bit LLMs


Official inference framework for 1-bit LLMs. Contribute to microsoft/BitNet development by creating an account on GitHub.

bitnet.cpp achieves speedups of 1.37x to 5.07x on ARM CPUs, with larger models experiencing greater performance gains. The tested models are dummy setups used in a research context to demonstrate the inference performance of bitnet.cpp. We also thank T-MAC team for the helpful discussion on the LUT method for low-bit LLM inference.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Microsoft

Microsoft

Photo of bit llms

bit llms

Photo of Microsoft BitNet

Microsoft BitNet

Related news:

News photo

Salesforce CEO Benioff Says Microsoft's Copilot Doesn't Work, Doesn't Offer 'Any Level of Accuracy' And Customers Are 'Left Cleaning Up the Mess'

News photo

The surprising way OpenAI could get out of its pact with Microsoft

News photo

Microsoft and OpenAI’s Close Partnership Shows Signs of Fraying