bit llms

Read news on bit llms with our app.

Read more in the app

Matrix-vector multiplication implemented in off-the-shelf DRAM for Low-Bit LLMs

Microsoft BitNet: inference framework for 1-bit LLMs

“Imprecise” language models are smaller, speedier, and nearly as accurate