
microsoft/BitNet: Official inference framework for 1-bit LLMs
bitnet.cpp is the official inference framework for 1-bit LLMs (e.g., BitNet b1.58). It offers a suite of optimized kernels, that support fast and lossless inference of 1.58-bit models on CPU (with NPU and GPU support coming next).
Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on …
13 hours ago · BitNet b1.58 2B4T is a native 1-bit LLM trained at scale; it only takes up 400MB, compared to other “small models” that can reach up to 4.8 GB. BitNet b1.58 2B4T model performance, purpose ...
How to Install Microsoft’s BitNet Locally on CPU or GPU
2 days ago · BitNet b1.58 2B4T, the first 1-bit model with 2 billion parameters, is developed to run efficiently even on CPUs like Apple’s M2. What makes BitNet special isn’t just its compact size, it’s the way it works. Instead of relying on full-precision weights, BitNet compresses them into just three values: -1, 0, and 1.
Microsoft researchers say they've developed a hyper-efficient AI …
2 days ago · Microsoft researchers claim they’ve developed the largest-scale 1-bit AI model, also known as a “bitnet,” to date. Called BitNet b1.58 2B4T, it’s openly available under an MIT license and ...
Releases · microsoft/BitNet - GitHub
Official inference framework for 1-bit LLMs. Contribute to microsoft/BitNet development by creating an account on GitHub.
[2504.12285] BitNet b1.58 2B4T Technical Report - arXiv.org
2 days ago · We introduce BitNet b1.58 2B4T, the first open-source, native 1-bit Large Language Model (LLM) at the 2-billion parameter scale. Trained on a corpus of 4 trillion tokens, the model has been rigorously evaluated across benchmarks covering language understanding, mathematical reasoning, coding proficiency, and conversational ability. Our results …
Microsoft researchers say new bitnet can run on CPUs, …
2 days ago · The Microsoft researchers say that BitNet b1.58 2B4T is the first bitnet with 2 billion parameters, “parameters” being largely synonymous with “weights.”
BitNet: Scaling 1-bit Transformers for Large Language Models
In this work, we introduce BitNet, a scalable and stable 1-bit Transformer architecture designed for large language models. Specifically, we introduce BitLinear as a drop-in replacement of the nn.Linear layer in order to train 1-bit weights from scratch.
BitNet: Scaling 1-bit Transformers for Large Language Models
Oct 17, 2023 · In this work, we introduce BitNet, a scalable and stable 1-bit Transformer architecture designed for large language models. Specifically, we introduce BitLinear as a drop-in replacement of the this http URL layer in order to train 1-bit weights from scratch.
BitNet/README.md at main · microsoft/BitNet - GitHub
bitnet.cpp is the official inference framework for 1-bit LLMs (e.g., BitNet b1.58). It offers a suite of optimized kernels, that support fast and lossless inference of 1.58-bit models on CPU (with NPU and GPU support coming next).
- Some results have been removed