The idea of simplifying model weights isn’t a completely new one in AI research. For years, researchers have been experimenting with quantization techniques that squeeze their neural network weights ...
Hosted on MSN
Microsoft researchers build 1-bit AI LLM with 2B parameters — model small enough to run on some CPUs
Microsoft researchers just created BitNet b1.58 2B4T, an open-source 1-bit large language model with two billion parameters and trained on four trillion tokens. But what makes this AI model unique is ...
Microsoft’s new large language model (LLM) puts significantly less strain on hardware than other LLMs—and it’s free to experiment with. The 1-bit LLM (1.58-bit, to be more precise) uses -1, 0, and 1 ...
One-bit large language models (LLMs) have emerged as a promising approach to making generative AI more accessible and affordable. By representing model weights with a very limited number of bits, ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results