Microsoft’s research division has unveiled an innovative development in artificial intelligence called BitNet b1.58 2B4T. This breakthrough model is designed to significantly reduce hardware dependency by operating exclusively on CPUs, even on widely-used hardware such as Apple’s M2 processors.
BitNet represents a hyper-efficient AI model, distinguishing itself by compressing model weights to three distinct values: -1, 0, and 1. This technique allows for high-level AI computations without the necessity of massive graphics cards. With 2 billion parameters, BitNet is the most extensive 1-bit AI model available, trained on an enormous dataset of approximately 4 trillion tokens, equivalent to about 33 million books.
Microsoft’s BitNet demonstrates remarkable superiority in benchmark comparisons, outperforming other AI models like Meta’s Llama 3.2 1B, Google’s Gemma 3 1B, and Alibaba’s Qwen 2.5 1.5B. Moreover, BitNet boasts a potential to operate twice as fast as its counterparts, with notably lower memory consumption.
Despite its advantages, BitNet’s full performance relies on Microsoft’s proprietary bitnet.cpp framework, currently compatible with a limited set of processors, indicating a need for hardware compatibility. As of now, the framework does not support the GPUs that predominantly power today’s AI infrastructure.
ENGLİSH
5 gün önceSİGORTA
5 gün önceSİGORTA
6 gün önceSİGORTA
9 gün önceSİGORTA
10 gün önceSİGORTA
10 gün önceDÜNYA
19 gün önceVeri politikasındaki amaçlarla sınırlı ve mevzuata uygun şekilde çerez konumlandırmaktayız. Detaylar için veri politikamızı inceleyebilirsiniz.