Large language models have entered the "1-bit era," where Microsoft and the University of Chinese Academy of Sciences propose the BitNet b1.58 method, which converts parameters into ternary representation, fundamentally reducing the model's memory footprint and simplifying the computational process. This method has been tested on models of various sizes, demonstrating improved speed and reduced memory usage, sparking heated discussions and debates among netizens.
Microsoft's 6-Page Paper Goes Viral: Ternary LLM, Truly Amazing!

量子位
This article is from AIbase Daily
Welcome to the [AI Daily] column! This is your daily guide to exploring the world of artificial intelligence. Every day, we present you with hot topics in the AI field, focusing on developers, helping you understand technical trends, and learning about innovative AI product applications.