Content deleted Content added
Artoria2e5 (talk | contribs) Fine-tuning LLMs to 1.58bit: extreme quantization made eas |
TokenByToken (talk | contribs) revise lead |
||
Line 1:
{{Short description|Large language model with ternary weights}}
A '''1.58-bit large language model''' (also known as a '''ternary LLM''') is a type of [[large language model]] (LLM) designed to be computationally efficient. It achieves this by using [[neural network#in machine learning|weights]] that are restricted to only three values: -1, 0, and +1. This restriction significantly reduces the model's memory footprint and allows for faster processing, as complex multiplication operations can be replaced with simpler additions. This contrasts with traditional models that use 16-bit floating-point numbers ([[FP16]] or [[BF16]]) for their weights.
Studies have shown that for models up to several billion parameters, the performance of 1.58-bit LLMs on various tasks is comparable to their full-precision counterparts.{{sfn|Ma|Wang|Ma|Wang|2024|p=1}}{{sfn|Hutson|2024}} This approach could enable powerful AI to run on less specialized and lower-power hardware.{{sfn|Friha|Amine Ferrag|Kantarci|Cakmak|2024|p=5822}}
The name "1.58-bit" comes from the fact that a system with three states contains <math>\log_2 3 \approx 1.58</math> [[bit|bits]] of [[information theory|information]]. These models are sometimes also referred to as '''1-bit LLMs''' in research papers, although this term can also refer to true binary models (with weights of -1 and +1).{{sfn|Ma|Wang|Ma|Wang|2024|p=1}}{{sfn|Morales|2025}}
== BitNet ==
{{redirect|BitNet|a computer network|BITNET}}
|