Détail du document
Identifiant

oai:arXiv.org:2405.14917

Sujet
Computer Science - Machine Learnin... Computer Science - Computation and...
Auteur
Huang, Wei Qin, Haotong Liu, Yangdong Li, Yawei Liu, Xianglong Benini, Luca Magno, Michele Qi, Xiaojuan
Catégorie

Computer Science

Année

2024

Date de référencement

29/05/2024

Mots clés
methods salience efficiency quantization language slim-llm llms
Métrique

Résumé

Large language models (LLMs) achieve remarkable performance in natural language understanding but require substantial computation and memory resources.

Post-training quantization (PTQ) is a powerful compression technique extensively investigated in LLMs.

However, existing PTQ methods are still not ideal in terms of accuracy and efficiency, especially with below 4 bit-widths.

Standard PTQ methods using group-wise quantization suffer difficulties in quantizing LLMs accurately to such low-bit, but advanced methods remaining high-precision weights element-wisely are hard to realize their theoretical hardware efficiency.

This paper presents a Salience-Driven Mixed-Precision Quantization scheme for LLMs, namely SliM-LLM.

The scheme exploits the salience distribution of weights to determine optimal bit-width and quantizers for accurate LLM quantization, while aligning bit-width partition to groups for compact memory usage and fast integer inference.

Specifically, the proposed SliM-LLM mainly relies on two novel techniques: (1) Salience-Determined Bit Allocation utilizes the clustering characteristics of salience distribution to allocate the bit-widths of each group, increasing the accuracy of quantized LLMs and maintaining the inference efficiency; (2) Salience-Weighted Quantizer Calibration optimizes the parameters of the quantizer by considering the element-wise salience within the group, balancing the maintenance of salient information and minimization of errors.

Comprehensive experiments show that SliM-LLM significantly improves the accuracy of LLMs at ultra-low bits, e.g., 2-bit LLaMA-7B achieves a 5.5-times memory-saving than original model on NVIDIA A800 GPUs, and 48% decrease of perplexity compared to the state-of-the-art gradient-free PTQ method.

Moreover, SliM-LLM+, which is integrated from the extension of SliM-LLM with gradient-based quantizers, further reduces perplexity by 35.1%.

;Comment: 22 pages

Huang, Wei,Qin, Haotong,Liu, Yangdong,Li, Yawei,Liu, Xianglong,Benini, Luca,Magno, Michele,Qi, Xiaojuan, 2024, SliM-LLM: Salience-Driven Mixed-Precision Quantization for Large Language Models

Document

Ouvrir

Partager

Source

Articles recommandés par ES/IODE IA

Lung cancer risk and exposure to air pollution: a multicenter North China case–control study involving 14604 subjects
lung cancer case–control air pollution never-smokers nomogram model controls lung-related 14604 subjects north polluted consistent smokers quit exposure lung cancer risk air people factor smoking pollution study history