Document detail
ID

oai:arXiv.org:2403.15447

Topic
Computer Science - Computation and... Computer Science - Artificial Inte...
Author
Hong, Junyuan Duan, Jinhao Zhang, Chenhui Li, Zhangheng Xie, Chulin Lieberman, Kelsey Diffenderfer, James Bartoldson, Brian Jaiswal, Ajay Xu, Kaidi Kailkhura, Bhavya Hendrycks, Dan Song, Dawn Wang, Zhangyang Li, Bo
Category

Computer Science

Year

2024

listing date

6/12/2024

Keywords
quantization compression llms trustworthiness
Metrics

Abstract

Compressing high-capability Large Language Models (LLMs) has emerged as a favored strategy for resource-efficient inferences.

While state-of-the-art (SoTA) compression methods boast impressive advancements in preserving benign task performance, the potential risks of compression in terms of safety and trustworthiness have been largely neglected.

This study conducts the first, thorough evaluation of three (3) leading LLMs using five (5) SoTA compression techniques across eight (8) trustworthiness dimensions.

Our experiments highlight the intricate interplay between compression and trustworthiness, revealing some interesting patterns.

We find that quantization is currently a more effective approach than pruning in achieving efficiency and trustworthiness simultaneously.

For instance, a 4-bit quantized model retains the trustworthiness of its original counterpart, but model pruning significantly degrades trustworthiness, even at 50% sparsity.

Moreover, employing quantization within a moderate bit range could unexpectedly improve certain trustworthiness dimensions such as ethics and fairness.

Conversely, extreme quantization to very low bit levels (3 bits) tends to reduce trustworthiness significantly.

This increased risk cannot be uncovered by looking at benign performance alone, in turn, mandating comprehensive trustworthiness evaluation in practice.

These findings culminate in practical recommendations for simultaneously achieving high utility, efficiency, and trustworthiness in LLMs.

Code and models are available at https://decoding-comp-trust.github.io.

;Comment: Accepted to ICML'24

Hong, Junyuan,Duan, Jinhao,Zhang, Chenhui,Li, Zhangheng,Xie, Chulin,Lieberman, Kelsey,Diffenderfer, James,Bartoldson, Brian,Jaiswal, Ajay,Xu, Kaidi,Kailkhura, Bhavya,Hendrycks, Dan,Song, Dawn,Wang, Zhangyang,Li, Bo, 2024, Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression

Document

Open

Share

Source

Articles recommended by ES/IODE AI

A Novel MR Imaging Sequence of 3D-ZOOMit Real Inversion-Recovery Imaging Improves Endolymphatic Hydrops Detection in Patients with Ménière Disease
ménière disease p < detection imaging sequences 3d-zoomit 3d endolymphatic real tse reconstruction ir inversion-recovery hydrops ratio
Successful omental flap coverage repair of a rectovaginal fistula after low anterior resection: a case report
rectovaginal fistula rectal cancer low anterior resection omental flap muscle flap rectal cancer pod initial repair rvf flap omental lar coverage