Model compression techniques, such as quantization, pruning, distillation, and low-rank adaptation, are widely used to reduce the deployment cost of language models while maintaining performance. However, compression can result in:
- Increased stereotype generation
- Reduced robustness to adversarial attacks
- Increased calibration errors
- Higher model uncertainty
- Overall safety risks in downstream applications
This repository curates papers on the evaluation and mitigation of compression-induced safety degradation in LLMs, VLMs, and multimodal models, covering robustness, calibration, and alignment.
Papers are currently listed in a single section. Subcategories will be introduced as the collection grows.
Contributions are welcome! Please open an issue or submit a pull request following the existing format.
| Date | Institute | Publication | Paper |
|---|---|---|---|
| 24.03 | UT Austin; Drexel; MIT; UIUC; Duke; LLNL; CAIS; UC Berkeley; UChicago | ICML 2024 | Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression |
| 24.12 | IBM Research Europe; Trinity College Dublin; Imperial College London | NeurIPS 2024 SafeGenAI | HarmLevelBench: Evaluating Harm-Level Compliance and the Impact of Quantization on Model Alignment |
| 25.02 | Skolkovo Institute of Science and Technology; AIRI; HSE University | arXiv | Investigating the Impact of Quantization Methods on the Safety and Reliability of Large Language Models |
| 25.02 | Harbin Institute of Technology (Shenzhen); Illinois Institute of Technology | arXiv | Benchmarking Post-Training Quantization in LLMs: A Comprehensive Taxonomy |
If you use this repository in your research, you can cite it as:
@misc{proskurina2026awesome_llm_compression_safety,
title = {Awesome LLM Compression Safety},
author = {Proskurina, Irina},
year = {2026},
howpublished = {\url{https://github.com/upunaprosk/Awesome-LLM-Compression-Safety}},
note = {GitHub repository, accessed 2026}
}