Haolei Bai1,2, Siyong Jian1,4, Tuo Liang3, Yu Yin3, Huan Wang1,*
1Westlake University, 2Nanyang Technological University, 3Case Western Reserve University, 4Nanjing University
conda create -n erc-svd python=3.10
conda activate erc-svd
pip install -r requirements.txt# loss analysis (compression ratio=20% for example)
python ERC-SVD_compression.py --model /yout/path/to/llama-2-7b-hf/ --step 0 --ratio 0.2 --whitening_nsamples 256 --dataset wikitext2 --seed 3 --model_seq_len 2048 --layer_num 16
python ERC-SVD_compression.py --model /yout/path/to/llama-2-7b-hf/ --step 1 --ratio 0.2 --whitening_nsamples 256 --dataset wikitext2 --seed 3 --model_seq_len 2048 --layer_num 16 --save_path /your/save/path/python ERC-SVD_compression.py --step 2 --model_path /your/path/to/compressed_model.ptpython evaluation_zero_acc.py --model_id /your/path/to/compressed_model.pt --pt_flagWe are grateful to the ASVD and SVD-LLM for releasing their code publicly, which greatly facilitated our work.
If you find ERC-SVD useful for your research or projects, please consider citing our work:
@inproceedings{bai2026ercsvd,
title={ERC-SVD: Error-Controlled SVD for Large Language Model Compression},
author={Bai, Haolei and Jian, Siyong and Liang, Tuo and Yin, Yu and Wang, Huan},
booktitle={CPAL},
year={2026}
}
@article{bai2025ressvd,
title={Ressvd: Residual compensated svd for large language model compression},
author={Bai, Haolei and Jian, Siyong and Liang, Tuo and Yin, Yu and Wang, Huan},
journal={arXiv preprint arXiv:2505.20112},
year={2025}
}