[go: up one dir, main page]

Skip to content

[ICLR 2024] Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models

Notifications You must be signed in to change notification settings

johnheo/adadim-llm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models

Set up

Environment & Data: sh setup.sh

For math QA data, download the raw file from here and place it in the ./data/ folder

Set your path to HF cache directory in each of the bash files below. E.g.,

  • export TRANSFORMERS_CACHE=/username/my_hf_cache
  • export HF_DATASETS_CACHE=/username/my_hf_cache

Usage

Running RTN-ada

  • To evaluate MMLU and Common Sense Reasoning (CSR): sh scripts/rtn/lmeval.sh
  • To evaluate math reasoning: sh scripts/rtn/math.sh
  • To evaluate code generation: sh scripts/rtn/code.sh

Running GPTQ-ada

  • To evaluate MMLU and Common Sense Reasoning (CSR): sh scripts/gptq/lmeval.sh
  • To evaluate math reasoning: sh scripts/gptq/math.sh
  • To evaluate code generation: sh scripts/gptq/code.sh

Citation

If you find AdaDim helpful or relevant, please kindly cite our paper:

@inproceedings{
heo2024rethinking,
title={Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models},
author={Jung Hwan Heo and Jeonghoon Kim and Beomseok Kwon and Byeongwook Kim and Se Jung Kwon and Dongsoo Lee},
booktitle={The Twelfth International Conference on Learning Representations},
year={2024},
url={https://openreview.net/forum?id=JzG7kSpjJk}
}

Acknowledgements

This code base is expanded upon wonderful works from

About

[ICLR 2024] Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published