You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: examples/pytorch/nlp/huggingface_models/language-modeling/quantization/autoround/README.md
+16-11Lines changed: 16 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,20 +6,24 @@ AutoRound is an advanced weight-only quantization algorithm, based on SignRound.
6
6
## Prerequisites
7
7
- Python 3.9 or higher
8
8
9
+
9
10
- The transformers version required varies across different types of models. Here, the transformers version used for running models during experiments is provided as a reference.
10
11
| Model | Transformers version |
11
12
| :----: | :----: |
12
-
| EleutherAI/gpt-j-6b | 4.28/4.30/4.34 |
13
-
| huggyllama/llama-7b | 4.28/4.30/4.34 |
14
-
| meta-llama/Llama-2-7b-hf | 4.30/4.34 |
15
-
| facebook/opt-6.7b | 4.28/4.30/4.34 |
16
-
| tiiuae/falcon-7b | 4.28/4.30/4.34 |
17
-
| mosaicml/mpt-7b | 4.28/4.30/4.34 |
18
-
| bigscience/bloom-7b1 | 4.28/4.30/4.34 |
13
+
| EleutherAI/gpt-j-6b | 4.28/4.30/4.34/4.36|
14
+
| huggyllama/llama-7b | 4.28/4.30/4.34/4.36|
15
+
| meta-llama/Llama-2-7b-hf | 4.30/4.34/4.36|
16
+
| facebook/opt-6.7b | 4.28/4.30/4.34/4.36|
17
+
| tiiuae/falcon-7b | 4.28/4.30/4.34/4.36|
18
+
| mosaicml/mpt-7b | 4.28/4.30/4.34/4.36|
19
+
| bigscience/bloom-7b1 | 4.28/4.30/4.34/4.36|
19
20
| baichuan-inc/Baichuan-7B | 4.28/4.30 |
20
-
| Qwen/Qwen-7B | 4.28/4.30/4.34 |
21
-
| THUDM/chatglm2-6b | 4.28/4.30 |
22
-
| mistralai/Mistral-7B-v0.1 | 4.34 |
21
+
| Qwen/Qwen-7B | 4.28/4.30/4.34/4.36 |
22
+
| THUDM/chatglm3-6b | 4.34/4.36 |
23
+
| mistralai/Mistral-7B-v0.1 | 4.34/4.36 |
24
+
25
+
Please note that all experiments in the SignRound+ technical report were conducted using transformers version 4.34.1.
26
+
23
27
24
28
25
29
## Installation
@@ -42,7 +46,7 @@ Include the flag `--adam`. Note that AdamW may be slightly less effective than S
Copy file name to clipboardExpand all lines: examples/pytorch/nlp/huggingface_models/language-modeling/quantization/signround/README.md
+15-4Lines changed: 15 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,11 +3,20 @@ This is a sample code for SignRound ([arxiv](https://arxiv.org/abs/2309.05516)),
3
3

4
4
5
5
6
-
7
6
# Prerequisite
8
-
python 3.9 or higher
7
+
-python 3.9 or higher
8
+
9
+
- The transformers version required varies across different types of models. Here, the transformers version used for running models during experiments is provided as a reference.
Please note that all experimental data in the paper is based on transformer version 3.28.1. the huggingface source for llama-7b-hf mentioned in the paper, 'decapoda-research/llama-7b-hf', is currently unavailable. You may opt for 'huggyllama/llama-7b' as an alternative, but please be aware that this replacement might yield slight differences in results.
To address the original lambada evaluation bug in the old version of lm-eval, we have incorporated the lm-eval from intel extension for transformers(ITREX). This discrepancy may lead to certain variations.
26
35
27
-
To reproduce our results in the paper, please install ITREX
36
+
To reproduce our results in the paper, please install ITREX
28
37
29
38
```bash
30
39
pip install intel-extension-for-transformers
@@ -41,3 +50,5 @@ If you find SignRound useful or relevant to your research, please kindly cite ou
0 commit comments