File tree Expand file tree Collapse file tree 2 files changed +2
-2
lines changed
examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/weight_only Expand file tree Collapse file tree 2 files changed +2
-2
lines changed Original file line number Diff line number Diff line change 10
10
include :
11
11
- neural_compressor/common
12
12
- neural_compressor/torch
13
- - examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/llm
13
+ - examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/weight_only
14
14
- setup.py
15
15
- requirements_pt.txt
16
16
- .azure-pipelines/scripts/models
Original file line number Diff line number Diff line change @@ -50,7 +50,7 @@ function run_tuning {
50
50
extra_cmd=$extra_cmd " --double_quant_type BNB_NF4"
51
51
elif [ " ${topology} " = " opt_125m_woq_gptq_int4_dq_ggml" ]; then
52
52
model_name_or_path=" facebook/opt-125m"
53
- extra_cmd=$extra_cmd " --woq_algo GPTQ --woq_bits 4 --woq_group_size 128 --woq_scheme asym --woq_use_mse_search --gptq_use_max_length --gptq_percdamp 0.1 --gptq_actorder"
53
+ extra_cmd=$extra_cmd " --woq_algo GPTQ --woq_bits 4 --woq_group_size 128 --woq_scheme asym --woq_use_mse_search --gptq_use_max_length --gptq_percdamp 0.8 --gptq_actorder"
54
54
extra_cmd=$extra_cmd " --double_quant_type GGML_TYPE_Q4_K"
55
55
elif [ " ${topology} " = " llama2_7b_gptq_int4" ]; then
56
56
model_name_or_path=" meta-llama/Llama-2-7b-hf"
You can’t perform that action at this time.
0 commit comments