Skip to content

Commit b99abae

Browse files
authored
Fix opt_125m_woq_gptq_int4_dq_ggml issue (#1965)
Signed-off-by: Kaihui-intel <[email protected]>
1 parent b35ff8f commit b99abae

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

.azure-pipelines/model-test-3x.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ pr:
1010
include:
1111
- neural_compressor/common
1212
- neural_compressor/torch
13-
- examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/llm
13+
- examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/weight_only
1414
- setup.py
1515
- requirements_pt.txt
1616
- .azure-pipelines/scripts/models

examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/weight_only/run_quant.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ function run_tuning {
5050
extra_cmd=$extra_cmd" --double_quant_type BNB_NF4"
5151
elif [ "${topology}" = "opt_125m_woq_gptq_int4_dq_ggml" ]; then
5252
model_name_or_path="facebook/opt-125m"
53-
extra_cmd=$extra_cmd" --woq_algo GPTQ --woq_bits 4 --woq_group_size 128 --woq_scheme asym --woq_use_mse_search --gptq_use_max_length --gptq_percdamp 0.1 --gptq_actorder"
53+
extra_cmd=$extra_cmd" --woq_algo GPTQ --woq_bits 4 --woq_group_size 128 --woq_scheme asym --woq_use_mse_search --gptq_use_max_length --gptq_percdamp 0.8 --gptq_actorder"
5454
extra_cmd=$extra_cmd" --double_quant_type GGML_TYPE_Q4_K"
5555
elif [ "${topology}" = "llama2_7b_gptq_int4" ]; then
5656
model_name_or_path="meta-llama/Llama-2-7b-hf"

0 commit comments

Comments
 (0)