Skip to content

Commit 0f8269a

Browse files
[SAVE] Save config files with empty state dict (#1293)
* Save model and config files with empty state dict * cleanup * cleanup
1 parent 48318ac commit 0f8269a

File tree

2 files changed

+7
-6
lines changed

2 files changed

+7
-6
lines changed

gptqmodel/models/base.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -952,7 +952,7 @@ def save(
952952
meta_quantizer: Optional[str] = None,
953953
**kwargs,
954954
):
955-
extra_json_file_names = ["preprocessor_config.json", "chat_template.json"]
955+
extra_json_file_names = ["preprocessor_config.json"]
956956
for name in extra_json_file_names:
957957
json_path = os.path.join(self.model_local_path, name)
958958
if os.path.exists(json_path):

gptqmodel/models/writer.py

Lines changed: 6 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -29,6 +29,7 @@
2929
from huggingface_hub import split_torch_state_dict_into_shards
3030
from huggingface_hub.constants import SAFETENSORS_WEIGHTS_FILE_PATTERN
3131
from safetensors.torch import save_file as safe_save
32+
from torch import nn
3233
from transformers import AutoConfig, PreTrainedTokenizerFast, GenerationConfig
3334
from transformers.modeling_utils import no_init_weights
3435
from transformers.models.auto.tokenization_auto import get_tokenizer_config
@@ -305,13 +306,13 @@ def save_quantized(
305306
logger.info(f"Size difference: {size_diff_mb:.2f}MB, {size_diff_gb:.2f}GB - {percent_diff:.2f}%")
306307

307308
config.quantization_config = quantize_config.to_dict()
308-
config.save_pretrained(save_dir)
309309

310-
quantize_config.save_pretrained(save_dir)
310+
self.model.config = config
311+
312+
# Save config files with empty state dict
313+
self.model.save_pretrained(save_dir, state_dict={})
311314

312-
# check for generation_config(.json)
313-
if hasattr(self, "generation_config") and isinstance(self.generation_config, GenerationConfig):
314-
self.generation_config.save_pretrained(save_dir)
315+
quantize_config.save_pretrained(save_dir)
315316

316317
# need to copy .py files for model/tokenizers not yet merged to HF transformers
317318
if self.trust_remote_code:

0 commit comments

Comments
 (0)