@@ -787,7 +787,7 @@ def __init__(self, load_config: LoadConfig):
787
787
with open (config_file_path , "r" ) as f :
788
788
config = json .load (f )
789
789
self .target_modules = config ["target_modules" ]
790
- self .wo_sharded_weights_modules : List [str ] = []
790
+ self .unsharded_weights_modules : List [str ] = []
791
791
792
792
def _get_config_file (self , qlora_adapter : str ) -> str :
793
793
is_local = os .path .isdir (qlora_adapter )
@@ -1010,7 +1010,7 @@ def _unquantized_generator(self, hf_weights_files, use_safetensors,
1010
1010
# Without sharding
1011
1011
if any (
1012
1012
weight_name .startswith (module )
1013
- for module in self .wo_sharded_weights_modules ):
1013
+ for module in self .unsharded_weights_modules ):
1014
1014
weight_sub_tensor = weight_tensor
1015
1015
# Shard by column
1016
1016
elif any (module in weight_name
@@ -1080,7 +1080,7 @@ def _load_weights(self, model_config: ModelConfig,
1080
1080
# static variable in the model implementation.
1081
1081
# TODO: Can we reduce the static variables needed for BNB based on
1082
1082
# model information?
1083
- self .wo_sharded_weights_modules = [
1083
+ self .unsharded_weights_modules = [
1084
1084
name for name , module in model .named_modules ()
1085
1085
if isinstance (module , (ReplicatedLinear , ))
1086
1086
]
0 commit comments