@@ -209,9 +209,9 @@ def _convert_kohya_lora_to_diffusers(state_dict, unet_name="unet", text_encoder_
209209
210210 if is_unet_dora_lora :
211211 dora_scale_key_to_replace = "_lora.down." if "_lora.down." in diffusers_name else ".lora.down."
212- unet_state_dict [diffusers_name . replace ( dora_scale_key_to_replace , ".lora_magnitude_vector." )] = (
213- state_dict . pop ( key . replace ("lora_down.weight" , "dora_scale" ) )
214- )
212+ unet_state_dict [
213+ diffusers_name . replace (dora_scale_key_to_replace , ".lora_magnitude_vector." )
214+ ] = state_dict . pop ( key . replace ( "lora_down.weight" , "dora_scale" ) )
215215
216216 elif lora_name .startswith (("lora_te_" , "lora_te1_" , "lora_te2_" )):
217217 if lora_name .startswith (("lora_te_" , "lora_te1_" )):
@@ -249,13 +249,13 @@ def _convert_kohya_lora_to_diffusers(state_dict, unet_name="unet", text_encoder_
249249 "_lora.down." if "_lora.down." in diffusers_name else ".lora_linear_layer."
250250 )
251251 if lora_name .startswith (("lora_te_" , "lora_te1_" )):
252- te_state_dict [diffusers_name . replace ( dora_scale_key_to_replace_te , ".lora_magnitude_vector." )] = (
253- state_dict . pop ( key . replace ("lora_down.weight" , "dora_scale" ) )
254- )
252+ te_state_dict [
253+ diffusers_name . replace (dora_scale_key_to_replace_te , ".lora_magnitude_vector." )
254+ ] = state_dict . pop ( key . replace ( "lora_down.weight" , "dora_scale" ) )
255255 elif lora_name .startswith ("lora_te2_" ):
256- te2_state_dict [diffusers_name . replace ( dora_scale_key_to_replace_te , ".lora_magnitude_vector." )] = (
257- state_dict . pop ( key . replace ("lora_down.weight" , "dora_scale" ) )
258- )
256+ te2_state_dict [
257+ diffusers_name . replace (dora_scale_key_to_replace_te , ".lora_magnitude_vector." )
258+ ] = state_dict . pop ( key . replace ( "lora_down.weight" , "dora_scale" ) )
259259
260260 # Rename the alphas so that they can be mapped appropriately.
261261 if lora_name_alpha in state_dict :
0 commit comments