@@ -160,13 +160,14 @@ def guessed(model: 'LazyModel') -> 'Params':
160
160
def loadHFTransformerJson (model : 'LazyModel' , config_path : 'Path' ) -> 'Params' :
161
161
config = json .load (open (config_path ))
162
162
163
- n_vocab = config ["vocab_size" ]
164
- n_embd = config ["hidden_size" ]
165
- n_layer = config ["num_hidden_layers" ]
166
- n_ff = config ["intermediate_size" ]
167
- n_head = config ["num_attention_heads" ]
168
- n_head_kv = config ["num_key_value_heads" ] if "num_key_value_heads" in config else n_head
169
- f_norm_eps = config ["rms_norm_eps" ]
163
+ n_vocab = config ["vocab_size" ]
164
+ n_embd = config ["hidden_size" ]
165
+ n_layer = config ["num_hidden_layers" ]
166
+ n_ff = config ["intermediate_size" ]
167
+ n_head = config ["num_attention_heads" ]
168
+ n_head_kv = config ["num_key_value_heads" ] if "num_key_value_heads" in config else n_head
169
+ f_norm_eps = config ["rms_norm_eps" ]
170
+ f_rope_freq_base = config ["rope_theta" ] if "rope_theta" in config else None
170
171
171
172
n_mult = Params .find_n_mult (n_ff , n_embd )
172
173
@@ -179,15 +180,16 @@ def loadHFTransformerJson(model: 'LazyModel', config_path: 'Path') -> 'Params':
179
180
"Suggestion: provide 'config.json' of the model in the same directory containing model files." )
180
181
181
182
return Params (
182
- n_vocab = n_vocab ,
183
- n_embd = n_embd ,
184
- n_mult = n_mult ,
185
- n_layer = n_layer ,
186
- n_ctx = n_ctx ,
187
- n_ff = n_ff ,
188
- n_head = n_head ,
189
- n_head_kv = n_head_kv ,
190
- f_norm_eps = f_norm_eps ,
183
+ n_vocab = n_vocab ,
184
+ n_embd = n_embd ,
185
+ n_mult = n_mult ,
186
+ n_layer = n_layer ,
187
+ n_ctx = n_ctx ,
188
+ n_ff = n_ff ,
189
+ n_head = n_head ,
190
+ n_head_kv = n_head_kv ,
191
+ f_norm_eps = f_norm_eps ,
192
+ f_rope_freq_base = f_rope_freq_base ,
191
193
)
192
194
193
195
# LLaMA v2 70B params.json
0 commit comments