@@ -145,7 +145,6 @@ def type_for_tensor(self, name: str, tensor: LazyTensor) -> DataType:
145
145
class Params :
146
146
n_vocab : int
147
147
n_embd : int
148
- n_mult : int
149
148
n_layer : int
150
149
n_ctx : int
151
150
n_ff : int
@@ -161,15 +160,6 @@ class Params:
161
160
# path to the directory containing the model files
162
161
path_model : Path | None = None
163
162
164
- @staticmethod
165
- def find_n_mult (n_ff : int , n_embd : int ) -> int :
166
- # hardcoded magic range
167
- for n_mult in range (8192 , 1 , - 1 ):
168
- calc_ff = (((8 * n_embd ) // 3 + n_mult - 1 ) // n_mult )* n_mult
169
- if calc_ff == n_ff :
170
- return n_mult
171
- raise Exception (f"failed to find n_mult for (n_ff={ n_ff } , n_embd={ n_embd } )." )
172
-
173
163
@staticmethod
174
164
def guessed (model : LazyModel ) -> Params :
175
165
# try transformer naming first
@@ -197,7 +187,6 @@ def guessed(model: LazyModel) -> Params:
197
187
return Params (
198
188
n_vocab = n_vocab ,
199
189
n_embd = n_embd ,
200
- n_mult = n_mult ,
201
190
n_layer = n_layer ,
202
191
n_ctx = - 1 ,
203
192
n_ff = n_ff ,
@@ -225,8 +214,6 @@ def loadHFTransformerJson(model: LazyModel, config_path: Path) -> Params:
225
214
else :
226
215
f_rope_scale = None
227
216
228
- n_mult = Params .find_n_mult (n_ff , n_embd )
229
-
230
217
if "max_sequence_length" in config :
231
218
n_ctx = config ["max_sequence_length" ]
232
219
elif "max_position_embeddings" in config :
@@ -238,7 +225,6 @@ def loadHFTransformerJson(model: LazyModel, config_path: Path) -> Params:
238
225
return Params (
239
226
n_vocab = n_vocab ,
240
227
n_embd = n_embd ,
241
- n_mult = n_mult ,
242
228
n_layer = n_layer ,
243
229
n_ctx = n_ctx ,
244
230
n_ff = n_ff ,
@@ -250,15 +236,14 @@ def loadHFTransformerJson(model: LazyModel, config_path: Path) -> Params:
250
236
)
251
237
252
238
# LLaMA v2 70B params.json
253
- # {"dim": 8192, "multiple_of": 4096, "ffn_dim_multiplier": 1.3, "n_heads": 64, "n_kv_heads": 8, "n_layers": 80, "norm_eps": 1e-05, "vocab_size": -1
239
+ # {"dim": 8192, "multiple_of": 4096, "ffn_dim_multiplier": 1.3, "n_heads": 64, "n_kv_heads": 8, "n_layers": 80, "norm_eps": 1e-05, "vocab_size": -1}
254
240
@staticmethod
255
241
def loadOriginalParamsJson (model : LazyModel , config_path : Path ) -> Params :
256
242
config = json .load (open (config_path ))
257
243
258
244
n_vocab = config ["vocab_size" ] if "vocab_size" in config else - 1
259
245
n_embd = config ["dim" ]
260
246
n_layer = config ["n_layers" ]
261
- n_mult = config ["multiple_of" ]
262
247
n_ff = - 1
263
248
n_head = config ["n_heads" ]
264
249
n_head_kv = config ["n_kv_heads" ] if "n_kv_heads" in config else n_head
@@ -285,7 +270,6 @@ def loadOriginalParamsJson(model: LazyModel, config_path: Path) -> Params:
285
270
return Params (
286
271
n_vocab = n_vocab ,
287
272
n_embd = n_embd ,
288
- n_mult = n_mult ,
289
273
n_layer = n_layer ,
290
274
n_ctx = n_ctx ,
291
275
n_ff = n_ff ,
0 commit comments