Closed
Description
convert-llama-ggmlv3-to-gguf.py
produces a model without errors, which causes unexpected behaviour in main
and server
I believe the model is this one from here
Source model md5 e87520b6393ea5ed6f9419e9fe6aba96 mythomax-l2-13b.ggmlv3.q5_K_M.bin
Resulting model md5 ce6cf60b707cb21fc04ac0e6cf6a147e mythomax-l2-13b.ggmlv3.q5_K_M.gguf
Exact command with output:
python3 convert-llama-ggmlv3-to-gguf.py -i /storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.bin -o /storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.gguf --eps 1e-5 -c 4096
* Using config: Namespace(input=PosixPath('/storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.bin'), output=PosixPath('/storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.gguf'), name=None, desc=None, gqa=1, eps='1e-5', context_length=4096, model_metadata_dir=None, vocab_dir=None, vocabtype='spm')
=== WARNING === Be aware that this conversion script is best-effort. Use a native GGUF model if possible. === WARNING ===
* Scanning GGML input file
* GGML model hyperparameters: <Hyperparameters: n_vocab=32000, n_embd=5120, n_mult=6912, n_head=40, n_layer=40, n_rot=128, n_ff=13824, ftype=17>
=== WARNING === Special tokens may not be converted correctly. Use --model-metadata-dir if possible === WARNING ===
* Preparing to save GGUF file
* Adding model parameters and KV items
* Adding 32000 vocab item(s)
* Adding 363 tensor(s)
gguf: write header
gguf: write metadata
gguf: write tensors
* Successful completion. Output saved to: /storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.gguf
Metadata
Metadata
Assignees
Labels
No labels