Skip to content

Silently failing ggml to gguf conversion #2697

Closed
@staviq

Description

@staviq

@KerfuffleV2
#2398

convert-llama-ggmlv3-to-gguf.py produces a model without errors, which causes unexpected behaviour in main and server

I believe the model is this one from here

Source model md5 e87520b6393ea5ed6f9419e9fe6aba96 mythomax-l2-13b.ggmlv3.q5_K_M.bin
Resulting model md5 ce6cf60b707cb21fc04ac0e6cf6a147e mythomax-l2-13b.ggmlv3.q5_K_M.gguf

Exact command with output:

python3 convert-llama-ggmlv3-to-gguf.py -i /storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.bin -o /storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.gguf --eps 1e-5 -c 4096
* Using config: Namespace(input=PosixPath('/storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.bin'), output=PosixPath('/storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.gguf'), name=None, desc=None, gqa=1, eps='1e-5', context_length=4096, model_metadata_dir=None, vocab_dir=None, vocabtype='spm')

=== WARNING === Be aware that this conversion script is best-effort. Use a native GGUF model if possible. === WARNING ===

* Scanning GGML input file
* GGML model hyperparameters: <Hyperparameters: n_vocab=32000, n_embd=5120, n_mult=6912, n_head=40, n_layer=40, n_rot=128, n_ff=13824, ftype=17>

=== WARNING === Special tokens may not be converted correctly. Use --model-metadata-dir if possible === WARNING ===

* Preparing to save GGUF file
* Adding model parameters and KV items
* Adding 32000 vocab item(s)
* Adding 363 tensor(s)
    gguf: write header
    gguf: write metadata
    gguf: write tensors
* Successful completion. Output saved to: /storage/models/mythomax-l2-13b.ggmlv3.q5_K_M.gguf

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions