Porting LeoLM instruct models to llama.cpp #3935
Unanswered
sorgfresser
asked this question in
Q&A
Replies: 1 comment 2 replies
-
I think what you might need to do is add pad tokens rather than trying to change the model's vocab size. You can try #3743 with the |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hey,
I'm very impressed by the speed and ease at which llama.cpp can deploy many models. I tried converting a German & English only model named LeoLM but did only manage to get it to work for the non-instruct finetuned variants which seems a bit odd to me.
First of all, if I just try to convert the
LeoLM/leo-hessianai-7b-chat
(available on hf) I getThe vocab size of the instruction ones does actually exceed
32000
(there are some special tokens above id31999
in the range[32000, 32006]
) but by simply modifying the config.json to vocab_size: 32000 similar to #3900 I at least managed to get the conversion itself to run through.But if I run
./main -m "models/leo-hessianai-7b-chat/ggml-model-f16.gguf"
it fails withwhich is fair since I edited the vocab size from
32128
to32000
but confuses me since I initially got the impression that the tokenizer only had 32000 tokens due to the error.Is there any param I could pass to convert.py that I'm overlooking? All the ones I get using -h do not seem helpful to me.
Thanks a lot for any advice!
Beta Was this translation helpful? Give feedback.
All reactions