-
Notifications
You must be signed in to change notification settings - Fork 11.9k
Silently failing ggml to gguf conversion #2697
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Thank you. Can you please try the fix here? #2698 Just change line 240 of the conversion script from vbytes = bytes(f'<0x{hv}>', encoding = 'UTF-8') to vbytes = bytes(f'<0x{vbytes[0]:02X}>', encoding = 'UTF-8') |
At first glance, this appears to have fixed it, tested with server:
Resultant model md5 after fix: fcf16d638dc53d4bec7e827ee71192de mythomax-l2-13b.ggmlv3.q5_K_M.gguf
|
Great, thanks for the repeat and testing! I will try to get the fix merged as soon as possible. |
Hmm, is it possible to convert ggml to gguf on mobile? or are you converting on PC? |
I think they just meant they were currently not at their computer and posting from mobile. But you can set up a Unix environment on Android phones pretty easily and run stuff like Python scripts, compilers. You can even compile/run llama.cpp on mobile device, though it's not gonna be super fast. |
Pretty much yes, to all of that. I was momentarily not in my house. You could actually run any part of llama.cpp on Android at least. If you have Termux or equivalent, you get plain old Linux shell and all that comes with it. |
We should be good now. Please let me know if you have any further issues with converted models! |
Yeah, probably.
I run llama.cpp in termux daily. I've got python/numpy installed, but got some strange error that doesn't recognise numpy:
It's not a big deal 'cuz I don't have a PC so it's probably too intensive for my little Android to handle. I'll be patient ❤️ |
@JackJollimore did you try |
Ah, I'm downloading updated llama.cpp and I'll try again shortly to see if I can actually finish a conversion. Edit: It converted! I didn't know my device could do that, thanks for the suggestion. |
@JackJollimore You need to install the gguf package for python. You can I'm not sure if this is documented anywhere. |
Right again @cebtenzzre, Thank you. |
Uh oh!
There was an error while loading. Please reload this page.
@KerfuffleV2
#2398
convert-llama-ggmlv3-to-gguf.py
produces a model without errors, which causes unexpected behaviour inmain
andserver
I believe the model is this one from here
Source model md5 e87520b6393ea5ed6f9419e9fe6aba96 mythomax-l2-13b.ggmlv3.q5_K_M.bin
Resulting model md5 ce6cf60b707cb21fc04ac0e6cf6a147e mythomax-l2-13b.ggmlv3.q5_K_M.gguf
Exact command with output:
The text was updated successfully, but these errors were encountered: