-
Notifications
You must be signed in to change notification settings - Fork 12k
models generating gibberish on hipblas #4503
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
i have some more info. the issue can be reproduced on release b1618 if i offload all or almost all layers to gpu. it also happens on the server version with b1618, the same way it does on b1619 and above. |
I had a similar problem once when I updated my checked out copy of llama.cpp and ran a build without cleaning the previous build artifacts first. Did you run |
yes, i did, multiple times |
Same issue
|
Are you actually having the same issue? Our machines are different by quite a lot, so this would prove it is in fact a bug |
Confirm |
I get gibberish when I use the updated chat prompt from https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF but good results also with GPU offload (with the UMA patch) on a gfx1035 when using the original prompt. |
i think it is fixed in the latest code now, but i need to make sure. i'll try later. |
This issue is stale because it has been open for 30 days with no activity. |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Uh oh!
There was an error while loading. Please reload this page.
Prerequisites
Expected Behavior
Should generate comprehensible responses accelerated using hipblas on my rx 6600.
example on release 1618 with mistral 7b instruct:
Current Behavior
Generates gibberish that you can't make any sense out of, but not on other frameworks, no matter the model or drivers that i use.
example on release b1619 with mistral 7b instruct:
Environment and Context
$ lscpu
$ uname -a
Linux PouPC 6.2.0-39-generic #40~22.04.1-Ubuntu SMP PREEMPT_DYNAMIC Thu Nov 16 10:53:04 UTC 2 x86_64 x86_64 x86_64 GNU/Linux
Failure Information (for bugs)
I noticed that this started happening in release b1619. this issue existed before that, but it was kind of rare, and only happened when i left llama.cpp doing nothing for at least 10 minutes and did other stuff on ubuntu.
Steps to Reproduce
make LLAMA_HIPBLAS=1
./main -m models/mistral-7b-instruct-v0.1.Q4_K_M.gguf -ngl 32 -ins
Failure logs
there doesn't seem to be any difference in logs between 1619 and 1618.
The text was updated successfully, but these errors were encountered: