-
-
Notifications
You must be signed in to change notification settings - Fork 128
LlamaCpp crash when embedding (in beta) #211
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Confirmed that the embedding functionality works when using model It might be worth noting that when loading |
I've run into the same problem with |
I found the cause for this issue and will release a new version with the fix in the next few days. |
🎉 This issue has been resolved in version 3.0.0-beta.18 🎉 The release is available on:
Your semantic-release bot 📦🚀 |
Edit: found this issue in llama.cpp: Hello, while I confirm this fixes Mac Os BGE models, this causes a crash on Windows. Running the test code above with
|
I'll release a new version in the next few days that will include prebuilt binaries with the fix for Vulkan |
On Mac (mac-arm64-metal): 3.0 beta 18 + Electron (with Electron-forge + vite) + BGE models run on Electron development ( EDIT: A non-crashing workaround (but effectively useless b/c gibberish is outputted) is using |
@bitterspeed Eventually, it's going to take more than a few days until I can publish a new beta version due to a bug in the In the meantime, you can run this command to download and build the most recent release of npx --no node-llama-cpp download --release b2952 |
@giladgd Thanks. I have run that command, and while that above Vulkan error does not show up anymore, there is now a crash on runtime (with no error message) when using Vulkan ( CUDA works fine). This applies to not only using BGE embeddings but also Llama 3 inference.
|
@bitterspeed I've managed to use BGE and run inference with Llama 3 on release |
Amazing. Thank you for the guidance, works perfectly! |
🎉 This PR is included in version 3.0.0 🎉 The release is available on: Your semantic-release bot 📦🚀 |
Uh oh!
There was an error while loading. Please reload this page.
Issue description
LlamaCpp crash when embedding
Expected Behavior
The code generates correct embedding vector.
Actual Behavior
LlamaCpp crashed with error code:
Steps to reproduce
Download the embedding model from https://huggingface.co/CompendiumLabs/bge-large-en-v1.5-gguf/tree/main
Run the following code:
My Environment
node-llama-cpp
versionAdditional Context
No response
Relevant Features Used
Are you willing to resolve this issue by submitting a Pull Request?
Yes, I have the time, but I don't know how to start. I would need guidance.
The text was updated successfully, but these errors were encountered: