Closed
Description
Hello,
I have followed the instructions provided for using the GPT-4ALL model. I used the convert-gpt4all-to-ggml.py
script to convert the gpt4all-lora-quantized.bin
model, as instructed. However, I encountered an error related to an invalid model file when running the example.
Here are the steps I followed, as described in the instructions:
- Convert the model using the
convert-gpt4all-to-ggml.py
script:
python3 convert-gpt4all-to-ggml.py models/gpt4all/gpt4all-lora-quantized.bin ./models/tokenizer.model
- Run the
interactive mode
example with the newly generatedgpt4all-lora-quantized.bin
model:
./main -m ./models/gpt4all/gpt4all-lora-quantized.bin -n 256 --repeat_penalty 1.0 --color -i -r "User:" -f prompts/chat-with-bob.txt
However, I encountered the following error:
./models/gpt4all/gpt4all-lora-quantized.bin: invalid model file (bad magic [got 0x67676d66 want 0x67676a74])
you most likely need to regenerate your ggml files
the benefit is you'll get 10-100x faster load times
see https://github.com/ggerganov/llama.cpp/issues/91
use convert-pth-to-ggml.py to regenerate from original pth
use migrate-ggml-2023-03-30-pr613.py if you deleted originals
llama_init_from_file: failed to load model
main: error: failed to load model './models/gpt4all/gpt4all-lora-quantized.bin'
Please let me know how to resolve this issue and correctly convert and use the GPT-4ALL model with the interactive mode
example.
Thank you.
Metadata
Metadata
Assignees
Labels
No labels