Skip to content

Error: Invalid model file when using converted GPT4ALL model after following provided instructions #655

Closed
@gaceladri

Description

@gaceladri

Hello,

I have followed the instructions provided for using the GPT-4ALL model. I used the convert-gpt4all-to-ggml.py script to convert the gpt4all-lora-quantized.bin model, as instructed. However, I encountered an error related to an invalid model file when running the example.

Here are the steps I followed, as described in the instructions:

  1. Convert the model using the convert-gpt4all-to-ggml.py script:
python3 convert-gpt4all-to-ggml.py models/gpt4all/gpt4all-lora-quantized.bin ./models/tokenizer.model
  1. Run the interactive mode example with the newly generated gpt4all-lora-quantized.bin model:
./main -m ./models/gpt4all/gpt4all-lora-quantized.bin -n 256 --repeat_penalty 1.0 --color -i -r "User:" -f prompts/chat-with-bob.txt

However, I encountered the following error:

./models/gpt4all/gpt4all-lora-quantized.bin: invalid model file (bad magic [got 0x67676d66 want 0x67676a74])
you most likely need to regenerate your ggml files
the benefit is you'll get 10-100x faster load times
see https://github.com/ggerganov/llama.cpp/issues/91
use convert-pth-to-ggml.py to regenerate from original pth
use migrate-ggml-2023-03-30-pr613.py if you deleted originals
llama_init_from_file: failed to load model
main: error: failed to load model './models/gpt4all/gpt4all-lora-quantized.bin'

Please let me know how to resolve this issue and correctly convert and use the GPT-4ALL model with the interactive mode example.

Thank you.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions