We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent d938e59 commit 37d5192Copy full SHA for 37d5192
docs/macos_install.md
@@ -38,14 +38,11 @@ llama-cpp-python 0.1.62
38
39
```
40
41
-**(4) Download a v3 ggml llama/vicuna/alpaca model**
+**(4) Download a v3 ggml model**
42
- **ggmlv3**
43
- file name ends with **q4_0.bin** - indicating it is 4bit quantized, with quantisation method 0
44
45
-https://huggingface.co/vicuna/ggml-vicuna-13b-1.1/blob/main/ggml-vic13b-q4_0.bin
46
-https://huggingface.co/vicuna/ggml-vicuna-13b-1.1/blob/main/ggml-vic13b-uncensored-q4_0.bin
47
-https://huggingface.co/TheBloke/LLaMa-7B-GGML/blob/main/llama-7b.ggmlv3.q4_0.bin
48
-https://huggingface.co/TheBloke/LLaMa-13B-GGML/blob/main/llama-13b.ggmlv3.q4_0.bin
+https://huggingface.co/TheBloke/open-llama-7b-open-instruct-GGML
49
50
51
**(6) run the llama-cpp-python API server with MacOS Metal GPU support**
0 commit comments