Skip to content

Commit 0c68033

Browse files
authored
Update README.md
1 parent f60fa9e commit 0c68033

File tree

1 file changed

+6
-1
lines changed

1 file changed

+6
-1
lines changed

README.md

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,6 +22,11 @@ The main goal is to run the model using 4-bit quantization on a MacBook.
2222
- Runs on the CPU
2323

2424
This was hacked in an evening - I have no idea if it works correctly.
25+
Please do not make conclusions about the models based on the results from this implementation.
26+
For all I know, it can be completely wrong. This project is for educational purposes and is not going to be maintained properly.
27+
New features will probably be added mostly through community contributions, if any.
28+
29+
---
2530

2631
Here is a typical run using LLaMA-7B:
2732

@@ -183,7 +188,7 @@ When running the larger models, make sure you have enough disk space to store al
183188
- x86 quantization support [not yet ready](https://github.com/ggerganov/ggml/pull/27). Basically, you want to run this
184189
on Apple Silicon. For now, on Linux and Windows you can use the F16 `ggml-model-f16.bin` model, but it will be much
185190
slower.
186-
- The Accelerate framework is actually currently unused since I found that for tensors shapes typical for the Decoder,
191+
- The Accelerate framework is actually currently unused since I found that for tensor shapes typical for the Decoder,
187192
there is no benefit compared to the ARM_NEON intrinsics implementation. Of course, it's possible that I simlpy don't
188193
know how to utilize it properly. But in any case, you can even disable it with `LLAMA_NO_ACCELERATE=1 make` and the
189194
performance will be the same, since no BLAS calls are invoked by the current implementation

0 commit comments

Comments
 (0)