Skip to content

Commit 921dcee

Browse files
authored
readme: add missing info (ggml-org#1324)
1 parent 2d13786 commit 921dcee

File tree

1 file changed

+4
-2
lines changed

1 file changed

+4
-2
lines changed

README.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -18,10 +18,12 @@ The main goal of `llama.cpp` is to run the LLaMA model using 4-bit integer quant
1818

1919
- Plain C/C++ implementation without dependencies
2020
- Apple silicon first-class citizen - optimized via ARM NEON and Accelerate framework
21-
- AVX2 support for x86 architectures
21+
- AVX, AVX2 and AVX512 support for x86 architectures
2222
- Mixed F16 / F32 precision
23-
- 4-bit integer quantization support
23+
- 4-bit, 5-bit and 8-bit integer quantization support
2424
- Runs on the CPU
25+
- OpenBLAS support
26+
- cuBLAS and CLBlast support
2527

2628
The original implementation of `llama.cpp` was [hacked in an evening](https://github.com/ggerganov/llama.cpp/issues/33#issuecomment-1465108022).
2729
Since then, the project has improved significantly thanks to many contributions. This project is for educational purposes and serves

0 commit comments

Comments
 (0)