File tree Expand file tree Collapse file tree 1 file changed +8
-1
lines changed Expand file tree Collapse file tree 1 file changed +8
-1
lines changed Original file line number Diff line number Diff line change @@ -5,10 +5,11 @@ Inference of [Facebook's LLaMA](https://github.com/facebookresearch/llama) model
5
5
** Hot topics**
6
6
7
7
- Running on Windows: https://github.com/ggerganov/llama.cpp/issues/22
8
+ - Fix Tokenizer / Unicode support: https://github.com/ggerganov/llama.cpp/issues/11
8
9
9
10
## Description
10
11
11
- The main goal is to run the model using 4-bit quantization on a MacBook.
12
+ The main goal is to run the model using 4-bit quantization on a MacBook
12
13
13
14
- Plain C/C++ implementation without dependencies
14
15
- Apple silicon first-class citizen - optimized via Arm Neon and Accelerate framework
@@ -22,6 +23,12 @@ Please do not make conclusions about the models based on the results from this i
22
23
For all I know, it can be completely wrong. This project is for educational purposes and is not going to be maintained properly.
23
24
New features will probably be added mostly through community contributions, if any.
24
25
26
+ Supported platformst:
27
+
28
+ - [X] Mac OS
29
+ - [X] Linux
30
+ - [ ] Windows (soon)
31
+
25
32
---
26
33
27
34
Here is a typical run using LLaMA-7B:
You can’t perform that action at this time.
0 commit comments