Skip to content

Commit 8a3e5ef

Browse files
authored
Move model section from issue template to README.md (#421)
* Update custom.md * Removed Model section as it is better placed in README.md * Updates to README.md model section * Inserted text that was removed from issue template about obtaining models from FB and links to papers describing the various models * Removed IPF down links for the Alpaca 7B models as these look to be in the old data format and probably shouldn't be directly linked to, anyway * Updated the perplexity section to point at Perplexity scores #406 discussion
1 parent 8eea5ae commit 8a3e5ef

File tree

2 files changed

+23
-36
lines changed

2 files changed

+23
-36
lines changed

.github/ISSUE_TEMPLATE/custom.md

Lines changed: 3 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -44,20 +44,6 @@ $ make --version
4444
$ g++ --version
4545
```
4646

47-
# Models
48-
49-
* The LLaMA models are officially distributed by Facebook and will never be provided through this repository. See this [pull request in Facebook's LLaMA repository](https://github.com/facebookresearch/llama/pull/73/files) if you need to obtain access to the model data.
50-
* If your issue is with model conversion please verify the `sha256sum` of each of your `consolidated*.pth` and `ggml-model-XXX.bin` files to confirm that you have the correct model data files before logging an issue. [Latest sha256 sums for your reference](https://github.com/ggerganov/llama.cpp/issues/238).
51-
* If your issue is with model generation quality then please at least scan the following links and papers to understand the limitations of LLaMA models. This is especially important when choosing an appropriate model size and appreciating both the significant and subtle differences between LLaMA models and ChatGPT:
52-
* LLaMA:
53-
* [Introducing LLaMA: A foundational, 65-billion-parameter large language model](https://ai.facebook.com/blog/large-language-model-llama-meta-ai/)
54-
* [LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)
55-
* GPT-3
56-
* [Language Models are Few-Shot Learners](https://arxiv.org/abs/2005.14165)
57-
* GPT-3.5 / InstructGPT / ChatGPT:
58-
* [Aligning language models to follow instructions](https://openai.com/research/instruction-following)
59-
* [Training language models to follow instructions with human feedback](https://arxiv.org/abs/2203.02155)
60-
6147
# Failure Information (for bugs)
6248

6349
Please help provide information about the failure if this is a bug. If it is not a bug, please remove the rest of this template.
@@ -75,8 +61,9 @@ Please provide detailed steps for reproducing the issue. We are not sitting in f
7561

7662
Please include any relevant log snippets or files. If it works under one configuration but not under another, please provide logs for both configurations and their corresponding outputs so it is easy to see where behavior changes.
7763

78-
Also, please try to **avoid using screenshots** if at all possible. Instead, copy/paste the console output and use [Github's markdown](https://docs.github.com/en/get-started/writing-on-github/getting-started-with-writing-and-formatting-on-github/basic-writing-and-formatting-syntax) to cleanly format your logs for easy readability. e.g.
64+
Also, please try to **avoid using screenshots** if at all possible. Instead, copy/paste the console output and use [Github's markdown](https://docs.github.com/en/get-started/writing-on-github/getting-started-with-writing-and-formatting-on-github/basic-writing-and-formatting-syntax) to cleanly format your logs for easy readability.
7965

66+
Example environment info:
8067
```
8168
llama.cpp$ git log | head -1
8269
commit 2af23d30434a677c6416812eea52ccc0af65119c
@@ -103,8 +90,8 @@ GNU Make 4.3
10390
$ md5sum ./models/65B/ggml-model-q4_0.bin
10491
dbdd682cce80e2d6e93cefc7449df487 ./models/65B/ggml-model-q4_0.bin
10592
```
106-
Here's a run with the Linux command [perf](https://www.brendangregg.com/perf.html)
10793

94+
Example run with the Linux command [perf](https://www.brendangregg.com/perf.html)
10895
```
10996
llama.cpp$ perf stat ./main -m ./models/65B/ggml-model-q4_0.bin -t 16 -n 1024 -p "Please close your issue when it has been answered."
11097
main: seed = 1679149377

README.md

Lines changed: 20 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -191,17 +191,8 @@ Note the use of `--color` to distinguish between user input and generated text.
191191

192192
### Instruction mode with Alpaca
193193

194-
First, download the `ggml` Alpaca model into the `./models` folder:
195-
196-
```
197-
# use one of these
198-
# TODO: add a script to simplify the download
199-
curl -o ./models/ggml-alpaca-7b-q4.bin -C - https://gateway.estuary.tech/gw/ipfs/QmUp1UGeQFDqJKvtjbSYPBiZZKRjLp8shVP9hT8ZB9Ynv1
200-
curl -o ./models/ggml-alpaca-7b-q4.bin -C - https://ipfs.io/ipfs/QmUp1UGeQFDqJKvtjbSYPBiZZKRjLp8shVP9hT8ZB9Ynv1
201-
curl -o ./models/ggml-alpaca-7b-q4.bin -C - https://cloudflare-ipfs.com/ipfs/QmUp1UGeQFDqJKvtjbSYPBiZZKRjLp8shVP9hT8ZB9Ynv1
202-
```
203-
204-
Now run the `main` tool like this:
194+
1. First, download the `ggml` Alpaca model into the `./models` folder
195+
2. Run the `main` tool like this:
205196

206197
```
207198
./main -m ./models/ggml-alpaca-7b-q4.bin --color -f ./prompts/alpaca.txt -ins
@@ -228,26 +219,34 @@ cadaver, cauliflower, cabbage (vegetable), catalpa (tree) and Cailleach.
228219

229220
### Obtaining and verifying the Facebook LLaMA original model and Stanford Alpaca model data
230221

231-
* The LLaMA models are officially distributed by Facebook and will never be provided through this repository. See this [Pull Request in Facebook's LLaMA repository](https://github.com/facebookresearch/llama/pull/73/files) if you need to obtain access to the model data.
232-
222+
* The LLaMA models are officially distributed by Facebook and will never be provided through this repository. See this [pull request in Facebook's LLaMA repository](https://github.com/facebookresearch/llama/pull/73/files) if you need to obtain access to the model data.
233223
* Please verify the sha256 checksums of all of your `consolidated*.pth` and corresponding converted `ggml-model-*.bin` model files to confirm that you have the correct model data files before creating an issue relating to your model files.
224+
* The following command will verify if you have all possible latest files in your self-installed `./models` subdirectory:
234225

235-
The following command will verify if you have all possible latest files in your self-installed `./models` subdirectory:
236-
237-
`sha256sum --ignore-missing -c SHA256SUMS` on Linux
226+
`sha256sum --ignore-missing -c SHA256SUMS` on Linux
238227

239-
or
228+
or
240229

241-
`shasum -a 256 --ignore-missing -c SHA256SUMS` on macOS
230+
`shasum -a 256 --ignore-missing -c SHA256SUMS` on macOS
242231

232+
* If your issue is with model generation quality then please at least scan the following links and papers to understand the limitations of LLaMA models. This is especially important when choosing an appropriate model size and appreciating both the significant and subtle differences between LLaMA models and ChatGPT:
233+
* LLaMA:
234+
* [Introducing LLaMA: A foundational, 65-billion-parameter large language model](https://ai.facebook.com/blog/large-language-model-llama-meta-ai/)
235+
* [LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)
236+
* GPT-3
237+
* [Language Models are Few-Shot Learners](https://arxiv.org/abs/2005.14165)
238+
* GPT-3.5 / InstructGPT / ChatGPT:
239+
* [Aligning language models to follow instructions](https://openai.com/research/instruction-following)
240+
* [Training language models to follow instructions with human feedback](https://arxiv.org/abs/2203.02155)
241+
243242
### Perplexity (Measuring model quality)
244243

245244
You can pass `--perplexity` as a command line option to measure perplexity over the given prompt. For more background,
246245
see https://huggingface.co/docs/transformers/perplexity. However, in general, lower perplexity is better for LLMs.
247246

248-
#### Measurements
247+
#### Latest measurements
249248

250-
https://github.com/ggerganov/llama.cpp/pull/270 is the unofficial tracking page for now. llama.cpp is measuring very well
249+
The latest perplexity scores for the various model sizes and quantizations are being tracked in [discussion #406](https://github.com/ggerganov/llama.cpp/discussions/406). `llama.cpp` is measuring very well
251250
compared to the baseline implementations. Quantization has a small negative impact to quality, but, as you can see, running
252251
13B at q4_0 beats the 7B f16 model by a significant amount.
253252

@@ -347,3 +346,4 @@ docker run -v /llama/models:/models ghcr.io/ggerganov/llama.cpp:light -m /models
347346
- There are no strict rules for the code style, but try to follow the patterns in the code (indentation, spaces, etc.). Vertical alignment makes things more readable and easier to batch edit
348347
- Clean-up any trailing whitespaces, use 4 spaces indentation, brackets on same line, `void * ptr`, `int & a`
349348
- See [good first issues](https://github.com/ggerganov/llama.cpp/issues?q=is%3Aissue+is%3Aopen+label%3A%22good+first+issue%22) for tasks suitable for first contributions
349+

0 commit comments

Comments
 (0)