You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Move model section from issue template to README.md (#421)
* Update custom.md
* Removed Model section as it is better placed in README.md
* Updates to README.md model section
* Inserted text that was removed from issue template about obtaining models from FB and links to papers describing the various models
* Removed IPF down links for the Alpaca 7B models as these look to be in the old data format and probably shouldn't be directly linked to, anyway
* Updated the perplexity section to point at Perplexity scores #406 discussion
Copy file name to clipboardExpand all lines: .github/ISSUE_TEMPLATE/custom.md
+3-16Lines changed: 3 additions & 16 deletions
Original file line number
Diff line number
Diff line change
@@ -44,20 +44,6 @@ $ make --version
44
44
$ g++ --version
45
45
```
46
46
47
-
# Models
48
-
49
-
* The LLaMA models are officially distributed by Facebook and will never be provided through this repository. See this [pull request in Facebook's LLaMA repository](https://github.com/facebookresearch/llama/pull/73/files) if you need to obtain access to the model data.
50
-
* If your issue is with model conversion please verify the `sha256sum` of each of your `consolidated*.pth` and `ggml-model-XXX.bin` files to confirm that you have the correct model data files before logging an issue. [Latest sha256 sums for your reference](https://github.com/ggerganov/llama.cpp/issues/238).
51
-
* If your issue is with model generation quality then please at least scan the following links and papers to understand the limitations of LLaMA models. This is especially important when choosing an appropriate model size and appreciating both the significant and subtle differences between LLaMA models and ChatGPT:
52
-
* LLaMA:
53
-
*[Introducing LLaMA: A foundational, 65-billion-parameter large language model](https://ai.facebook.com/blog/large-language-model-llama-meta-ai/)
54
-
*[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)
55
-
* GPT-3
56
-
*[Language Models are Few-Shot Learners](https://arxiv.org/abs/2005.14165)
57
-
* GPT-3.5 / InstructGPT / ChatGPT:
58
-
*[Aligning language models to follow instructions](https://openai.com/research/instruction-following)
59
-
*[Training language models to follow instructions with human feedback](https://arxiv.org/abs/2203.02155)
60
-
61
47
# Failure Information (for bugs)
62
48
63
49
Please help provide information about the failure if this is a bug. If it is not a bug, please remove the rest of this template.
@@ -75,8 +61,9 @@ Please provide detailed steps for reproducing the issue. We are not sitting in f
75
61
76
62
Please include any relevant log snippets or files. If it works under one configuration but not under another, please provide logs for both configurations and their corresponding outputs so it is easy to see where behavior changes.
77
63
78
-
Also, please try to **avoid using screenshots** if at all possible. Instead, copy/paste the console output and use [Github's markdown](https://docs.github.com/en/get-started/writing-on-github/getting-started-with-writing-and-formatting-on-github/basic-writing-and-formatting-syntax) to cleanly format your logs for easy readability. e.g.
64
+
Also, please try to **avoid using screenshots** if at all possible. Instead, copy/paste the console output and use [Github's markdown](https://docs.github.com/en/get-started/writing-on-github/getting-started-with-writing-and-formatting-on-github/basic-writing-and-formatting-syntax) to cleanly format your logs for easy readability.
### Obtaining and verifying the Facebook LLaMA original model and Stanford Alpaca model data
230
221
231
-
* The LLaMA models are officially distributed by Facebook and will never be provided through this repository. See this [Pull Request in Facebook's LLaMA repository](https://github.com/facebookresearch/llama/pull/73/files) if you need to obtain access to the model data.
232
-
222
+
* The LLaMA models are officially distributed by Facebook and will never be provided through this repository. See this [pull request in Facebook's LLaMA repository](https://github.com/facebookresearch/llama/pull/73/files) if you need to obtain access to the model data.
233
223
* Please verify the sha256 checksums of all of your `consolidated*.pth` and corresponding converted `ggml-model-*.bin` model files to confirm that you have the correct model data files before creating an issue relating to your model files.
224
+
* The following command will verify if you have all possible latest files in your self-installed `./models` subdirectory:
234
225
235
-
The following command will verify if you have all possible latest files in your self-installed `./models` subdirectory:
236
-
237
-
`sha256sum --ignore-missing -c SHA256SUMS` on Linux
226
+
`sha256sum --ignore-missing -c SHA256SUMS` on Linux
238
227
239
-
or
228
+
or
240
229
241
-
`shasum -a 256 --ignore-missing -c SHA256SUMS` on macOS
230
+
`shasum -a 256 --ignore-missing -c SHA256SUMS` on macOS
242
231
232
+
* If your issue is with model generation quality then please at least scan the following links and papers to understand the limitations of LLaMA models. This is especially important when choosing an appropriate model size and appreciating both the significant and subtle differences between LLaMA models and ChatGPT:
233
+
* LLaMA:
234
+
*[Introducing LLaMA: A foundational, 65-billion-parameter large language model](https://ai.facebook.com/blog/large-language-model-llama-meta-ai/)
235
+
*[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)
236
+
* GPT-3
237
+
*[Language Models are Few-Shot Learners](https://arxiv.org/abs/2005.14165)
238
+
* GPT-3.5 / InstructGPT / ChatGPT:
239
+
*[Aligning language models to follow instructions](https://openai.com/research/instruction-following)
240
+
*[Training language models to follow instructions with human feedback](https://arxiv.org/abs/2203.02155)
241
+
243
242
### Perplexity (Measuring model quality)
244
243
245
244
You can pass `--perplexity` as a command line option to measure perplexity over the given prompt. For more background,
246
245
see https://huggingface.co/docs/transformers/perplexity. However, in general, lower perplexity is better for LLMs.
247
246
248
-
#### Measurements
247
+
#### Latest measurements
249
248
250
-
https://github.com/ggerganov/llama.cpp/pull/270 is the unofficial tracking page for now. llama.cpp is measuring very well
249
+
The latest perplexity scores for the various model sizes and quantizations are being tracked in [discussion #406](https://github.com/ggerganov/llama.cpp/discussions/406). `llama.cpp` is measuring very well
251
250
compared to the baseline implementations. Quantization has a small negative impact to quality, but, as you can see, running
252
251
13B at q4_0 beats the 7B f16 model by a significant amount.
- There are no strict rules for the code style, but try to follow the patterns in the code (indentation, spaces, etc.). Vertical alignment makes things more readable and easier to batch edit
348
347
- Clean-up any trailing whitespaces, use 4 spaces indentation, brackets on same line, `void * ptr`, `int & a`
349
348
- See [good first issues](https://github.com/ggerganov/llama.cpp/issues?q=is%3Aissue+is%3Aopen+label%3A%22good+first+issue%22) for tasks suitable for first contributions
0 commit comments