-
Notifications
You must be signed in to change notification settings - Fork 12.5k
Description
Prerequisites
Please answer the following questions for yourself before submitting an issue.
- I am running the latest code. Development is very rapid so there are no tagged versions as of now.
- I carefully followed the README.md.
- I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed).
- I reviewed the Discussions, and have a new bug or useful enhancement to share.
Feature Description
The LLaMA 3 8b Scoreboard on the following link was computed against fp16. https://github.com/ggerganov/llama.cpp/tree/master/examples/perplexity
However, the model was released as bf16 weights. Is there a quantifiable negative impact on perplexity due to conversion between weight formats? Or a difference when compating perplexity against bf16 instead of fp16? It's unclear. Even a brief mention of this could bring clarity.
Motivation
Curiosity about the impact of bf16 versus fp16 on models, and subsequent training/merging.
Possible Implementation
If you have an idea as to how it can be implemented, please write a detailed description. Feel free to give links to external sources or share visuals that might be helpful to understand the details better.