Replies: 2 comments
-
|
Without knowing what you changed, as I recently found out PPL tests are impervious to batch_size = 1 bugs. |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
Thanks @am17an, I'll keep the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
While experimenting with GEMM kernels (CUDA), I introduced a bug. The perplexing part is that all the unit tests (
test-backend-ops) were passing, and the PPL score was exactly the same as before the change which introduced the bug, and the only way I noticed that something is wrong was playing withllama-cli.The unit tests gap is understandable, since the coverage is not 100%, and the only case that was affected was an fp16
MUL_MATwith non-contiguous tensors of "irregular" dimensions. I assumed that unit tests + PPL score would catch regressions with high confidence, but it doesn't appear to be the case.What would be a better way to catch functional regressions?
Beta Was this translation helpful? Give feedback.
All reactions