Bug: llama-parallel crashes when adding more tokens to llama_batch than context size #9667
Labels
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
Uh oh!
There was an error while loading. Please reload this page.
What happened?
Observation
llama-parallel
crashes with aSegmentation fault
when the number of tokens added to a batch exceeds the context size.Command to reproduce:
Explanation of error:
llama_batch
withn_ctx
size. If ourn-parallel
parameter is large enough (and/or the input prompts are long enough) to exceed then_ctx
, we observe aSegmentation fault
when attempting to callllama_batch_add
.Expected behavior:
1
with a clear explanation of the error (e.g. number of tokens being processed within a batch exceeds context size), and a recommendation of what to do (e.g. increase context size).Name and Version
What operating system are you seeing the problem on?
Linux
Relevant log output
The text was updated successfully, but these errors were encountered: