-
Notifications
You must be signed in to change notification settings - Fork 12k
scripts: fix pattern and get n_tokens in one go #10221
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
@slaren @ggerganov PTAL~ |
Can confirm that this solves the error message and exit 1 status reported in #10219 |
Thank you @lhpqaq ! |
@reinvantveer thanks for your review |
I'm not sure but I suspect there's some kind of regression though. The script appears to read input where I haven't given any |
A quote from the "interactive storytelling" I'm trying out with the model
|
I'm unsure on where these extra |
This is from llama-cli (main. cpp), this PR only modifies regular pattern matching |
I see, I switched back to the
... [etc] |
After each conversation, the script infers once in the background, but only redirects the standard error to the log, which
@reinvantveer There are indeed bugs here |
Implement a todo and fix #10219