Description
I first encountered this problem after upgrading to the latest llamaccp in silly tavern. It would generate gibberish no matter what model or settings I used, including models that used to work (like mistral based models). It was confusing because the models generate normally in kobold lite. I thought it was a SillyTavern problem.
Then I pasted a purely coherent text long chat into the prompt in kobold lite (over 6k tokens), and it gave gibberish like "(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr(rr" or meaningless strings of characters and words just like the bug in SillyTavern, and likewise if I deleted the character card, authors note, and strong string from SillyTavern it would generate coherently and normally.
So this appears to be llamacpp and something to do with a fuller or longer context.
I'm using vulkan with fairly standard settings on windows 11. Doesn't matter what kind of base model I use, or any other settings. Basically everything that previously worked before I updated, no longer works (unless I trim the context down to nothing, but that makes it pretty useless). I use a max context size of 6144 if that matters, so it's never larger than that.