-
Notifications
You must be signed in to change notification settings - Fork 401
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reprocessing Issue with Llama 3 #803
Comments
Did you by any chance enable "Trim Sentences" or "Author Note"? |
No, I use default settings without trimming. So, you can't reproduce it? |
Yes, I can reproduce it. Looking closer, the tokenizer is behaving weirdly. I think there is an issue with token merges. Relevant: ggerganov#6809 You should experience a small amount of reprocessing all the way back to the previous newline. This is a bug. |
Hi, Should be fixed in the latest version. Remember to get freshly reconverted GGUFs |
@LostRuins Thanks! Yes, it looks like it’s working now. Thank you for continuing to maintain this project, you’re awesome! |
When using Llama 3, I've noticed that unnecessary reprocessing occurs on previously generated text.
To reproduce this issue, try generating a short piece of text couple of times and see how the processing sometimes happens.
Latest concedo_experimental.
The text was updated successfully, but these errors were encountered: