-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
fix: tokenization of special characters: #850
fix: tokenization of special characters: #850
Conversation
It should behave like llama.cpp, where most out of the box usages treat special characters accordingly
@abetlen In case you missed this. |
What about removing the empty test.py file? |
@antoine-lizee looks good, I'm slightly hesistant to change the default behaviour of the |
Nope, that will not be sufficient. In my case, I'm infilling codes using |
I'll go ahead and merge this in as is for now, should have time in the next week to address any issues if this causes breaking changes. @antoine-lizee thank you for the contribution! |
It should behave like llama.cpp, where most out of the box usages treat special characters accordingly
* Add low-level batching notebook * fix: tokenization of special characters: (#850) It should behave like llama.cpp, where most out of the box usages treat special characters accordingly * Update CHANGELOG * Cleanup * Fix runner label * Update notebook * Use llama_decode and batch api * Support logits_all parameter --------- Co-authored-by: Antoine Lizee <[email protected]>
It should behave like llama.cpp, where most out of the box usages treat special characters accordingly. See #838 (comment) for more details.
I checked that with this fix, the vanilla call to
llm.create_completion(temperature=0)
leads to exactly the same results for a simple chat prompt than when using./main --temp 0
fromllama.cpp
- which it didn't before.I changed the behaviour also for the embeddings and the LlamaTokenizer. I'm missing context so might be wrong on those, but I figured it would be good to be consistent.