Fix prefix space issues with certain tokenizers #156
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This adds a heuristic to catch tokenizers (e.g., Llama) which treats a word starting a string has having a prefixed space, where previously the extra space added at the start of "continuations" in ranked classification task leads to a spurious space token (so for instance total probability mass over answer choices A/B/C/D in MC tasks drops from around 1 to near zero).