-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
support for llama3 in autoquant #67
Comments
Sorry for the slow response, thanks a lot for opening this issue. I saw a lot of comments about issues with the tokenization in GGUF, so I don't know if it's the right time to update AutoQuant. I like your improvements in the first notebook. Do you think I should transfer them or should I wait until the situation is fixed? |
indeed might be better to wait with regard to the pre-tokenizer. i am not completely sure i understood the procedure for new models like say llama3 merges. but my current understanding is illustrated by this updated kaggle script.
|
... would need vocab_type bpe, see here for illustration
https://colab.research.google.com/drive/1q1hTxLZOCRf9n0KdxSSu3tD0EI5QufrV?usp=sharing
(i also made a few adaptions for faster running for my use case)
thank you and keep up the great work!!
The text was updated successfully, but these errors were encountered: