-
Notifications
You must be signed in to change notification settings - Fork 364
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llama 3.1 update binaries #874
Llama 3.1 update binaries #874
Conversation
…hen referring to the C++ - Exposed `KvCacheMaxPosition` on `SafeLLamaContextHandle`
Tests passed on Linux CUDA. Not sure if we need to test Linux/MacOS CPU as this should be covered by CI. |
Does the MacOS CI runner have metal or is it CPU only? |
Only CPU. The metal emulation fails. |
It works on MacOS with metal |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tested
Excellent, thanks for testing that @SignalRT. I'll try to run through the release process tonight or tomorrow. |
Resolved conflicts, will merge once CI completes. |
Update to new binaries for llama.cpp 345c8c0c87a97c1595f9c8b14833d531c8c7d8df. Built with this build action. This should include the RoPE fixes for
llama 3.1
.Testing: