Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update llama to pickup "Avoid heavy V transpose operation + improvements" #11

Closed
deadprogram opened this issue Apr 10, 2023 · 2 comments

Comments

@deadprogram
Copy link
Contributor

Please see ggml-org/llama.cpp#775

I added issue in this repo since the https://github.com/go-skynet/llama repo doesn't have issues available.

Thanks! 😸

@deadprogram
Copy link
Contributor Author

Just discovered https://github.com/go-skynet/go-llama.cpp

Closing since all the action is there.

@mudler
Copy link
Owner

mudler commented Apr 10, 2023

Correct! Seems you already found out all the bits - I will tag soon a new release that should include that as well. Currently images tagged with latest already have the change in, however the API changed quite a bit, mimicking OpenAI api with multi model support (thus docs needs to be updated!)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants