We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hello,
Is anyone aware of the 4-bit quantized models for LLaVA-1.5 available on Hugging Face?
Thanks in advance!
The text was updated successfully, but these errors were encountered:
what are the system requirement for that? (running solely on CPU?)
Sorry, something went wrong.
https://huggingface.co/mys/ggml_llava-v1.5-7b https://huggingface.co/mys/ggml_llava-v1.5-13b
7b works better than 13b for now.
Use the llama.cpp repo and run the make command and use ./lava
Runs on CPU. More info on this PR
The bug seems to be fixed in ggml-org/llama.cpp#3645 Now the quality for both 7b and 13b are improved.
Closing this now.
No branches or pull requests
Question
Hello,
Is anyone aware of the 4-bit quantized models for LLaVA-1.5 available on Hugging Face?
Thanks in advance!
The text was updated successfully, but these errors were encountered: