-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add KV override field for llama.cpp loaders #4925
base: dev
Are you sure you want to change the base?
Conversation
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch (oobabooga#5257)
Does it work with updated llama-cpp-python 0.2.29? Could you merge the dev branch? |
On looking into this and latest llama-cpp-python changes I see other parameters that may be interesting to include in webui (eg. split_mode to adjust layer allocation with multi GPU). Perhaps a more generic way to pass in any experimental parameters directly to |
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
Merge dev branch
This sounds good, it would solve the 'change number of Mixtral experts' issue raised here: |
Changed to build a dict from the gradio field so this works with current llama-cpp-python. |
Adds a new text field for llama.cpp and llamacpp_HF loaders to implement llama.cpp's
--override-kv
.depends abetlen/llama-cpp-python#1011