-
Notifications
You must be signed in to change notification settings - Fork 10.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
changelog : libllama
API
#9289
Comments
* ggml : add support for dynamic loading of backends --------- Co-authored-by: Georgi Gerganov <[email protected]>
Looks like |
I didn't expect that this function is being used by anyone, so I skipped updating the changelog. It's updated now. Btw, what do you use this call for? |
I don't use it personally but the function was included in my Python code, I started to get |
Overview
This is a list of changes to the public interface of the
llama
library. Collaborators are encouraged to edit this post in order to reflect important changes to the API that end up merged into themaster
branch.If you are building a 3rd party project that relies on
libllama
, it is recommended to follow this issue and check it before upgrading to new versions.See also:
llama-server
REST APIRecent API changes (most recent at the top)
llama_model
API namingllama_model_get_tensor()
llama_sampler_init_penalties()
Q4_0_N_M
model files in favor of automatic repacking ofQ4_0
devices
tollama_model_params
softmax
sampler and updatedist
sampler`all_pos_0, all_pos_1, all_seq_id
fromllama_batch
LLAMA_POOLING_TYPE_RANK
llama_n_head()
llama_perf
API + param to disable internal profilingllama_sampler_chain_remove()
LLAMA_VOCAB_TYPE_RWKV
enum valuellama_threadpool
API + changeuint32_t
->int32_t
llama_model_is_recurrent
For older changes, use:
(For collaborators) To link between PR number vs Build number:
Upcoming API changes
The text was updated successfully, but these errors were encountered: