Skip to content

Releases: ggerganov/llama.cpp

b4426

06 Jan 13:05
96a1dc2
Compare
Choose a tag to compare
llama : prevent system info string accumulation across calls (#11101)

b4425

06 Jan 12:51
6369f86
Compare
Choose a tag to compare
llama : rename missed batch params/vars to ubatch (#10059)

This commit renames the `batch` parameter to `ubatch` in the
`llama_kv_cache_find_slot`, `llm_build_inp_embd`, and
`llm_build_mamba` functions.

The motivation for this is that this should have been done as part of
Commit 19d900a7565b8f6b0a708836a57d26966cb9efe2 ("llama : rename batch
to ubatch (#9950)") but for some reason I missed these functions in
that commit and only noticed them now (sorry).

b4424

06 Jan 12:49
47182dd
Compare
Choose a tag to compare
llama : update llama_model API names (#11063)

* llama : deprecate llama_free_model, add llama_model_free

ggml-ci

* llama : change `llama_load_model_from_file` -> `llama_model_load_from_file`

ggml-ci

b4423

06 Jan 12:22
3e6e7a6
Compare
Choose a tag to compare
tokenize : escape the prompt (#11058)

* tokenize : escape the prompt

* tokenize : update help

b4422

06 Jan 12:02
ae2f606
Compare
Choose a tag to compare
mmap : fix fileno macro clash (#11076)

* mmap : fix fileno macro clash

ggml-ci

* cont

ggml-ci

b4421

06 Jan 12:00
727368c
Compare
Choose a tag to compare
llama : use LLAMA_TOKEN_NULL (#11062)

ggml-ci

b4420

06 Jan 12:00
5047dd3
Compare
Choose a tag to compare
llama : use _impl suffix instead of _internal (#11060)

ggml-ci

b4419

06 Jan 02:18
46e3556
Compare
Choose a tag to compare
CUDA: add BF16 support (#11093)

* CUDA: add BF16 support

b4418

04 Jan 20:57
b56f079
Compare
Choose a tag to compare
Vulkan: Add device-specific blacklist for coopmat for the AMD proprie…

b4417

04 Jan 20:50
9394bbd
Compare
Choose a tag to compare
llama : Add support for DeepSeek V3 (#11049)

* convert : extend DEEPSEEK2 model architecture to support DeepseekV3ForCausalLM by adding EXPERT_WEIGHTS_NORM and EXPERT_GATING_FUNC model parameters and FFN_EXP_PROBS_B tensor type

* vocab : add DeepSeek V3 pre-tokenizer regexes

* unicode : handle ACCENT_MARK and SYMBOL categories in regex

* llama : add DeepSeek V3 chat template, handle new model parameters and tensor types

---------

Co-authored-by: Stanisław Szymczyk <[email protected]>