Releases: ggerganov/llama.cpp
Releases · ggerganov/llama.cpp
b4426
b4425
llama : rename missed batch params/vars to ubatch (#10059) This commit renames the `batch` parameter to `ubatch` in the `llama_kv_cache_find_slot`, `llm_build_inp_embd`, and `llm_build_mamba` functions. The motivation for this is that this should have been done as part of Commit 19d900a7565b8f6b0a708836a57d26966cb9efe2 ("llama : rename batch to ubatch (#9950)") but for some reason I missed these functions in that commit and only noticed them now (sorry).
b4424
llama : update llama_model API names (#11063) * llama : deprecate llama_free_model, add llama_model_free ggml-ci * llama : change `llama_load_model_from_file` -> `llama_model_load_from_file` ggml-ci
b4423
tokenize : escape the prompt (#11058) * tokenize : escape the prompt * tokenize : update help
b4422
mmap : fix fileno macro clash (#11076) * mmap : fix fileno macro clash ggml-ci * cont ggml-ci
b4421
llama : use LLAMA_TOKEN_NULL (#11062) ggml-ci
b4420
llama : use _impl suffix instead of _internal (#11060) ggml-ci
b4419
CUDA: add BF16 support (#11093) * CUDA: add BF16 support
b4418
Vulkan: Add device-specific blacklist for coopmat for the AMD proprie…
b4417
llama : Add support for DeepSeek V3 (#11049) * convert : extend DEEPSEEK2 model architecture to support DeepseekV3ForCausalLM by adding EXPERT_WEIGHTS_NORM and EXPERT_GATING_FUNC model parameters and FFN_EXP_PROBS_B tensor type * vocab : add DeepSeek V3 pre-tokenizer regexes * unicode : handle ACCENT_MARK and SYMBOL categories in regex * llama : add DeepSeek V3 chat template, handle new model parameters and tensor types --------- Co-authored-by: Stanisław Szymczyk <[email protected]>