Skip to content

Releases: ngxson/llama.cpp

b4424

06 Jan 09:45
47182dd
Compare
Choose a tag to compare
llama : update llama_model API names (#11063)

* llama : deprecate llama_free_model, add llama_model_free

ggml-ci

* llama : change `llama_load_model_from_file` -> `llama_model_load_from_file`

ggml-ci

b4422

06 Jan 09:45
ae2f606
Compare
Choose a tag to compare
mmap : fix fileno macro clash (#11076)

* mmap : fix fileno macro clash

ggml-ci

* cont

ggml-ci

b4419

06 Jan 02:17
46e3556
Compare
Choose a tag to compare
CUDA: add BF16 support (#11093)

* CUDA: add BF16 support

b4418

04 Jan 21:00
b56f079
Compare
Choose a tag to compare
Vulkan: Add device-specific blacklist for coopmat for the AMD proprie…

b4417

04 Jan 20:53
9394bbd
Compare
Choose a tag to compare
llama : Add support for DeepSeek V3 (#11049)

* convert : extend DEEPSEEK2 model architecture to support DeepseekV3ForCausalLM by adding EXPERT_WEIGHTS_NORM and EXPERT_GATING_FUNC model parameters and FFN_EXP_PROBS_B tensor type

* vocab : add DeepSeek V3 pre-tokenizer regexes

* unicode : handle ACCENT_MARK and SYMBOL categories in regex

* llama : add DeepSeek V3 chat template, handle new model parameters and tensor types

---------

Co-authored-by: Stanisław Szymczyk <[email protected]>

b4416

04 Jan 16:46
f922a9c
Compare
Choose a tag to compare
[GGML][RPC] Support for models with non-512-aligned tensors over RPC.…

b4415

04 Jan 15:14
46be942
Compare
Choose a tag to compare
llama : add support for the cohere2 model architecture (#10900)

b4414

04 Jan 14:47
Compare
Choose a tag to compare
sync : ggml

b4411

04 Jan 08:59
c31fc8b
Compare
Choose a tag to compare
fix: Vulkan shader gen binary path (#11037)

b4410

03 Jan 12:53
4b0c638
Compare
Choose a tag to compare
common : disable KV cache shifting automatically for unsupported mode…