Skip to content

Releases: ngxson/llama.cpp

b2251

23 Feb 20:42
fd43d66
Compare
Choose a tag to compare
server : add KV cache quantization options (#5684)

b2239

22 Feb 15:43
3a03541
Compare
Choose a tag to compare
minor : fix trailing whitespace (#5638)

b2220

21 Feb 09:47
6560bed
Compare
Choose a tag to compare
server : support llava 1.6 (#5553)

* server: init working 1.6

* move clip_image to header

* remove commented code

* remove c++ style from header

* remove todo

* expose llava_image_embed_make_with_clip_img

* fix zig build

b2203

19 Feb 15:09
9d679f0
Compare
Choose a tag to compare
examples : support minItems/maxItems in JSON grammar converter (#5039)

* support minLength and maxLength in JSON schema grammar converter

* Update examples/json-schema-to-grammar.py

---------

Co-authored-by: Georgi Gerganov <[email protected]>

b2168

17 Feb 19:47
d250c9d
Compare
Choose a tag to compare
gitignore : update for CLion IDE (#5544)

b2104

08 Feb 21:11
41f308f
Compare
Choose a tag to compare
llama : do not print "offloading layers" message in CPU-only builds (…

b2038

01 Feb 14:45
ce32060
Compare
Choose a tag to compare
llama : support InternLM2 (#5184)

* support InternLM2 inference
  * add add_space_prefix KV pair

b2034

31 Jan 14:58
efb7bdb
Compare
Choose a tag to compare
metal : add im2col F32 dst support (#5132)

b2026

30 Jan 22:27
e0085fd
Compare
Choose a tag to compare
Revert "server : change deps.sh xxd files to string literals (#5221)"

This reverts commit 4003be0e5feef320f3707786f22722b73cff9356.

b2002

29 Jan 10:11
e76627b
Compare
Choose a tag to compare
py : improve BPE tokenizer support (#5189)