Skip to content

Releases: 3Simplex/llama.cpp

b4295

10 Dec 15:06
26a8406
Compare
Choose a tag to compare
CUDA: fix shared memory access condition for mmv (#10740)

b4288

08 Dec 17:16
43ed389
Compare
Choose a tag to compare
llama : use cmake for swift build (#10525)

* llama : use cmake for swift build

* swift : <> -> ""

* ci : remove make

* ci : disable ios build

* Revert "swift : <> -> """

This reverts commit d39ffd9556482b77d4ea5b118b453fc1c097a31d.

* ci : try fix ios build

* ci : cont

* ci : cont

---------

Co-authored-by: Georgi Gerganov <[email protected]>

b4271

05 Dec 17:44
0cd182e
Compare
Choose a tag to compare
sync : ggml

b4265

04 Dec 14:37
59f4db1
Compare
Choose a tag to compare
ggml : add predefined list of CPU backend variants to build (#10626)

* ggml : add predefined list of CPU backend variants to build

* update CPU dockerfiles

b4254

03 Dec 19:41
91c36c2
Compare
Choose a tag to compare
server : (web ui) Various improvements, now use vite as bundler (#10599)

* hide buttons in dropdown menu

* use npm as deps manager and vite as bundler

* fix build

* fix build (2)

* fix responsive on mobile

* fix more problems on mobile

* sync build

* (test) add CI step for verifying build

* fix ci

* force rebuild .hpp files

* cmake: clean up generated files pre build

b4248

03 Dec 16:27
3b4f2e3
Compare
Choose a tag to compare
llama : add missing LLAMA_API for llama_chat_builtin_templates (#10636)

b4164

25 Nov 16:58
9ca2e67
Compare
Choose a tag to compare
server : add speculative decoding support (#10455)

* server : add speculative decoding support

ggml-ci

* server : add helper function slot.can_speculate()

ggml-ci

b4153

22 Nov 14:32
6dfcfef
Compare
Choose a tag to compare
ci: Update oneAPI runtime dll packaging (#10428)

This is the minimum runtime dll dependencies for oneAPI 2025.0

b4145

20 Nov 20:51
9abe9ee
Compare
Choose a tag to compare
vulkan: predicate max operation in soft_max shaders/soft_max (#10437)

Fixes #10434

b4132

19 Nov 15:20
3ee6382
Compare
Choose a tag to compare
cuda : fix CUDA_FLAGS not being applied (#10403)