Releases: 3Simplex/llama.cpp
Releases · 3Simplex/llama.cpp
b4295
b4288
llama : use cmake for swift build (#10525) * llama : use cmake for swift build * swift : <> -> "" * ci : remove make * ci : disable ios build * Revert "swift : <> -> """ This reverts commit d39ffd9556482b77d4ea5b118b453fc1c097a31d. * ci : try fix ios build * ci : cont * ci : cont --------- Co-authored-by: Georgi Gerganov <[email protected]>
b4271
sync : ggml
b4265
ggml : add predefined list of CPU backend variants to build (#10626) * ggml : add predefined list of CPU backend variants to build * update CPU dockerfiles
b4254
server : (web ui) Various improvements, now use vite as bundler (#10599) * hide buttons in dropdown menu * use npm as deps manager and vite as bundler * fix build * fix build (2) * fix responsive on mobile * fix more problems on mobile * sync build * (test) add CI step for verifying build * fix ci * force rebuild .hpp files * cmake: clean up generated files pre build
b4248
llama : add missing LLAMA_API for llama_chat_builtin_templates (#10636)
b4164
server : add speculative decoding support (#10455) * server : add speculative decoding support ggml-ci * server : add helper function slot.can_speculate() ggml-ci
b4153
ci: Update oneAPI runtime dll packaging (#10428) This is the minimum runtime dll dependencies for oneAPI 2025.0
b4145
vulkan: predicate max operation in soft_max shaders/soft_max (#10437) Fixes #10434
b4132
cuda : fix CUDA_FLAGS not being applied (#10403)