Logo
Explore Help
Sign In
CS348Project/llama.cpp
5
0
Fork 0
You've already forked llama.cpp
mirror of https://github.com/ggml-org/llama.cpp.git synced 2025-11-01 09:01:57 +00:00
Code Issues Packages Projects Releases Wiki Activity
Files
74b8fc17f92ada295a648e3c5eb28f46bca7d892
llama.cpp/tools
History
Georgi Gerganov df1b612e29 server : add /v1/health endpoint (#16461)
* server : add /v1/health endpoint

* cont : update readme
2025-10-07 15:57:14 +03:00
..
batched-bench
…
cvector-generator
…
export-lora
…
gguf-split
ci : use smaller model (#16168)
2025-09-22 09:11:39 +03:00
imatrix
cmake : Do not install tools on iOS targets (#15903)
2025-09-16 09:54:44 +07:00
llama-bench
llama : add --no-host to disable host buffers (#16310)
2025-10-06 19:55:53 +02:00
main
llama-cli: prevent spurious assistant token (#16202)
2025-09-29 10:03:12 +03:00
mtmd
chat : Granite Docling stopping (#16438)
2025-10-06 18:59:40 +02:00
perplexity
perplexity : show more kl-divergence data (#16321)
2025-09-29 09:30:45 +03:00
quantize
ci : use smaller model (#16168)
2025-09-22 09:11:39 +03:00
rpc
rpc : update documentation (#16441)
2025-10-07 06:59:13 +00:00
run
common: introduce http.h for httplib-based client (#16373)
2025-10-01 20:22:18 +03:00
server
server : add /v1/health endpoint (#16461)
2025-10-07 15:57:14 +03:00
tokenize
…
tts
model : Apertus model implementation (#15852)
2025-10-02 20:43:22 +03:00
CMakeLists.txt
…
Powered by Gitea Version: 1.25.0 Page: 3055ms Template: 591ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API