Eric Curtin 
							
						 
					 
					
						
						
							
						
						a4417ddda9 
					 
					
						
						
							
							Add new hf protocol for ollama ( #11449 )  
						
						... 
						
						
						
						https://huggingface.co/docs/hub/en/ollama 
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
					
						2025-01-27 19:36:10 +01:00 
						 
				 
			
				
					
						
							
							
								bandoti 
							
						 
					 
					
						
						
							
						
						19f65187cb 
					 
					
						
						
							
							cmake: add ggml find package ( #11369 )  
						
						... 
						
						
						
						* Add initial ggml cmake package
* Add build numbers to ggml find-package
* Expand variables with GGML_ prefix
* Guard against adding to cache variable twice
* Add git to msys2 workflow
* Handle ggml-cpu-* variants
* Link ggml/ggml-base libraries to their targets
* Replace main-cmake-pkg with simple-cmake-pkg
* Interface features require c_std_90
* Fix typo
* Removed unnecessary bracket from status message
* Update examples/simple-cmake-pkg/README.md
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com >
* Update examples/simple-cmake-pkg/README.md
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com >
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com > 
						
						
					 
					
						2025-01-26 12:07:48 -04:00 
						 
				 
			
				
					
						
							
							
								Xuan Son Nguyen 
							
						 
					 
					
						
						
							
						
						49b0e3cec4 
					 
					
						
						
							
							server : fix cleaning up stream task ( #11418 )  
						
						... 
						
						
						
						* server : fix cleaning up stream task
* one more spot 
						
						
					 
					
						2025-01-25 16:36:44 +01:00 
						 
				 
			
				
					
						
							
							
								Eric Curtin 
							
						 
					 
					
						
						
							
						
						01f37edf1a 
					 
					
						
						
							
							Update llama-run README.md ( #11386 )  
						
						... 
						
						
						
						For consistency
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
						
						
					 
					
						2025-01-24 09:39:24 +00:00 
						 
				 
			
				
					
						
							
							
								stduhpf 
							
						 
					 
					
						
						
							
						
						c07e87f38b 
					 
					
						
						
							
							server : (webui) put DeepSeek R1 CoT in a collapsible <details> element ( #11364 )  
						
						... 
						
						
						
						* webui : put DeepSeek R1 CoT in a collapsible <details> element
* webui: refactor split
* webui: don't use regex to split cot and response
* webui: format+qol
* webui: no loading icon if the model isn't generating
* ui fix, add configs
* add jsdoc types
* only filter </think> for assistant msg
* build
* update build
---------
Co-authored-by: Xuan Son Nguyen <son@huggingface.co > 
						
						
					 
					
						2025-01-24 09:02:38 +01:00 
						 
				 
			
				
					
						
							
							
								Eric Curtin 
							
						 
					 
					
						
						
							
						
						05f63cc9ee 
					 
					
						
						
							
							Update documentation ( #11373 )  
						
						... 
						
						
						
						To show -n, -ngl, --ngl is acceptable.
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
						
						
					 
					
						2025-01-23 20:04:31 +00:00 
						 
				 
			
				
					
						
							
							
								Eric Curtin 
							
						 
					 
					
						
						
							
						
						f7fb43cd0b 
					 
					
						
						
							
							Add -ngl ( #11372 )  
						
						... 
						
						
						
						Most other llama.cpp cli tools accept -ngl with a single dash.
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
						
						
					 
					
						2025-01-23 16:16:18 +00:00 
						 
				 
			
				
					
						
							
							
								Xuan Son Nguyen 
							
						 
					 
					
						
						
							
						
						5845661640 
					 
					
						
						
							
							server : add more clean up when cancel_tasks is called ( #11340 )  
						
						... 
						
						
						
						* server : add more clean up when cancel_tasks is called
* fix recv_with_timeout
* std::remove_if
* fix std::remove_if 
						
						
					 
					
						2025-01-23 13:56:05 +01:00 
						 
				 
			
				
					
						
							
							
								Eric Curtin 
							
						 
					 
					
						
						
							
						
						f211d1dc10 
					 
					
						
						
							
							Treat hf.co/ prefix the same as hf:// ( #11350 )  
						
						... 
						
						
						
						ollama uses hf.co/ to specify huggingface prefix, like RamaLama
uses hf://
Treat them similarly.
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
						
						
					 
					
						2025-01-23 10:38:20 +00:00 
						 
				 
			
				
					
						
							
							
								Diego Devesa 
							
						 
					 
					
						
						
							
						
						6152129d05 
					 
					
						
						
							
							main : update README documentation for batch size ( #11353 )  
						
						... 
						
						
						
						* main : update README documentation for batch size
* fix formatting
* minor 
						
						
					 
					
						2025-01-22 19:22:20 +01:00 
						 
				 
			
				
					
						
							
							
								Diego Devesa 
							
						 
					 
					
						
						
							
						
						12c2bdf2de 
					 
					
						
						
							
							server : fix draft context not being released ( #11354 )  
						
						
						
						
					 
					
						2025-01-22 17:44:40 +01:00 
						 
				 
			
				
					
						
							
							
								Jiří Podivín 
							
						 
					 
					
						
						
							
						
						96f4053934 
					 
					
						
						
							
							Adding logprobs to /v1/completions ( #11344 )  
						
						... 
						
						
						
						Signed-off-by: Jiri Podivin <jpodivin@redhat.com > 
						
						
					 
					
						2025-01-22 12:51:32 +01:00 
						 
				 
			
				
					
						
							
							
								tc-mb 
							
						 
					 
					
						
						
							
						
						3e3357fd77 
					 
					
						
						
							
							llava : support Minicpm-omni ( #11289 )  
						
						... 
						
						
						
						* init
* add readme
* update readme
* no use make
* update readme
* update fix code
* fix editorconfig-checker
* no change convert py
* use clip_image_u8_free 
						
						
					 
					
						2025-01-22 09:35:48 +02:00 
						 
				 
			
				
					
						
							
							
								Olivier Chafik 
							
						 
					 
					
						
						
							
						
						6171c9d258 
					 
					
						
						
							
							Add Jinja template support ( #11016 )  
						
						... 
						
						
						
						* Copy minja from 58f0ca6dd7https://github.com/google/minja/pull/22 )
* Apply suggestions from code review
Co-authored-by: Xuan Son Nguyen <thichthat@gmail.com >
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com >
* Finish suggested renamings
* Move chat_templates inside server_context + remove mutex
* Update --chat-template-file w/ recent change to --chat-template
* Refactor chat template validation
* Guard against missing eos/bos tokens (null token otherwise throws in llama_vocab::impl::token_get_attr)
* Warn against missing eos / bos tokens when jinja template references them
* rename: common_chat_template[s]
* reinstate assert on chat_templates.template_default
* Update minja to b8437df626https://github.com/google/minja/pull/25 
* Update minja from https://github.com/google/minja/pull/27 
* rm unused optional header
---------
Co-authored-by: Xuan Son Nguyen <thichthat@gmail.com >
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com > 
						
						
					 
					
						2025-01-21 13:18:51 +00:00 
						 
				 
			
				
					
						
							
							
								Xuan Son Nguyen 
							
						 
					 
					
						
						
							
						
						e28245f35f 
					 
					
						
						
							
							export-lora : fix tok_embd tensor ( #11330 )  
						
						
						
						
					 
					
						2025-01-21 14:07:12 +01:00 
						 
				 
			
				
					
						
							
							
								Eric Curtin 
							
						 
					 
					
						
						
							
						
						2e2f8f093c 
					 
					
						
						
							
							linenoise.cpp refactoring ( #11301 )  
						
						... 
						
						
						
						More RAII mainly
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
						
						
					 
					
						2025-01-21 09:32:35 +00:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						80d0d6b4b7 
					 
					
						
						
							
							common : add -hfd option for the draft model ( #11318 )  
						
						... 
						
						
						
						* common : add -hfd option for the draft model
* cont : fix env var
* cont : more fixes 
						
						
					 
					
						2025-01-20 22:29:43 +02:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						9f7add1cde 
					 
					
						
						
							
							examples : fix add_special conditions ( #11311 )  
						
						
						
						
					 
					
						2025-01-20 16:36:08 +02:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						92bc493917 
					 
					
						
						
							
							tests : increase timeout when sanitizers are enabled ( #11300 )  
						
						... 
						
						
						
						* tests : increase timeout when sanitizers are enabled
* tests : add DEFAULT_HTTP_TIMEOUT 
						
						
					 
					
						2025-01-19 20:22:30 +02:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						b9daaffe02 
					 
					
						
						
							
							simple-chat : fix BOS being added to each message ( #11278 )  
						
						
						
						
					 
					
						2025-01-19 18:12:09 +02:00 
						 
				 
			
				
					
						
							
							
								Eric Curtin 
							
						 
					 
					
						
						
							
						
						a1649cc13f 
					 
					
						
						
							
							Adding linenoise.cpp to llama-run ( #11252 )  
						
						... 
						
						
						
						This is a fork of linenoise that is C++17 compatible. I intend on
adding it to llama-run so we can do things like traverse prompt
history via the up and down arrows:
https://github.com/ericcurtin/linenoise.cpp 
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
						
						
					 
					
						2025-01-18 14:42:31 +00:00 
						 
				 
			
				
					
						
							
							
								Xuan Son Nguyen 
							
						 
					 
					
						
						
							
						
						f30f099228 
					 
					
						
						
							
							server : implement cancellable request ( #11285 )  
						
						... 
						
						
						
						* server : implement cancellable request
* fix typo
* httplib 0.18.5
* fix i underflow 
						
						
					 
					
						2025-01-18 14:12:05 +01:00 
						 
				 
			
				
					
						
							
							
								LostRuins Concedo 
							
						 
					 
					
						
						
							
						
						6390a998bf 
					 
					
						
						
							
							tts : add guide tokens support ( #11186 )  
						
						... 
						
						
						
						* Added the ability to use guide tokens for OuteTTS, greatly improving TTS recitation accuracy over long input sequences.
* applied linting suggestions, updated to latest llama_vocab changes, added a safety check, added newline to guide token start 
						
						
					 
					
						2025-01-18 12:20:57 +02:00 
						 
				 
			
				
					
						
							
							
								codezjx 
							
						 
					 
					
						
						
							
						
						3edfa7d375 
					 
					
						
						
							
							llama.android: add field formatChat to control whether to parse special tokens when send message ( #11270 )  
						
						
						
						
					 
					
						2025-01-17 14:57:56 +02:00 
						 
				 
			
				
					
						
							
							
								Radoslav Gerganov 
							
						 
					 
					
						
						
							
						
						667d72846c 
					 
					
						
						
							
							rpc : early register backend devices ( #11262 )  
						
						... 
						
						
						
						Early register RPC devices and do not propagate RPC specifics in the
llama model structures.
ref: #10609  
						
						
					 
					
						2025-01-17 10:57:09 +02:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						f11cfdfd7f 
					 
					
						
						
							
							ci : use -no-cnv in gguf-split tests ( #11254 )  
						
						... 
						
						
						
						* ci : use -no-cnv in gguf-split tests
ggml-ci
* ci : use -no-cnv in requantize tests
ggml-ci
* scripts : fix [no ci] 
						
						
					 
					
						2025-01-15 18:28:35 +02:00 
						 
				 
			
				
					
						
							
							
								Daniel Bevenius 
							
						 
					 
					
						
						
							
						
						0ccd7f3eb2 
					 
					
						
						
							
							examples : add embd_to_audio to tts-outetts.py [no ci] ( #11235 )  
						
						... 
						
						
						
						This commit contains a suggestion for adding the missing embd_to_audio
function from tts.cpp to tts-outetts.py. This introduces a depencency
numpy which I was not sure if that is acceptable or not (only PyTorch
was mentioned in referened PR).
Also the README has been updated with instructions to run the example
with llama-server and the python script.
Refs: https://github.com/ggerganov/llama.cpp/pull/10784#issuecomment-2548377734  
						
						
					 
					
						2025-01-15 05:44:38 +01:00 
						 
				 
			
				
					
						
							
							
								ebraminio 
							
						 
					 
					
						
						
							
						
						c5bf0d1bd7 
					 
					
						
						
							
							server : Improve code snippets direction between RTL text ( #11221 )  
						
						
						
						
					 
					
						2025-01-14 11:39:33 +01:00 
						 
				 
			
				
					
						
							
							
								ebraminio 
							
						 
					 
					
						
						
							
						
						504af20ee4 
					 
					
						
						
							
							server : (UI) Improve messages bubble shape in RTL ( #11220 )  
						
						... 
						
						
						
						I simply have overlooked message bubble's tail placement for RTL
text as I use the dark mode and that isn't visible there and this
fixes it. 
						
						
					 
					
						2025-01-13 20:23:31 +01:00 
						 
				 
			
				
					
						
							
							
								Xuan Son Nguyen 
							
						 
					 
					
						
						
							
						
						84a44815f7 
					 
					
						
						
							
							cli : auto activate conversation mode if chat template is available ( #11214 )  
						
						... 
						
						
						
						* cli : auto activate conversation mode if chat template is detected
* add warn on bad template
* update readme (writing with the help of chatgpt)
* update readme (2)
* do not activate -cnv for non-instruct models 
						
						
					 
					
						2025-01-13 20:18:12 +01:00 
						 
				 
			
				
					
						
							
							
								ebraminio 
							
						 
					 
					
						
						
							
						
						437e05f714 
					 
					
						
						
							
							server : (UI) Support for RTL text as models input or output ( #11208 )  
						
						
						
						
					 
					
						2025-01-13 14:46:39 +01:00 
						 
				 
			
				
					
						
							
							
								Eric Curtin 
							
						 
					 
					
						
						
							
						
						924518e2e5 
					 
					
						
						
							
							Reset color before we exit ( #11205 )  
						
						... 
						
						
						
						We don't want colors to leak post termination of llama-run.
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
						
						
					 
					
						2025-01-12 18:23:10 +00:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						afa8a9ec9b 
					 
					
						
						
							
							llama : add llama_vocab, functions -> methods, naming ( #11110 )  
						
						... 
						
						
						
						* llama : functions -> methods (#11110 )
* llama : add struct llama_vocab to the API (#11156 )
ggml-ci
* hparams : move vocab params to llama_vocab (#11159 )
ggml-ci
* vocab : more pimpl (#11165 )
ggml-ci
* vocab : minor tokenization optimizations (#11160 )
ggml-ci
Co-authored-by: Diego Devesa <slarengh@gmail.com >
* lora : update API names (#11167 )
ggml-ci
* llama : update API names to use correct prefix (#11174 )
* llama : update API names to use correct prefix
ggml-ci
* cont
ggml-ci
* cont
ggml-ci
* minor [no ci]
* vocab : llama_vocab_add_[be]os -> llama_vocab_get_add_[be]os (#11174 )
ggml-ci
* vocab : llama_vocab_n_vocab -> llama_vocab_n_tokens (#11174 )
ggml-ci
---------
Co-authored-by: Diego Devesa <slarengh@gmail.com > 
						
						
					 
					
						2025-01-12 11:32:42 +02:00 
						 
				 
			
				
					
						
							
							
								Daniel Bevenius 
							
						 
					 
					
						
						
							
						
						ba8a1f9c5b 
					 
					
						
						
							
							examples : add README.md to tts example [no ci] ( #11155 )  
						
						... 
						
						
						
						* examples : add README.md to tts example [no ci]
* squash! examples : add README.md to tts example [no ci]
Fix heading to be consistent with other examples, and add a quickstart
section to README.md.
* squash! examples : add README.md to tts example [no ci]
Fix spelling mistake. 
						
						
					 
					
						2025-01-10 13:16:16 +01:00 
						 
				 
			
				
					
						
							
							
								Daniel Bevenius 
							
						 
					 
					
						
						
							
						
						8eceb888d7 
					 
					
						
						
							
							server : add tooltips to settings and themes btn ( #11154 )  
						
						... 
						
						
						
						* server : add tooltips to settings and themes btn
This commit adds tooltips to the settings and themes buttons in the
webui. The tooltip will be displayed below the actual buttons when
hovered over.
The motivation for this change is to clarify the purpose of the themes
button.
* squash! server : add tooltips to settings and themes btn
This commit adds a tooltip to the '...' button when a chat has been
started. The tooltip is "Chat options" which think could be a good
description as the dropdown contains options to delete or download the
current chat.
* rm tooltip for 3 dots button
---------
Co-authored-by: Xuan Son Nguyen <son@huggingface.co > 
						
						
					 
					
						2025-01-09 11:28:29 +01:00 
						 
				 
			
				
					
						
							
							
								Eric Curtin 
							
						 
					 
					
						
						
							
						
						1bf839b1e8 
					 
					
						
						
							
							Enhance user input handling for llama-run ( #11138 )  
						
						... 
						
						
						
						The main motivation for this change is it was not handing
ctrl-c/ctrl-d correctly. Modify `read_user_input` to handle EOF,
"/bye" command, and empty input cases. Introduce `get_user_input`
function to manage user input loop and handle different return
cases.
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
						
						
					 
					
						2025-01-08 18:47:05 +00:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						a3c1232c3f 
					 
					
						
						
							
							arg : option to exclude arguments from specific examples ( #11136 )  
						
						... 
						
						
						
						* arg : option to exclude arguments from specific examples
ggml-ci
* readme : remove old args [no ci] 
						
						
					 
					
						2025-01-08 12:55:36 +02:00 
						 
				 
			
				
					
						
							
							
								Johannes Gäßler 
							
						 
					 
					
						
						
							
						
						53ff6b9b9f 
					 
					
						
						
							
							GGUF: C++ refactor, backend support, misc fixes ( #11030 )  
						
						... 
						
						
						
						* GGUF: C++ refactor, backend support, misc fixes
remove ggml_tensor.backend
update CODEOWNERS [no ci]
remove gguf_get_data from API
revise GGUF API data types 
						
						
					 
					
						2025-01-07 18:01:58 +01:00 
						 
				 
			
				
					
						
							
							
								Eric Curtin 
							
						 
					 
					
						
						
							
						
						dc7cef9f37 
					 
					
						
						
							
							llama-run : fix context size ( #11094 )  
						
						... 
						
						
						
						Set `n_ctx` equal to `n_batch` in `Opt` class. Now context size is
a more reasonable 2048.
Signed-off-by: Eric Curtin <ecurtin@redhat.com > 
						
						
					 
					
						2025-01-06 23:45:28 +01:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						e6e7c75d94 
					 
					
						
						
							
							server : fix extra BOS in infill endpoint ( #11106 )  
						
						... 
						
						
						
						* server : fix extra BOS in infill endpoing
ggml-ci
* server : update infill tests 
						
						
					 
					
						2025-01-06 15:36:08 +02:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						47182dd03f 
					 
					
						
						
							
							llama : update llama_model API names ( #11063 )  
						
						... 
						
						
						
						* llama : deprecate llama_free_model, add llama_model_free
ggml-ci
* llama : change `llama_load_model_from_file` -> `llama_model_load_from_file`
ggml-ci 
						
						
					 
					
						2025-01-06 10:55:18 +02:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						3e6e7a6bc2 
					 
					
						
						
							
							tokenize : escape the prompt ( #11058 )  
						
						... 
						
						
						
						* tokenize : escape the prompt
* tokenize : update help 
						
						
					 
					
						2025-01-06 10:54:25 +02:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						727368c60f 
					 
					
						
						
							
							llama : use LLAMA_TOKEN_NULL ( #11062 )  
						
						... 
						
						
						
						ggml-ci 
						
						
					 
					
						2025-01-06 10:52:15 +02:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						f66f582927 
					 
					
						
						
							
							llama : refactor src/llama.cpp ( #10902 )  
						
						... 
						
						
						
						* llama : scatter llama.cpp into multiple modules (wip)
* llama : control-vector -> adapter
* llama : arch
* llama : mmap
ggml-ci
* ci : remove BUILD_SHARED_LIBS=OFF
ggml-ci
* llama : arch (cont)
ggml-ci
* llama : chat
ggml-ci
* llama : model
ggml-ci
* llama : hparams
ggml-ci
* llama : adapter
ggml-ci
* examples : fix
ggml-ci
* rebase
ggml-ci
* minor
* llama : kv cache
ggml-ci
* llama : impl
ggml-ci
* llama : batch
ggml-ci
* cont
ggml-ci
* llama : context
ggml-ci
* minor
* llama : context (cont)
ggml-ci
* llama : model loader
ggml-ci
* common : update lora
ggml-ci
* llama : quant
ggml-ci
* llama : quant (cont)
ggml-ci
* minor [no ci] 
						
						
					 
					
						2025-01-03 10:18:53 +02:00 
						 
				 
			
				
					
						
							
							
								Pierrick Hymbert 
							
						 
					 
					
						
						
							
						
						2f0ee84b9b 
					 
					
						
						
							
							server: bench: minor fixes ( #10765 )  
						
						... 
						
						
						
						* server/bench:
- support openAI streaming standard output with [DONE]\n\n
- export k6 raw results in csv
- fix too many tcp idle connection in tcp_wait
- add metric time to emit first token
* server/bench:
- fix when prometheus not started
- wait for server to be ready before starting bench 
						
						
					 
					
						2025-01-02 18:06:12 +01:00 
						 
				 
			
				
					
						
							
							
								Xuan Son Nguyen 
							
						 
					 
					
						
						
							
						
						0da5d86026 
					 
					
						
						
							
							server : allow using LoRA adapters per-request ( #10994 )  
						
						... 
						
						
						
						* slot.can_batch_with
* lora per request
* test: force disable cache prompt
* move can_batch_with check
* fix condition
* add slow test with llama 8b
* update docs
* move lora change task to queue
* Apply suggestions from code review
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com >
* lora_base
* remove redundant check
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com > 
						
						
					 
					
						2025-01-02 15:05:18 +01:00 
						 
				 
			
				
					
						
							
							
								Xuan Son Nguyen 
							
						 
					 
					
						
						
							
						
						45095a61bf 
					 
					
						
						
							
							server : clean up built-in template detection ( #11026 )  
						
						... 
						
						
						
						* server : clean up built-in template detection
* fix compilation
* add chat template test
* fix condition 
						
						
					 
					
						2024-12-31 15:22:01 +01:00 
						 
				 
			
				
					
						
							
							
								Xuan Son Nguyen 
							
						 
					 
					
						
						
							
						
						5896c65232 
					 
					
						
						
							
							server : add OAI compat for /v1/completions ( #10974 )  
						
						... 
						
						
						
						* server : add OAI compat for /v1/completions
* add test
* add docs
* better docs 
						
						
					 
					
						2024-12-31 12:34:13 +01:00 
						 
				 
			
				
					
						
							
							
								Peter 
							
						 
					 
					
						
						
							
						
						6e1531aca5 
					 
					
						
						
							
							common, examples, ggml : fix MSYS2 GCC compiler errors and warnings when building with LLAMA_CURL=ON and GGML_OPENCL=ON ( #11013 )  
						
						... 
						
						
						
						In common/common.cpp:
* Convert usage of stat() function call to check if file exists to standard library function std::filesystem::exists (error unable to match to correct function signature)
* Additional conditions to check if PATH_MAX is already defined in WIN32 environment (warning it is already defined in MSYS2)
In examples/run/run.cpp:
* Add io.h header inclusion (error cannot find function _get_osfhandle)
* Change initialisers for OVERLAPPED to empty struct (warning about uninitialised members)
* Add initialiser for hFile (warning it may be uninitialised)
* Add cast for curl_off_t percentage value to long int in generate_progress_prefix function (warning that curl_off_t is long long int)
In ggml/src/ggml-opencl/ggml-opencl.cpp:
* Initialise certain declared cl_mem variables to nullptr for greater safety (warning about B_d variable possibly used unassigned) 
						
						
					 
					
						2024-12-31 01:46:06 +01:00 
						 
				 
			
				
					
						
							
							
								ag2s20150909 
							
						 
					 
					
						
						
							
						
						c250ecb315 
					 
					
						
						
							
							android : fix llama_batch free ( #11014 )  
						
						
						
						
					 
					
						2024-12-30 14:35:13 +02:00