mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-04 09:32:00 +00:00 
			
		
		
		
	Merge branch 'master' into gg/llama-kv-cache
ggml-ci
This commit is contained in:
		@@ -78,6 +78,7 @@ class ServerProcess:
 | 
			
		||||
    draft_max: int | None = None
 | 
			
		||||
    no_webui: bool | None = None
 | 
			
		||||
    jinja: bool | None = None
 | 
			
		||||
    reasoning_format: Literal['deepseek', 'none'] | None = None
 | 
			
		||||
    chat_template: str | None = None
 | 
			
		||||
    chat_template_file: str | None = None
 | 
			
		||||
 | 
			
		||||
@@ -172,6 +173,8 @@ class ServerProcess:
 | 
			
		||||
            server_args.append("--no-webui")
 | 
			
		||||
        if self.jinja:
 | 
			
		||||
            server_args.append("--jinja")
 | 
			
		||||
        if self.reasoning_format is not None:
 | 
			
		||||
            server_args.extend(("--reasoning-format", self.reasoning_format))
 | 
			
		||||
        if self.chat_template:
 | 
			
		||||
            server_args.extend(["--chat-template", self.chat_template])
 | 
			
		||||
        if self.chat_template_file:
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user