mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	tests : add unified cache server tests
This commit is contained in:
		| @@ -78,6 +78,7 @@ class ServerProcess: | ||||
|     server_embeddings: bool | None = False | ||||
|     server_reranking: bool | None = False | ||||
|     server_metrics: bool | None = False | ||||
|     kv_unified: bool | None = False | ||||
|     server_slots: bool | None = False | ||||
|     pooling: str | None = None | ||||
|     draft: int | None = None | ||||
| @@ -159,6 +160,8 @@ class ServerProcess: | ||||
|             server_args.append("--reranking") | ||||
|         if self.server_metrics: | ||||
|             server_args.append("--metrics") | ||||
|         if self.kv_unified: | ||||
|             server_args.append("--kv-unified") | ||||
|         if self.server_slots: | ||||
|             server_args.append("--slots") | ||||
|         else: | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov