mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	server : (refactor) no more json in server_task input (#10691)
* server : (refactor) no more json in server_task input * add test for slots endpoint * add tests for /props and /slots * remove task inf_type * fix CI by adding safe_json_to_str * add "model_path" to /props * update readme
This commit is contained in:
		| @@ -30,6 +30,7 @@ def test_chat_completion(model, system_prompt, user_prompt, max_tokens, re_conte | ||||
|         ], | ||||
|     }) | ||||
|     assert res.status_code == 200 | ||||
|     assert "cmpl" in res.body["id"] # make sure the completion id has the expected format | ||||
|     assert res.body["model"] == model if model is not None else server.model_alias | ||||
|     assert res.body["usage"]["prompt_tokens"] == n_prompt | ||||
|     assert res.body["usage"]["completion_tokens"] == n_predicted | ||||
| @@ -59,9 +60,13 @@ def test_chat_completion_stream(system_prompt, user_prompt, max_tokens, re_conte | ||||
|         "stream": True, | ||||
|     }) | ||||
|     content = "" | ||||
|     last_cmpl_id = None | ||||
|     for data in res: | ||||
|         choice = data["choices"][0] | ||||
|         assert "gpt-3.5" in data["model"] # DEFAULT_OAICOMPAT_MODEL, maybe changed in the future | ||||
|         if last_cmpl_id is None: | ||||
|             last_cmpl_id = data["id"] | ||||
|         assert last_cmpl_id == data["id"] # make sure the completion id is the same for all events in the stream | ||||
|         if choice["finish_reason"] in ["stop", "length"]: | ||||
|             assert data["usage"]["prompt_tokens"] == n_prompt | ||||
|             assert data["usage"]["completion_tokens"] == n_predicted | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Xuan Son Nguyen
					Xuan Son Nguyen