mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-04 09:32:00 +00:00 
			
		
		
		
	* server : refactor slot input data, move tokenizer to HTTP thread * move prompt_tokens.empty() check * fix incorrect if branch * fix infinite generation loop * bring back infill validation * add infill test * try fixing format_infill * fix test * remove redundant code * rename completion to inference * update docs * use llama_tokens everywhere
Server tests
Python based server tests scenario using BDD and behave:
- issues.feature Pending issues scenario
 - parallel.feature Scenario involving multi slots and concurrent requests
 - security.feature Security, CORS and API Key
 - server.feature Server base scenario: completion, embedding, tokenization, etc...
 
Tests target GitHub workflows job runners with 4 vCPU.
Requests are using aiohttp, asyncio based http client.
Note: If the host architecture inference speed is faster than GitHub runners one, parallel scenario may randomly fail.
To mitigate it, you can increase values in n_predict, kv_size.
Install dependencies
pip install -r requirements.txt
Run tests
- Build the server
 
cd ../../..
cmake -B build -DLLAMA_CURL=ON
cmake --build build --target llama-server
- Start the test: 
./tests.sh 
It's possible to override some scenario steps values with environment variables:
| variable | description | 
|---|---|
PORT | 
context.server_port to set the listening port of the server during scenario, default: 8080 | 
LLAMA_SERVER_BIN_PATH | 
to change the server binary path, default: ../../../build/bin/llama-server | 
DEBUG | 
"ON" to enable steps and server verbose mode --verbose | 
N_GPU_LAYERS | 
number of model layers to offload to VRAM -ngl --n-gpu-layers | 
Run @bug, @wip or @wrong_usage annotated scenario
Feature or Scenario must be annotated with @llama.cpp to be included in the default scope.
@bugannotation aims to link a scenario with a GitHub issue.@wrong_usageare meant to show user issue that are actually an expected behavior@wipto focus on a scenario working in progress@slowheavy test, disabled by default
To run a scenario annotated with @bug, start:
DEBUG=ON ./tests.sh --no-skipped --tags bug --stop
After changing logic in steps.py, ensure that @bug and @wrong_usage scenario are updated.
./tests.sh --no-skipped --tags bug,wrong_usage || echo "should failed but compile"