mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	 45abe0f74e
			
		
	
	45abe0f74e
	
	
	
		
			
			* server : replace behave with pytest * fix test on windows * misc * add more tests * more tests * styling * log less, fix embd test * added all sequential tests * fix coding style * fix save slot test * add parallel completion test * fix parallel test * remove feature files * update test docs * no cache_prompt for some tests * add test_cache_vs_nocache_prompt
		
			
				
	
	
		
			16 lines
		
	
	
		
			449 B
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			16 lines
		
	
	
		
			449 B
		
	
	
	
		
			Python
		
	
	
	
	
	
| import pytest
 | |
| from utils import *
 | |
| 
 | |
| 
 | |
| # ref: https://stackoverflow.com/questions/22627659/run-code-before-and-after-each-test-in-py-test
 | |
| @pytest.fixture(autouse=True)
 | |
| def stop_server_after_each_test():
 | |
|     # do nothing before each test
 | |
|     yield
 | |
|     # stop all servers after each test
 | |
|     instances = set(
 | |
|         server_instances
 | |
|     )  # copy the set to prevent 'Set changed size during iteration'
 | |
|     for server in instances:
 | |
|         server.stop()
 |