mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	 097e121e2f
			
		
	
	097e121e2f
	
	
	
		
			
			* llama : add benchmark example * add to examples CMakeLists.txt * fix msvc build * add missing include * add Bessel's correction to stdev calculation Co-authored-by: Johannes Gäßler <johannesg@5d6.de> * improve markdown formatting * add missing include * print warning is NDEBUG is not defined * remove n_prompt and n_gen from the matrix, use each value separately instead * better checks for non-optimized builds * llama.cpp : fix MEM_REQ_SCRATCH0 reusing the value of n_ctx of the first call * fix json formatting * add sql output * add basic cpu and gpu info (linx/cuda only) * markdown: also show values that differ from the default * markdown: add build id * cleanup * improve formatting * formatting --------- Co-authored-by: Johannes Gäßler <johannesg@5d6.de>
		
			
				
	
	
		
			9 lines
		
	
	
		
			306 B
		
	
	
	
		
			CMake
		
	
	
	
	
	
			
		
		
	
	
			9 lines
		
	
	
		
			306 B
		
	
	
	
		
			CMake
		
	
	
	
	
	
| set(TARGET llama-bench)
 | |
| add_executable(${TARGET} llama-bench.cpp)
 | |
| install(TARGETS ${TARGET} RUNTIME)
 | |
| target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT})
 | |
| target_compile_features(${TARGET} PRIVATE cxx_std_11)
 | |
| if(TARGET BUILD_INFO)
 | |
|   add_dependencies(${TARGET} BUILD_INFO)
 | |
| endif()
 |