mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	Merge branch 'master' into compilade/imatrix-batched-chunks
This commit is contained in:
		@@ -15,9 +15,7 @@
 | 
			
		||||
#pragma warning(disable: 4244 4267) // possible loss of data
 | 
			
		||||
#endif
 | 
			
		||||
 | 
			
		||||
static void print_usage(int argc, char ** argv, const gpt_params & params) {
 | 
			
		||||
    gpt_params_print_usage(argc, argv, params);
 | 
			
		||||
 | 
			
		||||
static void print_usage(int, char ** argv) {
 | 
			
		||||
    LOG_TEE("\nexample usage:\n");
 | 
			
		||||
    LOG_TEE("\n    %s \\\n"
 | 
			
		||||
            "       -m model.gguf -f some-text.txt [-o imatrix.gguf] [--process-output] [--verbosity 1] \\\n"
 | 
			
		||||
@@ -657,8 +655,8 @@ int main(int argc, char ** argv) {
 | 
			
		||||
    params.logits_all = true;
 | 
			
		||||
    params.verbosity = 1;
 | 
			
		||||
 | 
			
		||||
    if (!gpt_params_parse(argc, argv, params)) {
 | 
			
		||||
        print_usage(argc, argv, params);
 | 
			
		||||
    auto options = gpt_params_parser_init(params, LLAMA_EXAMPLE_IMATRIX, print_usage);
 | 
			
		||||
    if (!gpt_params_parse(argc, argv, params, options)) {
 | 
			
		||||
        return 1;
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -731,7 +729,8 @@ int main(int argc, char ** argv) {
 | 
			
		||||
 | 
			
		||||
    g_collector.save_imatrix();
 | 
			
		||||
 | 
			
		||||
    llama_print_timings(ctx);
 | 
			
		||||
    LOG_TEE("\n");
 | 
			
		||||
    llama_perf_print(ctx, LLAMA_PERF_TYPE_CONTEXT);
 | 
			
		||||
 | 
			
		||||
    llama_free(ctx);
 | 
			
		||||
    llama_free_model(model);
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user