mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	Merge branch 'master' into xsn/private_batch_api
This commit is contained in:
		| @@ -206,9 +206,6 @@ bool IMatrixCollector::collect_imatrix(struct ggml_tensor * t, bool ask, void * | ||||
|  | ||||
| void IMatrixCollector::save_imatrix(int ncall) const { | ||||
|     auto fname = m_params.out_file; | ||||
|     if (fname.empty()) { | ||||
|         fname = "imatrix.dat"; | ||||
|     } | ||||
|  | ||||
|     if (ncall > 0) { | ||||
|         fname += ".at_"; | ||||
| @@ -498,7 +495,7 @@ static bool compute_imatrix(llama_context * ctx, const common_params & params) { | ||||
|         const auto t_start = std::chrono::high_resolution_clock::now(); | ||||
|  | ||||
|         // clear the KV cache | ||||
|         llama_kv_cache_clear(ctx); | ||||
|         llama_kv_self_clear(ctx); | ||||
|  | ||||
|         llama_batch_ext * batch = llama_batch_ext_init(n_batch, 1); | ||||
|  | ||||
| @@ -584,6 +581,8 @@ static bool compute_imatrix(llama_context * ctx, const common_params & params) { | ||||
| int main(int argc, char ** argv) { | ||||
|     common_params params; | ||||
|  | ||||
|     params.out_file = "imatrix.dat" ; | ||||
|  | ||||
|     params.n_ctx = 512; | ||||
|     params.logits_all = true; | ||||
|     params.escape = false; | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Xuan Son Nguyen
					Xuan Son Nguyen