mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	ggml : hide ggml_object, ggml_cgraph, ggml_hash_set
ggml-ci
This commit is contained in:
		| @@ -183,7 +183,7 @@ int main(int argc, char ** argv)  { | ||||
|  | ||||
|     ggml_graph_compute_helper(work_buffer, gf, benchmark_params.n_threads); | ||||
|  | ||||
|     TENSOR_DUMP(gf->nodes[0]); | ||||
|     TENSOR_DUMP(ggml_graph_node(gf, 0)); | ||||
|  | ||||
|     printf("\n------ Test 2 - Matrix Mult via %s code\n", ggml_type_name(qtype)); | ||||
|  | ||||
| @@ -224,7 +224,7 @@ int main(int argc, char ** argv)  { | ||||
|  | ||||
|  | ||||
|     // Let's use the F32 result from above as a reference for the quantized multiplication | ||||
|     float sum_of_F32_reference = tensor_sum_elements(gf->nodes[0]); | ||||
|     float sum_of_F32_reference = tensor_sum_elements(ggml_graph_node(gf, 0)); | ||||
|  | ||||
|     printf("Iteration;NThreads; SizeX; SizeY; SizeZ; Required_FLOPS; Elapsed_u_Seconds; gigaFLOPS\n"); | ||||
|     printf("=====================================================================================\n"); | ||||
| @@ -252,7 +252,7 @@ int main(int argc, char ** argv)  { | ||||
|  | ||||
|         // Check that the matrix multiplication result is in the right ballpark | ||||
|         // We cannot use the exact value from the F32 multiplication because the quantizuation will be slightly different | ||||
|         float sum_of_Q4_result = tensor_sum_elements(gf31->nodes[0]); | ||||
|         float sum_of_Q4_result = tensor_sum_elements(ggml_graph_node(gf31, 0)); | ||||
|         float delta = std::abs(sum_of_Q4_result - sum_of_F32_reference); | ||||
|         float allowed_delta = (sum_of_F32_reference) / 1000 / 1000; //  Let's accept an epsilon of 10^-6 | ||||
|  | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov