mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-04 09:32:00 +00:00 
			
		
		
		
	mpi : fix output tensor after MPI compute (still not working)
This commit is contained in:
		@@ -141,8 +141,8 @@ void ggml_mpi_graph_compute(
 | 
			
		||||
        const int il0 =               (mpi_idx + 0) * n_per_node;
 | 
			
		||||
        const int il1 = MIN(n_layers, (mpi_idx + 1) * n_per_node);
 | 
			
		||||
 | 
			
		||||
        char name_l0[64];
 | 
			
		||||
        char name_l1[64];
 | 
			
		||||
        char name_l0[GGML_MAX_NAME];
 | 
			
		||||
        char name_l1[GGML_MAX_NAME];
 | 
			
		||||
 | 
			
		||||
        snprintf(name_l0, sizeof(name_l0), "layer_inp_%d", il0);
 | 
			
		||||
        snprintf(name_l1, sizeof(name_l1), "layer_inp_%d", il1);
 | 
			
		||||
@@ -175,6 +175,11 @@ void ggml_mpi_graph_compute(
 | 
			
		||||
 | 
			
		||||
    ggml_graph_compute(ctx, gf);
 | 
			
		||||
 | 
			
		||||
    //if (mpi_rank == 0) {
 | 
			
		||||
    //    ggml_graph_print(gf);
 | 
			
		||||
    //    ggml_graph_dump_dot(gf, NULL, "llama.dot");
 | 
			
		||||
    //}
 | 
			
		||||
 | 
			
		||||
    //fprintf(stderr, "%s: node %d: done\n", __func__, mpi_rank);
 | 
			
		||||
 | 
			
		||||
    // send the output data to the next node
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user