mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	llama : fix command-r inference when omitting outputs (#6367)
This commit is contained in:
		@@ -9154,6 +9154,7 @@ struct llm_build_context {
 | 
				
			|||||||
                struct ggml_tensor * inp_out_ids = build_inp_out_ids();
 | 
					                struct ggml_tensor * inp_out_ids = build_inp_out_ids();
 | 
				
			||||||
                cur     = ggml_get_rows(ctx0,     cur, inp_out_ids);
 | 
					                cur     = ggml_get_rows(ctx0,     cur, inp_out_ids);
 | 
				
			||||||
                inpL    = ggml_get_rows(ctx0,    inpL, inp_out_ids);
 | 
					                inpL    = ggml_get_rows(ctx0,    inpL, inp_out_ids);
 | 
				
			||||||
 | 
					                ffn_inp = ggml_get_rows(ctx0, ffn_inp, inp_out_ids);
 | 
				
			||||||
            }
 | 
					            }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            struct ggml_tensor * attn_out = cur;
 | 
					            struct ggml_tensor * attn_out = cur;
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user