mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	sched : offload_op also requires supports_op (#7977)
This commit is contained in:
		@@ -1172,7 +1172,7 @@ static int ggml_backend_sched_backend_id_from_cur(ggml_backend_sched_t sched, st
 | 
				
			|||||||
            // check if a backend with higher prio wants to offload the op
 | 
					            // check if a backend with higher prio wants to offload the op
 | 
				
			||||||
            if (src_backend_id == sched->n_backends - 1) {
 | 
					            if (src_backend_id == sched->n_backends - 1) {
 | 
				
			||||||
                for (int b = 0; b < src_backend_id; b++) {
 | 
					                for (int b = 0; b < src_backend_id; b++) {
 | 
				
			||||||
                    if (ggml_backend_offload_op(sched->backends[b], tensor)) {
 | 
					                    if (ggml_backend_supports_op(sched->backends[b], tensor) && ggml_backend_offload_op(sched->backends[b], tensor)) {
 | 
				
			||||||
                        SET_CAUSE(tensor, "1.off");
 | 
					                        SET_CAUSE(tensor, "1.off");
 | 
				
			||||||
                        return b;
 | 
					                        return b;
 | 
				
			||||||
                    }
 | 
					                    }
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user