mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	metal : disable graph concurrency optimization due to bug (#2413)
This commit is contained in:
		| @@ -1722,9 +1722,10 @@ static bool llama_eval_internal( | ||||
|  | ||||
| #ifdef GGML_USE_METAL | ||||
|     if (lctx.ctx_metal && N == 1) { | ||||
|         if (!ggml_metal_if_optimized(lctx.ctx_metal)) { | ||||
|             ggml_metal_graph_find_concurrency(lctx.ctx_metal, gf); | ||||
|         } | ||||
|         // TODO: disabled until #2413 is resolved | ||||
|         //if (!ggml_metal_if_optimized(lctx.ctx_metal)) { | ||||
|         //    ggml_metal_graph_find_concurrency(lctx.ctx_metal, gf); | ||||
|         //} | ||||
|         ggml_metal_set_n_cb     (lctx.ctx_metal, n_threads); | ||||
|         ggml_metal_graph_compute(lctx.ctx_metal, gf); | ||||
|         ggml_metal_get_tensor   (lctx.ctx_metal, cur); | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov