mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-04 09:32:00 +00:00 
			
		
		
		
	llama : fix buffer checks for mamba and rwk (#10111)
* llama : fix buffer checks for mamba and rwk * llama : fix missing worst case flag during reserve * cuda : fix supports_op for norm * disable sched SET_CAUSE
This commit is contained in:
		@@ -7272,6 +7272,7 @@ struct ggml_tensor * ggml_ssm_conv(
 | 
			
		||||
    const int64_t n_s     = sx->ne[2];
 | 
			
		||||
 | 
			
		||||
    // TODO: maybe support other strides than 1?
 | 
			
		||||
    // FIXME: this is always true?
 | 
			
		||||
    GGML_ASSERT(sx->ne[0] == d_conv - 1 + n_t);
 | 
			
		||||
    GGML_ASSERT(sx->ne[1] == d_inner);
 | 
			
		||||
    GGML_ASSERT(n_t >= 0);
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user