mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	CUDA: fix bad asserts for partial offload (#13337)
This commit is contained in:
		@@ -1299,6 +1299,10 @@ bool ggml_is_contiguous_2(const struct ggml_tensor * tensor) {
 | 
			
		||||
    return ggml_is_contiguous_n(tensor, 2);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
bool ggml_is_contiguously_allocated(const struct ggml_tensor * tensor) {
 | 
			
		||||
    return ggml_nbytes(tensor) == ggml_nelements(tensor) * ggml_type_size(tensor->type)/ggml_blck_size(tensor->type);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
bool ggml_is_permuted(const struct ggml_tensor * tensor) {
 | 
			
		||||
    static_assert(GGML_MAX_DIMS == 4, "GGML_MAX_DIMS is not 4 - update this function");
 | 
			
		||||
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user