mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-29 08:41:22 +00:00 
			
		
		
		
	CUDA: fix compilation with GGML_CUDA_F16 (#14837)
This commit is contained in:
		| @@ -31,8 +31,8 @@ static __global__ void dequantize_block(const void * __restrict__ vx, dst_t * __ | |||||||
|     dequantize_kernel(vx, ib, iqs, v); |     dequantize_kernel(vx, ib, iqs, v); | ||||||
|  |  | ||||||
|     const int64_t iy0 = ((i03*ne02 + i02)*ne01 + i01)*ne00 + iybs + iqs; |     const int64_t iy0 = ((i03*ne02 + i02)*ne01 + i01)*ne00 + iybs + iqs; | ||||||
|     y[iy0 + 0]        = v.x; |     y[iy0 + 0]        = float(v.x); | ||||||
|     y[iy0 + y_offset] = v.y; |     y[iy0 + y_offset] = float(v.y); | ||||||
| } | } | ||||||
|  |  | ||||||
| template <bool need_check> | template <bool need_check> | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Johannes Gäßler
					Johannes Gäßler