mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	HIP: remove the use of __HIP_PLATFORM_AMD__, explicitly support only AMD targets (#14945)
This commit is contained in:
		@@ -5,9 +5,9 @@
 | 
			
		||||
#define FATTN_KQ_STRIDE_TILE_F16 64
 | 
			
		||||
 | 
			
		||||
template<int D, int ncols, int nwarps, bool use_logit_softcap> // D == head size
 | 
			
		||||
#if !(defined(GGML_USE_HIP) && defined(__HIP_PLATFORM_AMD__))
 | 
			
		||||
#if !defined(GGML_USE_HIP)
 | 
			
		||||
__launch_bounds__(nwarps*WARP_SIZE, 2)
 | 
			
		||||
#endif // !(defined(GGML_USE_HIP) && defined(__HIP_PLATFORM_AMD__))
 | 
			
		||||
#endif // !defined(GGML_USE_HIP)
 | 
			
		||||
static __global__ void flash_attn_tile_ext_f16(
 | 
			
		||||
        const char * __restrict__ Q,
 | 
			
		||||
        const char * __restrict__ K,
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user