mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	* ggml: support CUDA's half type for aarch64(#1455) support CUDA's half type for aarch64 in ggml_fp16_t definition * ggml: use __CUDACC__ to recognise nvcc compiler
This commit is contained in:
		
							
								
								
									
										5
									
								
								ggml.h
									
									
									
									
									
								
							
							
						
						
									
										5
									
								
								ggml.h
									
									
									
									
									
								
							| @@ -259,8 +259,9 @@ | |||||||
| extern "C" { | extern "C" { | ||||||
| #endif | #endif | ||||||
|  |  | ||||||
| #ifdef __ARM_NEON | #if defined(__ARM_NEON) && defined(__CUDACC__) | ||||||
|     // we use the built-in 16-bit float type |     typedef half ggml_fp16_t; | ||||||
|  | #elif defined(__ARM_NEON) | ||||||
|     typedef __fp16 ggml_fp16_t; |     typedef __fp16 ggml_fp16_t; | ||||||
| #else | #else | ||||||
|     typedef uint16_t ggml_fp16_t; |     typedef uint16_t ggml_fp16_t; | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Kylin
					Kylin