mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	llama : add gpt-oss (#15091)
* oai moe * compat with new checkpoint * add attn sink impl * add rope scaling yarn * logits match with latest transformers code * wip chat template * rm trailing space * use ggml_scale_bias * rm redundant is_swa_all * convert interleaved gate_up * graph : fix activation function to match reference (#7) * vocab : handle o200k_harmony special tokens * ggml : add attention sinks support (#1) * llama : add attn sinks * ggml : add attn sinks * cuda : add attn sinks * vulkan : add support for sinks in softmax remove unnecessary return * ggml : add fused swiglu_oai op (#11) * ggml : add fused swiglu_oai op * Update ggml/src/ggml-cpu/ops.cpp Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * update CUDA impl * cont : metal impl * add vulkan impl * test-backend-ops : more test cases, clean up * llama : remove unfused impl * remove extra lines --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> --------- Co-authored-by: slaren <slarengh@gmail.com> * repack mxfp4 upon conversion * clean up a bit * enable thinking * add quick hack to render only some special tokens * fix bf16 conversion * remove vocab hack * webui ok * support chat parsing for gpt-oss * fix webui * direct mapping mxfp4, FINALLY * force using mxfp4 * properly use lazy tensor * ggml : add mxfp4 ggml : use e8m0 conversion instead of powf Co-authored-by: Diego Devesa <slarengh@gmail.com> change kvalues_mxfp4 table to match e2m1 (#6) metal : remove quantization for now (not used) cuda : fix disabled CUDA graphs due to ffn moe bias vulkan : add support for mxfp4 cont : add cm2 dequant * ggml : add ggml_add_id (#13) * ggml : add ggml_add_id * add cuda impl * llama : add weight support check for add_id * perf opt * add vulkan impl * rename cuda files * add metal impl * allow in-place ggml_add_id * llama : keep biases on CPU with --cpu-moe * llama : fix compile error ggml-ci * cuda : add fallback for __nv_cvt_e8m0_to_bf16raw ggml-ci * cleanup ggml-ci * sycl : fix supports_op for MXFP4 ggml-ci * fix Unknown reasoning format * ggml-cpu : fix AVX build ggml-ci * fix hip build ggml-ci * cuda : add mxfp4 dequantization support for cuBLAS ggml-ci * ggml-cpu : fix mxfp4 fallback definitions for some architectures ggml-ci * cuda : fix version required for __nv_cvt_e8m0_to_bf16raw --------- Co-authored-by: Xuan Son Nguyen <son@huggingface.co> Co-authored-by: slaren <slarengh@gmail.com>
This commit is contained in:
		| @@ -410,6 +410,67 @@ static inline ggml_fp16_t ggml_compute_fp32_to_fp16(float f) { | ||||
| #define GGML_FP16_TO_FP32(x) GGML_COMPUTE_FP16_TO_FP32(x) | ||||
| #define GGML_FP32_TO_FP16(x) GGML_COMPUTE_FP32_TO_FP16(x) | ||||
|  | ||||
| static inline float ggml_e8m0_to_fp32(uint8_t x) { | ||||
|     uint32_t bits;  // Stores the raw bit representation of the float | ||||
|  | ||||
|     // Handle special case for minimum exponent (denormalized float) | ||||
|     if (x == 0) { | ||||
|         // Bit pattern for 2^(-127): | ||||
|         // - Sign bit: 0 (positive) | ||||
|         // - Exponent: 0 (denormalized number) | ||||
|         // - Mantissa: 0x400000 (0.5 in fractional form) | ||||
|         // Value = 0.5 * 2^(-126) = 2^(-127) | ||||
|         bits = 0x00400000; | ||||
|     } | ||||
|     // note: disabled as we don't need to handle NaNs | ||||
|     //// Handle special case for NaN (all bits set) | ||||
|     //else if (x == 0xFF) { | ||||
|     //    // Standard quiet NaN pattern: | ||||
|     //    // - Sign bit: 0 | ||||
|     //    // - Exponent: all 1s (0xFF) | ||||
|     //    // - Mantissa: 0x400000 (quiet NaN flag) | ||||
|     //    bits = 0x7FC00000; | ||||
|     //} | ||||
|     // Normalized values (most common case) | ||||
|     else { | ||||
|         // Construct normalized float by shifting exponent into position: | ||||
|         // - Exponent field: 8 bits (positions 30-23) | ||||
|         // - Mantissa: 0 (implicit leading 1) | ||||
|         // Value = 2^(x - 127) | ||||
|         bits = (uint32_t) x << 23; | ||||
|     } | ||||
|  | ||||
|     float result;  // Final float value | ||||
|                    // Safely reinterpret bit pattern as float without type-punning issues | ||||
|     memcpy(&result, &bits, sizeof(float)); | ||||
|     return result; | ||||
| } | ||||
|  | ||||
| // Equal to ggml_e8m0_to_fp32/2 | ||||
| // Useful with MXFP4 quantization since the E0M2 values are doubled | ||||
| static inline float ggml_e8m0_to_fp32_half(uint8_t x) { | ||||
|     uint32_t bits; | ||||
|  | ||||
|     // For x < 2: use precomputed denormal patterns | ||||
|     if (x < 2) { | ||||
|         // 0x00200000 = 2^(-128), 0x00400000 = 2^(-127) | ||||
|         bits = 0x00200000 << x; | ||||
|     } | ||||
|     // For x >= 2: normalized exponent adjustment | ||||
|     else { | ||||
|         // 0.5 * 2^(x-127) = 2^(x-128) = normalized with exponent (x-1) | ||||
|         bits = (uint32_t)(x - 1) << 23; | ||||
|     } | ||||
|     // Note: NaNs are not handled here | ||||
|  | ||||
|     float result; | ||||
|     memcpy(&result, &bits, sizeof(float)); | ||||
|     return result; | ||||
| } | ||||
|  | ||||
| #define GGML_E8M0_TO_FP32(x) ggml_e8m0_to_fp32(x) | ||||
| #define GGML_E8M0_TO_FP32_HALF(x) ggml_e8m0_to_fp32_half(x) | ||||
|  | ||||
| /** | ||||
|  * Converts brain16 to float32. | ||||
|  * | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov