Files
llama.cpp/ggml
Sigbjørn Skjæret 4683cb402a CUDA: don't convert BF16 weights to FP32 (ggml/1174)
* add bf16 support

* use convert_from_bf16_cuda instead of convert_unary_cuda for f32

* revert 7ec5085

* move functionality into convert_unary with constexpr
2025-04-07 12:31:23 +03:00
..
2024-07-13 18:12:39 +02:00