This website requires JavaScript.
Explore
Help
Sign In
CS348Project
/
llama.cpp
Watch
5
Star
0
Fork
0
You've already forked llama.cpp
mirror of
https://github.com/ggml-org/llama.cpp.git
synced
2025-11-16 11:27:03 +00:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
Files
5d46babdc2d4675d96ebcf23cac098a02f0d30cc
llama.cpp
/
ggml
/
src
/
ggml-cann
History
Georgi Gerganov
ec68e84c32
ggml : support bcast ggml_soft_max_ext, ggml_flash_attn_ext (
#14435
)
...
ggml-ci
2025-07-02 15:48:33 +03:00
..
acl_tensor.cpp
CANN: Add the basic supports of Flash Attention kernel (
#13627
)
2025-05-26 10:20:18 +08:00
acl_tensor.h
CANN: Add the basic supports of Flash Attention kernel (
#13627
)
2025-05-26 10:20:18 +08:00
aclnn_ops.cpp
CANN: update aclnnGroupedMatmulV2 to aclnnGroupedMatmulV3 (
#14411
)
2025-07-01 16:47:30 +08:00
aclnn_ops.h
CANN: Add the basic supports of Flash Attention kernel (
#13627
)
2025-05-26 10:20:18 +08:00
CMakeLists.txt
CANN: Add SOC TYPE printing in cmake configuration (
#13837
)
2025-05-28 11:54:20 +08:00
common.h
fix async_mode bug (
#14432
)
2025-06-28 17:35:41 +08:00
Doxyfile
CANN: Add the basic supports of Flash Attention kernel (
#13627
)
2025-05-26 10:20:18 +08:00
ggml-cann.cpp
ggml : support bcast ggml_soft_max_ext, ggml_flash_attn_ext (
#14435
)
2025-07-02 15:48:33 +03:00