This website requires JavaScript.
Explore
Help
Sign In
CS348Project
/
llama.cpp
Watch
5
Star
0
Fork
0
You've already forked llama.cpp
mirror of
https://github.com/ggml-org/llama.cpp.git
synced
2025-11-17 11:37:10 +00:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
Files
05cd6e5036d72d0930de4d8f6be7bce09e8dda24
llama.cpp
/
ggml-metal.m
Georgi Gerganov
d7b800b8bc
llama : pad KV cache size (
#4280
)
...
* llama : pad KV cache size to 32 * metal : try to improve batched decoding
2023-12-03 10:58:16 +02:00
89 KiB
Raw
Blame
History
View Raw
Reference in New Issue
View Git Blame
Copy Permalink