mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-10 10:27:03 +00:00
* gemma : fix attn scale for 27B * cont : apply scale before attn * cont : consistent attention scaling
594 KiB
594 KiB