chore : fix models indent after refactor (#16992)

This commit is contained in:
Sigbjørn Skjæret
2025-11-04 12:29:15 +01:00
committed by GitHub
parent 1f5accb8d0
commit b164259bba
39 changed files with 4044 additions and 4055 deletions

View File

@@ -27,7 +27,6 @@ llm_build_grok::llm_build_grok(const llama_model & model, const llm_graph_params
LLM_NORM_RMS, il);
cb(cur, "attn_norm", il);
// self-attention
{
// compute Q and K and RoPE them

View File

@@ -1,6 +1,5 @@
#include "models.h"
llm_build_internlm2::llm_build_internlm2(const llama_model & model, const llm_graph_params & params) : llm_graph_context(params) {
const int64_t n_embd_head = hparams.n_embd_head_v;

View File

@@ -1,6 +1,5 @@
#include "models.h"
llm_build_jamba::llm_build_jamba(const llama_model & model, const llm_graph_params & params) : llm_graph_context_mamba(params) {
const int64_t n_embd_head = hparams.n_embd_head_v;

View File

@@ -120,4 +120,3 @@ llm_build_llada_moe::llm_build_llada_moe(const llama_model & model, const llm_gr
ggml_build_forward_expand(gf, cur);
}

View File

@@ -1,8 +1,6 @@
#include "models.h"
llm_build_llada::llm_build_llada(const llama_model & model, const llm_graph_params & params) :
llm_graph_context(params) {
llm_build_llada::llm_build_llada(const llama_model & model, const llm_graph_params & params) : llm_graph_context(params) {
// LLaDA is similar to LLaMA but uses non-causal attention for diffusion
const int64_t n_embd_head = hparams.n_embd_head_v;

View File

@@ -1,6 +1,5 @@
#include "models.h"
llm_build_llama::llm_build_llama(const llama_model & model, const llm_graph_params & params) : llm_graph_context(params) {
const int64_t n_embd_head = hparams.n_embd_head_v;

View File

@@ -1,6 +1,5 @@
#include "models.h"
llm_build_minicpm3::llm_build_minicpm3(const llama_model & model, const llm_graph_params & params) : llm_graph_context(params) {
//TODO: if the model varies, these parameters need to be read from the model
const int64_t n_embd_base = 256;

View File

@@ -1,6 +1,5 @@
#include "models.h"
template <bool iswa>
llm_build_olmo2<iswa>::llm_build_olmo2(const llama_model & model, const llm_graph_params & params) : llm_graph_context(params) {
const int64_t n_embd_head = hparams.n_embd_head_v;

View File

@@ -1,6 +1,5 @@
#include "models.h"
template<bool iswa>
llm_build_phi3<iswa>::llm_build_phi3(const llama_model & model, const llm_graph_params & params) : llm_graph_context(params) {
const int64_t n_embd_head = hparams.n_embd_head_v;

View File

@@ -1,6 +1,5 @@
#include "models.h"
llm_build_qwen2::llm_build_qwen2(const llama_model & model, const llm_graph_params & params) : llm_graph_context(params) {
const int64_t n_embd_head = hparams.n_embd_head_v;