mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-06 09:46:50 +00:00
examples : utilize new llama_get_logits_ith()
This commit is contained in:
@@ -150,7 +150,7 @@ int main(int argc, char ** argv) {
|
||||
}
|
||||
|
||||
auto n_vocab = llama_n_vocab(ctx);
|
||||
auto logits = llama_get_logits(ctx) + i_batch[i] * n_vocab;
|
||||
auto logits = llama_get_logits_ith(ctx, i_batch[i]);
|
||||
|
||||
std::vector<llama_token_data> candidates;
|
||||
candidates.reserve(n_vocab);
|
||||
|
||||
Reference in New Issue
Block a user