mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-10-27 08:21:30 +00:00
cont : add TODO
This commit is contained in:
@@ -956,6 +956,7 @@ extern "C" {
|
|||||||
// in the order they have appeared in the batch.
|
// in the order they have appeared in the batch.
|
||||||
// Rows: number of tokens for which llama_batch.logits[i] != 0
|
// Rows: number of tokens for which llama_batch.logits[i] != 0
|
||||||
// Cols: n_vocab
|
// Cols: n_vocab
|
||||||
|
// TODO: deprecate in favor of llama_get_logits_ith() (ref: https://github.com/ggml-org/llama.cpp/pull/14853#issuecomment-3113143522)
|
||||||
LLAMA_API float * llama_get_logits(struct llama_context * ctx);
|
LLAMA_API float * llama_get_logits(struct llama_context * ctx);
|
||||||
|
|
||||||
// Logits for the ith token. For positive indices, Equivalent to:
|
// Logits for the ith token. For positive indices, Equivalent to:
|
||||||
@@ -970,6 +971,7 @@ extern "C" {
|
|||||||
// in the order they have appeared in the batch.
|
// in the order they have appeared in the batch.
|
||||||
// shape: [n_outputs*n_embd]
|
// shape: [n_outputs*n_embd]
|
||||||
// Otherwise, returns NULL.
|
// Otherwise, returns NULL.
|
||||||
|
// TODO: deprecate in favor of llama_get_embeddings_ith() (ref: https://github.com/ggml-org/llama.cpp/pull/14853#issuecomment-3113143522)
|
||||||
LLAMA_API float * llama_get_embeddings(struct llama_context * ctx);
|
LLAMA_API float * llama_get_embeddings(struct llama_context * ctx);
|
||||||
|
|
||||||
// Get the embeddings for the ith token. For positive indices, Equivalent to:
|
// Get the embeddings for the ith token. For positive indices, Equivalent to:
|
||||||
|
|||||||
Reference in New Issue
Block a user