Files
llama.cpp/examples/server/server.cpp
Xuan Son Nguyen 99b71c068f Server: Use multi-task for embeddings endpoint (#6001)
* use multitask for embd endpoint

* specify types

* remove redundant {"n_predict", 0}
2024-03-13 11:39:11 +01:00

140 KiB