Files
llama.cpp/examples/server/server.cpp
Tobias Lütke 7a3895641c allow server to multithread
because web browsers send a lot of garbage requests we want the server
to multithread when serving 404s for favicon's etc. To avoid blowing up
llama we just take a mutex when it's invoked.
2023-07-04 09:14:49 -04:00

42 KiB