mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	readme : add llama-swap to infrastructure section (#11032)
* list llama-swap under tools in README * readme: add llama-swap to Infrastructure
This commit is contained in:
		| @@ -201,6 +201,7 @@ Instructions for adding support for new models: [HOWTO-add-model.md](docs/develo | ||||
| - [Paddler](https://github.com/distantmagic/paddler) - Stateful load balancer custom-tailored for llama.cpp | ||||
| - [GPUStack](https://github.com/gpustack/gpustack) - Manage GPU clusters for running LLMs | ||||
| - [llama_cpp_canister](https://github.com/onicai/llama_cpp_canister) - llama.cpp as a smart contract on the Internet Computer, using WebAssembly | ||||
| - [llama-swap](https://github.com/mostlygeek/llama-swap) - transparent proxy that adds automatic model switching with llama-server | ||||
|  | ||||
| </details> | ||||
|  | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Benson Wong
					Benson Wong