mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	more llama-cli(.exe)
This commit is contained in:
		| @@ -427,7 +427,7 @@ Otherwise, run the `win-build-sycl.bat` wrapper which encapsulates the former in | ||||
|  | ||||
| *Notes:* | ||||
|  | ||||
| - By default, calling `make` will build all target binary files. In case of a minimal experimental setup, the user can build the inference executable only through `make llama`. | ||||
| - By default, calling `make` will build all target binary files. In case of a minimal experimental setup, the user can build the inference executable only through `make llama-cli`. | ||||
|  | ||||
| ### III. Run the inference | ||||
|  | ||||
| @@ -488,13 +488,13 @@ Examples: | ||||
| - Use device 0: | ||||
|  | ||||
| ``` | ||||
| build\bin\llama.exe -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 33 -s 0 -sm none -mg 0 | ||||
| build\bin\llama-cli.exe -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 33 -s 0 -sm none -mg 0 | ||||
| ``` | ||||
|  | ||||
| - Use multiple devices: | ||||
|  | ||||
| ``` | ||||
| build\bin\llama.exe -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 33 -s 0 -sm layer | ||||
| build\bin\llama-cli.exe -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 33 -s 0 -sm layer | ||||
| ``` | ||||
| Otherwise, run the following wrapper script: | ||||
|  | ||||
|   | ||||
| @@ -26,7 +26,7 @@ To get started right away, run the following command, making sure to use the cor | ||||
| #### Windows: | ||||
|  | ||||
| ```powershell | ||||
| llama.exe -m models\7B\ggml-model.bin --prompt "Once upon a time" | ||||
| llama-cli.exe -m models\7B\ggml-model.bin --prompt "Once upon a time" | ||||
| ``` | ||||
|  | ||||
| For an interactive experience, try this command: | ||||
| @@ -45,7 +45,7 @@ User:' | ||||
| #### Windows: | ||||
|  | ||||
| ```powershell | ||||
| llama.exe -m models\7B\ggml-model.bin -n -1 --color -r "User:" --in-prefix " " -i -e -p "User: Hi\nAI: Hello. I am an AI chatbot. Would you like to talk?\nUser: Sure!\nAI: What would you like to talk about?\nUser:" | ||||
| llama-cli.exe -m models\7B\ggml-model.bin -n -1 --color -r "User:" --in-prefix " " -i -e -p "User: Hi\nAI: Hello. I am an AI chatbot. Would you like to talk?\nUser: Sure!\nAI: What would you like to talk about?\nUser:" | ||||
| ``` | ||||
|  | ||||
| The following command generates "infinite" text from a starting prompt (you can use `Ctrl-C` to stop it): | ||||
| @@ -59,7 +59,7 @@ The following command generates "infinite" text from a starting prompt (you can | ||||
| #### Windows: | ||||
|  | ||||
| ```powershell | ||||
| llama.exe -m models\7B\ggml-model.bin --ignore-eos -n -1 | ||||
| llama-cli.exe -m models\7B\ggml-model.bin --ignore-eos -n -1 | ||||
| ``` | ||||
|  | ||||
| ## Common Options | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Olivier Chafik
					Olivier Chafik