mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	more llama-cli(.exe)
This commit is contained in:
		| @@ -427,7 +427,7 @@ Otherwise, run the `win-build-sycl.bat` wrapper which encapsulates the former in | |||||||
|  |  | ||||||
| *Notes:* | *Notes:* | ||||||
|  |  | ||||||
| - By default, calling `make` will build all target binary files. In case of a minimal experimental setup, the user can build the inference executable only through `make llama`. | - By default, calling `make` will build all target binary files. In case of a minimal experimental setup, the user can build the inference executable only through `make llama-cli`. | ||||||
|  |  | ||||||
| ### III. Run the inference | ### III. Run the inference | ||||||
|  |  | ||||||
| @@ -488,13 +488,13 @@ Examples: | |||||||
| - Use device 0: | - Use device 0: | ||||||
|  |  | ||||||
| ``` | ``` | ||||||
| build\bin\llama.exe -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 33 -s 0 -sm none -mg 0 | build\bin\llama-cli.exe -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 33 -s 0 -sm none -mg 0 | ||||||
| ``` | ``` | ||||||
|  |  | ||||||
| - Use multiple devices: | - Use multiple devices: | ||||||
|  |  | ||||||
| ``` | ``` | ||||||
| build\bin\llama.exe -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 33 -s 0 -sm layer | build\bin\llama-cli.exe -m models\llama-2-7b.Q4_0.gguf -p "Building a website can be done in 10 simple steps:\nStep 1:" -n 400 -e -ngl 33 -s 0 -sm layer | ||||||
| ``` | ``` | ||||||
| Otherwise, run the following wrapper script: | Otherwise, run the following wrapper script: | ||||||
|  |  | ||||||
|   | |||||||
| @@ -26,7 +26,7 @@ To get started right away, run the following command, making sure to use the cor | |||||||
| #### Windows: | #### Windows: | ||||||
|  |  | ||||||
| ```powershell | ```powershell | ||||||
| llama.exe -m models\7B\ggml-model.bin --prompt "Once upon a time" | llama-cli.exe -m models\7B\ggml-model.bin --prompt "Once upon a time" | ||||||
| ``` | ``` | ||||||
|  |  | ||||||
| For an interactive experience, try this command: | For an interactive experience, try this command: | ||||||
| @@ -45,7 +45,7 @@ User:' | |||||||
| #### Windows: | #### Windows: | ||||||
|  |  | ||||||
| ```powershell | ```powershell | ||||||
| llama.exe -m models\7B\ggml-model.bin -n -1 --color -r "User:" --in-prefix " " -i -e -p "User: Hi\nAI: Hello. I am an AI chatbot. Would you like to talk?\nUser: Sure!\nAI: What would you like to talk about?\nUser:" | llama-cli.exe -m models\7B\ggml-model.bin -n -1 --color -r "User:" --in-prefix " " -i -e -p "User: Hi\nAI: Hello. I am an AI chatbot. Would you like to talk?\nUser: Sure!\nAI: What would you like to talk about?\nUser:" | ||||||
| ``` | ``` | ||||||
|  |  | ||||||
| The following command generates "infinite" text from a starting prompt (you can use `Ctrl-C` to stop it): | The following command generates "infinite" text from a starting prompt (you can use `Ctrl-C` to stop it): | ||||||
| @@ -59,7 +59,7 @@ The following command generates "infinite" text from a starting prompt (you can | |||||||
| #### Windows: | #### Windows: | ||||||
|  |  | ||||||
| ```powershell | ```powershell | ||||||
| llama.exe -m models\7B\ggml-model.bin --ignore-eos -n -1 | llama-cli.exe -m models\7B\ggml-model.bin --ignore-eos -n -1 | ||||||
| ``` | ``` | ||||||
|  |  | ||||||
| ## Common Options | ## Common Options | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Olivier Chafik
					Olivier Chafik