mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	Fix typo in llama_set_embeddings comment (#8077)
This commit is contained in:
		
							
								
								
									
										2
									
								
								llama.h
									
									
									
									
									
								
							
							
						
						
									
										2
									
								
								llama.h
									
									
									
									
									
								
							| @@ -786,7 +786,7 @@ extern "C" { | |||||||
|     // Get the number of threads used for prompt and batch processing (multiple token). |     // Get the number of threads used for prompt and batch processing (multiple token). | ||||||
|     LLAMA_API uint32_t llama_n_threads_batch(struct llama_context * ctx); |     LLAMA_API uint32_t llama_n_threads_batch(struct llama_context * ctx); | ||||||
|  |  | ||||||
|     // Set whether the model is in embeddings model or not |     // Set whether the model is in embeddings mode or not | ||||||
|     // If true, embeddings will be returned but logits will not |     // If true, embeddings will be returned but logits will not | ||||||
|     LLAMA_API void llama_set_embeddings(struct llama_context * ctx, bool embeddings); |     LLAMA_API void llama_set_embeddings(struct llama_context * ctx, bool embeddings); | ||||||
|  |  | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Daniel Bevenius
					Daniel Bevenius