mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	only rm when params.escape, rm space if possible which is added back or rm added space token
This commit is contained in:
		| @@ -234,10 +234,11 @@ int main(int argc, char ** argv) { | |||||||
|     LOG("add_bos: %d\n", add_bos); |     LOG("add_bos: %d\n", add_bos); | ||||||
|  |  | ||||||
|     bool suff_rm_leading_spc = params.escape; |     bool suff_rm_leading_spc = params.escape; | ||||||
|     if (suff_rm_leading_spc && params.input_suffix.find_first_of(" ") == 0 && params.input_suffix.size() > 1) { |     // if (suff_rm_leading_spc && params.input_suffix.find_first_of(" ") == 0 && params.input_suffix.size() > 1) { | ||||||
|         params.input_suffix.erase(0, 1); |     //     params.input_suffix.erase(0, 1); | ||||||
|         suff_rm_leading_spc = false; |     //     suff_rm_leading_spc = false; | ||||||
|     } |     // } | ||||||
|  |     suff_rm_leading_spc = true; | ||||||
|     std::vector<llama_token> embd_inp; |     std::vector<llama_token> embd_inp; | ||||||
|     std::vector<llama_token> inp_pfx = ::llama_tokenize(ctx, params.input_prefix, false); |     std::vector<llama_token> inp_pfx = ::llama_tokenize(ctx, params.input_prefix, false); | ||||||
|     std::vector<llama_token> inp_sfx = ::llama_tokenize(ctx, params.input_suffix, false); |     std::vector<llama_token> inp_sfx = ::llama_tokenize(ctx, params.input_suffix, false); | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 vvhg1
					vvhg1