mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-04 09:32:00 +00:00 
			
		
		
		
	Fix: `sentencepiece` tokenizers with added tokens failed with an incorrect assertion
This commit is contained in:
		@@ -8200,7 +8200,9 @@ int llama_token_to_piece(const struct llama_model * model, llama_token token, ch
 | 
			
		||||
                buf[0] = llama_token_to_byte(model->vocab, token);
 | 
			
		||||
                return 1;
 | 
			
		||||
            } else {
 | 
			
		||||
                GGML_ASSERT(false);
 | 
			
		||||
                // TODO: for now we accept all unsupported token types,
 | 
			
		||||
                // suppressing them like CONTROL tokens.
 | 
			
		||||
                // GGML_ASSERT(false);
 | 
			
		||||
            }
 | 
			
		||||
            break;
 | 
			
		||||
        }
 | 
			
		||||
@@ -8216,7 +8218,9 @@ int llama_token_to_piece(const struct llama_model * model, llama_token token, ch
 | 
			
		||||
            } else if (llama_is_control_token(model->vocab, token)) {
 | 
			
		||||
                ;
 | 
			
		||||
            } else {
 | 
			
		||||
                GGML_ASSERT(false);
 | 
			
		||||
                // TODO: for now we accept all unsupported token types,
 | 
			
		||||
                // suppressing them like CONTROL tokens.
 | 
			
		||||
                // GGML_ASSERT(false);
 | 
			
		||||
            }
 | 
			
		||||
            break;
 | 
			
		||||
        }
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user