mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-29 08:41:22 +00:00 
			
		
		
		
	model : add skt/A.X-4.0 model vocabulary (#14589)
This commit is contained in:
		| @@ -818,6 +818,9 @@ class TextModel(ModelBase): | |||||||
|         if chkhsh == "7e57df22b1fe23a7b1e1c7f3dc4e3f96d43a4eb0836d0c6bdc3436d7b2f1c664": |         if chkhsh == "7e57df22b1fe23a7b1e1c7f3dc4e3f96d43a4eb0836d0c6bdc3436d7b2f1c664": | ||||||
|             # ref: https://huggingface.co/tencent/Hunyuan-A13B-Instruct |             # ref: https://huggingface.co/tencent/Hunyuan-A13B-Instruct | ||||||
|             res = "hunyuan" |             res = "hunyuan" | ||||||
|  |         if chkhsh == "b0a6b1c0bd5998ebd9df08611efde34a4ff03faed45ae09c43e6b31ebd4b94cf": | ||||||
|  |             # ref: https://huggingface.co/skt/A.X-4.0 | ||||||
|  |             res = "a.x-4.0" | ||||||
|         if chkhsh == "a6b57017d60e6edb4d88ecc2845188e0eb333a70357e45dcc9b53964a73bbae6": |         if chkhsh == "a6b57017d60e6edb4d88ecc2845188e0eb333a70357e45dcc9b53964a73bbae6": | ||||||
|             # ref: https://huggingface.co/tiiuae/Falcon-H1-0.5B-Base |             # ref: https://huggingface.co/tiiuae/Falcon-H1-0.5B-Base | ||||||
|             res = "falcon-h1" |             res = "falcon-h1" | ||||||
|   | |||||||
| @@ -128,6 +128,7 @@ models = [ | |||||||
|     {"name": "llama4",           "tokt": TOKENIZER_TYPE.BPE, "repo": "https://huggingface.co/meta-llama/Llama-4-Scout-17B-16E-Instruct", }, |     {"name": "llama4",           "tokt": TOKENIZER_TYPE.BPE, "repo": "https://huggingface.co/meta-llama/Llama-4-Scout-17B-16E-Instruct", }, | ||||||
|     {"name": "pixtral",          "tokt": TOKENIZER_TYPE.BPE, "repo": "https://huggingface.co/mistral-community/pixtral-12b", }, |     {"name": "pixtral",          "tokt": TOKENIZER_TYPE.BPE, "repo": "https://huggingface.co/mistral-community/pixtral-12b", }, | ||||||
|     {"name": "seed-coder",       "tokt": TOKENIZER_TYPE.BPE, "repo": "https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Base", }, |     {"name": "seed-coder",       "tokt": TOKENIZER_TYPE.BPE, "repo": "https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Base", }, | ||||||
|  |     {"name": "a.x-4.0",          "tokt": TOKENIZER_TYPE.BPE, "repo": "https://huggingface.co/skt/A.X-4.0", }, | ||||||
| ] | ] | ||||||
|  |  | ||||||
| # some models are known to be broken upstream, so we will skip them as exceptions | # some models are known to be broken upstream, so we will skip them as exceptions | ||||||
|   | |||||||
| @@ -1556,7 +1556,8 @@ void llama_vocab::impl::load(llama_model_loader & ml, const LLM_KV & kv) { | |||||||
|                     tokenizer_pre == "jina-de" || |                     tokenizer_pre == "jina-de" || | ||||||
|                     tokenizer_pre == "gigachat"   || |                     tokenizer_pre == "gigachat"   || | ||||||
|                     tokenizer_pre == "jina-v2-es" || |                     tokenizer_pre == "jina-v2-es" || | ||||||
|                     tokenizer_pre == "jina-v2-de") { |                     tokenizer_pre == "jina-v2-de" || | ||||||
|  |                     tokenizer_pre == "a.x-4.0") { | ||||||
|                 pre_type = LLAMA_VOCAB_PRE_TYPE_GPT2; |                 pre_type = LLAMA_VOCAB_PRE_TYPE_GPT2; | ||||||
|             } else if ( |             } else if ( | ||||||
|                     tokenizer_pre == "jina-v1-en" || |                     tokenizer_pre == "jina-v1-en" || | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Dowon
					Dowon