mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-10-28 08:31:25 +00:00
convert : Support Qwen2_5_VLForConditionalGeneration (#12595)
This commit is contained in:
@@ -2269,7 +2269,7 @@ class Qwen2Model(Model):
|
|||||||
self.gguf_writer.add_rope_scaling_orig_ctx_len(self.hparams["rope_scaling"]["original_max_position_embeddings"])
|
self.gguf_writer.add_rope_scaling_orig_ctx_len(self.hparams["rope_scaling"]["original_max_position_embeddings"])
|
||||||
|
|
||||||
|
|
||||||
@Model.register("Qwen2VLForConditionalGeneration")
|
@Model.register("Qwen2VLForConditionalGeneration", "Qwen2_5_VLForConditionalGeneration")
|
||||||
class Qwen2VLModel(Model):
|
class Qwen2VLModel(Model):
|
||||||
model_arch = gguf.MODEL_ARCH.QWEN2VL
|
model_arch = gguf.MODEL_ARCH.QWEN2VL
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user