mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-01 09:01:57 +00:00
convert : fix conversion from FP8 for Deepseek-V3.1-Base
This commit is contained in:
@@ -254,6 +254,8 @@ class ModelBase:
|
|||||||
# TODO: make sure it's a list of integers
|
# TODO: make sure it's a list of integers
|
||||||
for i, size in enumerate(weight_block_size):
|
for i, size in enumerate(weight_block_size):
|
||||||
scale = scale.repeat_interleave(size, i)
|
scale = scale.repeat_interleave(size, i)
|
||||||
|
# unpad the scale (e.g. when the tensor size isn't a multiple of the block size)
|
||||||
|
scale = scale[tuple(slice(0, size) for size in weight.shape)]
|
||||||
|
|
||||||
return weight.float() * scale
|
return weight.float() * scale
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user