Kaynağa Gözat

convert : force patch_merger tensors to f16/f32 (#18124)

Sigbjørn Skjæret 1 ay önce
ebeveyn
işleme
5166aaf868
1 değiştirilmiş dosya ile 1 ekleme ve 1 silme
  1. 1 1
      convert_hf_to_gguf.py

+ 1 - 1
convert_hf_to_gguf.py

@@ -1838,7 +1838,7 @@ class MmprojModel(ModelBase):
 
 
     def tensor_force_quant(self, name, new_name, bid, n_dims):
     def tensor_force_quant(self, name, new_name, bid, n_dims):
         del bid, name, n_dims  # unused
         del bid, name, n_dims  # unused
-        if ".patch_embd.weight" in new_name:
+        if ".patch_embd.weight" in new_name or ".patch_merger.weight" in new_name:
             return gguf.GGMLQuantizationType.F16 if self.ftype == gguf.LlamaFileType.MOSTLY_F16 else gguf.GGMLQuantizationType.F32
             return gguf.GGMLQuantizationType.F16 if self.ftype == gguf.LlamaFileType.MOSTLY_F16 else gguf.GGMLQuantizationType.F32
         return False
         return False