Explorar o código

Remove .attention from skipped tensors to match more accurately (#7051)

Bartowski hai 1 ano
pai
achega
60325fa56f
Modificáronse 1 ficheiros con 1 adicións e 1 borrados
  1. 1 1
      convert-hf-to-gguf.py

+ 1 - 1
convert-hf-to-gguf.py

@@ -1427,7 +1427,7 @@ class LlamaModel(Model):
         experts = dict()
         experts = dict()
         for name, data_torch in self.get_tensors():
         for name, data_torch in self.get_tensors():
             # we don't need these
             # we don't need these
-            if name.endswith((".attention.masked_bias", ".attention.bias", ".attention.rotary_emb.inv_freq")):
+            if name.endswith((".attention.masked_bias", ".attention.bias", ".rotary_emb.inv_freq")):
                 continue
                 continue
 
 
             old_dtype = data_torch.dtype
             old_dtype = data_torch.dtype