|
|
@@ -1992,6 +1992,14 @@ class Qwen2Model(Model):
|
|
|
except FileNotFoundError:
|
|
|
self._set_vocab_gpt2()
|
|
|
|
|
|
+ def set_gguf_parameters(self):
|
|
|
+ super().set_gguf_parameters()
|
|
|
+ if self.hparams.get("rope_scaling") is not None and "factor" in self.hparams["rope_scaling"]:
|
|
|
+ if self.hparams["rope_scaling"].get("type") == "yarn":
|
|
|
+ self.gguf_writer.add_rope_scaling_type(gguf.RopeScalingType.YARN)
|
|
|
+ self.gguf_writer.add_rope_scaling_factor(self.hparams["rope_scaling"]["factor"])
|
|
|
+ self.gguf_writer.add_rope_scaling_orig_ctx_len(self.hparams["rope_scaling"]["original_max_position_embeddings"])
|
|
|
+
|
|
|
|
|
|
@Model.register("Qwen2MoeForCausalLM")
|
|
|
class Qwen2MoeModel(Model):
|