|
@@ -1095,6 +1095,7 @@ void llama_model::load_hparams(llama_model_loader & ml) {
|
|
|
ml.get_key(LLM_KV_ATTENTION_LAYERNORM_RMS_EPS, hparams.f_norm_rms_eps);
|
|
ml.get_key(LLM_KV_ATTENTION_LAYERNORM_RMS_EPS, hparams.f_norm_rms_eps);
|
|
|
|
|
|
|
|
switch (hparams.n_layer) {
|
|
switch (hparams.n_layer) {
|
|
|
|
|
+ case 18: type = LLM_TYPE_537M; break;
|
|
|
case 26: type = LLM_TYPE_1B; break;
|
|
case 26: type = LLM_TYPE_1B; break;
|
|
|
case 34: type = LLM_TYPE_4B; break;
|
|
case 34: type = LLM_TYPE_4B; break;
|
|
|
case 48: type = LLM_TYPE_12B; break;
|
|
case 48: type = LLM_TYPE_12B; break;
|