|
@@ -10,8 +10,8 @@ Proof of concept:
|
|
|
|
|
|
|
|
``` sh
|
|
``` sh
|
|
|
export model_name=llama_3.2-1b && export quantization=f32
|
|
export model_name=llama_3.2-1b && export quantization=f32
|
|
|
-./build/bin/finetune --file wikitext-2-raw/wiki.test.raw -ngl 999 --model models/${model_name}-${quantization}.gguf -c 512 -b 512 -ub 512
|
|
|
|
|
-./build/bin/perplexity --file wikitext-2-raw/wiki.test.raw -ngl 999 --model finetuned-model.gguf
|
|
|
|
|
|
|
+./build/bin/llama-finetune --file wikitext-2-raw/wiki.test.raw -ngl 999 --model models/${model_name}-${quantization}.gguf -c 512 -b 512 -ub 512
|
|
|
|
|
+./build/bin/llama-perplexity --file wikitext-2-raw/wiki.test.raw -ngl 999 --model finetuned-model.gguf
|
|
|
```
|
|
```
|
|
|
|
|
|
|
|
The perplexity value of the finetuned model should be lower after training on the test set for 2 epochs.
|
|
The perplexity value of the finetuned model should be lower after training on the test set for 2 epochs.
|