|
@@ -65,6 +65,16 @@ if [ ! -z ${GG_BUILD_CUDA} ]; then
|
|
|
fi
|
|
fi
|
|
|
fi
|
|
fi
|
|
|
|
|
|
|
|
|
|
+if [ ! -z ${GG_BUILD_ROCM} ]; then
|
|
|
|
|
+ CMAKE_EXTRA="${CMAKE_EXTRA} -DGGML_HIP=ON"
|
|
|
|
|
+ if [ -z ${GG_BUILD_AMDGPU_TARGETS} ]; then
|
|
|
|
|
+ echo "Missing GG_BUILD_AMDGPU_TARGETS, please set it to your GPU architecture (e.g. gfx90a, gfx1100, etc.)"
|
|
|
|
|
+ exit 1
|
|
|
|
|
+ fi
|
|
|
|
|
+
|
|
|
|
|
+ CMAKE_EXTRA="${CMAKE_EXTRA} -DAMDGPU_TARGETS=${GG_BUILD_AMDGPU_TARGETS}"
|
|
|
|
|
+fi
|
|
|
|
|
+
|
|
|
if [ ! -z ${GG_BUILD_SYCL} ]; then
|
|
if [ ! -z ${GG_BUILD_SYCL} ]; then
|
|
|
if [ -z ${ONEAPI_ROOT} ]; then
|
|
if [ -z ${ONEAPI_ROOT} ]; then
|
|
|
echo "Not detected ONEAPI_ROOT, please install oneAPI base toolkit and enable it by:"
|
|
echo "Not detected ONEAPI_ROOT, please install oneAPI base toolkit and enable it by:"
|
|
@@ -150,7 +160,7 @@ function gg_run_ctest_debug {
|
|
|
(time cmake -DCMAKE_BUILD_TYPE=Debug ${CMAKE_EXTRA} .. ) 2>&1 | tee -a $OUT/${ci}-cmake.log
|
|
(time cmake -DCMAKE_BUILD_TYPE=Debug ${CMAKE_EXTRA} .. ) 2>&1 | tee -a $OUT/${ci}-cmake.log
|
|
|
(time make -j$(nproc) ) 2>&1 | tee -a $OUT/${ci}-make.log
|
|
(time make -j$(nproc) ) 2>&1 | tee -a $OUT/${ci}-make.log
|
|
|
|
|
|
|
|
- (time ctest --output-on-failure -L main -E test-opt ) 2>&1 | tee -a $OUT/${ci}-ctest.log
|
|
|
|
|
|
|
+ (time ctest --output-on-failure -L main -E "test-opt|test-backend-ops" ) 2>&1 | tee -a $OUT/${ci}-ctest.log
|
|
|
|
|
|
|
|
set +e
|
|
set +e
|
|
|
}
|
|
}
|
|
@@ -249,15 +259,9 @@ function gg_sum_test_scripts_release {
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
function gg_get_model {
|
|
function gg_get_model {
|
|
|
- local gguf_0="$MNT/models/pythia/1.4B/ggml-model-f16.gguf"
|
|
|
|
|
- local gguf_1="$MNT/models/pythia/2.8B/ggml-model-f16.gguf"
|
|
|
|
|
- local gguf_2="$MNT/models/open-llama/7B-v2/ggml-model-f16.gguf"
|
|
|
|
|
|
|
+ local gguf_0="$MNT/models/qwen3/0.6B/ggml-model-bf16.gguf"
|
|
|
if [[ -s $gguf_0 ]]; then
|
|
if [[ -s $gguf_0 ]]; then
|
|
|
echo -n "$gguf_0"
|
|
echo -n "$gguf_0"
|
|
|
- elif [[ -s $gguf_1 ]]; then
|
|
|
|
|
- echo -n "$gguf_1"
|
|
|
|
|
- elif [[ -s $gguf_2 ]]; then
|
|
|
|
|
- echo -n "$gguf_2"
|
|
|
|
|
else
|
|
else
|
|
|
echo >&2 "No model found. Can't run gg_run_ctest_with_model."
|
|
echo >&2 "No model found. Can't run gg_run_ctest_with_model."
|
|
|
exit 1
|
|
exit 1
|
|
@@ -316,289 +320,22 @@ function gg_sum_ctest_with_model_release {
|
|
|
gg_printf '```\n'
|
|
gg_printf '```\n'
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
-# open_llama_7b_v2
|
|
|
|
|
|
|
+# qwen3_0_6b
|
|
|
|
|
|
|
|
-function gg_run_open_llama_7b_v2 {
|
|
|
|
|
|
|
+function gg_run_qwen3_0_6b {
|
|
|
cd ${SRC}
|
|
cd ${SRC}
|
|
|
|
|
|
|
|
- gg_wget models-mnt/open-llama/7B-v2/ https://huggingface.co/openlm-research/open_llama_7b_v2/raw/main/config.json
|
|
|
|
|
- gg_wget models-mnt/open-llama/7B-v2/ https://huggingface.co/openlm-research/open_llama_7b_v2/resolve/main/tokenizer.model
|
|
|
|
|
- gg_wget models-mnt/open-llama/7B-v2/ https://huggingface.co/openlm-research/open_llama_7b_v2/raw/main/tokenizer_config.json
|
|
|
|
|
- gg_wget models-mnt/open-llama/7B-v2/ https://huggingface.co/openlm-research/open_llama_7b_v2/raw/main/special_tokens_map.json
|
|
|
|
|
- gg_wget models-mnt/open-llama/7B-v2/ https://huggingface.co/openlm-research/open_llama_7b_v2/raw/main/pytorch_model.bin.index.json
|
|
|
|
|
- gg_wget models-mnt/open-llama/7B-v2/ https://huggingface.co/openlm-research/open_llama_7b_v2/resolve/main/pytorch_model-00001-of-00002.bin
|
|
|
|
|
- gg_wget models-mnt/open-llama/7B-v2/ https://huggingface.co/openlm-research/open_llama_7b_v2/resolve/main/pytorch_model-00002-of-00002.bin
|
|
|
|
|
- gg_wget models-mnt/open-llama/7B-v2/ https://huggingface.co/openlm-research/open_llama_7b_v2/raw/main/generation_config.json
|
|
|
|
|
-
|
|
|
|
|
- gg_wget models-mnt/wikitext/ https://huggingface.co/datasets/ggml-org/ci/resolve/main/wikitext-2-raw-v1.zip
|
|
|
|
|
- unzip -o models-mnt/wikitext/wikitext-2-raw-v1.zip -d models-mnt/wikitext/
|
|
|
|
|
-
|
|
|
|
|
- path_models="../models-mnt/open-llama/7B-v2"
|
|
|
|
|
- path_wiki="../models-mnt/wikitext/wikitext-2-raw"
|
|
|
|
|
-
|
|
|
|
|
- rm -rf build-ci-release && mkdir build-ci-release && cd build-ci-release
|
|
|
|
|
-
|
|
|
|
|
- set -e
|
|
|
|
|
-
|
|
|
|
|
- (time cmake -DCMAKE_BUILD_TYPE=Release ${CMAKE_EXTRA} .. ) 2>&1 | tee -a $OUT/${ci}-cmake.log
|
|
|
|
|
- (time make -j$(nproc) ) 2>&1 | tee -a $OUT/${ci}-make.log
|
|
|
|
|
-
|
|
|
|
|
- python3 ../examples/convert_legacy_llama.py ${path_models} --outfile ${path_models}/ggml-model-f16.gguf
|
|
|
|
|
-
|
|
|
|
|
- model_f16="${path_models}/ggml-model-f16.gguf"
|
|
|
|
|
- model_q8_0="${path_models}/ggml-model-q8_0.gguf"
|
|
|
|
|
- model_q4_0="${path_models}/ggml-model-q4_0.gguf"
|
|
|
|
|
- model_q4_1="${path_models}/ggml-model-q4_1.gguf"
|
|
|
|
|
- model_q5_0="${path_models}/ggml-model-q5_0.gguf"
|
|
|
|
|
- model_q5_1="${path_models}/ggml-model-q5_1.gguf"
|
|
|
|
|
- model_q2_k="${path_models}/ggml-model-q2_k.gguf"
|
|
|
|
|
- model_q3_k="${path_models}/ggml-model-q3_k.gguf"
|
|
|
|
|
- model_q4_k="${path_models}/ggml-model-q4_k.gguf"
|
|
|
|
|
- model_q5_k="${path_models}/ggml-model-q5_k.gguf"
|
|
|
|
|
- model_q6_k="${path_models}/ggml-model-q6_k.gguf"
|
|
|
|
|
-
|
|
|
|
|
- wiki_test="${path_wiki}/wiki.test.raw"
|
|
|
|
|
-
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q8_0} q8_0
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q4_0} q4_0
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q4_1} q4_1
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q5_0} q5_0
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q5_1} q5_1
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q2_k} q2_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q3_k} q3_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q4_k} q4_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q5_k} q5_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q6_k} q6_k
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_f16} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-f16.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q8_0} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q8_0.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q4_0} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_0.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q4_1} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_1.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q5_0} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_0.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q5_1} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_1.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q2_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q2_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q3_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q3_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q4_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q5_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q6_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q6_k.log
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_f16} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-f16.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q8_0} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q8_0.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q4_0} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_0.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q4_1} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_1.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q5_0} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q5_0.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q5_1} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q5_1.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q2_k} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q2_k.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q3_k} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q3_k.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q4_k} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_k.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q5_k} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q5_k.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q6_k} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q6_k.log
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-imatrix --model ${model_f16} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-imatrix.log
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 10 -c 0 -fa off ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 10 -c 0 -fa on ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 99 -c 0 -fa off ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 99 -c 0 -fa on ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
-
|
|
|
|
|
- function check_ppl {
|
|
|
|
|
- qnt="$1"
|
|
|
|
|
- ppl=$(echo "$2" | grep -oE "[0-9]+\.[0-9]+" | tail -n 1)
|
|
|
|
|
-
|
|
|
|
|
- if [ $(echo "$ppl > 20.0" | bc) -eq 1 ]; then
|
|
|
|
|
- printf ' - %s @ %s (FAIL: ppl > 20.0)\n' "$qnt" "$ppl"
|
|
|
|
|
- return 20
|
|
|
|
|
- fi
|
|
|
|
|
-
|
|
|
|
|
- printf ' - %s @ %s OK\n' "$qnt" "$ppl"
|
|
|
|
|
- return 0
|
|
|
|
|
- }
|
|
|
|
|
-
|
|
|
|
|
- check_ppl "f16" "$(cat $OUT/${ci}-tg-f16.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q8_0" "$(cat $OUT/${ci}-tg-q8_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q4_0" "$(cat $OUT/${ci}-tg-q4_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q4_1" "$(cat $OUT/${ci}-tg-q4_1.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q5_0" "$(cat $OUT/${ci}-tg-q5_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q5_1" "$(cat $OUT/${ci}-tg-q5_1.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q2_k" "$(cat $OUT/${ci}-tg-q2_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q3_k" "$(cat $OUT/${ci}-tg-q3_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q4_k" "$(cat $OUT/${ci}-tg-q4_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q5_k" "$(cat $OUT/${ci}-tg-q5_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q6_k" "$(cat $OUT/${ci}-tg-q6_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
-
|
|
|
|
|
- cat $OUT/${ci}-imatrix.log | grep "Final" >> $OUT/${ci}-imatrix-sum.log
|
|
|
|
|
-
|
|
|
|
|
- set +e
|
|
|
|
|
-}
|
|
|
|
|
-
|
|
|
|
|
-function gg_sum_open_llama_7b_v2 {
|
|
|
|
|
- gg_printf '### %s\n\n' "${ci}"
|
|
|
|
|
-
|
|
|
|
|
- gg_printf 'OpenLLaMA 7B-v2:\n'
|
|
|
|
|
- gg_printf '- status: %s\n' "$(cat $OUT/${ci}.exit)"
|
|
|
|
|
- gg_printf '- perplexity:\n%s\n' "$(cat $OUT/${ci}-ppl.log)"
|
|
|
|
|
- gg_printf '- imatrix:\n```\n%s\n```\n' "$(cat $OUT/${ci}-imatrix-sum.log)"
|
|
|
|
|
- gg_printf '- f16: \n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-f16.log)"
|
|
|
|
|
- gg_printf '- q8_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q8_0.log)"
|
|
|
|
|
- gg_printf '- q4_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_0.log)"
|
|
|
|
|
- gg_printf '- q4_1:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_1.log)"
|
|
|
|
|
- gg_printf '- q5_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q5_0.log)"
|
|
|
|
|
- gg_printf '- q5_1:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q5_1.log)"
|
|
|
|
|
- gg_printf '- q2_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q2_k.log)"
|
|
|
|
|
- gg_printf '- q3_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q3_k.log)"
|
|
|
|
|
- gg_printf '- q4_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_k.log)"
|
|
|
|
|
- gg_printf '- q5_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q5_k.log)"
|
|
|
|
|
- gg_printf '- q6_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q6_k.log)"
|
|
|
|
|
- gg_printf '- save-load-state: \n```\n%s\n```\n' "$(cat $OUT/${ci}-save-load-state.log)"
|
|
|
|
|
-}
|
|
|
|
|
|
|
+ gg_wget models-mnt/qwen3/0.6B/ https://huggingface.co/Qwen/Qwen3-0.6B-Base/raw/main/config.json
|
|
|
|
|
+ gg_wget models-mnt/qwen3/0.6B/ https://huggingface.co/Qwen/Qwen3-0.6B-Base/raw/main/tokenizer.json
|
|
|
|
|
+ gg_wget models-mnt/qwen3/0.6B/ https://huggingface.co/Qwen/Qwen3-0.6B-Base/raw/main/tokenizer_config.json
|
|
|
|
|
+ #gg_wget models-mnt/qwen3/0.6B/ https://huggingface.co/Qwen/Qwen3-0.6B-Base/raw/main/special_tokens_map.json
|
|
|
|
|
+ gg_wget models-mnt/qwen3/0.6B/ https://huggingface.co/Qwen/Qwen3-0.6B-Base/resolve/main/model.safetensors
|
|
|
|
|
|
|
|
-# pythia_1.4b
|
|
|
|
|
-
|
|
|
|
|
-function gg_run_pythia_1_4b {
|
|
|
|
|
- cd ${SRC}
|
|
|
|
|
-
|
|
|
|
|
- gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/raw/main/config.json
|
|
|
|
|
- gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/raw/main/tokenizer.json
|
|
|
|
|
- gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/raw/main/tokenizer_config.json
|
|
|
|
|
- gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/raw/main/special_tokens_map.json
|
|
|
|
|
- gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/resolve/main/pytorch_model.bin
|
|
|
|
|
-
|
|
|
|
|
- gg_wget models-mnt/wikitext/ https://huggingface.co/datasets/ggml-org/ci/resolve/main/wikitext-2-raw-v1.zip
|
|
|
|
|
- unzip -o models-mnt/wikitext/wikitext-2-raw-v1.zip -d models-mnt/wikitext/
|
|
|
|
|
- head -n 60 models-mnt/wikitext/wikitext-2-raw/wiki.test.raw > models-mnt/wikitext/wikitext-2-raw/wiki.test-60.raw
|
|
|
|
|
-
|
|
|
|
|
- path_models="../models-mnt/pythia/1.4B"
|
|
|
|
|
- path_wiki="../models-mnt/wikitext/wikitext-2-raw"
|
|
|
|
|
-
|
|
|
|
|
- rm -rf build-ci-release && mkdir build-ci-release && cd build-ci-release
|
|
|
|
|
-
|
|
|
|
|
- set -e
|
|
|
|
|
-
|
|
|
|
|
- (time cmake -DCMAKE_BUILD_TYPE=Release ${CMAKE_EXTRA} .. ) 2>&1 | tee -a $OUT/${ci}-cmake.log
|
|
|
|
|
- (time make -j$(nproc) ) 2>&1 | tee -a $OUT/${ci}-make.log
|
|
|
|
|
-
|
|
|
|
|
- python3 ../convert_hf_to_gguf.py ${path_models} --outfile ${path_models}/ggml-model-f16.gguf
|
|
|
|
|
-
|
|
|
|
|
- model_f16="${path_models}/ggml-model-f16.gguf"
|
|
|
|
|
- model_q8_0="${path_models}/ggml-model-q8_0.gguf"
|
|
|
|
|
- model_q4_0="${path_models}/ggml-model-q4_0.gguf"
|
|
|
|
|
- model_q4_1="${path_models}/ggml-model-q4_1.gguf"
|
|
|
|
|
- model_q5_0="${path_models}/ggml-model-q5_0.gguf"
|
|
|
|
|
- model_q5_1="${path_models}/ggml-model-q5_1.gguf"
|
|
|
|
|
- model_q2_k="${path_models}/ggml-model-q2_k.gguf"
|
|
|
|
|
- model_q3_k="${path_models}/ggml-model-q3_k.gguf"
|
|
|
|
|
- model_q4_k="${path_models}/ggml-model-q4_k.gguf"
|
|
|
|
|
- model_q5_k="${path_models}/ggml-model-q5_k.gguf"
|
|
|
|
|
- model_q6_k="${path_models}/ggml-model-q6_k.gguf"
|
|
|
|
|
-
|
|
|
|
|
- wiki_test_60="${path_wiki}/wiki.test-60.raw"
|
|
|
|
|
-
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q8_0} q8_0
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q4_0} q4_0
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q4_1} q4_1
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q5_0} q5_0
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q5_1} q5_1
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q2_k} q2_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q3_k} q3_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q4_k} q4_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q5_k} q5_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q6_k} q6_k
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_f16} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-f16.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q8_0} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q8_0.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q4_0} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_0.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q4_1} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_1.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q5_0} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_0.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q5_1} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_1.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q2_k} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q2_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q3_k} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q3_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q4_k} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q5_k} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q6_k} -ngl 99 -c 0 -s 1234 -n 64 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q6_k.log
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_f16} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-f16.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q8_0} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q8_0.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q4_0} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_0.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q4_1} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_1.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q5_0} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q5_0.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q5_1} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q5_1.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q2_k} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q2_k.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q3_k} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q3_k.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q4_k} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_k.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q5_k} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q5_k.log
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_q6_k} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-tg-q6_k.log
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-imatrix --model ${model_f16} -f ${wiki_test_60} -ngl 99 -c 128 -b 128 --chunks 1 ) 2>&1 | tee -a $OUT/${ci}-imatrix.log
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 99 -c 0 -fa off ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 99 -c 0 -fa on ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
-
|
|
|
|
|
- function check_ppl {
|
|
|
|
|
- qnt="$1"
|
|
|
|
|
- ppl=$(echo "$2" | grep -oE "[0-9]+\.[0-9]+" | tail -n 1)
|
|
|
|
|
-
|
|
|
|
|
- if [ $(echo "$ppl > 20.0" | bc) -eq 1 ]; then
|
|
|
|
|
- printf ' - %s @ %s (FAIL: ppl > 20.0)\n' "$qnt" "$ppl"
|
|
|
|
|
- return 20
|
|
|
|
|
- fi
|
|
|
|
|
-
|
|
|
|
|
- printf ' - %s @ %s OK\n' "$qnt" "$ppl"
|
|
|
|
|
- return 0
|
|
|
|
|
- }
|
|
|
|
|
-
|
|
|
|
|
- check_ppl "f16" "$(cat $OUT/${ci}-tg-f16.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q8_0" "$(cat $OUT/${ci}-tg-q8_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q4_0" "$(cat $OUT/${ci}-tg-q4_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q4_1" "$(cat $OUT/${ci}-tg-q4_1.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q5_0" "$(cat $OUT/${ci}-tg-q5_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q5_1" "$(cat $OUT/${ci}-tg-q5_1.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- #check_ppl "q2_k" "$(cat $OUT/${ci}-tg-q2_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log # note: ppl > 20.0 for this quant and model
|
|
|
|
|
- check_ppl "q3_k" "$(cat $OUT/${ci}-tg-q3_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q4_k" "$(cat $OUT/${ci}-tg-q4_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q5_k" "$(cat $OUT/${ci}-tg-q5_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
- check_ppl "q6_k" "$(cat $OUT/${ci}-tg-q6_k.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
-
|
|
|
|
|
- cat $OUT/${ci}-imatrix.log | grep "Final" >> $OUT/${ci}-imatrix-sum.log
|
|
|
|
|
-
|
|
|
|
|
- set +e
|
|
|
|
|
-}
|
|
|
|
|
-
|
|
|
|
|
-function gg_sum_pythia_1_4b {
|
|
|
|
|
- gg_printf '### %s\n\n' "${ci}"
|
|
|
|
|
-
|
|
|
|
|
- gg_printf 'Pythia 1.4B:\n'
|
|
|
|
|
- gg_printf '- status: %s\n' "$(cat $OUT/${ci}.exit)"
|
|
|
|
|
- gg_printf '- perplexity:\n%s\n' "$(cat $OUT/${ci}-ppl.log)"
|
|
|
|
|
- gg_printf '- imatrix:\n```\n%s\n```\n' "$(cat $OUT/${ci}-imatrix-sum.log)"
|
|
|
|
|
- gg_printf '- f16: \n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-f16.log)"
|
|
|
|
|
- gg_printf '- q8_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q8_0.log)"
|
|
|
|
|
- gg_printf '- q4_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_0.log)"
|
|
|
|
|
- gg_printf '- q4_1:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_1.log)"
|
|
|
|
|
- gg_printf '- q5_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q5_0.log)"
|
|
|
|
|
- gg_printf '- q5_1:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q5_1.log)"
|
|
|
|
|
- gg_printf '- q2_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q2_k.log)"
|
|
|
|
|
- gg_printf '- q3_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q3_k.log)"
|
|
|
|
|
- gg_printf '- q4_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_k.log)"
|
|
|
|
|
- gg_printf '- q5_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q5_k.log)"
|
|
|
|
|
- gg_printf '- q6_k:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q6_k.log)"
|
|
|
|
|
- gg_printf '- save-load-state: \n```\n%s\n```\n' "$(cat $OUT/${ci}-save-load-state.log)"
|
|
|
|
|
-}
|
|
|
|
|
-
|
|
|
|
|
-# pythia_2_8b
|
|
|
|
|
-
|
|
|
|
|
-function gg_run_pythia_2_8b {
|
|
|
|
|
- cd ${SRC}
|
|
|
|
|
-
|
|
|
|
|
- gg_wget models-mnt/pythia/2.8B/ https://huggingface.co/EleutherAI/pythia-2.8b/raw/main/config.json
|
|
|
|
|
- gg_wget models-mnt/pythia/2.8B/ https://huggingface.co/EleutherAI/pythia-2.8b/raw/main/tokenizer.json
|
|
|
|
|
- gg_wget models-mnt/pythia/2.8B/ https://huggingface.co/EleutherAI/pythia-2.8b/raw/main/tokenizer_config.json
|
|
|
|
|
- gg_wget models-mnt/pythia/2.8B/ https://huggingface.co/EleutherAI/pythia-2.8b/raw/main/special_tokens_map.json
|
|
|
|
|
- gg_wget models-mnt/pythia/2.8B/ https://huggingface.co/EleutherAI/pythia-2.8b/resolve/main/pytorch_model.bin
|
|
|
|
|
|
|
|
|
|
gg_wget models-mnt/wikitext/ https://huggingface.co/datasets/ggml-org/ci/resolve/main/wikitext-2-raw-v1.zip
|
|
gg_wget models-mnt/wikitext/ https://huggingface.co/datasets/ggml-org/ci/resolve/main/wikitext-2-raw-v1.zip
|
|
|
unzip -o models-mnt/wikitext/wikitext-2-raw-v1.zip -d models-mnt/wikitext/
|
|
unzip -o models-mnt/wikitext/wikitext-2-raw-v1.zip -d models-mnt/wikitext/
|
|
|
|
|
|
|
|
- path_models="../models-mnt/pythia/2.8B"
|
|
|
|
|
|
|
+ path_models="../models-mnt/qwen3/0.6B"
|
|
|
path_wiki="../models-mnt/wikitext/wikitext-2-raw"
|
|
path_wiki="../models-mnt/wikitext/wikitext-2-raw"
|
|
|
|
|
|
|
|
rm -rf build-ci-release && mkdir build-ci-release && cd build-ci-release
|
|
rm -rf build-ci-release && mkdir build-ci-release && cd build-ci-release
|
|
@@ -608,9 +345,9 @@ function gg_run_pythia_2_8b {
|
|
|
(time cmake -DCMAKE_BUILD_TYPE=Release ${CMAKE_EXTRA} .. ) 2>&1 | tee -a $OUT/${ci}-cmake.log
|
|
(time cmake -DCMAKE_BUILD_TYPE=Release ${CMAKE_EXTRA} .. ) 2>&1 | tee -a $OUT/${ci}-cmake.log
|
|
|
(time make -j$(nproc) ) 2>&1 | tee -a $OUT/${ci}-make.log
|
|
(time make -j$(nproc) ) 2>&1 | tee -a $OUT/${ci}-make.log
|
|
|
|
|
|
|
|
- python3 ../convert_hf_to_gguf.py ${path_models} --outfile ${path_models}/ggml-model-f16.gguf
|
|
|
|
|
|
|
+ python3 ../convert_hf_to_gguf.py ${path_models} --outfile ${path_models}/ggml-model-bf16.gguf --outtype bf16
|
|
|
|
|
|
|
|
- model_f16="${path_models}/ggml-model-f16.gguf"
|
|
|
|
|
|
|
+ model_bf16="${path_models}/ggml-model-bf16.gguf"
|
|
|
model_q8_0="${path_models}/ggml-model-q8_0.gguf"
|
|
model_q8_0="${path_models}/ggml-model-q8_0.gguf"
|
|
|
model_q4_0="${path_models}/ggml-model-q4_0.gguf"
|
|
model_q4_0="${path_models}/ggml-model-q4_0.gguf"
|
|
|
model_q4_1="${path_models}/ggml-model-q4_1.gguf"
|
|
model_q4_1="${path_models}/ggml-model-q4_1.gguf"
|
|
@@ -624,30 +361,30 @@ function gg_run_pythia_2_8b {
|
|
|
|
|
|
|
|
wiki_test="${path_wiki}/wiki.test.raw"
|
|
wiki_test="${path_wiki}/wiki.test.raw"
|
|
|
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q8_0} q8_0
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q4_0} q4_0
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q4_1} q4_1
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q5_0} q5_0
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q5_1} q5_1
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q2_k} q2_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q3_k} q3_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q4_k} q4_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q5_k} q5_k
|
|
|
|
|
- ./bin/llama-quantize ${model_f16} ${model_q6_k} q6_k
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_f16} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-f16.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q8_0} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q8_0.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q4_0} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_0.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q4_1} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_1.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q5_0} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_0.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q5_1} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_1.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q2_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q2_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q3_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q3_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q4_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q5_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_k.log
|
|
|
|
|
- (time ./bin/llama-cli -no-cnv --model ${model_q6_k} -t 1 -ngl 99 -c 0 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q6_k.log
|
|
|
|
|
-
|
|
|
|
|
- (time ./bin/llama-perplexity --model ${model_f16} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-f16.log
|
|
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q8_0} q8_0
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q4_0} q4_0
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q4_1} q4_1
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q5_0} q5_0
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q5_1} q5_1
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q2_k} q2_k
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q3_k} q3_k
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q4_k} q4_k
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q5_k} q5_k
|
|
|
|
|
+ ./bin/llama-quantize ${model_bf16} ${model_q6_k} q6_k
|
|
|
|
|
+
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_bf16} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-bf16.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q8_0} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q8_0.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q4_0} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_0.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q4_1} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_1.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q5_0} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_0.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q5_1} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_1.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q2_k} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q2_k.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q3_k} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q3_k.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q4_k} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q4_k.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q5_k} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q5_k.log
|
|
|
|
|
+ (time ./bin/llama-cli -no-cnv --model ${model_q6_k} -t 1 -ngl 99 -c 2048 -s 1234 -n 256 --ignore-eos -p "I believe the meaning of life is" ) 2>&1 | tee -a $OUT/${ci}-tg-q6_k.log
|
|
|
|
|
+
|
|
|
|
|
+ (time ./bin/llama-perplexity --model ${model_bf16} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-bf16.log
|
|
|
(time ./bin/llama-perplexity --model ${model_q8_0} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q8_0.log
|
|
(time ./bin/llama-perplexity --model ${model_q8_0} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q8_0.log
|
|
|
(time ./bin/llama-perplexity --model ${model_q4_0} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_0.log
|
|
(time ./bin/llama-perplexity --model ${model_q4_0} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_0.log
|
|
|
(time ./bin/llama-perplexity --model ${model_q4_1} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_1.log
|
|
(time ./bin/llama-perplexity --model ${model_q4_1} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q4_1.log
|
|
@@ -659,12 +396,12 @@ function gg_run_pythia_2_8b {
|
|
|
(time ./bin/llama-perplexity --model ${model_q5_k} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q5_k.log
|
|
(time ./bin/llama-perplexity --model ${model_q5_k} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q5_k.log
|
|
|
(time ./bin/llama-perplexity --model ${model_q6_k} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q6_k.log
|
|
(time ./bin/llama-perplexity --model ${model_q6_k} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-tg-q6_k.log
|
|
|
|
|
|
|
|
- (time ./bin/llama-imatrix --model ${model_f16} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-imatrix.log
|
|
|
|
|
|
|
+ (time ./bin/llama-imatrix --model ${model_bf16} -f ${wiki_test} -t 1 -ngl 99 -c 2048 -b 512 --chunks 4 ) 2>&1 | tee -a $OUT/${ci}-imatrix.log
|
|
|
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 10 -c 0 -fa off ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 10 -c 0 -fa on ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 99 -c 0 -fa off ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
- (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 99 -c 0 -fa on ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
|
|
+ (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 10 -c 2048 -fa off ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
+ (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 10 -c 2048 -fa on ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
+ (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 99 -c 2048 -fa off ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
+ (time ./bin/llama-save-load-state --model ${model_q4_0} -ngl 99 -c 2048 -fa on ) 2>&1 | tee -a $OUT/${ci}-save-load-state.log
|
|
|
|
|
|
|
|
function check_ppl {
|
|
function check_ppl {
|
|
|
qnt="$1"
|
|
qnt="$1"
|
|
@@ -679,7 +416,7 @@ function gg_run_pythia_2_8b {
|
|
|
return 0
|
|
return 0
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
- check_ppl "f16" "$(cat $OUT/${ci}-tg-f16.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
|
|
|
|
+ check_ppl "bf16" "$(cat $OUT/${ci}-tg-bf16.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
check_ppl "q8_0" "$(cat $OUT/${ci}-tg-q8_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
check_ppl "q8_0" "$(cat $OUT/${ci}-tg-q8_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
check_ppl "q4_0" "$(cat $OUT/${ci}-tg-q4_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
check_ppl "q4_0" "$(cat $OUT/${ci}-tg-q4_0.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
|
check_ppl "q4_1" "$(cat $OUT/${ci}-tg-q4_1.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
check_ppl "q4_1" "$(cat $OUT/${ci}-tg-q4_1.log | grep "^\[1\]")" | tee -a $OUT/${ci}-ppl.log
|
|
@@ -696,14 +433,14 @@ function gg_run_pythia_2_8b {
|
|
|
set +e
|
|
set +e
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
-function gg_sum_pythia_2_8b {
|
|
|
|
|
|
|
+function gg_sum_qwen3_0_6b {
|
|
|
gg_printf '### %s\n\n' "${ci}"
|
|
gg_printf '### %s\n\n' "${ci}"
|
|
|
|
|
|
|
|
gg_printf 'Pythia 2.8B:\n'
|
|
gg_printf 'Pythia 2.8B:\n'
|
|
|
gg_printf '- status: %s\n' "$(cat $OUT/${ci}.exit)"
|
|
gg_printf '- status: %s\n' "$(cat $OUT/${ci}.exit)"
|
|
|
gg_printf '- perplexity:\n%s\n' "$(cat $OUT/${ci}-ppl.log)"
|
|
gg_printf '- perplexity:\n%s\n' "$(cat $OUT/${ci}-ppl.log)"
|
|
|
gg_printf '- imatrix:\n```\n%s\n```\n' "$(cat $OUT/${ci}-imatrix-sum.log)"
|
|
gg_printf '- imatrix:\n```\n%s\n```\n' "$(cat $OUT/${ci}-imatrix-sum.log)"
|
|
|
- gg_printf '- f16: \n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-f16.log)"
|
|
|
|
|
|
|
+ gg_printf '- bf16:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-bf16.log)"
|
|
|
gg_printf '- q8_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q8_0.log)"
|
|
gg_printf '- q8_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q8_0.log)"
|
|
|
gg_printf '- q4_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_0.log)"
|
|
gg_printf '- q4_0:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_0.log)"
|
|
|
gg_printf '- q4_1:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_1.log)"
|
|
gg_printf '- q4_1:\n```\n%s\n```\n' "$(cat $OUT/${ci}-tg-q4_1.log)"
|
|
@@ -882,16 +619,10 @@ if [ -z ${GG_BUILD_LOW_PERF} ]; then
|
|
|
test $ret -eq 0 && gg_run test_scripts_release
|
|
test $ret -eq 0 && gg_run test_scripts_release
|
|
|
fi
|
|
fi
|
|
|
|
|
|
|
|
- if [ -z ${GG_BUILD_VRAM_GB} ] || [ ${GG_BUILD_VRAM_GB} -ge 8 ]; then
|
|
|
|
|
- if [ -z ${GG_BUILD_CUDA} ] && [ -z ${GG_BUILD_VULKAN} ]; then
|
|
|
|
|
- test $ret -eq 0 && gg_run pythia_1_4b
|
|
|
|
|
- else
|
|
|
|
|
- test $ret -eq 0 && gg_run pythia_2_8b
|
|
|
|
|
- #test $ret -eq 0 && gg_run open_llama_7b_v2
|
|
|
|
|
- fi
|
|
|
|
|
- test $ret -eq 0 && gg_run ctest_with_model_debug
|
|
|
|
|
- test $ret -eq 0 && gg_run ctest_with_model_release
|
|
|
|
|
- fi
|
|
|
|
|
|
|
+ test $ret -eq 0 && gg_run qwen3_0_6b
|
|
|
|
|
+
|
|
|
|
|
+ test $ret -eq 0 && gg_run ctest_with_model_debug
|
|
|
|
|
+ test $ret -eq 0 && gg_run ctest_with_model_release
|
|
|
fi
|
|
fi
|
|
|
|
|
|
|
|
exit $ret
|
|
exit $ret
|