1
0

tests.sh 5.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164
  1. #!/usr/bin/env bash
  2. # make sure we are in the right directory
  3. SCRIPT_DIR=$( cd -- "$( dirname -- "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )
  4. cd $SCRIPT_DIR
  5. #export LLAMA_CACHE="$SCRIPT_DIR/tmp"
  6. set -eux
  7. mkdir -p $SCRIPT_DIR/output
  8. PROJ_ROOT="$SCRIPT_DIR/../.."
  9. cd $PROJ_ROOT
  10. # Check if the first argument is "big", then run test with big models
  11. # This is useful if we're running the script on a larger machine, so we can test the big models
  12. RUN_BIG_TESTS=false
  13. if [ "${1:-}" = "big" ]; then
  14. RUN_BIG_TESTS=true
  15. echo "Include BIG models..."
  16. fi
  17. RUN_HUGE_TESTS=false
  18. if [ "${1:-}" = "huge" ]; then
  19. RUN_HUGE_TESTS=true
  20. RUN_BIG_TESTS=true
  21. echo "Include BIG and HUGE models..."
  22. fi
  23. ###############
  24. arr_prefix=()
  25. arr_hf=()
  26. arr_tmpl=() # chat template
  27. arr_file=()
  28. add_test_vision() {
  29. local hf=$1
  30. local tmpl=${2:-""} # default to empty string if not provided
  31. arr_prefix+=("[vision]")
  32. arr_hf+=("$hf")
  33. arr_tmpl+=("$tmpl")
  34. arr_file+=("test-1.jpeg")
  35. }
  36. add_test_audio() {
  37. local hf=$1
  38. arr_prefix+=("[audio] ")
  39. arr_hf+=("$hf")
  40. arr_tmpl+=("") # no need for chat tmpl
  41. arr_file+=("test-2.mp3")
  42. }
  43. add_test_vision "ggml-org/SmolVLM-500M-Instruct-GGUF:Q8_0"
  44. add_test_vision "ggml-org/SmolVLM2-2.2B-Instruct-GGUF:Q4_K_M"
  45. add_test_vision "ggml-org/SmolVLM2-500M-Video-Instruct-GGUF:Q8_0"
  46. add_test_vision "ggml-org/gemma-3-4b-it-GGUF:Q4_K_M"
  47. add_test_vision "THUDM/glm-edge-v-5b-gguf:Q4_K_M"
  48. add_test_vision "second-state/Llava-v1.5-7B-GGUF:Q2_K" "vicuna"
  49. add_test_vision "cjpais/llava-1.6-mistral-7b-gguf:Q3_K_M" "vicuna"
  50. add_test_vision "ibm-research/granite-vision-3.2-2b-GGUF:Q4_K_M"
  51. add_test_vision "second-state/MiniCPM-Llama3-V-2_5-GGUF:Q2_K" # model from openbmb is corrupted
  52. add_test_vision "openbmb/MiniCPM-V-2_6-gguf:Q2_K"
  53. add_test_vision "openbmb/MiniCPM-o-2_6-gguf:Q4_0"
  54. add_test_vision "bartowski/Qwen2-VL-2B-Instruct-GGUF:Q4_K_M"
  55. add_test_vision "ggml-org/Qwen2.5-VL-3B-Instruct-GGUF:Q4_K_M"
  56. add_test_vision "ggml-org/InternVL2_5-1B-GGUF:Q8_0"
  57. add_test_vision "ggml-org/InternVL3-1B-Instruct-GGUF:Q8_0"
  58. add_test_vision "ggml-org/Qwen2.5-Omni-3B-GGUF:Q4_K_M"
  59. add_test_vision "ggml-org/LFM2-VL-450M-GGUF:Q8_0"
  60. add_test_vision "ggml-org/granite-docling-258M-GGUF:Q8_0"
  61. add_test_audio "ggml-org/ultravox-v0_5-llama-3_2-1b-GGUF:Q8_0"
  62. add_test_audio "ggml-org/Qwen2.5-Omni-3B-GGUF:Q4_K_M"
  63. add_test_audio "ggml-org/Voxtral-Mini-3B-2507-GGUF:Q4_K_M"
  64. # to test the big models, run: ./tests.sh big
  65. if [ "$RUN_BIG_TESTS" = true ]; then
  66. add_test_vision "ggml-org/pixtral-12b-GGUF:Q4_K_M"
  67. add_test_vision "ggml-org/Mistral-Small-3.1-24B-Instruct-2503-GGUF" "mistral-v7"
  68. add_test_vision "ggml-org/Qwen2-VL-2B-Instruct-GGUF:Q4_K_M"
  69. add_test_vision "ggml-org/Qwen2-VL-7B-Instruct-GGUF:Q4_K_M"
  70. add_test_vision "ggml-org/Qwen2.5-VL-3B-Instruct-GGUF:Q4_K_M"
  71. add_test_vision "ggml-org/Qwen2.5-VL-7B-Instruct-GGUF:Q4_K_M"
  72. add_test_vision "ggml-org/InternVL3-8B-Instruct-GGUF:Q4_K_M"
  73. add_test_vision "ggml-org/InternVL3-14B-Instruct-GGUF:Q4_K_M"
  74. add_test_vision "ggml-org/Qwen2.5-Omni-7B-GGUF:Q4_K_M"
  75. # add_test_vision "ggml-org/Qwen2.5-VL-32B-Instruct-GGUF:Q4_K_M" # does not work on my mac M3 Ultra
  76. add_test_vision "ggml-org/Kimi-VL-A3B-Thinking-2506-GGUF:Q4_K_M"
  77. add_test_audio "ggml-org/ultravox-v0_5-llama-3_1-8b-GGUF:Q4_K_M"
  78. add_test_audio "ggml-org/Qwen2.5-Omni-7B-GGUF:Q4_K_M"
  79. fi
  80. # to test the huge models, run: ./tests.sh huge
  81. # this will run both the big and huge models
  82. # huge models are > 32B parameters
  83. if [ "$RUN_HUGE_TESTS" = true ]; then
  84. add_test_vision "ggml-org/Qwen2.5-VL-72B-Instruct-GGUF:Q4_K_M"
  85. add_test_vision "ggml-org/Llama-4-Scout-17B-16E-Instruct-GGUF:IQ1_S"
  86. fi
  87. # these models always give the wrong answer, not sure why
  88. # add_test_vision "ggml-org/SmolVLM-Instruct-GGUF:Q4_K_M"
  89. # add_test_vision "ggml-org/SmolVLM-256M-Instruct-GGUF:Q8_0"
  90. # add_test_vision "ggml-org/SmolVLM2-256M-Video-Instruct-GGUF:Q8_0"
  91. # this model has broken chat template, not usable
  92. # add_test_vision "cmp-nct/Yi-VL-6B-GGUF:Q5_K"
  93. # add_test_vision "guinmoon/MobileVLM-3B-GGUF:Q4_K_M" "deepseek"
  94. ###############
  95. cmake --build build -j --target llama-mtmd-cli
  96. arr_res=()
  97. for i in "${!arr_hf[@]}"; do
  98. bin="llama-mtmd-cli"
  99. prefix="${arr_prefix[$i]}"
  100. hf="${arr_hf[$i]}"
  101. tmpl="${arr_tmpl[$i]}"
  102. inp_file="${arr_file[$i]}"
  103. echo "Running test with binary: $bin and HF model: $hf"
  104. echo ""
  105. echo ""
  106. output=$(\
  107. "$PROJ_ROOT/build/bin/$bin" \
  108. -hf "$hf" \
  109. --image $SCRIPT_DIR/$inp_file \
  110. -p "what is the publisher name of the newspaper?" \
  111. --temp 0 -n 128 \
  112. ${tmpl:+--chat-template "$tmpl"} \
  113. 2>&1 | tee /dev/tty)
  114. echo "$output" > $SCRIPT_DIR/output/$bin-$(echo "$hf" | tr '/' '-').log
  115. if echo "$output" | grep -iq "new york"; then
  116. result="$prefix \033[32mOK\033[0m: $bin $hf"
  117. else
  118. result="$prefix \033[31mFAIL\033[0m: $bin $hf"
  119. fi
  120. echo -e "$result"
  121. arr_res+=("$result")
  122. echo ""
  123. echo ""
  124. echo ""
  125. echo "#################################################"
  126. echo "#################################################"
  127. echo ""
  128. echo ""
  129. done
  130. set +x
  131. for i in "${!arr_res[@]}"; do
  132. echo -e "${arr_res[$i]}"
  133. done
  134. echo ""
  135. echo "Output logs are saved in $SCRIPT_DIR/output"