| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183 |
- #!/usr/bin/env bash
- # make sure we are in the right directory
- SCRIPT_DIR=$( cd -- "$( dirname -- "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )
- cd $SCRIPT_DIR
- #export LLAMA_CACHE="$SCRIPT_DIR/tmp"
- set -eux
- mkdir -p $SCRIPT_DIR/output
- PROJ_ROOT="$SCRIPT_DIR/../.."
- cd $PROJ_ROOT
- # Check if the first argument is "big", then run test with big models
- # This is useful if we're running the script on a larger machine, so we can test the big models
- RUN_BIG_TESTS=false
- if [ "${1:-}" = "big" ]; then
- RUN_BIG_TESTS=true
- echo "Include BIG models..."
- fi
- RUN_HUGE_TESTS=false
- if [ "${1:-}" = "huge" ]; then
- RUN_HUGE_TESTS=true
- RUN_BIG_TESTS=true
- echo "Include BIG and HUGE models..."
- fi
- ###############
- arr_prefix=()
- arr_hf=()
- arr_extra_args=()
- arr_file=()
- add_test_vision() {
- local hf=$1
- shift
- local extra_args=""
- if [ $# -gt 0 ]; then
- extra_args=$(printf " %q" "$@")
- fi
- arr_prefix+=("[vision]")
- arr_hf+=("$hf")
- arr_extra_args+=("$extra_args")
- arr_file+=("test-1.jpeg")
- }
- add_test_audio() {
- local hf=$1
- shift
- local extra_args=""
- if [ $# -gt 0 ]; then
- extra_args=$(printf " %q" "$@")
- fi
- arr_prefix+=("[audio] ")
- arr_hf+=("$hf")
- arr_extra_args+=("$extra_args")
- arr_file+=("test-2.mp3")
- }
- add_test_vision "ggml-org/SmolVLM-500M-Instruct-GGUF:Q8_0"
- add_test_vision "ggml-org/SmolVLM2-2.2B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/SmolVLM2-500M-Video-Instruct-GGUF:Q8_0"
- add_test_vision "ggml-org/gemma-3-4b-it-GGUF:Q4_K_M"
- add_test_vision "THUDM/glm-edge-v-5b-gguf:Q4_K_M" -p "name of the newspaper?<__media__>"
- add_test_vision "second-state/Llava-v1.5-7B-GGUF:Q2_K" --chat-template vicuna
- add_test_vision "cjpais/llava-1.6-mistral-7b-gguf:Q3_K_M" --chat-template vicuna
- add_test_vision "ibm-research/granite-vision-3.2-2b-GGUF:Q4_K_M"
- add_test_vision "second-state/MiniCPM-Llama3-V-2_5-GGUF:Q2_K" # model from openbmb is corrupted
- add_test_vision "openbmb/MiniCPM-V-2_6-gguf:Q2_K"
- add_test_vision "openbmb/MiniCPM-o-2_6-gguf:Q4_0"
- add_test_vision "bartowski/Qwen2-VL-2B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/Qwen2.5-VL-3B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/InternVL2_5-1B-GGUF:Q8_0"
- add_test_vision "ggml-org/InternVL3-1B-Instruct-GGUF:Q8_0"
- add_test_vision "ggml-org/Qwen2.5-Omni-3B-GGUF:Q4_K_M"
- add_test_vision "ggml-org/LFM2-VL-450M-GGUF:Q8_0"
- add_test_vision "ggml-org/granite-docling-258M-GGUF:Q8_0"
- add_test_vision "ggml-org/LightOnOCR-1B-1025-GGUF:Q8_0"
- add_test_audio "ggml-org/ultravox-v0_5-llama-3_2-1b-GGUF:Q8_0"
- add_test_audio "ggml-org/Qwen2.5-Omni-3B-GGUF:Q4_K_M"
- add_test_audio "ggml-org/Voxtral-Mini-3B-2507-GGUF:Q4_K_M"
- add_test_audio "ggml-org/LFM2-Audio-1.5B-GGUF:Q8_0"
- # to test the big models, run: ./tests.sh big
- if [ "$RUN_BIG_TESTS" = true ]; then
- add_test_vision "ggml-org/pixtral-12b-GGUF:Q4_K_M"
- add_test_vision "ggml-org/Mistral-Small-3.1-24B-Instruct-2503-GGUF" --chat-template mistral-v7
- add_test_vision "ggml-org/Qwen2-VL-2B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/Qwen2-VL-7B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/Qwen2.5-VL-3B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/Qwen2.5-VL-7B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/Qwen3-VL-2B-Instruct-GGUF:Q8_0"
- add_test_vision "ggml-org/InternVL3-8B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/InternVL3-14B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/Qwen2.5-Omni-7B-GGUF:Q4_K_M"
- # add_test_vision "ggml-org/Qwen2.5-VL-32B-Instruct-GGUF:Q4_K_M" # does not work on my mac M3 Ultra
- # add_test_vision "ggml-org/Kimi-VL-A3B-Thinking-2506-GGUF:Q4_K_M" # not always working
- add_test_audio "ggml-org/ultravox-v0_5-llama-3_1-8b-GGUF:Q4_K_M"
- add_test_audio "ggml-org/Qwen2.5-Omni-7B-GGUF:Q4_K_M"
- fi
- # to test the huge models, run: ./tests.sh huge
- # this will run both the big and huge models
- # huge models are > 32B parameters
- if [ "$RUN_HUGE_TESTS" = true ]; then
- add_test_vision "ggml-org/Qwen2.5-VL-72B-Instruct-GGUF:Q4_K_M"
- add_test_vision "ggml-org/Llama-4-Scout-17B-16E-Instruct-GGUF:IQ1_S"
- fi
- # these models always give the wrong answer, not sure why
- # add_test_vision "ggml-org/SmolVLM-Instruct-GGUF:Q4_K_M"
- # add_test_vision "ggml-org/SmolVLM-256M-Instruct-GGUF:Q8_0"
- # add_test_vision "ggml-org/SmolVLM2-256M-Video-Instruct-GGUF:Q8_0"
- # this model has broken chat template, not usable
- # add_test_vision "cmp-nct/Yi-VL-6B-GGUF:Q5_K"
- # add_test_vision "guinmoon/MobileVLM-3B-GGUF:Q4_K_M" "deepseek"
- ###############
- cmake --build build -j --target llama-mtmd-cli
- arr_res=()
- for i in "${!arr_hf[@]}"; do
- bin="llama-mtmd-cli"
- prefix="${arr_prefix[$i]}"
- hf="${arr_hf[$i]}"
- extra_args="${arr_extra_args[$i]}"
- inp_file="${arr_file[$i]}"
- echo "Running test with binary: $bin and HF model: $hf"
- echo ""
- echo ""
- cmd="$(printf %q "$PROJ_ROOT/build/bin/$bin") \
- -hf $(printf %q "$hf") \
- --image $(printf %q "$SCRIPT_DIR/$inp_file") \
- --temp 0 -n 128 \
- ${extra_args}"
- # if extra_args does not contain -p, we add a default prompt
- if ! [[ "$extra_args" =~ "-p" ]]; then
- cmd+=" -p \"what is the publisher name of the newspaper?\""
- fi
- output=$(eval "$cmd" 2>&1 | tee /dev/tty)
- echo "$output" > $SCRIPT_DIR/output/$bin-$(echo "$hf" | tr '/' '-').log
- # either contains "new york" or both "men" and "walk"
- if echo "$output" | grep -iq "new york" \
- || (echo "$output" | grep -iq "men" && echo "$output" | grep -iq "walk")
- then
- result="$prefix \033[32mOK\033[0m: $hf"
- else
- result="$prefix \033[31mFAIL\033[0m: $hf"
- fi
- echo -e "$result"
- arr_res+=("$result")
- echo ""
- echo ""
- echo ""
- echo "#################################################"
- echo "#################################################"
- echo ""
- echo ""
- done
- set +x
- for i in "${!arr_res[@]}"; do
- echo -e "${arr_res[$i]}"
- done
- echo ""
- echo "Output logs are saved in $SCRIPT_DIR/output"
|