CMakeLists.txt 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244
  1. llama_add_compile_flags()
  2. function(llama_build source)
  3. set(TEST_SOURCES ${source} ${ARGN})
  4. if (DEFINED LLAMA_TEST_NAME)
  5. set(TEST_TARGET ${LLAMA_TEST_NAME})
  6. else()
  7. get_filename_component(TEST_TARGET ${source} NAME_WE)
  8. endif()
  9. add_executable(${TEST_TARGET} ${TEST_SOURCES})
  10. target_link_libraries(${TEST_TARGET} PRIVATE common)
  11. install(TARGETS ${TEST_TARGET} RUNTIME)
  12. endfunction()
  13. function(llama_test target)
  14. include(CMakeParseArguments)
  15. set(options)
  16. set(oneValueArgs NAME LABEL WORKING_DIRECTORY)
  17. set(multiValueArgs ARGS)
  18. cmake_parse_arguments(LLAMA_TEST "${options}" "${oneValueArgs}" "${multiValueArgs}" ${ARGN})
  19. if (NOT DEFINED LLAMA_TEST_LABEL)
  20. set(LLAMA_TEST_LABEL "main")
  21. endif()
  22. if (NOT DEFINED LLAMA_TEST_WORKING_DIRECTORY)
  23. set(LLAMA_TEST_WORKING_DIRECTORY .)
  24. endif()
  25. if (DEFINED LLAMA_TEST_NAME)
  26. set(TEST_NAME ${LLAMA_TEST_NAME})
  27. else()
  28. set(TEST_NAME ${target})
  29. endif()
  30. set(TEST_TARGET ${target})
  31. add_test(
  32. NAME ${TEST_NAME}
  33. WORKING_DIRECTORY ${LLAMA_TEST_WORKING_DIRECTORY}
  34. COMMAND $<TARGET_FILE:${TEST_TARGET}>
  35. ${LLAMA_TEST_ARGS})
  36. set_property(TEST ${TEST_NAME} PROPERTY LABELS ${LLAMA_TEST_LABEL})
  37. endfunction()
  38. function(llama_test_cmd target)
  39. include(CMakeParseArguments)
  40. set(options)
  41. set(oneValueArgs NAME LABEL WORKING_DIRECTORY)
  42. set(multiValueArgs ARGS)
  43. cmake_parse_arguments(LLAMA_TEST "${options}" "${oneValueArgs}" "${multiValueArgs}" ${ARGN})
  44. if (NOT DEFINED LLAMA_TEST_LABEL)
  45. set(LLAMA_TEST_LABEL "main")
  46. endif()
  47. if (NOT DEFINED LLAMA_TEST_WORKING_DIRECTORY)
  48. set(LLAMA_TEST_WORKING_DIRECTORY .)
  49. endif()
  50. if (DEFINED LLAMA_TEST_NAME)
  51. set(TEST_NAME ${LLAMA_TEST_NAME})
  52. else()
  53. set(TEST_NAME ${target})
  54. endif()
  55. add_test(
  56. NAME ${TEST_NAME}
  57. WORKING_DIRECTORY ${LLAMA_TEST_WORKING_DIRECTORY}
  58. COMMAND ${target}
  59. ${LLAMA_TEST_ARGS})
  60. set_property(TEST ${TEST_NAME} PROPERTY LABELS ${LLAMA_TEST_LABEL})
  61. endfunction()
  62. # Builds and runs a test source file.
  63. # Optional args:
  64. # - NAME: name of the executable & test target (defaults to the source file name without extension)
  65. # - LABEL: label for the test (defaults to main)
  66. # - ARGS: arguments to pass to the test executable
  67. # - WORKING_DIRECTORY
  68. function(llama_build_and_test source)
  69. include(CMakeParseArguments)
  70. set(options)
  71. set(oneValueArgs NAME LABEL WORKING_DIRECTORY)
  72. set(multiValueArgs ARGS)
  73. cmake_parse_arguments(LLAMA_TEST "${options}" "${oneValueArgs}" "${multiValueArgs}" ${ARGN})
  74. set(TEST_SOURCES ${source} ${LLAMA_TEST_UNPARSED_ARGUMENTS} get-model.cpp)
  75. if (NOT DEFINED LLAMA_TEST_LABEL)
  76. set(LLAMA_TEST_LABEL "main")
  77. endif()
  78. if (NOT DEFINED LLAMA_TEST_WORKING_DIRECTORY)
  79. set(LLAMA_TEST_WORKING_DIRECTORY .)
  80. endif()
  81. if (DEFINED LLAMA_TEST_NAME)
  82. set(TEST_TARGET ${LLAMA_TEST_NAME})
  83. else()
  84. get_filename_component(TEST_TARGET ${source} NAME_WE)
  85. endif()
  86. add_executable(${TEST_TARGET} ${TEST_SOURCES})
  87. install(TARGETS ${TEST_TARGET} RUNTIME)
  88. target_link_libraries(${TEST_TARGET} PRIVATE common)
  89. add_test(
  90. NAME ${TEST_TARGET}
  91. WORKING_DIRECTORY ${LLAMA_TEST_WORKING_DIRECTORY}
  92. COMMAND $<TARGET_FILE:${TEST_TARGET}>
  93. ${LLAMA_TEST_ARGS})
  94. set_property(TEST ${TEST_TARGET} PROPERTY LABELS ${LLAMA_TEST_LABEL})
  95. endfunction()
  96. # build test-tokenizer-0 target once and add many tests
  97. llama_build(test-tokenizer-0.cpp)
  98. llama_test(test-tokenizer-0 NAME test-tokenizer-0-bert-bge ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-bert-bge.gguf)
  99. llama_test(test-tokenizer-0 NAME test-tokenizer-0-command-r ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-command-r.gguf)
  100. llama_test(test-tokenizer-0 NAME test-tokenizer-0-deepseek-coder ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-deepseek-coder.gguf)
  101. llama_test(test-tokenizer-0 NAME test-tokenizer-0-deepseek-llm ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-deepseek-llm.gguf)
  102. llama_test(test-tokenizer-0 NAME test-tokenizer-0-falcon ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-falcon.gguf)
  103. llama_test(test-tokenizer-0 NAME test-tokenizer-0-gpt-2 ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-gpt-2.gguf)
  104. llama_test(test-tokenizer-0 NAME test-tokenizer-0-llama-bpe ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-llama-bpe.gguf)
  105. llama_test(test-tokenizer-0 NAME test-tokenizer-0-llama-spm ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-llama-spm.gguf)
  106. llama_test(test-tokenizer-0 NAME test-tokenizer-0-mpt ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-mpt.gguf)
  107. llama_test(test-tokenizer-0 NAME test-tokenizer-0-phi-3 ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-phi-3.gguf)
  108. llama_test(test-tokenizer-0 NAME test-tokenizer-0-qwen2 ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-qwen2.gguf)
  109. llama_test(test-tokenizer-0 NAME test-tokenizer-0-refact ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-refact.gguf)
  110. llama_test(test-tokenizer-0 NAME test-tokenizer-0-starcoder ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-starcoder.gguf)
  111. if (NOT WIN32)
  112. llama_test_cmd(
  113. ${CMAKE_CURRENT_SOURCE_DIR}/test-tokenizers-repo.sh
  114. NAME test-tokenizers-ggml-vocabs
  115. WORKING_DIRECTORY ${CMAKE_RUNTIME_OUTPUT_DIRECTORY}
  116. ARGS https://huggingface.co/ggml-org/vocabs ${PROJECT_SOURCE_DIR}/models/ggml-vocabs
  117. )
  118. endif()
  119. if (LLAMA_LLGUIDANCE)
  120. llama_build_and_test(test-grammar-llguidance.cpp ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-llama-bpe.gguf)
  121. endif ()
  122. if (NOT WIN32 OR NOT BUILD_SHARED_LIBS)
  123. # these tests are disabled on Windows because they use internal functions not exported with LLAMA_API (when building with shared libraries)
  124. llama_build_and_test(test-sampling.cpp)
  125. llama_build_and_test(test-grammar-parser.cpp)
  126. llama_build_and_test(test-grammar-integration.cpp)
  127. llama_build_and_test(test-llama-grammar.cpp)
  128. llama_build_and_test(test-chat.cpp)
  129. # TODO: disabled on loongarch64 because the ggml-ci node lacks Python 3.8
  130. if (NOT ${CMAKE_SYSTEM_PROCESSOR} MATCHES "loongarch64")
  131. llama_build_and_test(test-json-schema-to-grammar.cpp WORKING_DIRECTORY ${PROJECT_SOURCE_DIR})
  132. target_include_directories(test-json-schema-to-grammar PRIVATE ${PROJECT_SOURCE_DIR}/tools/server)
  133. endif()
  134. if (NOT GGML_BACKEND_DL)
  135. llama_build(test-quantize-stats.cpp)
  136. endif()
  137. llama_build(test-gbnf-validator.cpp)
  138. # build test-tokenizer-1-bpe target once and add many tests
  139. llama_build(test-tokenizer-1-bpe.cpp)
  140. # TODO: disabled due to slowness
  141. #llama_test(test-tokenizer-1-bpe NAME test-tokenizer-1-aquila ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-aquila.gguf)
  142. #llama_test(test-tokenizer-1-bpe NAME test-tokenizer-1-falcon ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-falcon.gguf)
  143. #llama_test(test-tokenizer-1-bpe NAME test-tokenizer-1-gpt-2 ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-gpt-2.gguf)
  144. #llama_test(test-tokenizer-1-bpe NAME test-tokenizer-1-gpt-neox ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-gpt-neox.gguf)
  145. #llama_test(test-tokenizer-1-bpe NAME test-tokenizer-1-llama-bpe ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-llama-bpe.gguf --ignore-merges)
  146. #llama_test(test-tokenizer-1-bpe NAME test-tokenizer-1-mpt ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-mpt.gguf)
  147. #llama_test(test-tokenizer-1-bpe NAME test-tokenizer-1-refact ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-refact.gguf)
  148. #llama_test(test-tokenizer-1-bpe NAME test-tokenizer-1-starcoder ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-starcoder.gguf)
  149. # build test-tokenizer-1-spm target once and add many tests
  150. llama_build(test-tokenizer-1-spm.cpp)
  151. llama_test(test-tokenizer-1-spm NAME test-tokenizer-1-llama-spm ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-llama-spm.gguf)
  152. #llama_test(test-tokenizer-1-spm NAME test-tokenizer-1-baichuan ARGS ${PROJECT_SOURCE_DIR}/models/ggml-vocab-baichuan.gguf)
  153. # llama_build_and_test(test-double-float.cpp) # SLOW
  154. endif()
  155. llama_build_and_test(test-chat-parser.cpp)
  156. llama_build_and_test(test-chat-peg-parser.cpp peg-parser/simple-tokenize.cpp)
  157. llama_build_and_test(test-chat-template.cpp)
  158. llama_build_and_test(test-json-partial.cpp)
  159. llama_build_and_test(test-log.cpp)
  160. llama_build_and_test(
  161. test-peg-parser.cpp
  162. peg-parser/simple-tokenize.cpp
  163. peg-parser/test-basic.cpp
  164. peg-parser/test-gbnf-generation.cpp
  165. peg-parser/test-json-parser.cpp
  166. peg-parser/test-json-serialization.cpp
  167. peg-parser/test-unicode.cpp
  168. peg-parser/testing.h
  169. peg-parser/tests.h
  170. )
  171. llama_build_and_test(test-regex-partial.cpp)
  172. if (NOT ${CMAKE_SYSTEM_PROCESSOR} MATCHES "s390x")
  173. llama_build_and_test(test-thread-safety.cpp ARGS -hf ggml-org/models -hff tinyllamas/stories15M-q4_0.gguf -ngl 99 -p "The meaning of life is" -n 128 -c 256 -ub 32 -np 4 -t 2)
  174. else()
  175. llama_build_and_test(test-thread-safety.cpp ARGS -hf ggml-org/models -hff tinyllamas/stories15M-be.Q4_0.gguf -ngl 99 -p "The meaning of life is" -n 128 -c 256 -ub 32 -np 4 -t 2)
  176. endif()
  177. # this fails on windows (github hosted runner) due to curl DLL not found (exit code 0xc0000135)
  178. if (NOT WIN32)
  179. llama_build_and_test(test-arg-parser.cpp)
  180. endif()
  181. if (NOT LLAMA_SANITIZE_ADDRESS AND NOT GGML_SCHED_NO_REALLOC)
  182. # TODO: repair known memory leaks
  183. llama_build_and_test(test-opt.cpp)
  184. endif()
  185. llama_build_and_test(test-gguf.cpp)
  186. llama_build_and_test(test-backend-ops.cpp)
  187. llama_build_and_test(test-model-load-cancel.cpp LABEL "model")
  188. llama_build_and_test(test-autorelease.cpp LABEL "model")
  189. if (NOT GGML_BACKEND_DL)
  190. # these tests use the backends directly and cannot be built with dynamic loading
  191. llama_build_and_test(test-barrier.cpp)
  192. llama_build_and_test(test-quantize-fns.cpp)
  193. llama_build_and_test(test-quantize-perf.cpp)
  194. llama_build_and_test(test-rope.cpp)
  195. endif()
  196. # libmtmd
  197. set(LLAMA_TEST_NAME test-mtmd-c-api)
  198. llama_build_and_test(test-mtmd-c-api.c)
  199. target_link_libraries(${LLAMA_TEST_NAME} PRIVATE mtmd)
  200. # dummy executable - not installed
  201. get_filename_component(TEST_TARGET test-c.c NAME_WE)
  202. add_executable(${TEST_TARGET} test-c.c)
  203. target_link_libraries(${TEST_TARGET} PRIVATE llama)
  204. llama_build_and_test(test-alloc.cpp)
  205. target_include_directories(test-alloc PRIVATE ${PROJECT_SOURCE_DIR}/ggml/src)