| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739 |
- name: Release
- on:
- workflow_dispatch: # allows manual triggering
- inputs:
- create_release:
- description: 'Create new release'
- required: true
- type: boolean
- push:
- branches:
- - master
- paths: ['.github/workflows/release.yml', '**/CMakeLists.txt', '**/.cmake', '**/*.h', '**/*.hpp', '**/*.c', '**/*.cpp', '**/*.cu', '**/*.cuh', '**/*.swift', '**/*.m', '**/*.metal', '**/*.comp']
- concurrency:
- group: ${{ github.workflow }}-${{ github.head_ref && github.ref || github.run_id }}
- cancel-in-progress: true
- env:
- BRANCH_NAME: ${{ github.head_ref || github.ref_name }}
- CMAKE_ARGS: "-DLLAMA_BUILD_EXAMPLES=OFF -DLLAMA_BUILD_TESTS=OFF -DLLAMA_BUILD_TOOLS=ON -DLLAMA_BUILD_SERVER=ON -DGGML_RPC=ON"
- jobs:
- macOS-arm64:
- runs-on: macos-14
- steps:
- - name: Clone
- id: checkout
- uses: actions/checkout@v4
- with:
- fetch-depth: 0
- - name: ccache
- uses: hendrikmuhs/ccache-action@v1.2.16
- with:
- key: macOS-latest-cmake-arm64
- evict-old-files: 1d
- - name: Dependencies
- id: depends
- continue-on-error: true
- run: |
- brew update
- brew install curl
- - name: Build
- id: cmake_build
- run: |
- sysctl -a
- cmake -B build \
- -DCMAKE_BUILD_RPATH="@loader_path" \
- -DLLAMA_FATAL_WARNINGS=ON \
- -DGGML_METAL_USE_BF16=ON \
- -DGGML_METAL_EMBED_LIBRARY=ON \
- -DGGML_RPC=ON \
- ${{ env.CMAKE_ARGS }}
- cmake --build build --config Release -j $(sysctl -n hw.logicalcpu)
- - name: Determine tag name
- id: tag
- uses: ./.github/actions/get-tag-name
- - name: Pack artifacts
- id: pack_artifacts
- run: |
- cp LICENSE ./build/bin/
- zip -r llama-${{ steps.tag.outputs.name }}-bin-macos-arm64.zip ./build/bin/*
- - name: Upload artifacts
- uses: actions/upload-artifact@v4
- with:
- path: llama-${{ steps.tag.outputs.name }}-bin-macos-arm64.zip
- name: llama-bin-macos-arm64.zip
- macOS-x64:
- runs-on: macos-13
- steps:
- - name: Clone
- id: checkout
- uses: actions/checkout@v4
- with:
- fetch-depth: 0
- - name: ccache
- uses: hendrikmuhs/ccache-action@v1.2.16
- with:
- key: macOS-latest-cmake-x64
- evict-old-files: 1d
- - name: Dependencies
- id: depends
- continue-on-error: true
- run: |
- brew update
- brew install curl
- - name: Build
- id: cmake_build
- run: |
- sysctl -a
- # Metal is disabled due to intermittent failures with Github runners not having a GPU:
- # https://github.com/ggml-org/llama.cpp/actions/runs/8635935781/job/23674807267#step:5:2313
- cmake -B build \
- -DCMAKE_BUILD_RPATH="@loader_path" \
- -DLLAMA_FATAL_WARNINGS=ON \
- -DGGML_METAL=OFF \
- -DGGML_RPC=ON
- cmake --build build --config Release -j $(sysctl -n hw.logicalcpu)
- - name: Determine tag name
- id: tag
- uses: ./.github/actions/get-tag-name
- - name: Pack artifacts
- id: pack_artifacts
- run: |
- cp LICENSE ./build/bin/
- zip -r llama-${{ steps.tag.outputs.name }}-bin-macos-x64.zip ./build/bin/*
- - name: Upload artifacts
- uses: actions/upload-artifact@v4
- with:
- path: llama-${{ steps.tag.outputs.name }}-bin-macos-x64.zip
- name: llama-bin-macos-x64.zip
- ubuntu-22-cpu:
- strategy:
- matrix:
- include:
- - build: 'x64'
- os: ubuntu-22.04
- - build: 'arm64'
- os: ubuntu-22.04-arm
- runs-on: ${{ matrix.os }}
- steps:
- - name: Clone
- id: checkout
- uses: actions/checkout@v4
- with:
- fetch-depth: 0
- - name: ccache
- uses: hendrikmuhs/ccache-action@v1.2.16
- with:
- key: ubuntu-cpu-cmake
- evict-old-files: 1d
- - name: Dependencies
- id: depends
- run: |
- sudo apt-get update
- sudo apt-get install build-essential libcurl4-openssl-dev
- - name: Build
- id: cmake_build
- run: |
- cmake -B build \
- -DLLAMA_FATAL_WARNINGS=ON \
- ${{ env.CMAKE_ARGS }}
- cmake --build build --config Release -j $(nproc)
- - name: Determine tag name
- id: tag
- uses: ./.github/actions/get-tag-name
- - name: Pack artifacts
- id: pack_artifacts
- run: |
- cp LICENSE ./build/bin/
- zip -r llama-${{ steps.tag.outputs.name }}-bin-ubuntu-${{ matrix.build }}.zip ./build/bin/*
- - name: Upload artifacts
- uses: actions/upload-artifact@v4
- with:
- path: llama-${{ steps.tag.outputs.name }}-bin-ubuntu-${{ matrix.build }}.zip
- name: llama-bin-ubuntu-${{ matrix.build }}.zip
- ubuntu-22-vulkan:
- runs-on: ubuntu-22.04
- steps:
- - name: Clone
- id: checkout
- uses: actions/checkout@v4
- with:
- fetch-depth: 0
- - name: ccache
- uses: hendrikmuhs/ccache-action@v1.2.16
- with:
- key: ubuntu-22-cmake-vulkan
- evict-old-files: 1d
- - name: Dependencies
- id: depends
- run: |
- wget -qO - https://packages.lunarg.com/lunarg-signing-key-pub.asc | sudo apt-key add -
- sudo wget -qO /etc/apt/sources.list.d/lunarg-vulkan-jammy.list https://packages.lunarg.com/vulkan/lunarg-vulkan-jammy.list
- sudo apt-get update -y
- sudo apt-get install -y build-essential mesa-vulkan-drivers vulkan-sdk libcurl4-openssl-dev
- - name: Build
- id: cmake_build
- run: |
- cmake -B build \
- -DGGML_VULKAN=ON \
- ${{ env.CMAKE_ARGS }}
- cmake --build build --config Release -j $(nproc)
- - name: Determine tag name
- id: tag
- uses: ./.github/actions/get-tag-name
- - name: Pack artifacts
- id: pack_artifacts
- run: |
- cp LICENSE ./build/bin/
- zip -r llama-${{ steps.tag.outputs.name }}-bin-ubuntu-vulkan-x64.zip ./build/bin/*
- - name: Upload artifacts
- uses: actions/upload-artifact@v4
- with:
- path: llama-${{ steps.tag.outputs.name }}-bin-ubuntu-vulkan-x64.zip
- name: llama-bin-ubuntu-vulkan-x64.zip
- windows-cpu:
- runs-on: windows-latest
- strategy:
- matrix:
- include:
- - arch: 'x64'
- - arch: 'arm64'
- steps:
- - name: Clone
- uses: actions/checkout@v4
- with:
- fetch-depth: 0
- - name: ccache
- uses: hendrikmuhs/ccache-action@v1.2.16
- with:
- key: windows-latest-cmake-cpu-${{ matrix.arch }}
- variant: ccache
- evict-old-files: 1d
- - name: Install Ninja
- run: |
- choco install ninja
- - name: libCURL
- id: get_libcurl
- uses: ./.github/actions/windows-setup-curl
- with:
- architecture: ${{ matrix.arch == 'x64' && 'win64' || 'win64a' }}
- - name: Build
- env:
- CURL_PATH: ${{ steps.get_libcurl.outputs.curl_path }}
- run: |
- cmake -S . -B build -G "Ninja Multi-Config" `
- -D CMAKE_TOOLCHAIN_FILE=cmake/${{ matrix.arch }}-windows-llvm.cmake `
- -DGGML_NATIVE=OFF `
- -DGGML_BACKEND_DL=ON `
- -DGGML_CPU_ALL_VARIANTS=ON `
- -DGGML_OPENMP=OFF `
- -DCURL_LIBRARY="$env:CURL_PATH/lib/libcurl.dll.a" -DCURL_INCLUDE_DIR="$env:CURL_PATH/include" `
- ${{ env.CMAKE_ARGS }}
- cmake --build build --config Release
- - name: Pack artifacts
- id: pack_artifacts
- env:
- CURL_PATH: ${{ steps.get_libcurl.outputs.curl_path }}
- run: |
- Copy-Item $env:CURL_PATH\bin\libcurl-${{ matrix.arch }}.dll .\build\bin\Release\
- 7z a llama-bin-win-cpu-${{ matrix.arch }}.zip .\build\bin\Release\*
- - name: Upload artifacts
- uses: actions/upload-artifact@v4
- with:
- path: llama-bin-win-cpu-${{ matrix.arch }}.zip
- name: llama-bin-win-cpu-${{ matrix.arch }}.zip
- windows:
- runs-on: windows-latest
- env:
- OPENBLAS_VERSION: 0.3.23
- VULKAN_VERSION: 1.4.309.0
- strategy:
- matrix:
- include:
- - backend: 'vulkan'
- arch: 'x64'
- defines: '-DGGML_VULKAN=ON'
- target: 'ggml-vulkan'
- - backend: 'opencl-adreno'
- arch: 'arm64'
- defines: '-G "Ninja Multi-Config" -D CMAKE_TOOLCHAIN_FILE=cmake/arm64-windows-llvm.cmake -DCMAKE_PREFIX_PATH="$env:RUNNER_TEMP/opencl-arm64-release" -DGGML_OPENCL=ON -DGGML_OPENCL_USE_ADRENO_KERNELS=ON'
- target: 'ggml-opencl'
- steps:
- - name: Clone
- id: checkout
- uses: actions/checkout@v4
- - name: ccache
- uses: hendrikmuhs/ccache-action@v1.2.16
- with:
- key: windows-latest-cmake-${{ matrix.backend }}-${{ matrix.arch }}
- variant: ccache
- evict-old-files: 1d
- - name: Install Vulkan SDK
- id: get_vulkan
- if: ${{ matrix.backend == 'vulkan' }}
- run: |
- curl.exe -o $env:RUNNER_TEMP/VulkanSDK-Installer.exe -L "https://sdk.lunarg.com/sdk/download/${env:VULKAN_VERSION}/windows/VulkanSDK-${env:VULKAN_VERSION}-Installer.exe"
- & "$env:RUNNER_TEMP\VulkanSDK-Installer.exe" --accept-licenses --default-answer --confirm-command install
- Add-Content $env:GITHUB_ENV "VULKAN_SDK=C:\VulkanSDK\${env:VULKAN_VERSION}"
- Add-Content $env:GITHUB_PATH "C:\VulkanSDK\${env:VULKAN_VERSION}\bin"
- - name: Install Ninja
- id: install_ninja
- run: |
- choco install ninja
- - name: Install OpenCL Headers and Libs
- id: install_opencl
- if: ${{ matrix.backend == 'opencl-adreno' && matrix.arch == 'arm64' }}
- run: |
- git clone https://github.com/KhronosGroup/OpenCL-Headers
- cd OpenCL-Headers
- cmake -B build `
- -DBUILD_TESTING=OFF `
- -DOPENCL_HEADERS_BUILD_TESTING=OFF `
- -DOPENCL_HEADERS_BUILD_CXX_TESTS=OFF `
- -DCMAKE_INSTALL_PREFIX="$env:RUNNER_TEMP/opencl-arm64-release"
- cmake --build build --target install
- git clone https://github.com/KhronosGroup/OpenCL-ICD-Loader
- cd OpenCL-ICD-Loader
- cmake -B build-arm64-release `
- -A arm64 `
- -DCMAKE_PREFIX_PATH="$env:RUNNER_TEMP/opencl-arm64-release" `
- -DCMAKE_INSTALL_PREFIX="$env:RUNNER_TEMP/opencl-arm64-release"
- cmake --build build-arm64-release --target install --config release
- - name: Build
- id: cmake_build
- run: |
- cmake -S . -B build ${{ matrix.defines }} -DGGML_NATIVE=OFF -DGGML_CPU=OFF -DGGML_BACKEND_DL=ON -DLLAMA_CURL=OFF
- cmake --build build --config Release --target ${{ matrix.target }}
- - name: Pack artifacts
- id: pack_artifacts
- run: |
- 7z a llama-bin-win-${{ matrix.backend }}-${{ matrix.arch }}.zip .\build\bin\Release\${{ matrix.target }}.dll
- - name: Upload artifacts
- uses: actions/upload-artifact@v4
- with:
- path: llama-bin-win-${{ matrix.backend }}-${{ matrix.arch }}.zip
- name: llama-bin-win-${{ matrix.backend }}-${{ matrix.arch }}.zip
- windows-cuda:
- runs-on: windows-2019
- strategy:
- matrix:
- cuda: ['12.4', '11.7']
- steps:
- - name: Clone
- id: checkout
- uses: actions/checkout@v4
- - name: Install ccache
- uses: hendrikmuhs/ccache-action@v1.2.16
- with:
- key: windows-cuda-${{ matrix.cuda }}
- variant: ccache
- evict-old-files: 1d
- - name: Install Cuda Toolkit
- uses: ./.github/actions/windows-setup-cuda
- with:
- cuda_version: ${{ matrix.cuda }}
- - name: Install Ninja
- id: install_ninja
- run: |
- choco install ninja
- - name: Build
- id: cmake_build
- shell: cmd
- run: |
- call "C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\VC\Auxiliary\Build\vcvars64.bat"
- cmake -S . -B build -G "Ninja Multi-Config" ^
- -DGGML_BACKEND_DL=ON ^
- -DGGML_NATIVE=OFF ^
- -DGGML_CPU=OFF ^
- -DGGML_CUDA=ON ^
- -DLLAMA_CURL=OFF
- set /A NINJA_JOBS=%NUMBER_OF_PROCESSORS%-1
- cmake --build build --config Release -j %NINJA_JOBS% --target ggml-cuda
- - name: Pack artifacts
- id: pack_artifacts
- run: |
- 7z a llama-bin-win-cuda-${{ matrix.cuda }}-x64.zip .\build\bin\Release\ggml-cuda.dll
- - name: Upload artifacts
- uses: actions/upload-artifact@v4
- with:
- path: llama-bin-win-cuda-${{ matrix.cuda }}-x64.zip
- name: llama-bin-win-cuda-${{ matrix.cuda }}-x64.zip
- - name: Copy and pack Cuda runtime
- run: |
- echo "Cuda install location: ${{ env.CUDA_PATH }}"
- $dst='.\build\bin\cudart\'
- robocopy "${{env.CUDA_PATH}}\bin" $dst cudart64_*.dll cublas64_*.dll cublasLt64_*.dll
- robocopy "${{env.CUDA_PATH}}\lib" $dst cudart64_*.dll cublas64_*.dll cublasLt64_*.dll
- 7z a cudart-llama-bin-win-cuda-${{ matrix.cuda }}-x64.zip $dst\*
- - name: Upload Cuda runtime
- uses: actions/upload-artifact@v4
- with:
- path: cudart-llama-bin-win-cuda-${{ matrix.cuda }}-x64.zip
- name: cudart-llama-bin-win-cuda-${{ matrix.cuda }}-x64.zip
- windows-sycl:
- runs-on: windows-latest
- defaults:
- run:
- shell: bash
- env:
- WINDOWS_BASEKIT_URL: https://registrationcenter-download.intel.com/akdlm/IRC_NAS/7cd9bba0-7aab-4e30-b3ae-2221006a4a05/intel-oneapi-base-toolkit-2025.1.1.34_offline.exe
- WINDOWS_DPCPP_MKL: intel.oneapi.win.cpp-dpcpp-common:intel.oneapi.win.mkl.devel:intel.oneapi.win.dnnl:intel.oneapi.win.tbb.devel
- ONEAPI_ROOT: "C:/Program Files (x86)/Intel/oneAPI"
- steps:
- - name: Clone
- id: checkout
- uses: actions/checkout@v4
- - name: ccache
- uses: hendrikmuhs/ccache-action@v1.2.16
- with:
- key: windows-latest-cmake-sycl
- variant: ccache
- evict-old-files: 1d
- - name: Install
- run: |
- scripts/install-oneapi.bat $WINDOWS_BASEKIT_URL $WINDOWS_DPCPP_MKL
- - name: Build
- id: cmake_build
- shell: cmd
- run: |
- call "C:\Program Files (x86)\Intel\oneAPI\setvars.bat" intel64 --force
- cmake -G "Ninja" -B build ^
- -DCMAKE_C_COMPILER=cl -DCMAKE_CXX_COMPILER=icx ^
- -DCMAKE_BUILD_TYPE=Release ^
- -DGGML_BACKEND_DL=ON -DBUILD_SHARED_LIBS=ON ^
- -DGGML_CPU=OFF -DGGML_SYCL=ON ^
- -DLLAMA_CURL=OFF
- cmake --build build --target ggml-sycl -j
- - name: Build the release package
- id: pack_artifacts
- run: |
- echo "cp oneAPI running time dll files in ${{ env.ONEAPI_ROOT }} to ./build/bin"
- cp "${{ env.ONEAPI_ROOT }}/mkl/latest/bin/mkl_sycl_blas.5.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/mkl/latest/bin/mkl_core.2.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/mkl/latest/bin/mkl_tbb_thread.2.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/compiler/latest/bin/ur_adapter_level_zero.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/compiler/latest/bin/ur_adapter_opencl.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/compiler/latest/bin/ur_loader.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/compiler/latest/bin/ur_win_proxy_loader.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/compiler/latest/bin/sycl8.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/compiler/latest/bin/svml_dispmd.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/compiler/latest/bin/libmmd.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/compiler/latest/bin/libiomp5md.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/dnnl/latest/bin/dnnl.dll" ./build/bin
- cp "${{ env.ONEAPI_ROOT }}/tbb/latest/bin/tbb12.dll" ./build/bin
- echo "cp oneAPI running time dll files to ./build/bin done"
- 7z a llama-bin-win-sycl-x64.zip ./build/bin/*
- - name: Upload the release package
- uses: actions/upload-artifact@v4
- with:
- path: llama-bin-win-sycl-x64.zip
- name: llama-bin-win-sycl-x64.zip
- windows-hip:
- runs-on: windows-latest
- strategy:
- matrix:
- include:
- - name: "radeon"
- gpu_targets: "gfx1100;gfx1101;gfx1102;gfx1030;gfx1031;gfx1032"
- steps:
- - name: Clone
- id: checkout
- uses: actions/checkout@v4
- - name: Clone rocWMMA repository
- id: clone_rocwmma
- run: |
- git clone https://github.com/rocm/rocwmma --branch rocm-6.2.4 --depth 1
- - name: ccache
- uses: hendrikmuhs/ccache-action@v1.2.16
- with:
- key: windows-latest-cmake-hip-${{ matrix.name }}-x64
- evict-old-files: 1d
- - name: Install
- id: depends
- run: |
- $ErrorActionPreference = "Stop"
- write-host "Downloading AMD HIP SDK Installer"
- Invoke-WebRequest -Uri "https://download.amd.com/developer/eula/rocm-hub/AMD-Software-PRO-Edition-24.Q3-WinSvr2022-For-HIP.exe" -OutFile "${env:RUNNER_TEMP}\rocm-install.exe"
- write-host "Installing AMD HIP SDK"
- Start-Process "${env:RUNNER_TEMP}\rocm-install.exe" -ArgumentList '-install' -NoNewWindow -Wait
- write-host "Completed AMD HIP SDK installation"
- - name: Verify ROCm
- id: verify
- run: |
- & 'C:\Program Files\AMD\ROCm\*\bin\clang.exe' --version
- - name: Build
- id: cmake_build
- run: |
- $env:HIP_PATH=$(Resolve-Path 'C:\Program Files\AMD\ROCm\*\bin\clang.exe' | split-path | split-path)
- $env:CMAKE_PREFIX_PATH="${env:HIP_PATH}"
- cmake -G "Unix Makefiles" -B build -S . `
- -DCMAKE_C_COMPILER="${env:HIP_PATH}\bin\clang.exe" `
- -DCMAKE_CXX_COMPILER="${env:HIP_PATH}\bin\clang++.exe" `
- -DCMAKE_CXX_FLAGS="-I$($PWD.Path.Replace('\', '/'))/rocwmma/library/include/ -Wno-ignored-attributes -Wno-nested-anon-types" `
- -DCMAKE_BUILD_TYPE=Release `
- -DGGML_BACKEND_DL=ON `
- -DGGML_NATIVE=OFF `
- -DGGML_CPU=OFF `
- -DAMDGPU_TARGETS="${{ matrix.gpu_targets }}" `
- -DGGML_HIP_ROCWMMA_FATTN=ON `
- -DGGML_HIP=ON `
- -DLLAMA_CURL=OFF
- cmake --build build --target ggml-hip -j ${env:NUMBER_OF_PROCESSORS}
- md "build\bin\rocblas\library\"
- cp "${env:HIP_PATH}\bin\hipblas.dll" "build\bin\"
- cp "${env:HIP_PATH}\bin\rocblas.dll" "build\bin\"
- cp "${env:HIP_PATH}\bin\rocblas\library\*" "build\bin\rocblas\library\"
- - name: Pack artifacts
- id: pack_artifacts
- run: |
- 7z a llama-bin-win-hip-${{ matrix.name }}-x64.zip .\build\bin\*
- - name: Upload artifacts
- uses: actions/upload-artifact@v4
- with:
- path: llama-bin-win-hip-${{ matrix.name }}-x64.zip
- name: llama-bin-win-hip-${{ matrix.name }}-x64.zip
- ios-xcode-build:
- runs-on: macos-latest
- steps:
- - name: Checkout code
- uses: actions/checkout@v4
- with:
- fetch-depth: 0
- - name: Build
- id: cmake_build
- run: |
- sysctl -a
- cmake -B build -G Xcode \
- -DGGML_METAL_USE_BF16=ON \
- -DGGML_METAL_EMBED_LIBRARY=ON \
- -DLLAMA_CURL=OFF \
- -DLLAMA_BUILD_EXAMPLES=OFF \
- -DLLAMA_BUILD_TOOLS=OFF \
- -DLLAMA_BUILD_TESTS=OFF \
- -DLLAMA_BUILD_SERVER=OFF \
- -DCMAKE_SYSTEM_NAME=iOS \
- -DCMAKE_OSX_DEPLOYMENT_TARGET=14.0 \
- -DCMAKE_XCODE_ATTRIBUTE_DEVELOPMENT_TEAM=ggml
- cmake --build build --config Release -j $(sysctl -n hw.logicalcpu) -- CODE_SIGNING_ALLOWED=NO
- - name: xcodebuild for swift package
- id: xcodebuild
- run: |
- ./build-xcframework.sh
- - name: Build Xcode project
- run: xcodebuild -project examples/llama.swiftui/llama.swiftui.xcodeproj -scheme llama.swiftui -sdk iphoneos CODE_SIGNING_REQUIRED=NO CODE_SIGN_IDENTITY= -destination 'generic/platform=iOS' FRAMEWORK_FOLDER_PATH=./build-ios build
- - name: Determine tag name
- id: tag
- uses: ./.github/actions/get-tag-name
- - name: Pack artifacts
- id: pack_artifacts
- run: |
- zip --symlinks -r llama-${{ steps.tag.outputs.name }}-xcframework.zip build-apple/llama.xcframework
- - name: Upload artifacts
- uses: actions/upload-artifact@v4
- with:
- path: llama-${{ steps.tag.outputs.name }}-xcframework.zip
- name: llama-${{ steps.tag.outputs.name }}-xcframework
- release:
- if: ${{ ( github.event_name == 'push' && github.ref == 'refs/heads/master' ) || github.event.inputs.create_release == 'true' }}
- # Fine-grant permission
- # https://docs.github.com/en/actions/security-for-github-actions/security-guides/automatic-token-authentication#modifying-the-permissions-for-the-github_token
- permissions:
- contents: write # for creating release
- runs-on: ubuntu-latest
- needs:
- - windows
- - windows-cpu
- - windows-cuda
- - windows-sycl
- - windows-hip
- - ubuntu-22-cpu
- - ubuntu-22-vulkan
- - macOS-arm64
- - macOS-x64
- - ios-xcode-build
- steps:
- - name: Clone
- id: checkout
- uses: actions/checkout@v4
- with:
- fetch-depth: 0
- - name: Determine tag name
- id: tag
- uses: ./.github/actions/get-tag-name
- - name: Download artifacts
- id: download-artifact
- uses: actions/download-artifact@v4
- with:
- path: ./artifact
- merge-multiple: true
- - name: Move artifacts
- id: move_artifacts
- run: |
- mkdir -p release
- echo "Adding CPU backend files to existing zips..."
- for arch in x64 arm64; do
- cpu_zip="artifact/llama-bin-win-cpu-${arch}.zip"
- temp_dir=$(mktemp -d)
- echo "Extracting CPU backend for $arch..."
- unzip "$cpu_zip" -d "$temp_dir"
- echo "Adding CPU files to $arch zips..."
- for target_zip in artifact/llama-bin-win-*-${arch}.zip; do
- if [[ "$target_zip" == "$cpu_zip" ]]; then
- continue
- fi
- echo "Adding CPU backend to $(basename "$target_zip")"
- realpath_target_zip=$(realpath "$target_zip")
- (cd "$temp_dir" && zip -r "$realpath_target_zip" .)
- done
- rm -rf "$temp_dir"
- done
- echo "Renaming and moving zips to release..."
- for zip_file in artifact/llama-bin-win-*.zip; do
- base_name=$(basename "$zip_file" .zip)
- zip_name="llama-${{ steps.tag.outputs.name }}-${base_name#llama-}.zip"
- echo "Moving $zip_file to release/$zip_name"
- mv "$zip_file" "release/$zip_name"
- done
- echo "Moving other artifacts..."
- mv -v artifact/*.zip release
- - name: Create release
- id: create_release
- uses: ggml-org/action-create-release@v1
- env:
- GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
- with:
- tag_name: ${{ steps.tag.outputs.name }}
- - name: Upload release
- id: upload_release
- uses: actions/github-script@v3
- with:
- github-token: ${{secrets.GITHUB_TOKEN}}
- script: |
- const path = require('path');
- const fs = require('fs');
- const release_id = '${{ steps.create_release.outputs.id }}';
- for (let file of await fs.readdirSync('./release')) {
- if (path.extname(file) === '.zip') {
- console.log('uploadReleaseAsset', file);
- await github.repos.uploadReleaseAsset({
- owner: context.repo.owner,
- repo: context.repo.repo,
- release_id: release_id,
- name: file,
- data: await fs.readFileSync(`./release/${file}`)
- });
- }
- }
|