%FILENAME%
mingw-w64-ucrt-x86_64-llama.cpp-1~b9097-1-any.pkg.tar.zst

%NAME%
mingw-w64-ucrt-x86_64-llama.cpp

%BASE%
mingw-w64-llama.cpp

%VERSION%
1~b9097-1

%DESC%
Library and tools for running inference with Meta's LLaMA model (and derivatives) in C/C++ (mingw-w64)

%CSIZE%
5722162

%ISIZE%
30607337

%SHA256SUM%
23419f6138f32309abb44333104b73928b96cf525e8e6142ef288596a231ed8f

%URL%
https://github.com/ggml-org/llama.cpp

%LICENSE%
spdx:MIT

%ARCH%
any

%BUILDDATE%
1778561242

%PACKAGER%
CI (https://github.com/msys2/msys2-autobuild/actions/runs/25711575475/job/75492781387)

%DEPENDS%
mingw-w64-ucrt-x86_64-cc-libs
mingw-w64-ucrt-x86_64-curl
mingw-w64-ucrt-x86_64-ggml
mingw-w64-ucrt-x86_64-openssl

%MAKEDEPENDS%
mingw-w64-ucrt-x86_64-cc
mingw-w64-ucrt-x86_64-cmake
mingw-w64-ucrt-x86_64-ninja

