%FILENAME%
mingw-w64-clang-aarch64-llama.cpp-1~b6646-1-any.pkg.tar.zst

%NAME%
mingw-w64-clang-aarch64-llama.cpp

%BASE%
mingw-w64-llama.cpp

%VERSION%
1~b6646-1

%DESC%
Library and tools for running inference with Meta's LLaMA model (and derivatives) in C/C++ (mingw-w64)

%CSIZE%
6572584

%ISIZE%
61553864

%SHA256SUM%
8429daee413e0cc4494f6e5cc0eb28e6282fb15c7d8713c07c2344f8f431a6b2

%URL%
https://github.com/ggml-org/llama.cpp

%LICENSE%
spdx:MIT

%ARCH%
any

%BUILDDATE%
1759301574

%PACKAGER%
CI (msys2/msys2-autobuild/ecd1d51f/18153827405)

%DEPENDS%
mingw-w64-clang-aarch64-curl
mingw-w64-clang-aarch64-cc-libs
mingw-w64-clang-aarch64-ggml
mingw-w64-clang-aarch64-openssl

%MAKEDEPENDS%
mingw-w64-clang-aarch64-cc
mingw-w64-clang-aarch64-cmake
mingw-w64-clang-aarch64-ninja

