%FILENAME%
llama-cpp-0.0.0.b8969-0-aarch64.pkg.tar.xz

%NAME%
llama-cpp

%BASE%
llama-cpp

%VERSION%
0.0.0.b8969-0

%DESC%
LLM inference in C/C++

%CSIZE%
4962464

%ISIZE%
24241949

%MD5SUM%
9029c5cf6f7084f3a8c47b8ec878ee23

%SHA256SUM%
b442ecb31dcddb0796da49bec018becfa4a7076c2402bcb9171996dba3774050

%PGPSIG%
iQGzBAABCgAdFiEEmY3icxjoZ+qXa6h3OJzu1kVz38oFAmnyEqoACgkQOJzu1kVz38rRIwwAm8VrdDCMsbCPIbBr+jqGYlPQfeZg/p7JEIBwg9nybuEBWUGjd/tMxCHRz9tSn4Iif1PKHpF2tGK5XqFSu4BRYJPn8qz5gdvEili9iwUqNN9Wtz8Qfu0ywkVy7P49axLtYEX72Ol5BMBNjN0xVXupLu1L9jtvxZgwEGNnk7DQdPd3lvSGoF4pf6ZoC2kFFrzkpMqH0o7DRCn0AKb00rL5s/C/TZtkn+zmbG2ST7bGWhw4K4CxZOBcPgw4GQSfboFJfQTA5SnXqpvcUFO67j/KUX6bwrEsDlQdoO0z2u5FhkK4+DHtq1tIAtZCW9Q5wTcnjqvSTPfDQJMm/XXd01qlJpTxEE6CtenXRf8K/mEWfJuzd5ImeGp8IvB783m1395FXv5qVpvDaB8Cfl8/Xz6f8uTgLm7iyGy6bOXjHxh36uuPfdxKLNq4B+OFjSGlouq+I111cNnU+iauNGpZYJ/1rn7pa2xCCwLkN1tEBCsOsc9cgwm9RfCTAYRENgvRwzKu

%URL%
https://github.com/ggml-org/llama.cpp

%LICENSE%
MIT

%ARCH%
aarch64

%BUILDDATE%
1777469770

%PACKAGER%
@termux

%DEPENDS%
libandroid-spawn
libc++
libcurl

%OPTDEPENDS%
llama-cpp-backend-vulkan
llama-cpp-backend-opencl

%MAKEDEPENDS%
ocl-icd
opencl-headers
spirv-headers
vulkan-headers

