%FILENAME%
llama-cpp-0.0.0.b8969-0-x86_64.pkg.tar.xz

%NAME%
llama-cpp

%BASE%
llama-cpp

%VERSION%
0.0.0.b8969-0

%DESC%
LLM inference in C/C++

%CSIZE%
5535444

%ISIZE%
25835590

%MD5SUM%
1fd81c53b4baebf6fd25c42677be86b3

%SHA256SUM%
c4d54753a93dcd41e7dc62c7922d9c242f8604349e2cc3b31458c0b5f7ef99ae

%PGPSIG%
iQGzBAABCgAdFiEEmY3icxjoZ+qXa6h3OJzu1kVz38oFAmnyEloACgkQOJzu1kVz38qtRAv9FafUR32bjv/Xr6m7zue4rWrRhAhxaw1gsDv7nwie5l5CNf8UABHFdz9ByZSSNbKEEJB49+PbcuMxxwguooVc/HdIYG0XjNg+7q5mzfjbjiQw7Az3R6XsP5ttM+AslU9ARghaVRCktYHZ/2u4tAnuT1ts41mLk8ybUnLKccTigPerHTW/tXmtEvXPoCTdJpSj8rs6SfAWfSI6xFZVTqUXQa4EOdZ5HjL1jMhBCnG0rBCkoPZRiMkZMzvJpfAYRBUrqSdvS4y0uAi2KPyGVkrDnzWjiJZOOy8xzzt5Wqct2Ol0HsDPE8vJRjdzjd4d0B4V1Me2A34EoNaB93jeLc+R02vRb7ezCtlVoGlHLWYHaw81FMy3pVuf6VPW5QkxCL1sT0Xfv/63F9+7ExBQAqnQswPrNwwN4sNqqIOh4I21+mC9cuRAJM6V9npuiVCh2oOX80UaW5/qMb+5pdNG4YoDbR44GtHoDpmboiYMawutSbmq0w7JnZUw/FD1NOmB6c88

%URL%
https://github.com/ggml-org/llama.cpp

%LICENSE%
MIT

%ARCH%
x86_64

%BUILDDATE%
1777469770

%PACKAGER%
@termux

%DEPENDS%
libandroid-spawn
libc++
libcurl

%OPTDEPENDS%
llama-cpp-backend-vulkan
llama-cpp-backend-opencl

%MAKEDEPENDS%
ocl-icd
opencl-headers
spirv-headers
vulkan-headers

