%FILENAME%
llama-cpp-backend-opencl-0.0.0.b8969-0-aarch64.pkg.tar.xz

%NAME%
llama-cpp-backend-opencl

%BASE%
llama-cpp

%VERSION%
0.0.0.b8969-0

%DESC%
OpenCL Backend for LLaMA.cpp

%CSIZE%
178492

%ISIZE%
1240656

%MD5SUM%
004764ced8ee44fc379f99283c71e50b

%SHA256SUM%
4e4b951a0d1671c28292f3aded92bc6dc2708ffcc975d9c0af9ac180ace1ef96

%PGPSIG%
iQGzBAABCgAdFiEEmY3icxjoZ+qXa6h3OJzu1kVz38oFAmnyErMACgkQOJzu1kVz38relAwAmiOl2xwRP/YxURAQ4DZMPSt8HjmzH2f7FngvcFsr6Dg8CYfC40sRicc2fWiOrRhwBHYnXarD0JzW7amQ+JAMzrUtjKpmqB87xftDg31bZiV+43zOZVkcVkbrpcISnTRRXMuIgKz8YFG0ISgwQyB8QP1/aBBZs8RXcmOLElzGx9zxPpHGUMkcgQ/+RKtsYGkOzWZS/09OkftlMOtT/4hvfE6ZaKzEaTn43QWamQS00YJwncseAMpUfXApR4YIINvM7EXLRSeDaQB+8xrseYq6hjslo9iYxdiOnMFymqDEkTfJMQuLSoMLulhSota2iX7bLnBcIBEUhf5kPa7Z0cDGr2aFHzeHXWOvs+h4IABPf5EbtDeEOWlUrVKB+jDwoCPSSw5GjMbDOhufpu5iO5Cg/N8LPOiqxx4llHdLhaDnhZWZ1MYDX8ePeQoCACbutKOhueiIb4BBe6NNUFWPXV9jJXxDmGieepbubobcgq8elnrMSpyCsVDPu7jyueADiYbj

%URL%
https://github.com/ggml-org/llama.cpp

%ARCH%
aarch64

%BUILDDATE%
1777469770

%PACKAGER%
@termux

%DEPENDS%
llama-cpp

%OPTDEPENDS%
ocl-icd
opencl-vendor-driver

