%FILENAME%
llama-cpp-backend-vulkan-0.0.0.b8969-0-aarch64.pkg.tar.xz

%NAME%
llama-cpp-backend-vulkan

%BASE%
llama-cpp

%VERSION%
0.0.0.b8969-0

%DESC%
Vulkan Backend for LLaMA.cpp

%CSIZE%
3279084

%ISIZE%
32942008

%MD5SUM%
7e1f05cac5c7ef2950f0184938256ab4

%SHA256SUM%
3fa186977d380e8eed06a02aea869feb02e74d63f52213c9ad513379ef489f5a

%PGPSIG%
iQGzBAABCgAdFiEEmY3icxjoZ+qXa6h3OJzu1kVz38oFAmnyErwACgkQOJzu1kVz38qIFQwArmvXPwQkoIAcmt4k1Z2/u2z9XYPtW63HyJVQACOAGSXgILyYOAlnvcgg97L1VWit0mgr3Sz9ONmKpwTOF97AupKTChOJOzwXIT2m6MvNaSHXyVHaw6jc8SgPCi5H7/FasZtTSuCqkHHM1jmwwFQcycdSxA0RmvjP2AdMfu3qrWHGmbhUedAiyWBokrYDw1wVhkJzeZvihSoDiyZ77DYpoU8DQi1W5U6zYgBbOZwbP6aJffd5/ZXXnul20Mgqk1T4rz9fMuPPlQIFFvAijFMqIvdlIkTZrbMj0TstsTv/zFwQzUkdc1rymkQzXmM1V5YbAs1BPQbWWnED61z91MZ5vUUWlEQYPxqYf3WciOGCOB122IxZktkCo6q009Jceg1kgOUuaGS7/tc2qGg1pBL4cynL6kAN1HMdSJ+IPsZwdnrUyX2KmphqA2ZahGIWpihGNEPT4GOg9+K+fWJ2AowA+XlSOTls8QVjqK2fl3iy/wp0neLTMG4f14NJQX+Dxcqd

%URL%
https://github.com/ggml-org/llama.cpp

%ARCH%
aarch64

%BUILDDATE%
1777469770

%PACKAGER%
@termux

%DEPENDS%
llama-cpp

%OPTDEPENDS%
vulkan-loader
vulkan-loader-generic
mesa-vulkan-icd-freedreno

