%FILENAME%
llama-cpp-backend-vulkan-0.0.0.b8969-0-x86_64.pkg.tar.xz

%NAME%
llama-cpp-backend-vulkan

%BASE%
llama-cpp

%VERSION%
0.0.0.b8969-0

%DESC%
Vulkan Backend for LLaMA.cpp

%CSIZE%
3334672

%ISIZE%
33124944

%MD5SUM%
812f665e5d0e57ee2b55daea385941ec

%SHA256SUM%
a3ef2bcd5f6fc09dd42bfa6b58642cd2164342060c0b74e694254769d5b5db30

%PGPSIG%
iQGzBAABCgAdFiEEmY3icxjoZ+qXa6h3OJzu1kVz38oFAmnyEmoACgkQOJzu1kVz38rywwwAho05vHNCj9q6wEDxgWm9Uq5/xddQ4Yo1KbMYl70C8jEVKyj/QhGPObQOwOUDEXJhEUWV1zxVKo7bq1Ha2Yv0UFW/RfCri2J1wBctZakyQzsNVJzaOKLYL0udI87u+nGN2zwnPyxBmyjetlXbvIxr++shkJRnwiApaMUaE+D7lNGmKxwbPveGIXH+NZzX3Nlg4dXA/5EPN82d6ZvMIz4mH9HNfo/wZNSCG10ORpvAFacSP7TWCzZ7ERGaphdF1IxoF7iTtdwbJ0SRaKfCCa8YJAPKtpcBM2+pubup8UW9oYcqX0SBOZGcQy4DH8sdFSXPv+Uv6sFKQ49ybzjOP07mV7mSbtv/2vJITpz0BIWIA54MLZJFwvBnvk/vliowxg9NvNS15xK+Ro5fZ5aQTsrKn6kbsR6uvbm4PXpmDc+FAEwjKDd5KKMz8D/sI54T128cJ5lO1Wn11EdFkisFXVk47BfPzuDDtBaLD0+z00BN5oVxNi1+cX+13oGnH2zkXo1n

%URL%
https://github.com/ggml-org/llama.cpp

%ARCH%
x86_64

%BUILDDATE%
1777469770

%PACKAGER%
@termux

%DEPENDS%
llama-cpp

%OPTDEPENDS%
vulkan-loader
vulkan-loader-generic
mesa-vulkan-icd-freedreno

