%FILENAME%
llama.cpp-b7708-1-x86_64.pkg.tar.zst

%NAME%
llama.cpp

%BASE%
llama.cpp

%VERSION%
b7708-1

%DESC%
Port of Facebook's LLaMA model in C/C++ (with system ggml support)

%CSIZE%
6838580

%ISIZE%
28913603

%SHA256SUM%
6b4891025c91a0c904c8c5f9fa8fba414ad4053350e7c7842abfc62aba8fd0e4

%URL%
https://github.com/ggerganov/llama.cpp

%LICENSE%
MIT

%ARCH%
x86_64

%BUILDDATE%
1768193583

%PACKAGER%
lilac (on behalf of 依云) <lilydjwg@gmail.com>

%DEPENDS%
curl
gcc-libs
python
ggml

%OPTDEPENDS%
python-numpy: needed for convert_hf_to_gguf.py
python-safetensors: needed for convert_hf_to_gguf.py
python-sentencepiece: needed for convert_hf_to_gguf.py
python-pytorch: needed for convert_hf_to_gguf.py
python-transformers: needed for convert_hf_to_gguf.py

%MAKEDEPENDS%
cmake
git
vulkan-radeon

