%FILENAME%
llama.cpp-b7710-1-x86_64.pkg.tar.zst

%NAME%
llama.cpp

%BASE%
llama.cpp

%VERSION%
b7710-1

%DESC%
Port of Facebook's LLaMA model in C/C++ (with system ggml support)

%CSIZE%
6838437

%ISIZE%
28913603

%SHA256SUM%
e426f4d8bc67368834fce3a13a38f4b9b79c80d1c95fdb17f32e2923ecf1bab1

%URL%
https://github.com/ggerganov/llama.cpp

%LICENSE%
MIT

%ARCH%
x86_64

%BUILDDATE%
1768223113

%PACKAGER%
lilac (on behalf of 依云) <lilydjwg@gmail.com>

%DEPENDS%
curl
gcc-libs
python
ggml

%OPTDEPENDS%
python-numpy: needed for convert_hf_to_gguf.py
python-safetensors: needed for convert_hf_to_gguf.py
python-sentencepiece: needed for convert_hf_to_gguf.py
python-pytorch: needed for convert_hf_to_gguf.py
python-transformers: needed for convert_hf_to_gguf.py

%MAKEDEPENDS%
cmake
git
vulkan-radeon

