%FILENAME%
llama.cpp-b7574-1-x86_64.pkg.tar.zst

%NAME%
llama.cpp

%BASE%
llama.cpp

%VERSION%
b7574-1

%DESC%
Port of Facebook's LLaMA model in C/C++ (with system ggml support)

%CSIZE%
6964185

%ISIZE%
29342390

%SHA256SUM%
7adb9ec8aa9778895ad99c94a627dae0b970f75d5a4883a5e089f0ed9acc628f

%URL%
https://github.com/ggerganov/llama.cpp

%LICENSE%
MIT

%ARCH%
x86_64

%BUILDDATE%
1767041079

%PACKAGER%
lilac (on behalf of 依云) <lilydjwg@gmail.com>

%DEPENDS%
curl
gcc-libs
python
ggml

%OPTDEPENDS%
python-numpy: needed for convert_hf_to_gguf.py
python-safetensors: needed for convert_hf_to_gguf.py
python-sentencepiece: needed for convert_hf_to_gguf.py
python-pytorch: needed for convert_hf_to_gguf.py
python-transformers: needed for convert_hf_to_gguf.py

%MAKEDEPENDS%
cmake
git
vulkan-radeon

