# Generated by makepkg 7.0.0
# using fakeroot version 1.37.1.2
pkgname = llama.cpp
pkgbase = llama.cpp
xdata = pkgtype=pkg
pkgver = b6960-1
pkgdesc = Port of Facebook's LLaMA model in C/C++ (with system ggml support)
url = https://github.com/ggerganov/llama.cpp
builddate = 1762377184
packager = lilac (on behalf of 依云) <lilydjwg@gmail.com>
size = 21231807
arch = x86_64
license = MIT
depend = curl
depend = gcc-libs
depend = python
depend = ggml
optdepend = python-numpy: needed for convert_hf_to_gguf.py
optdepend = python-safetensors: needed for convert_hf_to_gguf.py
optdepend = python-sentencepiece: needed for convert_hf_to_gguf.py
optdepend = python-pytorch: needed for convert_hf_to_gguf.py
optdepend = python-transformers: needed for convert_hf_to_gguf.py
makedepend = cmake
makedepend = git
makedepend = vulkan-radeon
