aboutsummaryrefslogtreecommitdiffstats
path: root/llama.cpp-vulkan/README
diff options
context:
space:
mode:
authorDanilo M. <danix@danix.xyz>2026-04-01 22:56:07 +0200
committerDanilo M. <danix@danix.xyz>2026-04-01 22:58:53 +0200
commit1045963959ddfb697898fa90476f837aae4e2881 (patch)
treef4f93fcd99ea0f2dfa1342d79baef10d5f66cc7c /llama.cpp-vulkan/README
parentfbb8e8f558f541d6a573105dc369ba7c563f78eb (diff)
downloadmy-slackbuilds-1045963959ddfb697898fa90476f837aae4e2881.tar.gz
my-slackbuilds-1045963959ddfb697898fa90476f837aae4e2881.zip
repo: move all packages under SlackBuilds/ subfolder
Reorganize repository by moving all 11 package directories into a new SlackBuilds/ subfolder. Update README.md, CLAUDE.md, and hooks/post-commit to reflect the new path structure. Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
Diffstat (limited to 'llama.cpp-vulkan/README')
-rw-r--r--llama.cpp-vulkan/README22
1 files changed, 0 insertions, 22 deletions
diff --git a/llama.cpp-vulkan/README b/llama.cpp-vulkan/README
deleted file mode 100644
index 5509d44..0000000
--- a/llama.cpp-vulkan/README
+++ /dev/null
@@ -1,22 +0,0 @@
-llama.cpp
-
-LLM inference in C/C++
-
-The main goal of llama.cpp is to enable LLM inference with minimal
-setup and state-of-the-art performance on a wide range of hardware
-locally and in the cloud.
-
- - Plain C/C++ implementation without any dependencies
- - Apple silicon is a first-class citizen - optimized via ARM NEON,
- Accelerate and Metal frameworks
- - AVX, AVX2, AVX512 and AMX support for x86 architectures
- - RVV, ZVFH, ZFH, ZICBOP and ZIHINTPAUSE support for RISC-V
- architectures
- - 1.5-bit, 2-bit, 3-bit, 4-bit, 5-bit, 6-bit, and 8-bit integer
- quantization for faster inference and reduced memory use
- - Custom CUDA kernels for running LLMs on NVIDIA GPUs (support for
- AMD GPUs via HIP and Moore Threads GPUs via MUSA)
- - Vulkan and SYCL backend support
- - CPU+GPU hybrid inference to partially accelerate models larger than
- the total VRAM capacity
-