The b9064 release of llama.cpp has been announced, featuring expanded support across various platforms. Notably, the update includes KleidiAI integration for Apple Silicon, ROCm 7.2 support for Ubuntu x64, and CUDA 12 and 13 compatibility for Windows. These enhancements aim to improve performance and accessibility for developers working with different hardware setups. While no new models are introduced, the release strengthens llama.cpp's role as a versatile inference runtime.
Read originalThe latest b9056 release of llama.cpp continues its trend of broadening platform compatibility, now including support for macOS Apple Silicon with KleidiAI enabled and a variety of Linux configurations such as Ubuntu with Vulkan and ROCm 7.2. This update also enhances Windows support with CUDA 12 and 13 DLLs, making it more versatile for developers working across different environments. While there are no groundbreaking new features, the release solidifies llama.cpp's position as a flexible inference runtime across diverse hardware setups. Developers can now leverage these updates to optimize performance on their specific systems, whether they're using Apple Silicon, AMD, or NVIDIA GPUs.
The latest b9057 release of llama.cpp continues its trend of broadening platform compatibility, now optimizing for RISC-V CPUs with q1_0 dot support. This update enhances performance across a wide array of systems, including macOS, Linux, Windows, and Android, with specific builds for Apple Silicon, Vulkan, and CUDA environments. Notably, the inclusion of ROCm 7.2 for Ubuntu x64 and CUDA 13 for Windows x64 signifies a commitment to supporting diverse hardware configurations. While no new models are introduced, this release solidifies llama.cpp's position as a versatile inference runtime across multiple architectures.
The v0.18.2rc0 release includes a fix for handling the max_pixels parameter in the PaddleOCR-VL image processor across transformations.