The b9056 release of llama.cpp has been announced, featuring expanded support across multiple platforms. This update includes macOS Apple Silicon with KleidiAI enabled, various Linux configurations such as Ubuntu with Vulkan and ROCm 7.2, and enhanced Windows support with CUDA 12 and 13 DLLs. These additions aim to provide developers with more flexibility and performance optimization options across different hardware environments. The release underscores llama.cpp's commitment to being a versatile inference runtime for a wide range of systems.
Read originalThe latest b9057 release of llama.cpp continues its trend of broadening platform compatibility, now optimizing for RISC-V CPUs with q1_0 dot support. This update enhances performance across a wide array of systems, including macOS, Linux, Windows, and Android, with specific builds for Apple Silicon, Vulkan, and CUDA environments. Notably, the inclusion of ROCm 7.2 for Ubuntu x64 and CUDA 13 for Windows x64 signifies a commitment to supporting diverse hardware configurations. While no new models are introduced, this release solidifies llama.cpp's position as a versatile inference runtime across multiple architectures.
The b9058 release of llama.cpp significantly enhances its reach by supporting more platforms, making it a versatile tool for developers. It now includes KleidiAI support for macOS Apple Silicon, which optimizes performance on Apple's ARM architecture. The update also brings Vulkan support to both Ubuntu and Windows, boosting graphics processing capabilities. With the integration of ROCm 7.2 for Ubuntu, AMD GPU users see improved compatibility, narrowing the gap with NVIDIA. Additionally, Windows users benefit from CUDA 12 and 13 DLLs, catering to NVIDIA GPU needs. This release positions llama.cpp as a more adaptable solution for developers working with diverse hardware setups.
The latest b9060 release of llama.cpp introduces several new SYCL operations, including FILL, CUMSUM, and DIAG, which expand the library's computational capabilities. This update also addresses a critical issue that caused aborts during test-backend-ops, ensuring more stable performance. With the addition of scope_dbg_print to both new and existing SYCL operations, developers gain enhanced debugging tools. This release continues to broaden llama.cpp's platform support, making it a more versatile tool for developers working across different environments.
© Matt WolfeDeepSeek V4 is an open-source AI model offering near state-of-the-art capabilities at a significantly lower cost than competitors.
The v0.18.2rc0 release includes a fix for handling the max_pixels parameter in the PaddleOCR-VL image processor across transformations.
© Lev SelectorAnthropic has released a suite of plugins that enhance the Claude ecosystem.