The latest b9008 release of llama.cpp has expanded its platform support, offering builds for macOS, Linux, Windows, and Android. This update includes support for Vulkan on both Ubuntu and Windows, as well as ROCm 7.2 on Ubuntu, enhancing its compatibility with various hardware architectures. The release aims to make llama.cpp a more versatile tool for developers by supporting a wide range of systems, including Apple Silicon, Intel, and CUDA. This update reinforces llama.cpp's position as a flexible inference runtime across different platforms.
Read originalThe b9010 release of llama.cpp tackles a crucial bug in CUDA device PCI bus ID detection, which previously caused out-of-memory errors by failing to recognize multiple GPUs. This update significantly improves multi-GPU support, especially for Windows users leveraging CUDA. The release also brings enhancements for macOS, Linux, and Windows, with specific improvements for Apple Silicon and Vulkan integration. While it doesn't introduce groundbreaking new features, this update strengthens llama.cpp's reliability and compatibility across different hardware setups, including ROCm 7.2 and KleidiAI on Apple Silicon.
The b9002 version of Llama.cpp has been released, supporting multiple platforms.
The b9004 release of llama.cpp introduces support for various platforms including macOS, Linux, Android, and Windows.
© Matt WolfeDeepSeek V4 is an open-source AI model offering near state-of-the-art capabilities at a significantly lower cost than competitors.
The v0.18.2rc0 release includes a fix for handling the max_pixels parameter in the PaddleOCR-VL image processor across transformations.
© Lev SelectorAnthropic has released a suite of plugins that enhance the Claude ecosystem.