The b9015 release of llama.cpp has been announced, featuring expanded support across various platforms. This update includes builds for macOS Apple Silicon, Ubuntu with ROCm 7.2, and Windows with CUDA 12 and 13. The release also extends Vulkan support to multiple systems, enhancing compatibility and performance. This iteration focuses on broadening the software's reach, making it a more versatile tool for developers working with different hardware setups.
Read originalThe b9018 release of llama.cpp continues its trend of broadening platform compatibility, now supporting a wide array of systems including macOS, Linux, Windows, and Android. Notably, it introduces Vulkan support on Ubuntu and Windows, and adds ROCm 7.2 for AMD GPUs, which is a significant step for users seeking alternatives to NVIDIA's CUDA. This release doesn't bring new models or quantization methods, but it solidifies llama.cpp's position as a versatile inference runtime across diverse hardware configurations. Users can now leverage these enhancements to optimize performance on their specific setups.
The b9019 release of llama.cpp brings notable changes by relocating functions like load_hparams and load_tensors to be defined per model, enhancing the flexibility for developers. This structural shift is complemented by the introduction of build_graph and refined switch case logic, which collectively improve the system's modularity. These updates facilitate easier adaptation to various hardware setups, including macOS, Linux, and Windows environments. Although no new model architectures are introduced, the release sets a foundation for more efficient development and deployment, particularly with support for configurations like KleidiAI on Apple Silicon and ROCm 7.2 on AMD GPUs.
The latest b9025 release of llama.cpp continues its trend of broadening platform compatibility, now supporting a wide array of systems including macOS, Linux, Windows, and Android. Notably, it introduces Vulkan support on Ubuntu and Windows, and adds ROCm 7.2 for Ubuntu, enhancing GPU performance options. This release doesn't introduce new models but focuses on making llama.cpp a versatile tool across different hardware configurations. By expanding its reach, llama.cpp is positioning itself as a go-to runtime for diverse computing environments, ensuring developers can leverage its capabilities regardless of their platform choice.
© Matt WolfeDeepSeek V4 is an open-source AI model offering near state-of-the-art capabilities at a significantly lower cost than competitors.
The v0.18.2rc0 release includes a fix for handling the max_pixels parameter in the PaddleOCR-VL image processor across transformations.
© Lev SelectorAnthropic has released a suite of plugins that enhance the Claude ecosystem.