16 × AIAI signal, amplified
AI newsAboutSources
TelegramFollow on Telegram
AI newsAboutSources
16 × AIAI signal, amplified

An AI news engine that ingests trusted sources, scores with Claude, and posts only what clears the bar.

Follow on Telegram →

Subscribe

  • Telegram
  • RSS
  • All channels

Legal

  • Privacy
  • Imprint
© 2026 16 × AI. All rights reserved.Curated by Claude. Posts every 6 hours. No newsletter, no funnel.
Home/Models & Labs
Models & Labs

llama.cpp b9010 Release Fixes CUDA Multi-GPU Issue

llama.cpp Releases·May 3, 2026·high confidence

Why it matters

  • →Fixes critical multi-GPU support issue, enhancing performance for CUDA users.
  • →Expands compatibility across various platforms, including Apple Silicon and Vulkan.
  • →Improves stability and reliability of llama.cpp for diverse hardware setups.

The b9010 release of llama.cpp resolves a significant issue with CUDA device PCI bus ID detection that led to out-of-memory errors by ignoring additional GPUs. This fix enhances multi-GPU support, particularly benefiting users on Windows platforms. The update also includes platform-specific improvements for macOS, Linux, and Windows, with notable support for Apple Silicon and Vulkan. This release focuses on improving stability and compatibility rather than introducing new features.

Read original

More from llama.cpp Releases

Open Sourcemodels

llama.cpp b9008 Release Expands Platform Support

The b9008 release of llama.cpp continues its trend of broadening platform support, making it a versatile tool for developers across various systems. This update includes new builds for macOS, Linux, Windows, and Android, with notable additions like Vulkan support on Ubuntu and Windows, and ROCm 7.2 on Ubuntu. By enhancing compatibility with different architectures, including Apple Silicon and Intel on macOS, and CUDA on Windows, llama.cpp is positioning itself as a go-to runtime for diverse hardware environments. While there are no groundbreaking new features, the release solidifies llama.cpp's role as a flexible and accessible inference tool for developers.

llama.cpp Releases·May 3, 2026
Models & Labsother

b9002 Release for Llama.cpp

The b9002 version of Llama.cpp has been released, supporting multiple platforms.

llama.cpp Releases·May 2, 2026
Open Sourceother

b9004 Release for Multiple Platforms

The b9004 release of llama.cpp introduces support for various platforms including macOS, Linux, Android, and Windows.

llama.cpp Releases·May 2, 2026

More in Models & Labs

DeepSeek V4 Pro Launches with 1.6T Parameters© Lev Selector
Models & Labsmodels

DeepSeek V4 Pro Launches with 1.6T Parameters

DeepSeek V4 Pro is a new AI model with 1.6 trillion parameters.

Lev Selector·May 1, 2026
DeepSeek V4 Preview Released© Matt Wolfe
Models & Labsmodels

DeepSeek V4 Preview Released

DeepSeek has launched a preview of its V4 model.

Matt Wolfe·May 1, 2026
NVIDIA Launches Nemotron 3 Nano Omni© Matt Wolfe
Models & Labsagents

NVIDIA Launches Nemotron 3 Nano Omni

NVIDIA has introduced the Nemotron 3 Nano Omni multimodal AI agent.

Matt Wolfe·May 1, 2026