
DeepSeek V4 has been introduced as a nearly state-of-the-art AI model that is open-sourced and offers substantial cost savings. It charges $1.74 per million input tokens, which is significantly cheaper than the $5 charged by competitors like GPT-5.5 and Claude Opus 4.7. Additionally, DeepSeek V4 can be run locally, providing users with enhanced privacy. This combination of affordability, capability, and privacy could potentially disrupt the current AI market landscape.
Read originalThe b9008 release of llama.cpp continues its trend of broadening platform support, making it a versatile tool for developers across various systems. This update includes new builds for macOS, Linux, Windows, and Android, with notable additions like Vulkan support on Ubuntu and Windows, and ROCm 7.2 on Ubuntu. By enhancing compatibility with different architectures, including Apple Silicon and Intel on macOS, and CUDA on Windows, llama.cpp is positioning itself as a go-to runtime for diverse hardware environments. While there are no groundbreaking new features, the release solidifies llama.cpp's role as a flexible and accessible inference tool for developers.
The b9004 release of llama.cpp introduces support for various platforms including macOS, Linux, Android, and Windows.
DeepSeek has launched a preview of its V4 model.