
GitHub has introduced team-level Copilot usage metrics through its API, allowing enterprise administrators and organization owners to track Copilot adoption and activity by team. The new feature provides detailed insights into active users, completions, and other metrics, broken down by language, IDE, feature, and model. This development enables organizations to identify which teams are leading in adoption and which may need further support. The metrics are accessible via REST API endpoints, though no dashboard is available for these team-level insights.
Read original
© GitHub ChangelogGitHub has launched a technical preview of its Copilot app, offering a desktop experience for agentic development directly within GitHub. This app allows developers to start sessions from existing work artifacts like issues or pull requests, maintaining context and isolation across tasks. It supports automation of routine tasks and integrates seamlessly with GitHub's existing review and merge processes. This development marks a step towards more integrated and efficient workflows for developers using GitHub's ecosystem.
© GitHub ChangelogThe latest llama.cpp release, b9145, tackles a significant issue with SYCL's memory allocation on multi-GPU systems, particularly those using Intel Arc Pro GPUs. By replacing sycl::malloc_device with zeMemAllocDevice, the update drastically reduces system RAM usage from 60 GiB to just 6.7 GiB for a 15.6 GiB model, preventing out-of-memory crashes without sacrificing performance. This change is crucial for developers working with large models on multi-GPU setups, as it ensures more efficient memory management. The update also includes several improvements and bug fixes, enhancing the robustness of the SYCL backend.
Llama.cpp's latest release enhances its capabilities with a non-backtracking tokenizer handler specifically designed for Qwen3.5. This update significantly improves Unicode tokenization, addressing stack overflow issues that occur with long inputs. By adapting the previous Qwen2 fix to meet Qwen3.5's regex requirements, including support for accent marks, the update ensures more reliable text processing. Developers can now expect more stable performance when handling complex Unicode inputs, benefiting from the robust tokenization across different operating systems and hardware configurations. This means smoother operations on platforms like macOS with KleidiAI, Ubuntu with ROCm 7.2, and Windows with CUDA 12 and 13.
GitHub's Copilot cloud agent now includes an auto model selection feature, which intelligently determines the most suitable model based on system health and performance. This advancement not only enhances the user experience by ensuring optimal model usage but also provides a 10% discount on the normal model multiplier. Users will also enjoy the benefit of not being subject to weekly rate limits, making the service more efficient and accessible. This development highlights GitHub's ongoing efforts to refine AI-driven coding tools for developers, offering a more seamless and cost-effective solution.