MIT researchers developed a method to better identify overconfident large language models (LLMs) by measuring cross-model disagreement, enhancing the reliability of predictions. This new approach aims to reduce the risk of misleading outputs in critical fields like healthcare and finance.
Read original
© Microsoft ResearchMicrosoft Research explores vulnerabilities in networks of AI agents, highlighting risks that emerge only through interaction. Their tests reveal how malicious messages can propagate and manipulate agent behavior.
Google DeepMind is researching the development of an AI co-clinician aimed at augmenting healthcare delivery. This initiative focuses on integrating AI into clinical settings to enhance patient care.