Posted by Alumni from Nature
February 6, 2026
Large language models (LLMs) can pass postgraduate medical examinations and help clinicians to make diagnoses, at least in controlled benchmarking tests. But are they useful in real-world settings, which have too few physicians to check the answers, as well as long patient lists and limited resources' Two studies published in Nature Health on 6 February suggest that they are up to the task. The work reveals that cheap-to-use LLMs can boost diagnostic success rates, even outperforming trained clinicians, in health-care settings in Rwanda1 and Pakistan2. In Rwanda, chatbot answers outscored those of local clinicians across every metric assessed. And in Pakistan, physicians using LLMs to aid their diagnosis achieved a mean diagnostic reasoning score of 71%, versus 43% for those using conventional resources. 'The papers highlight how LLMs might be able to support clinicians in lower- and middle-income countries to improve the level of care,' says Caroline Green, director of research at... learn more