Here’s what you need to know
Babylon Health has announced a world-first during a presentation streamed live from London’s Royal College of Physicians: The company’s AI, in a series of robust tests (including the relevant sections of the MRCGP exam), has demonstrated its ability to provide heath advice which is on-par with practicing clinicians.
The MRCGP exam is the final test for trainee General Practitioners (GPs), set by the Royal College of General Practitioners (RCGP). Trainee GPs who pass this assessment have demonstrated their competence and clinical skills to a level which is sufficiently high enough for them to undertake independent practice.
A key part of this exam tests a doctor’s ability to diagnose.
Babylon took a representative sample-set of questions testing diagnostic skills from publicly available RCGP sources, as well as independently published examination preparation materials, and mapped these to the current RCGP curriculum in order to ensure the questions resembled actual MRCGP questions as closely as possible.
The average pass mark over the past five years for real-life doctors was 72 per cent. In sitting the exam for the first time, Babylon’s AI scored 81 per cent. As the AI is continues to learn and accumulate knowledge, Babylon expects that subsequent testing will produce significant improvements in terms of results.
Important though exams are, doctors are presented with a much wider range of illnesses and conditions in their daily practice. Therefore, to further test the AI’s capabilities, Babylon’s team of scientists, clinicians and engineers next collaborated with the Royal College of Physicians, Dr Megan Mahoney (Chief of General Primary Care, Division of Primary Care and Population Health, Stanford University), and Dr Arnold DoRosario (Chief Population Health Officer, Yale New Haven Health) to test Babylon’s AI alongside seven highly-experienced primary care doctors using 100 independently-devised symptom sets (or ‘vignettes’).
Babylon’s AI scored 80 per cent for accuracy, while the seven doctors achieved an accuracy range of 64-94 per cent.
The accuracy of the AI was 98 per cent when assessed against conditions seen most frequently in primary care medicine. In comparison, when Babylon’s research team assessed experienced clinicians using the same measure, their accuracy ranged from 52-99 per cent.
Crucially, the safety of the AI was 97 per cent. This compares favourably to the doctors, whose average was 93.1 per cent.