AI assists radiologists in detecting fractures, improves workflow

Research published Tuesday in Radiology found comparable performance between artificial intelligence and clinicians for fracture detection. 

Experts examined 42 studies that compared the diagnostic performance of clinicians and AI in fracture detection and found no statistical differences in sensitivity between the two. In fact, a difference of only 1% in sensitivity was observed. 

“We found that AI performed with a high degree of accuracy comparable to clinician performance,” lead author Rachel Kuo, from the Botnar Research Center, Nuffield Department of Orthopaedics, Rheumatology and Musculoskeletal Sciences in Oxford, England, and co-authors discussed. “Importantly, we found this to be the case when AI was validated using independent external datasets, suggesting that the results may be generalizable to the wider population.” 

The team’s meta-analysis of peer-reviewed publications included studies that developed and/or validated AI algorithms for the specific purpose of fracture detection. This included more than 55,000 images from 37 studies that used radiographs and five that used computed tomography scans to identify fractures. 

For the internal validation sets in those studies, the pooled sensitivity and specificity of AI was 92% and 91%, compared to 91% and 92% for clinicians. These numbers were similar in the external validation sets for both readers as well. 

The experts noted that their findings point to a promising future for educational and clinical application of AI. In addition to improving workflows, AI could help to reduce the rate of early fracture misdiagnoses in emergency settings, especially in circumstances when patients have sustained multiple fractures. 

“The results from this meta-analysis cautiously suggest that AI is noninferior to clinicians in terms of diagnostic performance in fracture detection, showing promise as a useful diagnostic tool,” the authors wrote. “Current artificial intelligence is designed as a diagnostic adjunct and may improve workflow through screening or prioritizing images on work lists and highlighting regions of interest for a reporting radiologist. AI may also improve diagnostic certainty through acting as a second reader for clinicians or as an interim report prior to radiologist interpretation.” 

The researchers did note, however, that AI is not a replacement for clinicians, and that readers must still practice caution when applying algorithms as an adjunct. 

“It remains important for clinicians to continue to exercise their own judgment,” Kuo said. “AI is not infallible and is subject to bias and error.” 

More on orthopedic imaging:

Radiomics-clinical model accurately predicts osteoporotic spinal fracture timeline on CT images

FDA greenlights AI software that detects fractures and traumatic injuries

Patient discussions—not X-ray results—should guide common forearm fracture treatment in older adults

AI helps radiologists and non-specialists detect fractures on X-rays, but experts remain dubious

Hannah murhphy headshot

In addition to her background in journalism, Hannah also has patient-facing experience in clinical settings, having spent more than 12 years working as a registered rad tech. She joined Innovate Healthcare in 2021 and has since put her unique expertise to use in her editorial role with Health Imaging.

Around the web

The new technology shows early potential to make a significant impact on imaging workflows and patient care. 

Richard Heller III, MD, RSNA board member and senior VP of policy at Radiology Partners, offers an overview of policies in Congress that are directly impacting imaging.
 

The two companies aim to improve patient access to high-quality MRI scans by combining their artificial intelligence capabilities.