AI As Malpractice Safety Net

One of the emerging use cases for AI in radiology is as a safety net that could help hospitals avoid malpractice cases by catching errors made by radiologists before they can cause patient harm. The topic was reviewed in a Sunday presentation at RSNA 2024

Clinical AI adoption has been held back by economic factors such as limited reimbursement and the lack of strong return on investment. 

  • Healthcare providers want to know that their AI investments will pay off, either through direct reimbursement from payors or improved operational efficiency.

At the same time, providers face rising malpractice risk, with a number of recent high-profile legal cases.

  • For example, a New York hospital was hit with a $120M verdict after a resident physician working the night shift missed a pulmonary embolism. 

Could AI limit risk by acting as a backstop to radiologists? 

  • At RSNA 2024, Benjamin Strong, MD, chief medical officer at vRad, described how they have deployed AI as a QA safety net. 

vRad mostly develops its own AI algorithms, with the first algorithm deployed in 2015. 

  • vRad is running AI algorithms as a backstop for 13 critical pathologies, from aortic dissection to superior mesenteric artery occlusion.

vRad’s QA workflow begins after the radiologist issues a final report (without using AI), and an algorithm then reviews the report automatically. 

  • If discrepancies are found the report is sent to a second radiologist, who can kick the study back to the original radiologist if they believe an error has occurred. The entire process takes 20 minutes. 

In a review of the program over one year, vRad found …

  • Corrections were made for about 1.5k diagnoses out of 6.7M exams.
  • The top five AI models accounted for over $8M in medical malpractice savings. 
  • Three pathologies – spinal epidural abscess, aortic dissection, and ischemic bowel due to SMA occlusion – would have amounted to $18M in payouts over four years.
  • Adding intracranial hemorrhage and pulmonary embolism creates what Strong called the “Big Five” of pathologies that are either the most frequently missed or the most expensive when missed.

The Takeaway

The findings offer an intriguing new use case for AI adoption. Avoiding just one malpractice verdict or settlement would more than pay for the cost of AI installation, in most cases many times over. How’s that for return on investment?

How Are Doctors Using AI?

How are healthcare providers who have adopted AI really using it? A new Medscape/HIMSS survey found that most providers are using AI for administrative tasks, while medical image analysis is also one of the top AI use cases. 

AI has the potential to revolutionize healthcare, but many industry observers have been frustrated with the slow pace of clinical adoption. 

  • Implementation challenges, regulatory issues, and lack of reimbursement are among the reasons keeping more healthcare providers from embracing the technology.

But the Medscape/HIMSS survey shows some early successes for AI … as well as lingering questions. 

  • Researchers surveyed a total of 846 people in the U.S. who were either executive or clinical leaders, practicing physicians or nurses, or IT professionals, and whose practices were already using AI in some way.

The top four tasks for which AI is being used were administrative rather than clinical, with image analysis occupying the fifth spot … 

  1. Transcribing patient notes (36%). 
  2. Transcribing business meetings (32%).
  3. Creating routine patient communications (29%).
  4. Performing patient record-keeping (27%).
  5. Analyzing medical images (26%).

The survey also analyzed attitudes toward AI, finding …

  • 57% said AI helped them be more efficient and productive.
  • But lower marks were given for reducing staff hours (10%) and lowering costs (31%).
  • AI got the highest marks for helping with transcription of business meetings (77%) and patient notes (73%), reviewing medical literature (72%), and medical image analysis (70%).

The findings track well with developments at last week’s RSNA 2024, where AI algorithms dedicated to non-clinical tasks like radiology report generation, scheduling, and operation analysis showed growing prominence. 

  • Indeed, many AI developers have specifically targeted the non-clinical space, both because commercialization is easier (FDA authorization is not typically needed) and because doctors often say they need more help with administrative rather than clinical tasks.

The Takeaway

While it’s easy to be impatient with AI’s slow uptake, the Medscape/HIMSS survey shows that AI adoption is indeed occurring at medical practices. And while image analysis was radiology’s first AI use case, speeding up workflow and administrative tasks may end up being the technology’s most impactful application.

How Should AI Be Monitored?

Once an AI algorithm has been approved and moves into clinical use, how should its performance be monitored? This question was top of mind at last week’s meeting of the FDA’s new Digital Health Advisory Committee.

AI has the potential to radically reshape healthcare and help clinicians manage more patients with fewer staff and other resources. 

  • But AI also represents a regulatory challenge because it’s constantly learning, such that after a few years an AI algorithm might be operating much differently from the version first approved by the FDA – especially with generative AI. 

This conundrum was a point of discussion at last week’s DHAC meeting, which was called specifically to focus on regulation of generative AI, and could result in new rules covering all AI algorithms. (An executive summary that outlines the FDA’s thinking is available for download.)

Radiology was well-represented at DHAC, understandable given it has the lion’s share of authorized algorithms (73% of 950 devices at last count). 

  • A half-dozen radiology AI experts gave presentations over two days, including Parminder Bhatia of GE HealthCare; Nina Kottler, MD, of Radiology Partners; Pranav Rajpurkar, PhD, of Harvard; and Keith Dreyer, DO, PhD, and Bernardo Bizzo, MD, PhD, both of Mass General Brigham and the ACR’s Data Science Institute.  

Dreyer and Bizzo directly addressed the question of post-market AI surveillance, discussing ongoing efforts to track AI performance, including … 

The Takeaway

Last week’s DHAC meeting offers a fascinating glimpse at the issues the FDA is wrestling with as it contemplates stronger regulation of generative AI. Fortunately, radiology has blazed a trail in setting up structures like ARCH-AI and Assess-AI to monitor AI performance, and the FDA is likely to follow the specialty’s lead as it develops a regulatory framework.

Low-Dose CT Confounds CAD in Kids

When it comes to pediatric CT scans, clinicians should make every effort to reduce dose as much as possible. But a new study in AJR indicates that lower CT radiation dose can affect the performance of software tools like computer-aided detection. 

Initiatives like the Image Wisely and Image Gently projects have succeeded in raising awareness of radiation dose and have helped radiologists find ways to reduce it.

But every little bit counts in pediatric dose reduction, especially given that one CT exam can raise the risk of developing cancer by 0.35%. 

  • Imaging tools like AI and CAD could help, but there have been few studies examining the performance of pulmonary CAD software developed for adults in analyzing scans of children.

To address that gap, researchers including radiologists from Cincinnati Children’s Hospital Medical Center investigated the performance of two open-source CAD algorithms trained on adults for detecting lung nodules in 73 patients with a mean age of 14.7 years. 

  • The algorithms included FlyerScan, a CAD developed by the authors, and MONAI, an open-source project for deep learning in medical imaging. 

Scans were acquired at standard-dose (mean effective dose=1.77 mSv) and low-dose (mean effective dose=0.32 mSv) levels, with the results showing that both algorithms turned in lower performance at lower radiation dose for nodules 3-30 mm … 

  • FlyerScan saw its sensitivity decline (77% vs. 67%) and detected fewer 3mm lung nodules (33 vs. 24).
  • MONAI also saw lower sensitivity (68% vs. 62%) and detected fewer 3mm lung nodules (16 vs. 13).
  • Reduced sensitivity was more pronounced for nodules less than 5 mm.

The findings should be taken with a grain of salt, as the open-source algorithms were not originally trained on pediatric data.

  • But the results do underscore the challenge in developing image analysis software optimized for pediatric applications.

The Takeaway

With respect to low radiation dose and high AI accuracy in CT scans of kids, radiologists may not be able to have their cake and eat it too – yet. More work will be needed before AI solutions developed for adults can be used in children.

Mammography AI Predicts Cancer Before It’s Detected

A new study highlights the predictive power of AI for mammography screening – before cancers are even detected. Researchers in a study JAMA Network Open found that risk scores generated by Lunit’s Insight MMG algorithm predicted which women would develop breast cancer – years before radiologists found it on mammograms. 

Mammography image analysis has always been one of the most promising use cases for AI – even dating back to the days of computer-aided detection in the early 2000s. 

  • Most mammography AI developers have focused on helping radiologists identify suspicious lesions on mammograms, or triage low-risk studies so they don’t require extra review.

But a funny thing has happened during clinical use of these algorithms – radiologists found that AI-generated risk scores appeared to predict future breast cancers before they could be seen on mammograms. 

  • Insight MMG marks areas of concern and generates a risk score of 0-100 for the presence of breast cancer (higher numbers are worse). 

Researchers decided to investigate the risk scores’ predictive power by applying Insight MMG to screening mammography exams acquired in the BreastScreen Norway program over three biennial rounds of screening from 2004 to 2018. 

  • They then correlated AI risk scores to clinical outcomes in exams for 116k women for up to six years after the initial screening round.

Major findings of the study included … 

  • AI risk scores were higher for women who later developed cancer, 4-6 years before the cancer was detected.
  • The difference in risk scores increased over three screening rounds, from 21 points in the first round to 79 points in the third round.
  • Risk scores had very high accuracy by the third round (AUC=0.93).
  • AI scores were more accurate than existing risk tools like the Tyrer-Cuzick model.

How could AI risk scores be used in clinical practice? 

  • Women without detectable cancer but with high scores could be directed to shorter screening intervals or screening with supplemental modalities like ultrasound or MRI.

The Takeaway
It’s hard to overstate the significance of the new results. While AI for direct mammography image interpretation still seems to be having trouble catching on (just like CAD did), risk prediction is a use case that could direct more effective breast screening. The study is also a major coup for Lunit, continuing a string of impressive clinical results with the company’s technology.

AI Recon Cuts CT Radiation Dose

Artificial intelligence got its start in radiology as a tool to help medical image interpretation, but much of AI’s recent progress is in data reconstruction: improving images before radiologists even get to see them. Two new studies underscore the potential of AI-based reconstruction to reduce CT radiation dose while preserving image quality. 

Radiology vendors and clinicians have been remarkably successful in reducing CT radiation dose over the past two decades, but there’s always room for improvement. 

  • In addition to adjusting CT scanning protocols like tube voltage and current, data reconstruction protocols have been introduced to take images acquired at lower radiation levels and “boost” them to look like full-dose images. 

The arrival of AI and other deep learning-based technologies has turbocharged these efforts. 

They compared DLIR operating at high strength to GE’s older ASiR-V protocol in CCTA scans with lower tube voltage (80 kVp), finding that deep learning reconstruction led to …

  • 42% reduction in radiation dose (2.36 mSv vs. 4.07)
  • 13% reduction in contrast dose (50 mL vs. 58 mL).
  • Better signal- and contrast-to-noise ratios.
  • Higher image quality ratings.

In the second study, researchers from China including two employees of United Imaging Healthcare used a deep learning reconstruction algorithm to test ultralow-dose CT scans for coronary artery calcium scoring. 

  • They wanted to see if CAC scoring could be performed with lower tube voltage and current (80 kVp/20 mAs) and how the protocol compared to existing low-dose scans.

In tests with 156 patients, they found the ultralow-dose protocol produced …

  • Lower radiation dose (0.09 vs. 0.49 mSv).
  • No difference in CAC scoring or risk categorization. 
  • Higher contrast-to-noise ratio.

The Takeaway

AI-based data reconstruction gives radiologists the best of both worlds: lower radiation dose with better-quality images. These two new studies illustrate AI’s potential for lowering CT dose to previously unheard-of levels, with major benefits for patients.

AI Detects Interval Cancer on Mammograms

In yet another demonstration of AI’s potential to improve mammography screening, a new study in Radiology shows that Lunit’s Insight MMG algorithm detected nearly a quarter of interval cancers missed by radiologists on regular breast screening exams. 

Breast screening is one of healthcare’s most challenging cancer screening exams, and for decades has been under attack by skeptics who question its life-saving benefit relative to “harms” like false-positive biopsies.  

  • But AI has the potential to change the cost-benefit equation by detecting a higher percentage of early-stage cancers and improving breast cancer survival rates. 

Indeed, 2024 has been a watershed year for mammography AI. 

U.K. researchers used Insight MMG (also used in the BreastScreen Norway trial) to analyze 2.1k screening mammograms, of which 25% were interval cancers (cancers occurring between screening rounds) and the rest normal. 

  • The AI algorithm generates risk scores from 0-100, with higher scores indicating likelihood of malignancy, and this study was set at a 96% specificity threshold, equivalent to the average 4% recall rate in the U.K. national breast screening program.

In analyzing the results, researchers found … 

  • AI flagged 24% of the interval cancers and correctly localized 77%.
  • AI localized a higher proportion of node-positive than node-negative cancers (24% vs. 16%).
  • Invasive tumors had higher median risk scores than noninvasive (62 vs. 33), with median scores of 26 for normal mammograms.

Researchers also tested AI at a lower specificity threshold of 90%. 

  • AI detected more interval cancers at this level, but in real-world practice this would bump up recall rates.  

It’s also worth noting that Insight MMG is designed for the analysis of 2D digital mammography, which is more common in Europe than DBT. 

  • For the U.S., Lunit is emphasizing its recently cleared Insight DBT algorithm, which may perform differently.  

The Takeaway

As with the MASAI and BreastScreen Norway results, the new study points to an exciting role for AI in making mammography screening more accurate with less drain on radiologist resources. But as with those studies, the new results must be interpreted against Europe’s double-reading paradigm, which differs from the single-reading protocol used in the U.S. 

FDA Keeps Pace on AI Approvals

The FDA has updated its list of AI- and machine learning-enabled medical devices that have received regulatory authorization. The list is a closely watched barometer of the health of the AI sector, and the update shows the FDA is keeping a brisk pace of authorizations.

The FDA has maintained double-digit growth of AI authorizations for the last several years, a pace that reflects the growing number of submissions it’s getting from AI developers. 

  • Indeed, data compiled by regulatory expert Bradley Merrill Thompson show how the number of FDA authorizations has been growing rapidly since the dawn of the medical AI era in around 2016 (see also our article on AI safety below). 

The new FDA numbers show that …

  • The FDA has now authorized 950 AI/ML-enabled devices since it began keeping track
  • Device authorizations are up 15% for the first half of 2024 compared to the same period the year before (107 vs. 93)
  • The pace could grow even faster in late 2024 – in 2023, FDA in the second half authorized 126 devices, up 35% over the first half
  • At that pace, the FDA should hit just over 250 total authorizations in 2024 
  • This would represent 14% growth over 220 authorizations in 2023, and compares to growth of 14% in 2022 and 15% in 2021
  • As with past updates, radiology makes up the lion’s share of AI/ML authorizations, but had a 73% share in the first half, down from 80% for all of 2023
  • Siemens Healthineers led in all H1 2024 clearances with 11, bringing its total to 70 (66 for Siemens and four for Varian). GE HealthCare remains the leader with 80 total clearances after adding three in H1 2024 (GE’s total includes companies it has acquired, like Caption Health and MIM Software). There’s a big drop off after GE and Siemens, including Canon Medical (30), Aidoc (24), and Philips (24).

The FDA’s list includes both software-only algorithms as well as hardware devices like scanners that have built-in AI capabilities, such as a mobile X-ray unit that can alert users to emergent conditions. 

  • Indeed, many of the authorizations on the FDA’s list are for updated versions of already-cleared products rather than brand-new solutions – a trend that tends to inflate radiology’s share of approvals.

The Takeaway

The new FDA numbers on AI/ML regulatory authorizations are significant not only for revealing the growth in approvals, but also because the agency appears to be releasing the updates more frequently – perhaps a sign it is practicing what it preaches when it comes to AI openness and transparency. 

Better Prostate MRI with AI

A homegrown AI algorithm was able to detect clinically significant prostate cancer on MRI scans with the same accuracy as experienced radiologists. In a new study in Radiology, researchers say the algorithm could improve radiologists’ ability to detect prostate cancer on MRI, with fewer false positives.

In past issues of The Imaging Wire, we’ve discussed the need to improve on existing tools like PSA tests to make prostate cancer screening more precise with fewer false positives and less need for patient work-up.

  • Adding MRI to prostate screening protocols is a step forward, but MRI is an expensive technology that requires experienced radiologists to interpret.

Could AI help? In the new study, researchers tested a deep learning algorithm developed at the Mayo Clinic to detect clinically significant prostate cancer on multiparametric (mpMRI) scans.

  • In an interesting wrinkle, the Mayo algorithm does not indicate tumor location, so a second algorithm – called Grad-CAM – was employed to localize tumors.

The Mayo algorithm was trained on a population of 5k patients with a cancer prevalence similar to a screening population, then tested in an external test set of 204 patients, finding …

  • No statistically significant difference in performance between the Mayo algorithm and radiologists based on AUC (0.86 vs. 0.84, p=0.68)
  • The highest AUC was with the combination of AI and radiologists (0.89, p<0.001)
  • The Grad-CAM algorithm was accurate in localizing 56 of 58 true-positive exams

An editorial noted that the study employed the Mayo algorithm on multiparametric MRI exams.

  • Prostate cancer imaging is moving from mpMRI toward biparametric MRI (bpMRI) due to its faster scan times and lack of contrast, and if validated on bpMRI, AI’s impact could be even more dramatic.

The Takeaway
The current study illustrates the exciting developments underway to make prostate imaging more accurate and easier to perform. They also support the technology evolution that could one day make prostate cancer screening a more widely accepted test.

US + Mammo vs. Mammo + AI for Dense Breasts

Artificial intelligence may represent radiology’s future, but for at least one clinical application traditional imaging seems to be the present. In a new study in Radiology, ultrasound was more effective than AI for supplemental imaging of women with dense breast tissue. 

Dense breast tissue has long presented problems for breast imaging specialists. 

  • Women with dense breasts are at higher risk of breast cancer, but traditional screening modalities like X-ray mammography don’t work very well (sensitivity of 30-48%), creating the need for supplemental imaging tools like ultrasound and MRI.

In the new study, researchers from South Korea tested the use of Lunit’s Insight MMG mammography AI algorithm in 5.7k women without symptoms who had breast tissue classified as heterogeneously (63%) or extremely dense (37%). 

  • AI’s performance was compared to both mammography alone as well as to mammography with ultrasound, one of the gold-standard modalities for imaging women with dense breasts. 

All in all, researchers found …

  • Mammography with AI had lower sensitivity than mammography with ultrasound but slightly better than mammography alone (61% vs. 97% vs. 58%)
  • Mammography with AI had a lower cancer detection rate per 1k women but higher than mammography alone (3.5 vs. 5.6 vs. 3.3)
  • Mammography with AI missed 12 cancers detected with mammography with ultrasound
  • Mammography with AI had the highest specificity (95% vs. 78% vs. 94%)
  • And the lowest abnormal interpretation rate (5% vs. 23% vs. 6%)

The results show that while AI can help radiologists interpret screening mammography for most women, at present it can’t compensate for mammography’s low sensitivity in women with dense breast tissue.

In an editorial, breast radiologists Gary Whitman, MD, and Stamatia Destounis, MD, observed that supplemental imaging of women with dense breasts is getting more attention as the FDA prepares to implement breast density notification rules in September. 

  • They recommended follow-up studies with other AI algorithms, more patients, and a longer follow-up period. 

The Takeaway

As with a recent study on AI and teleradiology, the current research is a good step toward real-world evaluation of AI for a specific use case. While AI in this instance didn’t improve mammography’s sensitivity in women with dense breast tissue, it could carve out a role reducing false positives for these women who get mammography and ultrasound.

Get every issue of The Imaging Wire, delivered right to your inbox.

You might also like..

Select All

You're signed up!

It's great to have you as a reader. Check your inbox for a welcome email.

-- The Imaging Wire team

You're all set!