About the Session

When a healthcare center develops its own medical AI models, it assumes long-term responsibility for ensuring that performance remains consistent over time. This session will explore best practices for pre-deployment validation, post-deployment monitoring, and strategies to bridge the gap between theoretical AI performance and real-world clinical outcomes. We will discuss how rigorous local validation can uncover potential diagnostic improvements before deployment and how ongoing performance tracking can help mitigate biases, optimize radiologist-AI collaboration, and ensure sustainable AI impact. Through case studies from a large radiology practice, we will share insights on key performance indicators (KPIs), monitoring methods, and lessons learned in scaling AI in clinical practice. 

 

Objectives: 

  • Describe key performance metrics (KPIs) for tracking AI model performance over time.
  • Compare pre-deployment Enhanced Detection Rate (EDR) predictions with real-world clinical outcomes.
  • Identify strategies for addressing performance deviations in AI-assisted radiology.
Session Number

2018

Format

Education Session

Learning Topic
Artificial Intelligence (AI)Machine Learning (ML)
Credit Type
ASRT-RTCAMPEP-MPCECSIIM IIP-CIIP

Presented By

 

Tim Kline, PhD

Assistant Professor of Radiology
Mayo Clinic

Nina Kottler, MD, MS, FSIIM

Associate CMO, Clinical AI
Radiology Partners

Jason A. Poff, MD

Director of Innovation Deployment
Radiology Partners

Vera Sorin, MD

Radiology Informatics Fellow
Radiology, Mayo Clinic, Rochester

Christoph Wald, MD, PhD, MBA

Chair, Professor of Radiology
Lahey Clinic

Walter F. Wiggins, MD, PhD

Director, Clinical AI
Radiology Partners