Proactive Learning Strategies Improve AI Model Safety and Effectiveness in Toronto Hospitals
In a groundbreaking study published today in JAMA Network Open, researchers from York University have identified key strategies to enhance the effectiveness and safety of AI models in hospital settings, particularly in the Greater Toronto Area (GTA). The study focuses on addressing data shifts, a common issue where the training data for AI models does not accurately represent the real-world data encountered during the model's operation. This discrepancy can lead to inaccurate predictions and, in some cases, patient harm. Core Characters and Context Led by senior author Dr. Elham Dolatabadi, an Assistant Professor at York University's School of Health Policy and Management, and first author Vallijah Subasri, an AI scientist at University Health Network, the research team built and evaluated an early-warning system designed to predict the risk of in-hospital patient mortality. This system was assessed across seven large hospitals in the GTA, leveraging GEMINI, Canada's largest hospital data-sharing network. Key Findings and Strategies The study utilized data from 143,049 patient encounters, encompassing various aspects such as lab results, transfusions, imaging reports, and administrative features. The researchers found significant data shifts between the training and real-world application phases, which included changes in patient demographics, hospital types, admission sources, and critical laboratory assays. These shifts can occur due to multiple factors, such as differing healthcare practices, staff changes, resource availability, and unexpected events like pandemics. To address these issues, the researchers proposed two primary strategies: transfer learning and continual learning. Transfer learning allows an AI model to retain knowledge from one domain and apply it to another related domain, while continual learning involves updating the model using a continuous stream of new data in response to detected drifts. The study showed that models specific to hospital types, employing transfer learning, performed better than those trained on data from all available hospitals. Practical Application and Outcomes The early-warning system was tested for its ability to predict patient mortality risk and enhance triaging efficiency. The researchers discovered that harmful data shifts occurred when models trained on community hospital patient visits were deployed in academic hospitals, but not vice versa. This highlights the importance of tailoring AI models to specific hospital environments to avoid potential harm. Additionally, the study incorporated a drift-triggered continual learning approach, which helped prevent harmful data shifts caused by the COVID-19 pandemic. This method not only improved model performance over time but also ensured that the AI system remained adaptable and reliable despite evolving conditions. Biases and Fairness The research also addressed the issue of biases within AI models. These biases can arise if the training data disproportionately represents certain patient subgroups, leading to unfair or discriminatory outcomes. The team demonstrated how to detect these shifts and assess their negative impact on model performance. They proposed practical mitigation strategies to ensure that the AI models remain fair and equitable across all patient groups. Significance and Impact This study is pivotal in the journey towards safe and effective deployment of clinical AI models. It provides a comprehensive framework for monitoring data shifts and adapting AI systems in real-time, thereby bridging the gap between the theoretical promise of AI in healthcare and its practical implementation in clinical settings. Industry Insights and Company Profiles Industry experts have hailed the study for its practical approach to a critical problem in the field of clinical AI. Dr. Dolatabadi, a faculty affiliate at the Vector Institute, emphasizes the need for reliable and robust machine learning models in healthcare. According to her, the strategies outlined in the study offer a feasible path to achieving this goal. The Vector Institute, known for its leading research in AI, supports this endeavor by fostering collaboration between academia and industry. Dr. Vallijah Subasri, the study’s first author, underscores the importance of a proactive, label-agnostic monitoring pipeline. This pipeline, integrating transfer and continual learning, can effectively detect and mitigate harmful data shifts, ensuring that AI deployment in the GTA's hospitals is both robust and equitable. Overall, the research provides a solid foundation for the future development and deployment of AI in healthcare, emphasizing the need for ongoing adaptation and refinement of these models to match the dynamic nature of real-world clinical settings.
