Exposure to AI models may reshape clinician's decisions and improve care outcomes

What if just observing Artificial Intelligence (AI) could make a clinician more efficient at their job and improve care outcomes for patients?

After repeated use of an AI model, a nurse practitioner at The Hospital for Sick Children (SickKids) was able to analyze ultrasounds more efficiently, ultimately reducing the need for invasive scans and illuminating a previously undescribed impact of AI in clinical settings.

Mandy Rickard, a nurse practitioner in the Urology Clinic at SickKids, performs and analyzes kidney ultrasounds almost daily as lead of the pre- and postnatal hydronephrosis clinic. In the clinic, she cares for up to 100 children a month with hydronephrosis, a condition caused by a urinary tract blockage that can result in kidney obstruction.

While analyzing the ultrasounds, Rickard looks at the degree of hydronephrosis and other characteristics that may suggest an obstruction that needs to be operated on. She also oversees an AI model developed to predict the severity of hydronephrosis cases and help providers identify appropriate care pathways. Over time, Rickard found that she was able to anticipate the AI model's behavior and intuit how it would categorize an ultrasound.

In new research published in the New England Journal of Medicine AI, the research team noted that Rickard's improved ability to interpret kidney ultrasounds was a direct result of her repeated exposure to the AI model. Over a three year period, Rickard reduced the number of children with hydronephrosis sent for invasive nuclear scans from 80 to 58 per cent.

When the model trains the user

The AI model, created by Lauren Erdman, a former PhD student at SickKids, was in a silent trial phase when Rickard began to predict the model's outcomes. During a silent trial, researchers use prospective patients and data to observe the function of a model without informing care decisions. Even in this early stage, and without any other changes to her practice, Rickard was already predicting the model's outcomes – but why?

The research team classified Rickard's developed intuition as a new type of bias: induced belief revision. Data-induced belief revision occurs when clinical end-users, like Rickard, develop an unconscious intuition based on collection and labeling of data. Model-induced belief revision arises from repeated exposures to an AI model's inputs and outputs, resulting in clinicians anticipating model predictions based on unseen data.

With induced belief revision, clinical behavioral change can happen even when AI is not in use. While this type of bias proved beneficial to both care providers like Rickard and patients with hydronephrosis, the team is cognizant that the bias may also unintentionally influence clinical practice in less positive ways if not properly observed and documented.

"This model is meant to augment, not to replace us as experts," Rickard says.

To help address this bias, Rickard's team suggests minimizing a clinical user's chance of being influenced by a model, especially when one person provides the bulk of critical care. Doing so could help maintain the important distinction between true clinical judgment and a model's predictions.

AI research at SickKids

AI has the power to improve patient experience and reduce provider fatigue when used responsibly. At SickKids, using AI in an ethical manner is possible due largely in part to a framework developed by bioethicist Dr. Melissa McCradden.

The framework established the silent trial to clinical trial pipeline and led to the development of an AI Regulatory Advisement Board that helps research move from ideation to clinical trial in a smooth process.

"At SickKids, we're actively shaping guidelines for the responsible development and safe implementation of AI in health care," says Jethro Kwong, a Urology resident at SickKids and first author on the paper.

With the success of the model, the team is moving to clinical trial and implementing the model in a patient-first way by prioritizing informed consent.

Our ultimate goal is to put this tool into communities where there may not be a urology expert, to preserve hospital resources and potentially save families from taking repeated, sometimes lengthy trips to the hospital. By embedding our model into care, we can hopefully reduce the number of low-risk hydronephrosis cases that escalate to a renal scan, ultimately streamlining testing and surgery for those who it would benefit."

Mandy Rickard, nurse practitioner in the Urology Clinic at SickKids

Source:
Journal reference:

Kwong, J. C. C., et al. (2024) When the Model Trains You: Induced Belief Revision and Its Implications on Artificial Intelligence Research and Patient Care — A Case Study on Predicting Obstructive Hydronephrosis in Children. NEJM -AI. doi.org/10.1056/AIcs2300004.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Research pinpoints critical age for improving children's cardiovascular health