Clinician AI

Feedback System

Partner with UPMC clinicians, we aim to develop a clinician feedback interface for evaluating AI-generated Prograf dosage recommendations.



Our goal is to:


Develop a clinician feedback interface that displays essential information for evaluating AI-generated Prograf dosage recommendations, enhancing model reliability and facilitate clinical adoption.

Lead Designer:

ux design

User Testing

Research


Team


Saera Kim

Ananya Chaudhary

Guanjie Cheng

Xinfei Cen

Tools


Figma, Adobe CC, Miro

Duration


Spring 2025

4 months

Project Introduction

Clinicians + AI: Closing the Gap in Clinical Dosing


Artificial intelligence holds promise for improving patient care—but many models stall before making it into real clinical settings. Clinician concerns about reliability, interpretability, and workflow fit often block adoption.


Our project centers on empowering clinicians to co-refine AI-generated dosing suggestions. We partnered with UPMC’s kidney transplant team to design a feedback interface for Prograf (an immunosuppressant drug). Clinicians review AI-recommended doses and provide structured feedback, enabling the model to learn and improve in real-world use.

Why It’s Hard: The Implementation Gap

  • AI research often remains decoupled from real-world clinical demands.

  • Models trained on retrospective datasets fail to capture workflow constraints, emergent risk factors, or nuanced clinician intuition.

  • In high-stakes areas like transplant dosing, clinicians must consider many dynamic variables (infection risk, prior patient response, comorbidities) that AI may not account for.

  • Without transparency, control, or adaptability, many clinicians distrust AI recommendations.

What We Aimed to Build


Primary Goal
To build a feedback-driven clinician interface that lets doctors evaluate and fine-tune AI-recommended Prograf doses—thus improving model trust and readiness for clinical deployment.


Subsidiary Objectives

  • Understand clinicians’ existing workflows & pain points

  • Map the mental model of how they choose Prograf dosing

  • Identify which clinical data/context is essential for safe dose adjustment

  • Design feedback interactions that minimize cognitive load

  • Structure the interface to support asynchronous, thoughtful feedback


Stakeholders / Partners

  • Academic & Technical: Holly Wiberg, Woody Zhu, Satyam Verma (AI / modeling leads)

  • Clinical: Dr. Amit Tevar (Transplant Director), UPMC kidney transplant clinical team

  • This cross-disciplinary collaboration ensures we ground the tool in real clinical realities.

Research

Understanding Clinicians, Workflows, and AI Gaps

Mixed-method research process combining expert interviews, think-aloud sessions, and literature review.


Our focus: uncover how clinicians make dosing decisions, what information they rely on, and how AI recommendations fit, or conflict, with real workflows.


  • Primary Research: Interviews with 1 doctor, 2 clinicians, and 1 medical student to trace their real-time dosing logic.

  • Secondary Research: Six academic studies on clinician feedback systems, decision support, and usability in AI healthcare tools.

  • Synthesis: Collaborative affinity mapping to extract insights and define design opportunities.


What We Learned About Clinical Decision-Making


Our findings revealed that dosing decisions are highly dynamic and intuitive, shaped by both data and experience.


  1. Dosing is a feedback loop.
    Clinicians adjust doses based on patient trends over time (e.g., creatinine, urine output) —much like “tasting and adjusting a recipe.”

  2. Clinicians input orders, not explanations.
    Current systems lack space for reasoning, orders are entered quickly via structured fields. Any new system must balance structured familiarity with flexibility for rationale.

  3. Dosage decisions are subjective.
    Experience and patient variability mean each clinician may decide differently. This highlighted the need for a tool that supports transparency and nuance, not uniformity.

Prototyping

Entering the prototyping phase, we focused on a key question:

“How might clinicians best document their rationale when accepting, rejecting, or modifying AI-suggested Prograf dosages?”

Mapping the Clinician Journey


We created storyboards to visualize how clinicians might interact with the AI feedback system during their workflow — from reviewing AI recommendations to providing feedback and updating patient records.


This helped us identify the key interaction moments where context, clarity, and reasoning support were critical.

Defining the Feedback Spectrum


Through storyboarding and early prototyping, we identified a spectrum of feedback methods—from open, flexible input to tightly structured formats.
Each option reflected a trade-off between expressiveness and machine readability.


We mapped six approaches along this continuum to test how clinicians naturally communicate their reasoning while ensuring the AI model could interpret it effectively.

Lo-Fi Prototype Testing


We conducted think-aloud testing with nursing practitioners from inpatient and outpatient settings.
Clinicians walked through mock patient cases, reviewed AI-generated Prograf doses, and verbalized how they’d accept, adjust, or reject recommendations.


What We Explored:

  • How feedback structure impacts cognitive flow

  • Whether context and reasoning are preserved

  • How decision patterns differ between care settings


Key Takeaways:

  • Inpatient clinicians favor rapid, structured feedback.

  • Outpatient clinicians prefer flexibility to explain nuanced decisions.

  • Feedback must adapt to both pace and context of each environment.

Lo-fi Prototype Insights + Refinements

Mid-fi Prototype Insights + Refinements

We shifted our goal to a more pressing need after refining our feedback interaction model:

"How can we display patient data to enable clinicians to make informed adjustments to AI-suggested Prograf dosages with ease?"


The shift in focus was driven by usability feedback indicating that timely access to patient data is essential for safe and effective dosage decisions.

We created four mid-fidelity interfaces — table and graph versions for both inpatient and outpatient settings — to test how clinicians interpret information differently.


  • Table View: prioritizes precision and quick scanning

  • Graph View: emphasizes visual patterns and longitudinal trends

  • Inpatient Mode: highlights granular, real-time metrics

  • Outpatient Mode: emphasizes longer-term adherence and history


Each variation explored how data representation affects confidence, speed, and reasoning.

Finale

Final prototype- A cleaner, more integrated experience that minimizes cognitive load and enables fast, high-quality feedback.

Success Metrics


We evaluated usability by measuring time-to-decision for dose evaluation versus the current EHR workflow.


Result-


~50% Faster Decision Time
Clinicians reached decisions in 0:47s (vs. 1:33s) while maintaining clinical accuracy using our software. This indicates improved efficiency and lower mental effort when reviewing cases.


Although decisions were made 50% faster, speed is not the primary goal for clinicians—accuracy remains the priority. The design emphasizes clarity and efficiency, reducing cognitive load and encouraging sustained engagement. Faster, lower-effort decisions help minimize fatigue and build long-term trust in AI-supported workflows.


The platform allows clinicians to provide structured feedback on AI-generated dosing recommendations, improving model accuracy and confidence in AI-assisted care. This feedback loop enables continuous model refinement and supports the broader adoption of AI in clinical practice.

CLINICIAN AI FEEDBACK SYSTEM [HCI RESEARCH]