Documentation

Methodology

Understanding how PrismDX detects and surfaces AI diagnostic bias

What is PrismDX?

PrismDX is a point-of-care demographic bias overlay designed to fill a critical gap in current MLOps tools. While most AI monitoring focuses on aggregate model performance, PrismDX surfaces potential bias flags directly to clinicians at the moment of diagnosis—when it matters most.

Research has consistently shown that AI diagnostic models can exhibit reduced accuracy for underrepresented demographic groups, particularly in dermatology and radiology. PrismDX provides real-time visibility into these disparities, empowering clinicians to make informed decisions about when additional verification is warranted.

Core Capabilities

Point-of-Care Bias Overlay

PrismDX acts as a transparent layer over existing AI diagnostic tools, flagging potential demographic biases in real-time without disrupting clinical workflow.

Research-Backed Baselines

Our bias detection algorithms are calibrated against peer-reviewed research from Nature Medicine, JAMA, and other leading medical journals documenting AI performance disparities.

Proactive Alerting

When confidence levels fall below demographic-specific baselines, clinicians receive immediate alerts with actionable recommendations for additional verification.

Audit Trail

Every scan generates a complete audit trail documenting the bias evaluation, enabling compliance with emerging AI healthcare regulations.

How Bias Detection Works
A step-by-step overview of the PrismDX pipeline
1

Image Analysis

The medical image is processed through the primary diagnostic AI model, generating a diagnosis and raw confidence score.

2

Demographic Contextualization

Patient demographic data (Fitzpatrick scale, age, gender) is cross-referenced against our baseline database derived from published research.

3

Bias Evaluation

If the confidence score deviates significantly from expected baselines for the patient's demographic group, a bias flag is triggered.

4

Clinical Decision Support

The clinician receives the diagnosis alongside bias evaluation data, enabling informed decisions about additional verification or specialist consultation.

Fitzpatrick Scale Baselines
Expected vs. observed AI confidence by skin phototype (dermatology applications)
Fitzpatrick TypeExpected BaselineObserved Avg.Status
Type I95%94%Normal
Type II94%92%Normal
Type III92%89%Normal
Type IV90%85%Monitor
Type V90%78%Bias Alert
Type VI90%72%Bias Alert

Data derived from: Adamson AS, Smith A. Machine Learning and Health Care Disparities in Dermatology. JAMA Dermatology. 2018; Daneshjou R et al. Disparities in dermatology AI performance. Nature Medicine. 2024.

Research References
Peer-reviewed studies informing our bias baselines

Disparities in dermatology AI performance across skin tones

Daneshjou R, et al. • Nature Medicine, 2024

Machine Learning and Health Care Disparities in Dermatology

Adamson AS, Smith A. • JAMA Dermatology, 2018

Underdiagnosis bias of artificial intelligence algorithms

Seyyed-Kalantari L, et al. • Nature Medicine, 2021

Continuous Updates

Our baseline database is continuously updated as new research is published. PrismDX monitors leading journals for relevant studies on AI diagnostic disparities.