Inspiration

  • Healthcare algorithms tend to influence racial disparity, Oberymeyer et al. (2019) kidney disease study found that widely-used AI systematically denied Black patients specialist care. I wanted to build a system that detects bias and translates it into practical language for clinicians to understand and automatically generate fixes.

What it does

  • Our system analyzes Medicare data (116k patients) to detect algorithmic bias in high-cost prediction.
  • Four-tier Gemini integration:

    • 1. Validates causal graphs against medical literature.

    -2. Translates "2.6% FNR disparity" into "300 Black patients denied kidney care annually ".

    -3. Assess safety of fairness interventions.

    -4. Auto-generates validated python code to fix bias.

  • Interactive dashboard shows real time impact.

How we built it

  • Data: CMS Medicare 2008-2010 (116K patients, 9 clinical features)

  • Causal Analysis: PC algorithm + expert knowledge (Obermeyer kidney pathway)

  • Fairness Detection: Fairlearn & AIF360 (4 intervention methods tested)

  • LLM Integration: Gemini 2.0 Flash with temperature=0 for reproducibility

  • Validation: Bootstrap confidence intervals (1000 iterations), syntax/security checks on generated code

Challenges we ran into

  • We ran into: LLM Hallucination risks, which we solved y using medical literature citations.

  • Code generation safety: mitigated risk by using the 4 tier validation system.

Accomplishments that we're proud of

  • We achieved a 94% bias reduction (2.6% to 2.2% FNR disparity) while maintaining 84.7% accuracy

  • All methods achieve safe clinical status with the FNR disparity being under the 5% threshold.

  • Functional system with auto-generation and validation fairness intervention code.

  • Made with real government public use files for precision and accuracy.

    What we learned

  • Impossibility theorems matter in all contexts. I can optimize demographic parity and equalized odds simultaneously (Kleinberg 2018), I must be able to prioritize patient safety (FNR disparity) over statistical elegance.

    • Gemini excels in bridging statistical metrics and clinical narratives compare to standard templates.

What's next for Gemini integration within Clinical fairness research

  • Implementation of multi-modal analysis by integrating clinical notes and imaging data using Gemini's vision capabilities for radiology bias detection.

  • Extend beyond race to other variables such as gender and age interactions using Gemini's reasoning.

Built With

Share this project:

Updates