Inspiration

Healthcare today is burdened by clinician burnout and slow triage workflows — doctors often spend up to half their time on documentation, and thousands of patient calls are manually screened every day. Tele-Triage was inspired by the need to automate clinical intake, early risk detection, and documentation so clinicians can focus on care, not paperwork. By leveraging cutting-edge multimodal medical AI (like Google’s MedGemma) and acoustic biomarkers from audio, we envisioned a system that can hear, interpret, reason, and act autonomously in a clinical context.

What it does

  1. The Agentic AI Tele-Triage System is an autonomous AI agent that:

  2. Listens to patient speech + cough/respiratory sounds.

  3. Converts spoken complaints into clinical text (transcription + analysis).

  4. Uses MedGemma for medical reasoning and structured note drafting.

  5. Synthesizes standard clinical SOAP notes (Subjective, Objective, Assessment, Plan).

  6. Prioritizes high-risk cases in a clinician’s queue using acoustic and clinical cues.

  7. Operates end-to-end — from patient input to triage outcome and documentation delivery.

How we built it

  1. Reasoning Engine: MedGemma 4B via Ollama for medical semantic understanding.

  2. Speech to Text (ASR): Whisper (Python Transformers) for robust transcription of medical speech.

  3. Acoustic Analysis: Custom audio heuristics (e.g., cough/respiratory distress signals).

  4. Frontend: Next.js 14 + Tailwind CSS for interactive Patient & Clinician portals.

  5. Backend: Python FastAPI serving AI agents, logic, and triage APIs.

  6. Deployment: Local & privacy-first stack running entirely on user control

Challenges we ran into

  • Multimodal integration: Combining audio biomarkers with semantic AI reasoning was non-trivial — aligning acoustic features meaningfully with medical context required custom signal processing and heuristic design.

  • Domain accuracy vs privacy: Running MedGemma locally meant balancing medical reasoning power with resource constraints while ensuring patient data never left the local environment.

  • Clinical fidelity: Prompt engineering to elicit medically credible outputs from MedGemma required significant iteration and safety guards to avoid hallucination.

Accomplishments that we're proud of

  1. Built a true agentic pipeline — system perceives input, reasons clinically, and acts to generate structured medical documentation.

  2. Multimodal triage: First prototype combining speech + acoustic signals + LLM reasoning for clinical prioritization.

  3. Functional end-to-end interface for both patient input and clinician output — from raw speech to SOAP note with prioritized queue.

What we learned

  1. Integrating acoustic biomarkers with LLM reasoning opens rich avenues for non-text clinical signals in AI triage.

  2. Running advanced models like MedGemma locally (via Ollama) ensures data privacy and control, which is critical in healthcare applications.

  3. Prompt design and domain alignment are just as important as model choice — clinical fidelity depends on how we structure and guard the AI’s reasoning process.

What's next for AI Tele-Triage System

Clinical validation: Pilot deployments with real clinician feedback loops to improve safety and trust.

Built With

Share this project:

Updates