Inspiration
Emergency response is one of the few domains where technology failure can directly cost lives. While cloud-based AI has advanced rapidly, it assumes a constant, reliable internet connection—an assumption that breaks down in the real world.
- We were inspired by scenarios that are common but underrepresented in AI design:
- Paramedics operating in rural or mountainous regions
- Disaster response teams working after earthquakes, floods, or cyclones
- Emergency situations in subways, tunnels, or remote highways
In these environments, connectivity is unreliable or completely unavailable, yet decisions must be made within seconds. At the same time, medical data is among the most sensitive categories of personal information, making cloud transmission ethically and legally problematic.
This led us to a central question
- Why are we still “renting intelligence” from the cloud for situations that demand instant, private, and resilient decision-making?
- EDGE-MEDIC was inspired by the belief that AI should move closer to the human, not farther away into data centers.
What it Does
EDGE-MEDIC is an offline, on-device AI assistant designed to support emergency responders with instant medical guidance when cloud AI is unavailable or inappropriate.
The system enables responders to
- Speak patient symptoms using a push-to-talk voice interface
- Receive real-time, spoken medical guidance
- Perform basic triage reasoning based on symptoms
- Generate structured emergency notes, stored locally
- Operate entirely offline, with zero data transmission
All processing occurs on the device itself, ensuring
- Zero latency
- Complete data privacy
- Operational reliability in no-signal environments
EDGE-MEDIC does not attempt to replace medical professionals. Instead, it acts as a decision-support co-pilot, providing structured, context-aware assistance during high-stress moments.
How We Built It
- EDGE-MEDIC is architected around the constraints of mobile hardware, prioritizing efficiency, privacy, and feasibility.
- On-Device AI Pipeline
The complete data flow is executed locally
User Voice → Local STT → RunAnywhere Core → SLM Reasoning → Local TTS User Voice→Local STT→RunAnywhere Core→SLM Reasoning→Local TTS Core Components
Speech-to-Text
- Lightweight local Whisper (Tiny/Base) models are used for offline transcription.
Reasoning Engine
- A quantized DeepSeek-R1-Distill (4–6B) Small Language Model is used for medical reasoning. Quantization (INT4) reduces memory usage by approximately: -Memory Reduction≈75%
RunAnywhere SDK
- Acts as the orchestration layer, handling:
- Hardware-aware inference (CPU / NPU)
- Secure model execution
- Memory-efficient loading
- On-device lifecycle management
Text-to-Speech
- A lightweight local TTS engine provides immediate spoken responses.
Design Philosophy
- Inference-only (no on-device training)
- No external APIs
- No network calls
- Optimized for 6–8GB RAM devices
- This ensures EDGE-MEDIC is feasible today, not speculative.
Challenges We Ran Into
Balancing Model Capability vs. Device Constraints
Medical reasoning benefits from strong reasoning models, but mobile devices have strict memory and power limits. Choosing an SLM that was capable yet deployable required careful evaluation and quantization planning.
Avoiding Over-Promise
In healthcare-related projects, it is easy to overstate AI capabilities. We deliberately constrained EDGE-MEDIC to decision support, not diagnosis, ensuring ethical and realistic positioning.
Designing for Offline-First UX
Most AI experiences assume retries, loading states, or cloud fallbacks. Designing an interface that must always work instantly required a fundamentally different mindset.
Ensuring Privacy by Design
We treated privacy as an architectural requirement, not a feature. This meant eliminating analytics, logging, and telemetry that are common in modern apps but unacceptable in this context.
Accomplishments That We’re Proud Of
- Designing a fully offline AI system that addresses a real-world, high-impact problem
- Demonstrating how Small Language Models can outperform cloud systems in critical scenarios
- Creating a clear, feasible architecture aligned with modern mobile hardware
- Building a concept that showcases the true strengths of edge AI, not just cost savings
- Most importantly, we proved that privacy, speed, and intelligence do not require the cloud.
What We Learned
- Latency is not a luxury problem—in some domains, it is a safety issue
- On-device AI changes product design, not just infrastructure
- Smaller, well-optimized models can be more valuable than massive cloud-based systems
- Privacy-first design often leads to better engineering decisions overall
- This project deepened our understanding of edge computing, model optimization, and responsible AI design.
What’s Next for EDGE-MEDIC: Offline Emergency Medical Intelligence
- Looking ahead, EDGE-MEDIC can evolve into a broader offline emergency response platform:
- Multilingual offline support for global responders
- Integration with local medical protocols and guidelines
- Optional secure device-to-device communication (mesh networks)
- Expanded support for disaster relief and humanitarian missions
- Our long-term vision is to help redefine how AI is deployed in critical environments—from centralized clouds to resilient devices at the edge.
Final Note
EDGE-MEDIC is not just an idea—it is a statement
- The future of AI is not always online. Sometimes, the most powerful intelligence is the one that works when everything else fails.
Log in or sign up for Devpost to join the conversation.