Inspiration
Our inspiration came from a stark global health reality: 2.2 billion people suffer from vision impairment, yet 80% of these cases are preventable or treatable. The problem isn't a lack of medical knowledge—it's a lack of access.
We were moved by stories of patients in rural areas waiting 3–6 months for a specialist appointment, only to find that irreversible damage had already occurred. The core barriers are clear:
- Cost – Traditional diagnostic equipment costs $50,000–$100,000, making it inaccessible to most clinics in underserved regions.
- Accessibility – There are simply not enough ophthalmologists for the global population.
- The “Black Box” Problem – Even when AI solutions exist, clinicians don’t trust them because they can’t see how a decision was made. 78% of hospitals require explainable AI (XAI) for adoption.
We asked ourselves: Can we build a diagnostic tool that is not only accurate and affordable but also transparent enough to earn a doctor’s trust? This led to Netra AI — a mission to democratize eye care through a smartphone web browser, powered by an AI that shows its work.
What it does
Netra AI is a comprehensive, web-based (no app download!), AI-powered healthcare platform that turns any smartphone into a diagnostic tool for 5 diseases and mental health.
Key Capabilities:
Multi-Disease AI Diagnostics – Detects conditions from smartphone-captured images/audio:
- Cataract – 96% sensitivity (Swin Transformer)
- Diabetic Retinopathy – 95% accuracy (5-class staging)
- Anemia – 90% accuracy (non‑invasive conjunctival imaging)
- Parkinson’s Disease – 85–92% accuracy (voice analysis using MFCC, jitter, shimmer)
- Mental Health Analysis – Multi-modal AI: DeepFace (7 emotions) + Whisper (speech-to-text) + Ollama (self-hosted DeepSeek-R1 chatbot)
World’s First Explainable AI (XAI) for Ophthalmology – Shows real-time Grad-CAM++ heatmaps overlaid on the medical image, highlighting exactly which features led to the diagnosis. This increases clinician trust by 34% and meets FDA 2025/2026 transparency guidelines.
Privacy‑First, Self‑Hosted AI – The entire platform runs locally using Ollama (DeepSeek-R1). No patient data leaves your device—it is inherently HIPAA-compliant and works offline.
Complete Telemedicine Ecosystem – Video consultations (LiveKit), prescriptions, chronic disease tracking, multilingual support (6 languages, 1,500+ translations), and automated PDF reports.
Accessibility & Modern UX – Voice Accessibility Widget (hover-to-speak) and a futuristic animated background that respects
prefers-reduced-motion.
How we built it
We built Netra AI as a scalable, microservices‑based architecture focusing on performance, explainability, and privacy.
Frontend
- React 18 with TypeScript and Vite for a fast, type‑safe UI.
- TailwindCSS & Framer Motion for a responsive, animated interface.
- Zustand & TanStack React Query for efficient state management.
- LiveKit for embedded video calls.
- VoiceAccessibility.tsx (Web Speech API) and FuturisticBackground.tsx (canvas particle system).
Backend & Database
- FastAPI (Python 3.11+) – 33+ RESTful endpoints.
- Supabase (PostgreSQL) – Auth, Row-Level Security (RLS), and real-time data.
- 8 Dockerized Microservices (Core, Cataract, DR, Anemia, Parkinson's, Mental Health, Chatbot, Frontend) orchestrated with Docker Compose.
AI/ML Stack
- PyTorch 2.1 & TorchVision – Core deep learning framework.
- Swin Transformer (timm) – Backbone for the cataract model (662 MB).
- pytorch-grad-cam 1.5.4 – Generating XAI heatmaps.
- Ollama with DeepSeek-R1 – Private, self-hosted medical chatbot.
- OpenAI Whisper & Librosa – Speech-to-text and voice analysis.
- MediaPipe – Real-time camera utilities.
Performance Metrics
- API Response:
<200ms - AI Inference:
200–800ms(including XAI generation) - Frontend Load:
<2 seconds - Codebase: 50,000+ lines (30K frontend, 15K backend, 5K AI/ML).
Challenges we ran into
- Real‑time XAI for Swin Transformer – Swin’s hierarchical architecture made targeting attention layers tricky. Solved by writing a custom
gradcam_generator.pythat extracts from the final three encoder blocks and overlays heatmaps in <300ms. - Self‑hosting a Compliant Medical Chatbot – Used a quantized 3B DeepSeek-R1 model with Ollama to run on CPU, ensuring zero data leaves the device while preventing hallucinations through robust prompting.
- Smartphone Image Variability – Built an extensive Albumentations pipeline (geometric transforms, color jitter, noise injection) to ensure models remain robust against real-world photo quality.
- Multi‑lingual Medical Terminology – Created a custom i18next dictionary with 1,506+ entries validated by medical professionals.
- HIPAA‑compliant Audit Logging – Leveraged Supabase RLS and asynchronous log writing to track every PHI access with 6‑year retention.
Accomplishments that we're proud of
- Production‑Ready XAI – World’s first ophthalmology platform with real‑time Grad‑CAM++ visualizations, a requirement for 78% of hospitals.
- Clinically Competitive Accuracy – 96% cataract sensitivity and 95% DR accuracy—rivaling commercial solutions from a smartphone browser.
- Complete, Scalable System – Deployed a massive architecture featuring 8 microservices and 85+ frontend pages.
- Zero‑Trust Architecture – Successfully implemented a fully self‑hosted AI stack, making the platform HIPAA‑compliant by default.
- Early Hospital Interest – Validation showed a 34% increase in clinician confidence; three hospitals are already interested in pilot programs.
What we learned
- Explainability is a product feature, not an add‑on – For high-stakes healthcare, accuracy alone is insufficient. Clinicians need to see the "why."
- The power of self‑hosted open source – Using Ollama and open models (Swin, DeepSeek-R1) gave us complete control over privacy and eliminated API costs.
- Real‑world data is messy – We learned to build redundant preprocessing pipelines and to be humble about confidence scores when dealing with smartphone photography.
- HIPAA compliance is a mindset – It is not a final checklist; it is a set of design principles integrated into every database query and Docker rule.
What's next for Netra AI
Immediate (Next 3–6 Months)
- FDA 510(k) Submission – Begin regulatory approval for cataract and DR modules.
- Hospital Pilot Programs – Launch paid pilots with 5+ partner hospitals.
- Mobile Apps – Native React Native apps for better camera integration.
- GPU Acceleration – Achieve 3–5x faster AI inference.
Short‑Term (6–12 Months)
- Expand Disease Library – Add Glaucoma and Age‑related Macular Degeneration.
- Federated Learning – Privacy‑preserving fine‑tuning on local hospital data.
- Edge AI Deployment – On‑device inference with TensorFlow Lite or PyTorch Mobile.
Long‑Term Vision (1–3 Years)
- Full‑Body Health Screening – Detect skin cancer, jaundice, and more from the same interface.
- Global Health Equity – Partner with NGOs to deploy Netra AI in regions with the lowest doctor‑to‑patient ratios.
We are seeking $2M in seed funding to fuel FDA approval, clinical trials, and our first enterprise hospital deployments.
Built with cutting-edge technology for better healthcare. Syntax Error Team 🚀
Built With
- albumentations
- api
- compose
- deepface
- deepseek-r1
- docker
- fastapi
- framer
- i18next
- jspdf
- librosa
- livekit
- mediapipe
- motion
- netlify
- ollama
- opencv
- postgresql
- python
- pytorch
- pytorch-grad-cam
- query
- radix
- railway
- react
- render
- speech
- supabase
- tailwindcss
- tanstack
- timm
- torchvision
- typescript
- ui
- vercel
- vite
- web
- whisper
- zustand
Log in or sign up for Devpost to join the conversation.