💡 Inspiration

Access to timely medical advice remains a challenge for millions, particularly in remote and underserved areas. We wanted to build something that mimics a conversation with a doctor — simple, fast, and empathetic. The idea for MedAI came from observing how many people rely on the internet to self-diagnose, often without context or accuracy. We aimed to build a human-centric, voice-first medical AI that combines image and symptom-based diagnosis into one unified experience, lowering barriers to early medical intervention.

⚙️ What it does

MedAI Diagnosis is an AI-powered healthcare platform that allows users to:

  • Describe their symptoms using natural voice input
  • Upload medical images such as X-rays, CT scans, or rashes
  • Receive AI-generated preliminary health assessments
  • Hear spoken feedback from the AI for a more natural experience
  • Track their status and generate professional-grade medical reports

MedAI bridges the gap between people and preliminary medical guidance using multimodal AI, helping users take the right next steps in their healthcare journey.

🛠️ How we built it

  • Frontend: React 18 with TypeScript, Tailwind CSS, and Framer Motion
  • Voice Input: GROQ Whisper API for real-time speech-to-text
  • Image Analysis: Google Gemini Vision API for processing medical images
  • Voice Output: ElevenLabs API for natural text-to-speech responses
  • Backend: Supabase for database, storage, and authentication
  • Security: Row-level security, audit logging, and encrypted storage

🚧 Challenges we ran into

  • Ensuring accurate voice input capture in noisy environments
  • Integrating multiple APIs with real-time performance
  • Designing a natural and fluid voice conversation flow
  • Handling multiple file formats and sizes during image upload
  • Implementing secure data practices for handling medical images

🏆 Accomplishments that we're proud of

  • Built a fully functional and secure AI-powered diagnosis platform
  • Combined voice and image input for a rich multimodal experience
  • Delivered a responsive and accessible user interface
  • Supported end-to-end patient flow from diagnosis to report generation

📚 What we learned

  • How to design and implement voice-first user interfaces
  • Techniques for combining multimodal AI capabilities effectively
  • The importance of intuitive design in healthcare technology
  • Best practices for managing sensitive health data securely

🚀 What's next for Med-AI: AI-Powered Preliminary Health Diagnosis Tool

  • Train domain-specific AI models using medical datasets
  • Support for multilingual voice input and responses
  • Medical professional collaboration for improved accuracy
  • Mobile-first app for wider accessibility
  • Integration with telemedicine services for doctor follow-up

Built With

Share this project:

Updates