Inspiration

1 in 3 stroke survivors develops aphasia, losing reliable speech despite full cognitive ability. Existing AAC tools are either expensive ($15,000+ hardware) or clinically inaccessible. We built Clarivo to close that gap with a free, intelligent system that restores communication—not just through speech synthesis, but through personalized, adaptive expression that evolves with the user.

What it does

Clarivo is an AI-powered AAC platform that helps non-speaking or speech-impaired users communicate through icons and structured intent building.

Icon-Based Communication: Decision tree + free-form drawing “Icon Composer” for flexible expression Voice Cloning: ElevenLabs generates speech in the patient’s own voice, preserving it if they lose it fully later on AI Sentence Engine: Converts icon sequences into natural, context-aware speech Caregiver Intelligence Loop: Caregivers provide small feedback signals that continuously improve personalization Analytics Dashboard: Tracks usage patterns, urgency signals, and daily AI-generated summaries

How we built it

Frontend: Next.js 14, TypeScript, Tailwind CSS, Phosphor Icons, Recharts AI Backend: FastAPI with GPT-4o-mini for real-time intent generation and GPT-4o for deeper insights Voice Layer: ElevenLabs for voice cloning and real-time synthesis Data Layer: MongoDB Atlas with structured path-based session tracking and caching Architecture: Streaming via SSE with a shared type-safe API contract across services

Challenges we ran into

Achieving sub-400ms latency for real-time AI responses Managing limited LLM context while preserving medical and behavioral relevance Handling complex state synchronization between patient and caregiver interfaces Ensuring stable and consistent voice cloning across sessions

Accomplishments that we're proud of

Real-time speech generation in a cloned patient voice A working “Knowledge Score” that reflects system understanding over time Adaptive shortcuts driven by real usage frequency A fully functional caregiver feedback loop that improves AI behavior Building a system that meaningfully replaces expensive AAC hardware

What we learned

Breaking AI into separate roles (intent, clarification, summarization) improves reliability Accessibility-first design requires extreme simplicity in interaction patterns Streaming LLM outputs dramatically improves perceived responsiveness Hierarchical “path-based” data structures work better than traditional relational models for AAC flows

What's next for Clarivo

Offline-first mobile app with edge AI support Smart home integration for environmental control Voice-based messaging to external platforms (WhatsApp, iMessage) Clinical pilot studies with speech therapy professionals

Built With

  • elevenlabs
  • fastapi
  • mongodb-atlas
  • motor
  • next.js-14
  • openai-gpt-4o
  • openai-gpt-4o-mini
  • phosphor-icons
  • python
  • recharts
  • sse-(server-sent-events)
  • tailwind-css
  • typescript
Share this project:

Updates