ObservAI
Open-source AI observability platform designed to make large language models (LLMs) transparent, reliable, and actionable in real-world systems.
🧠 Inspiration
While building production AI applications with Gemini / Vertex AI, I realized something scary:
LLMs fail silently.
Prompts silently consume thousands of extra tokens, costs spike without warning, hallucinations creep in, and semantic drift happens — yet most teams ship without any visibility. Traditional observability tools stop at infra metrics like latency and error codes. LLMs are probabilistic systems, and they need a new layer of observability that understands AI behavior, not just server metrics.
That gap is what inspired ObservAI.
This challenge became even more relevant in the context of global mobility and digital inclusion — where reliable, efficient AI can help build tools for translation, documentation, collaboration, and intelligent workflows that make it easier for people to work and interact across borders.
🚀 What ObservAI Does
ObservAI is an enterprise-grade LLM observability platform for production AI systems.
It provides comprehensive tracking across your AI infrastructure:
- 📊 Token usage & latency metrics
- 💰 Real-time cost attribution per request
- 🎯 Quality metrics (coherence, toxicity, hallucination risk)
- 🔍 40 AI/ML detection rules for anomalies
- 🤖 Lyra — AI-powered prompt optimizer
All data is processed through a secure backend for real-time analytics, alerts, and actionable insights to help teams build AI services that are trustworthy and scalable.
🏗️ How I Built It
Architecture overview:
$$ \text{User App} \rightarrow \text{ObservAI SDK} \rightarrow \text{Supabase Edge Function} \rightarrow \text{PostgreSQL} \rightarrow \text{Dashboard} $$
This architecture ensures zero developer friction while enabling deep LLM observability.
1. SDK Layer
- Lightweight TypeScript/JavaScript SDK (npm:
@observai/sdk) - Wraps AI calls and intercepts all LLM interactions
- Collects telemetry (tokens, latency, cost, quality scores)
- Added batching + async sending for minimal overhead
- Cross-platform support: frontend, backend, AWS Lambda, Docker
2. Quality Analysis Engine
- Lightweight heuristic + NLP scoring
- Semantic awareness for coherence and hallucination risk
- Normalized scores between (0 \le x \le 1)
- Toxicity and prompt injection detection
3. Ingestion Backend
- Supabase Edge Functions for serverless ingestion
- PostgreSQL with strong row-level security
- Batched writes with validation
- Real-time anomaly detection hooks
4. Real-Time Dashboard
- Modern React/Vite/Tailwind web interface
- Live metrics visualization per request
- Token usage, latency percentiles, cost attribution
- Alerts and root-cause insights
5. AI-Specific Detection Rules & Lyra
- Suite of 40 AI/ML detection rules that catch:
- Hallucination risk
- Prompt injection attempts
- Cost spikes and token inefficiencies
- Latency anomalies
- Semantic drift over time
- Toxic output patterns
- Lyra — data-driven prompt optimizer using live metrics to suggest improved prompts
⚠️ Challenges Faced
Defining meaningful AI quality metrics
Traditional monitoring tools lack semantic awareness. I had to design new metrics (coherence, hallucination risk, toxicity) that correlate with real AI quality, not just infrastructure health.Balancing performance and overhead
Collecting telemetry without slowing down AI inference required efficient batching and optional async sending — making observability invisible to developers.Security & Privacy
Ensuring telemetry doesn't leak sensitive user data while still providing useful insights required careful design of row-level security and data handling.SDK Usability
Making the SDK drop-in and cross-platform required careful API design and strong TypeScript typing to ensure developers could adopt it in minutes.
Each challenge pushed me to think about AI observability as fundamentally different from traditional observability — it's behavioral first, not infrastructure first.
📚 What I Learned
- AI behavior must be monitored at a semantic level, not just as requests and responses
- LLMs should be treated as production systems, not just APIs
- Developers need not just alerts, but actionable insights — a key reason why Lyra is part of this system
- Open-source tools must be simple to adopt: this is why ObservAI's SDK installation can be done in minutes
- Small SDK design decisions massively affect developer adoption
I also dove deeply into AI evaluation techniques, client-side instrumentation patterns, and secure serverless backend design.
Most importantly, I learned how to design AI-native infrastructure.
🧩 Built With
Languages
- TypeScript
- JavaScript
Frontend
- React
- Vite
- Tailwind CSS
LLM Platform
- Google Gemini on Vertex AI
- Compatible with multiple LLM providers
Backend / Infrastructure
- Supabase Edge Functions
- PostgreSQL
- Serverless architecture
SDK & Tooling
- ObservAI SDK (
@observai/sdk) - Node.js
- npm package distribution
🌍 Impact & Global Relevance
ObservAI is designed to help the next generation of AI-powered tools that are:
- More cost-efficient through prompt optimization and usage insights
- More reliable in production with real-time anomaly detection
- Safer by detecting toxicity and prompt injections
- Better at providing meaningful responses through quality monitoring
These improvements matter especially for tools that support global mobility, cross-cultural communication, and collaborative workflows. By providing developers with real insights into how their AI behaves in the real world, ObservAI enables creation of AI services that are trustworthy and scalable — critical for global accessibility and inclusion.
🧾 What's Next
- Automated AI anomaly mitigation with smart alerting
- Plugin ecosystem for custom observability rules
- Expanded LLM support (OpenAI, Anthropic, and more)
- Predictive AI failure warnings using historical patterns
- Live dashboards & real-time alerts for production teams
- Prompt regression detection to catch quality degradation
- Team & org-level analytics for enterprise use cases
ObservAI is built for teams shipping real AI to real users.
📦 Installation
npm install @observai/sdk
📄 License
Open Source(MIT)
Built with ❤️ for the VisaVerse AI Hackathon
Built With
- datadog
- gemini
- javascript
- observai/sdk
- postgresql
- tailwind/css
- typescript
- vite
Log in or sign up for Devpost to join the conversation.