💡 Inspiration
"Netra" means Eye in Sanskrit. I asked myself a single question: Can one AI architecture solve two completely different human problems?
- The Darkness: Millions of visually impaired people need "Digital Eyes" to navigate safely.
- The Burnout: Developers need a "Digital Brain" to understand complex legacy codebases instantly.
CodeNetra-AI is the answer. It is not just an app; it is a Dual-Mode Multimodal Bridge powered by the sheer raw power of Gemini 3.0.
🚀 What it does
CodeNetra is a Super-App with two distinct personalities:
👁️ Mode 1: Netra Vision (Accessibility Suite)
- The "Jarvis" Loop: A completely hands-free, continuous voice loop. You speak, it answers, it listens again. No buttons.
- Live Hazard Defense: Detects cars, pits, and obstacles in real-time and screams "Caution!" before you hit them.
- The "Aura": Ask "Who made you?" and experience a unique AI identity that defines its own existence.
💻 Mode 2: Developer God-Mode (Productivity Suite)
- Talk to your Repo (ZIP Intelligence): Upload an entire project
.zip. CodeNetra reads the whole architecture and explains how the login flow works. - UI to Code: Show it a screenshot, and it writes the clean Flutter/React code instantly.
- AI Interviewer: It takes your technical mock interview to prep you for FAANG.
⚙️ How we built it
We pushed Gemini 3.0 Flash to its limits using Flutter.
- The Brain: Used
gemini-3-flash-previewfor ultra-low latency. Speed was our #1 priority for blind safety. - The Ears: Built a custom Voice Loop Engine that manages state between Listening -> Thinking -> Speaking to create a seamless conversation flow (1.5s latency).
- The Context: Leveraged Gemini's Long Context Window to ingest massive ZIP files without crashing.
🚧 Challenges we ran into
- The "Infinite Loop" Bug: Initially, the AI would listen to its own voice and reply to itself forever. I fixed this with a smart
SpeechStoplogic. - Token Limits: Uploading large codebases hit the limit fast. I wrote a filter to strip out assets/build folders before sending data to Gemini.
- Git Security: My API keys got exposed in history. Learned the hard way how to use
.gitignoreandgit rm --cachedto secure the project.
🏅 Accomplishments that I'm proud of
- The "Killer Aura": Successfully giving the AI a personality. It feels alive, not robotic.
- Latency Optimization: Bringing the Vision-to-Voice response time down to under 2 seconds.
- Solo Development: Building a Dual-Mode Enterprise-level app alone.
🧠 What I learned
Multimodality is the future. Text is not enough. The ability of Gemini to "See" (Vision) and "Hear" (Voice) simultaneously is what makes inclusive apps possible.
🔮 What's next for CodeNetra-AI
- Smart Glasses: Moving "Netra Mode" from phone to wearable glasses.
- Offline Mode: Using Gemini Nano for on-device object detection without internet.


Log in or sign up for Devpost to join the conversation.