π Language Lens β Breaking Language Barriers with AI & AR
π About the Project
Language Lens is an innovative AI-powered mobile app that uses real-time image recognition and translation to help users instantly understand the world around them. By simply pointing their phone camera at an object, users can identify it and translate its name into any language, making learning more effortless, immersive, and inclusive.
Inspired by the challenges of language barriers in everyday life, we set out to create a tool that empowers usersβwhether travelers, students, or multilingual familiesβto engage with the world in a seamless and intuitive way.
π How We Built It
We developed Language Lens using modern Android technologies, leveraging the power of AI and real-time processing to deliver an intuitive and high-performance experience.
Tech Stack
- Jetpack Compose β For a modern, declarative UI
- Kotlin β The backbone of our Android development
- CameraX β For real-time image capture and processing
- Retrofit β To communicate with the Google Gemini AI API
- Google Gemini AI β AI-powered object detection and text translation
- Coroutines β For efficient asynchronous API calls
- Base64 Encoding β To convert captured images into a format that the AI can process
- Logging (Log.e & Timber) β For debugging and performance tracking
β‘ Challenges We Faced
π Navigating Time Zones & Team Coordination
As an international team of two project managers, one designer, and one developer, we faced challenges coordinating across three different time zones. Quick decision-making and async communication became essential for us to stay on track.
π©βπ» Adapting to Team Changes
Initially, our team had an additional developer, but we had to quickly adapt after they had to step away from the project. This left the entire development effort in the hands of a single developer, requiring rapid problem-solving and strategic prioritization to meet the hackathon deadline.
βοΈ Balancing AI, Performance & Real-Time Processing
Integrating AI-powered object detection and translation on a mobile device posed technical challenges, especially when handling real-time image processing and API efficiency. We had to optimize memory usage, caching, and API request handling to ensure smooth performance.
π What We Learned
- Agility matters β We had to quickly adapt our workflow to changing team dynamics and project constraints.
- AI integration is powerful but complex β Real-time AI processing requires careful performance tuning and optimization.
- Collaboration is key β Even across different time zones, communication and teamwork made this project possible.
π― What's Next?
This is just the beginning! Future iterations of Language Lens could include:
β
Offline Translation Mode β Reducing dependency on API calls for faster results
β
Augmented Reality (AR) Labels β Overlaying translated text on real-world objects
β
Expanded Language Support β Enhancing accessibility for diverse communities
Through Language Lens, we aim to make the world more connected, inclusive, and accessibleβone translation at a time. πβ¨
Built With
- accompanist-permissions
- camerax
- figma
- google-gemini-api
- jetpack-compose
- meshy
- okhttp


Log in or sign up for Devpost to join the conversation.