Inspiration
We were inspired by the medical condition of David's father who is blind. Our app was built to help him see the world. Our app helps visually impaired people understand what's going on around them.
What it does
Look Assistive Technology is an application that describes its surrounding environment via audio to assist visually impaired individuals. The application is controlled through voice commands or simple whole page buttons. Our customers can use Look when crossing the road, looking for an object like a bag, or keys. We use computer vision to translate camera photos into speech. We can recognize a pedestrian crosswalk and help visually impaired people translate vision into speech.
How we built it
We use Google's cloud vision API and OpenCV for computer vision and TensorFlow for Machine Intelligence running on AWS.
Challenges we ran into
Android Studio build problems.
Accomplishments that we're proud of
We are proud of a product that has real potential to do positive change, and help the visually impaired.
What we learned
We learned a few new technologies and algorithms, including image recognition, machine learning, Google cloud platform API, and android app development.
What's next for Look
The app would allow users to request their location, walking directions to any spoken address, and facial recognition. We want to use this technology for self driving cars and autonomous robots.
Built With
- amazon-web-services
- android-studio
- google-cloud-vision
- google-text-to-speech
- python
- tensorflow
Log in or sign up for Devpost to join the conversation.