Inspiration
Presently, a visually impaired person has limited choices when it comes to moving about in known or unknown territory or travel. The person has to either employ the services of another person who can see, or use the help of a seeing-eye or guide dog if the person is unfamiliar with the surroundings.The third eye is an assistant for visually impaired which narrates the description of the scene.
What it does
It generates a caption describing the content of an image with the help of a module called Caption Generator, consists of Convolution Neural Networks and Recurrent Neural Network, which has been trained on a large scale datasets. Third Eye is an assistant for visually impaired which narrates the description of the scene. Third Eye system aims to bring the beautiful world as a narrative to the visually impaired. The narrative is generated by converting the scenes in front of them to text which describes the important objects in the scene. Examples of text include 'A group of people playing a game of football', 'yellow truck parked next to the car', a bowl of salad kept on a table'. Our system then converts this text into a speech using a speech synthesizer.
How we built it
We have built with,
- tensorflow
- Python /flask
- IBM Watson API
- Deep Learning
- Raspberry-Pi / Camera Module
Challenges we ran into
- we are planning to make AI hardware system for visually impaired people(virtual assistant) but due hardware failure(pi camera) we have to move it to software(we are not good with UI/UX).
Accomplishments that we're proud of
- we have made it work(somehow)!
What we learned
- AI in Production
What's next for Third Eye
- We gonna make it more stable product and soon available
Built With
- ai
- aws-gpu
- deep-learning
- flask
- ibm-watson
- python
- tensorflow

Log in or sign up for Devpost to join the conversation.