The idea was to help people who are blind, to be able to discretely gather context during social interactions and general day-to-day activities
What it does
The glasses take a picture and analyze them using Microsoft, Google, and IBM Watson's Vision Recognition APIs and try to understand what is happening. They then form a sentence and let the user know. There's also a neural network at play that discerns between the two dens and can tell who is in the frame
How I built it
We took a RPi Camera and increased the length of the cable. We then made a hole in our lens of the glasses and fit it in there. We added a touch sensor to discretely control the camera as well.
Challenges I ran into
The biggest challenge we ran into was Natural Language Processing, as in trying to parse together a human-sounding sentence that describes the scene.
What I learned
I learnt a lot about the different vision APIs out there and creating/trainingyour own Neural Network.
What's next for Let Me See
We want to further improve our analysis and reduce our analyzing time.
Log in or sign up for Devpost to join the conversation.