The idea was to help people who are blind, to be able to discretely gather context during social interactions and general day-to-day activities

What it does

The glasses take a picture and analyze them using Microsoft, Google, and IBM Watson's Vision Recognition APIs and try to understand what is happening. They then form a sentence and let the user know. There's also a neural network at play that discerns between the two dens and can tell who is in the frame

How I built it

We took a RPi Camera and increased the length of the cable. We then made a hole in our lens of the glasses and fit it in there. We added a touch sensor to discretely control the camera as well.

Challenges I ran into

The biggest challenge we ran into was Natural Language Processing, as in trying to parse together a human-sounding sentence that describes the scene.

What I learned

I learnt a lot about the different vision APIs out there and creating/trainingyour own Neural Network.

What's next for Let Me See

We want to further improve our analysis and reduce our analyzing time.

Share this project: