Blind walking canes are only capable of spacial awareness. They cannot easily convey the objects that are in the environment in front of a person.
What it does
An Android application that anyone can download.
Uses computer vision to detect objects that are near the user and give them cues (audio and haptic) to let them know about these objects. For example, if a person is getting close it will give a heads up to the user.
How we built it
Built off of TensorFlow and Android. Uses pre-trained models to do real-time processing of what the camera is currently seeing.
Challenges we ran into
Getting computer vision working quick enough on a smartphone so it can give real-time updates to users.
Accomplishments that we're proud of
Computer vision on a mobile phone.
What we learned
We used CruzHacks as an opportunity to learn computer vision. We learned the TensorFlow library and the different types of neural networks and explored other computer vision libraries (OpenCV).
What's next for EyeDog
Training a model that is based off a data set of imagery from what cameras normally see when walking.
Integration with Google maps to provide better walking directions
Improved haptic feedback (vibrations)