The idea of this project is to enhance the senses of the visually impaired through technology, specifically with Microsoft’s Computer Vision Technology.

What it does

Our system is composed of a number of wearables that are connected to a pair of smart glasses. Based on what the camera on the glasses sees, it sends a specific signal to the wearables. This signal is then transformed into not only sound but vibrations, electric signals or heat. The idea is that after a time of adaptation, the user becomes to used to this signals that they feel like an extension of his own senses, having a huge impact on the user’s lifestyle. This is a highly modular system, and new modules can be easily added. With enough wearables you can achieve a significant combination of signals which gives the user a lot of information about his environment.

A simple implementation of the system has been a platform that runs on the user’s smart glasses. When the user interacts with another person, the system detects his facial expression and sends a signal to the user. The user receives the signal as a sound, but it would perform better if the signal was a vibration to the wristband of the user.

How we built it

Python 3.6 and various libraries Speaker instead of a vibrating wristband

Challenges we ran into

Libraries compatibility

Accomplishments that we're proud of

Fully functional product

What we learned

How to use Microsoft’s Face API.

What's next for Cogniton Assistant

Adding support for more modules and wearables

Built With

Share this project: