We wanted to create something good for people with visual disabilities(or just very bad at taking pictures) so we came up with a app that helped blind people take pictures.
What it does
The app provides audio tips to the user about the optimal angle, distance and height to take a good picture. It also uses Nuance's natural language processing to understand the user's intentions and make the navigation process more seamless. Also used Google's TTS and STT to make the audio work.
How we built it
We built it using the standard android-sdk using java as a programming language.
Challenges we ran into
We had a lot of trouble with the Microsoft Azure platform, setting it up and getting the pictures to correctly store on their blob services and then using it to do sentiment analysis on it.
Accomplishments that we're proud of
Being able to complete a project :)
What we learned
Working with Azure is hard
What's next for blindSpot
Adding functionality such as voice recording for images.