We wanted to provide some medium through which people with communication disorders could indicate their emotions and have fun doing it.
What it does
It uses the Microsoft Band sensors as a modified polygraph in conjunction with our neural network implementation
How I built it
I created a neural network abstraction to create neurons dynamically for every new song played. The android app reads in sensor input from the Microsoft Band, gets a user's saved songs through the Spotify Web API, runs the sensors through the neural network which outputs a song name or random, plays the song with the Spotify Android SDK, and then waits for the user to let the song play or skip in order to train the neural network for that song.
Challenges I ran into
Fuck Spotify, fuck Microsoft, fuck Android. That is all.
Accomplishments that I'm proud of
Spotify, Microsoft, Android.
What I learned
Don't use Spotify, don't use Microsoft, don't use Android