We were inspired by the demo given to us of Google's ML Vision API in which a face was analyzed to determine emotion. Since as a team we were already focused on education, we decided to adapt this tool to the classroom.
What it does
Our web app takes continuous image data of students in class to analyze in Google's ML Vision. The algorithms send back data such as happiness and attentiveness. In addition, a text transcription of the lecture is taken in real time with relevant topics highlighted using Google's Speech API and Text Analytics.
How we built it
Our web app makes use of NodeJS, React, MongoDB, and Google Speech and Vision APIs
Challenges we ran into
Adapting available documentation to our personal use cases.
Accomplishments that we're proud of
This was our first full-stack application that we were able to create from the ground up with a database, server, and front-end.
What we learned
We learned how to properly make use of MongoDB as well as Google's powerful APIs utilizing machine learning.