Our hackathon entry didn't begin with Know-It-All, we were contemplating running different machine learning models and had almost decided on a completely different project until we learned that that we were to be provided with google cloud credits. Although this was an entire new ecosystem for the both of us, we decided to utilize google-compute along with natural language processing, speech to text, text to speech using wavenext (a state of the art text to speech network) in order to use the model GPT-2 by OpenAI.
GPT-2 is a heavyweight architecture composed of over 754 Million parameters, and for this reason we needed to run the backend on a google compute virtual machine utilizing thousands of dollars worth of hardware more than we could personally afford. This model has an interesting past as the largest, most intelligent version, has been held back from release due to the potential for abuse to create political controversy and fake facts.
In conclusion we spent 28 of the last 36 hours awake, debugging, and tweaking our model in order to create an entertaining live example of the power of modern machine learning.