Inspiration

I wanted to do something interesting with the Twitter API and I thought the recent breakthroughs in face animation based on voice would be interesting to learn about

What it does

A bot that grabs mentions and posts a video with a face saying the tweet's content with a sentiment analysis

How I built it

I used a google cloud vm to first grab tweets that mentioned the bot, then the tweets' would be fed into a text to speech engine. Then this audio file would be fed into a research project that allows for face animation generation from an audio file. Finally a sentiment analysis would be done of the tweet and then the video would be uploaded saying the tweet sounds either positive, negative, or neutral

Challenges I ran into

There were a lot of dependency issues and not being familiar with any of the libraries I worked with made development a lot more of reading documentation rather than coding.

Accomplishments that I'm proud of

I am proud I was able to parse through a research project, integrate different unfamiliar technologies together, and finally have something that works(ish).

What I learned

I feel like I starting to get a grasp about how animations are generated using audio files.

What's next for Twitter talker

There are three main features that I would like to add

  • Add real human faces instead of just a mesh, this should not be too difficult as there are ways to map the generated mesh to a real photo
  • Use the sentiment analysis to choose different text to speech voices (e.g. an angry tweet will have an angry text to speech voice)
  • Use the sentiment analysis to choose different faces instead of just a wire (e.g. a happy tweet will have a happy face or emoji saying the text)

Built With

Share this project:

Updates