Music is a shared human experience that everybody can enjoy. However, getting good at it takes tons of time and many people are simply to busy to master an instrument. What if you could compose the perfect soundtrack for your life with just your hands?

What it does

Employing two myo wristbands we recognize gestures of both hands and link them to music samples. You can compose your own tunes step-by-step by selecting the samples that you feel belong there. In order to make things even more interesting, you can also dynamically control the volume and the speed of your creation by simply moving your arm up and down. But I am with friends?! Even better, friends that jam together stay together. Just give one of the controllers to your friend and you're all set for a unique collaborative experience. You are always in control of what's going on: It's all just one glance onto the dashboard away.

How we built it

All the low-level stuff with gestures and processing tracking data etc. is done in C++. This interacts with the Cocoa app via an Objective-C wrapper.

Challenges we ran into

Getting the audio output to work

Turns out it is not as easy to mix Objective-C and C++ as Objective C++ input file suggests - (ever heard of PIMPL?)

MacOS sandboxing only applies to GUI applications and not to console ones

A quaternion ain't your usual (X,Y,Z) coordiantes

Accomplishments that we're proud of

Our intuitive user experience

How well our samples go with each other making it easy to get an amazing tune

What we learned

We figured out how to built our first MacOS X app ever

Objective-C sure loves brackets :]

Your neighbors don't appreciate you debugging an application that makes sounds all the time that much

What's next for Quaternion

We'd love to let even more people be part of your experience. All we need are more controllers (How about a hundred?).

Built With

Share this project: