Disabled persons have always been limited in their methods to communicate with the world. Due to the rapidly growing popularity and devleopment towards VR, we wanted to implement a method to allow mute people to communicate in VR using their "natural" language.

What it does

Leap motion detects sign language gestures converting it into command line text. Implementation in reality would be to allow mute people to sign in VR and communicate with those people who don't understand sign language.

How we built it

We used Leap Motion SDK and built it using python 2.7.

Challenges we ran into

Lots of technological difficulties. The most trouble we had was the intimidating factor of trying to understand Leap Motion API in a short amount of time. This was the first time we worked with Leap Motion.

Accomplishments that we're proud of

Managing to produce a minimum viable product despite having switched project ideas multiple times in 24 hours.

What we learned

We learned how to use LeapMotion API as well as python.

What's next for LeapLanguage

Add text to voice feature to allow mute people to sign in VR and having the ASL being interpreted as voice. This will have many applications such as in conference calls, mute people will be able to dynamically voice their opinions.

Built With

Share this project: