Inspiration
After a video promoting ALS appeared in our feed, we became acutely aware of a demographic unable to use technology due deteriorating or deteriorated motor function. We decided we would build a project to help combat this.
What it does
Our program uses only a webcam to track a users eye movement and blinks. Simultaneously it is listening for any audio input. This way a person can demonstrate complete control of a computer by simply moving their head to move the cursor, blinking and keeping eyes shut to click and drag and talking to replace the use of a keyboard.
How we built it
We built our project using exclusively python. We separated into two groups. One group would take advantage of libraries such as openCL, media pipes and speech recognition, to build they eye tracker and input device as a whole. Whilst the other group would work on developing a demo, leveraging artificial intelligence and image recognition to fully promote the capabilities of our device in a unique and fun way.
Challenges we ran into
When it came to developing the input device we faced two unique challenges. The first was tracking eye movement in an efficient way, we constantly overloaded the computer leading to heavy lag. The second challenge we faced was actually developing an intuitive way to actually click and move. We were constantly adjust sensitivity of eye movement the time it takes to register a click and more until we created something that worked.
Accomplishments that we're proud of
We are very proud of creating a product that can actually help others, it provides accessibility and functionality that will ultimately improve people live. Best of all it is uncostly as the program works with a basic web cam. We couldn't be more happy about this.
What we learned
Our main take aways in completing this project are abundantly clear. We have learn how track not only head and eyes, but really any part of the body using openCL and mediapipes. Whereas on the demo development side of things we learnt of the difficulty that comes with integrating artificial intelligence into our work, just how tedious and demanding this can be.
What's next for Project VISION
Without there is more tinkering to be done. We could make our clicking system more intuitive, the camera could track our eyes better and be more stable too. However most importantly we could make the program more efficient, and therefore run more smoothly.
Log in or sign up for Devpost to join the conversation.