The project of our team is to build the Human-computer interacted Virtual piano using the camera. Our team uses the openCv to track the object and then play music.


This idea is inspired by the HCI glove, the link of the demo is here:

What it does:

The virtual piano has 11 keys and can play different tunes with digital chord music.

How we built it:

Using python-openCV.

Challenges we ran into:

The problem of recognizing the gestures. We write the code to track the colored objects to implement the interaction.

Accomplishments that we're proud of:

The virtual environment and the hand-free interaction.

What we learned:

Programming of openCV.

What's next:

Apply deep learning technique to recognize the hand gesture

Built with:

Python, openCV.

Prizes we're going for:

$100 Amazon Gift Cards

Team Members

Yu Li, Jiaming Nie, Ruojun Li
View on Github