Tony’s Status Report 02/15

Progress

I read some action recognition papers (https://paperswithcode.com/search?q_meta=&q=action+recognition), many of which involve temporal in addition to spatial approaches in their models. After discussing this with Marios and Emily, Chris and I have opted instead to start with a spatial approach and then include temporal approaches if we cannot achieve the necessary accuracy. To start off, we will model claps and other gestures in terms of a change in the coordinates of the relevant joints.

Deliverables next week

By next week, I want to have a basic spatial algorithm that will recognize claps and stomps (taking videos of more actions as necessary).

Schedule

I feel slightly behind schedule, since I should have started coding algorithms by this point. I spent time on researching very complex algorithms when I should have been thinking of simpler approaches first.

Chris’ Status Update 02/15

Progress

I was able to get OpenPose running on the computer and I started to work on the gesture recognition using pose data from a video that was processed using OpenPose. I’m devising a way to use the pose of the hands to recognize a large clap.

I’m also looking into different action recognition “libraries” (other github projects) to do action recognition on a play pause wave.

I’ve also been able to get the Kinect working with OpenPose using libfreenect2.

Deliverables next week

Hopefully by the end of this week I’ll be able to get some basic gesture recognition using distance and frame numbers. I want to be able to recognize claps and stomps. This will still be on processed video, and when we get the Xavier Board we will be able to see if these work in real time, even with high latency.

Schedule

I think that I’m still on schedule, although I think that using machine learning classification to recognize claps and stomps will be a little more challenging than I initially thought.