Tuesday, February 23, 2010

Computer Vision Based Gesture Recognition for an Augmented Reality Interface

Comments
Franck Norman
Drew

Summary
In this paper, they try to create a vision based gesture interface for an Augmented Reality system. It can recognize a 3D pointing gesture, a click gesture and 5 static gestures.

To define the gestures, they just use a closed fist, and various numbers of fingers open. They asked the users to perform the gestures in the same plane, which reduced the recognition problem to a 2D issue. The recognition method relies on a pre-segmented image of the hand. They use a color pixel-based approach to account for varying size and forms from image to image. After segmenting the hand pixes from the image, the next step is to detect the number of fingers that are outstretched. It does this by measuring the smallest and largest radii from the center of the palm where there is a non finger pixel. After this, they would recognize the point if there is only one finger detected and click was done by the thumb.

The user study was done by several users. They just said that the users adapted quickly.

------------------------------------
This paper wasn't that good because it didn't go into more detail about the results. We don't know what they meant by several, and they didn't give any quantitative analysis of the results or mention what kind of tests the users went through. Overall, it was a good idea. The results were lacking.

1 comment:

  1. I agree, more results would have been helpful. I still find their approach for finding the number of fingers quite interesting.

    ReplyDelete