TY - GEN
T1 - Extending the interaction area for view-invariant 3D gesture recognition
AU - Caon, Maurizio
AU - Tscherrig, Julien
AU - Yue, Yong
AU - Khaled, Omar Abou
AU - Mugellini, Elena
PY - 2012
Y1 - 2012
N2 - This paper presents a non-intrusive approach for view-invariant hand gesture recognition. In fact, the representation of gestures changes dynamically depending on camera viewpoints. Therefore, the different positions of the user between the training phase and the evaluation phase can severely compromise the recognition process. The proposed approach involves the calibration of two Microsoft Kinect depth cameras to allow the 3D modeling of the dynamic hands movements. The gestures are modeled as 3D trajectories and the classification is based on Hidden Markov Models. The approach is trained on data from one viewpoint and tested on data from other very different viewpoints with an angular variation of 180°. The average recognition rate is always higher than 94%. Since it is similar to the recognition rate when training and testing on gestures from the same viewpoint, hence the approach is indeed view-invariant. Comparing these results with those deriving from the test of a one depth camera approach demonstrates that the adoption of two calibrated cameras is crucial.
AB - This paper presents a non-intrusive approach for view-invariant hand gesture recognition. In fact, the representation of gestures changes dynamically depending on camera viewpoints. Therefore, the different positions of the user between the training phase and the evaluation phase can severely compromise the recognition process. The proposed approach involves the calibration of two Microsoft Kinect depth cameras to allow the 3D modeling of the dynamic hands movements. The gestures are modeled as 3D trajectories and the classification is based on Hidden Markov Models. The approach is trained on data from one viewpoint and tested on data from other very different viewpoints with an angular variation of 180°. The average recognition rate is always higher than 94%. Since it is similar to the recognition rate when training and testing on gestures from the same viewpoint, hence the approach is indeed view-invariant. Comparing these results with those deriving from the test of a one depth camera approach demonstrates that the adoption of two calibrated cameras is crucial.
KW - 3D gesture recognition
KW - HMM
KW - Image processing application
KW - Kinect
KW - depth cameras calibration
KW - view-invariant
UR - http://www.scopus.com/inward/record.url?scp=84875861126&partnerID=8YFLogxK
U2 - 10.1109/IPTA.2012.6469542
DO - 10.1109/IPTA.2012.6469542
M3 - Conference Proceeding
AN - SCOPUS:84875861126
SN - 9781467325837
T3 - 2012 3rd International Conference on Image Processing Theory, Tools and Applications, IPTA 2012
SP - 293
EP - 298
BT - 2012 3rd International Conference on Image Processing Theory, Tools and Applications, IPTA 2012
T2 - 2012 3rd International Conference on Image Processing Theory, Tools and Applications, IPTA 2012
Y2 - 15 October 2012 through 18 October 2012
ER -