TY - CHAP
T1 - Visual modeling of dynamic gestures using 3D appearance and motion features
AU - Ye, Guangqi
AU - Corso, Jason J.
AU - Hager, Gregory D.
PY - 2005
Y1 - 2005
N2 - We present a novel 3D gesture recognition scheme that combines the 3D appearance of the hand and the motion dynamics of the gesture to classify manipulative and controlling gestures. Our method does not directly track the hand. Instead, we take an object-centered approach that efficiently computes 3D appearance using a region-based coarse stereo matching algorithm. Motion cues are captured by differentiating the appearance feature with respect to time. An unsupervised learning scheme is carried out to capture the cluster structure of these features. Then, the image sequence of a gesture is converted to a series of symbols that indicate the cluster identities of each image pair. Two schemes, i.e., forward HMMs and neural networks, are used to model the dynamics of the gestures. We implemented a real-time system and performed gesture recognition experiments to analyze the performance with different combinations of the appearance and motion features. The system achieves recognition accuracy of over 96% using both the appearance and motion cues.
AB - We present a novel 3D gesture recognition scheme that combines the 3D appearance of the hand and the motion dynamics of the gesture to classify manipulative and controlling gestures. Our method does not directly track the hand. Instead, we take an object-centered approach that efficiently computes 3D appearance using a region-based coarse stereo matching algorithm. Motion cues are captured by differentiating the appearance feature with respect to time. An unsupervised learning scheme is carried out to capture the cluster structure of these features. Then, the image sequence of a gesture is converted to a series of symbols that indicate the cluster identities of each image pair. Two schemes, i.e., forward HMMs and neural networks, are used to model the dynamics of the gestures. We implemented a real-time system and performed gesture recognition experiments to analyze the performance with different combinations of the appearance and motion features. The system achieves recognition accuracy of over 96% using both the appearance and motion cues.
UR - http://www.scopus.com/inward/record.url?scp=34548271039&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=34548271039&partnerID=8YFLogxK
U2 - 10.1007/0-387-27890-7_7
DO - 10.1007/0-387-27890-7_7
M3 - Chapter
AN - SCOPUS:34548271039
SN - 0387276971
SN - 9780387276977
SP - 103
EP - 120
BT - Real-Time Vision for Human-Computer Interaction
ER -