Untethered gesture acquisition and recognition for virtual world manipulation
D Demirdjian, T Ko, T Darrell - Virtual Reality, 2005 - Springer
D Demirdjian, T Ko, T Darrell
Virtual Reality, 2005•SpringerHumans use a combination of gesture and speech to interact with objects and usually do so
more naturally without holding a device or pointer. We present a system that incorporates
user body-pose estimation, gesture recognition and speech recognition for interaction in
virtual reality environments. We describe a vision-based method for tracking the pose of a
user in real time and introduce a technique that provides parameterized gesture recognition.
More precisely, we train a support vector classifier to model the boundary of the space of …
more naturally without holding a device or pointer. We present a system that incorporates
user body-pose estimation, gesture recognition and speech recognition for interaction in
virtual reality environments. We describe a vision-based method for tracking the pose of a
user in real time and introduce a technique that provides parameterized gesture recognition.
More precisely, we train a support vector classifier to model the boundary of the space of …
Abstract
Humans use a combination of gesture and speech to interact with objects and usually do so more naturally without holding a device or pointer. We present a system that incorporates user body-pose estimation, gesture recognition and speech recognition for interaction in virtual reality environments. We describe a vision-based method for tracking the pose of a user in real time and introduce a technique that provides parameterized gesture recognition. More precisely, we train a support vector classifier to model the boundary of the space of possible gestures, and train Hidden Markov Models (HMM) on specific gestures. Given a sequence, we can find the start and end of various gestures using a support vector classifier, and find gesture likelihoods and parameters with a HMM. A multimodal recognition process is performed using rank-order fusion to merge speech and vision hypotheses. Finally we describe the use of our multimodal framework in a virtual world application that allows users to interact using gestures and speech.
Springer
顯示最佳搜尋結果。 查看所有結果