scispace - formally typeset
Search or ask a question

Showing papers by "Rajeev Sharma published in 2000"


Journal ArticleDOI
TL;DR: A speech/gesture interface that uses visual hand-gesture analysis and speech recognition to control a 3D display in VMD, a virtual environment for structural biology, to simplify model manipulation and rendering to make biomolecular modeling more playful.
Abstract: We developed a speech/gesture interface that uses visual hand-gesture analysis and speech recognition to control a 3D display in VMD, a virtual environment for structural biology. The reason we used a particular virtual environment context was to set the necessary constraints to make our analysis robust and to develop a command language that optimally combines speech and gesture inputs. Our interface uses: automatic speech recognition (ASR), aided by a microphone, to recognize voice commands; two strategically positioned cameras to detect hand gestures; and automatic gesture recognition (AGR), a set of computer vision techniques to interpret those hand gestures. The computer vision algorithms can extract the user's hand from the background, detect different finger positions, and distinguish meaningful gestures from unintentional hand movements. Our main goal was to simplify model manipulation and rendering to make biomolecular modeling more playful. Researchers can explore variations of their model and concentrate on biomolecular aspects of their task without undue distraction by computational aspects. They can view simulations of molecular dynamics, play with different combinations of molecular structures, and better understand the molecules' important properties. A potential benefit, for example, might be reducing the time to discover new compounds for new drugs.

64 citations


Journal ArticleDOI
TL;DR: Results of user studies revealed that gesture primitives, originally extracted from weather map narration, form patterns of co-occurrence with speech parts in association with their meaning in a visual display control system, defining a direction in approaching interpretation in natural gesture-speech interfaces.
Abstract: In recent years because of the advances in computer vision research, free hand gestures have been explored as a means of human-computer interaction (HCI). Gestures in combination with speech can be an important step toward natural, multimodal HCI. However, interpretation of gestures in a multimodal setting can be a particularly challenging problem. In this paper, we propose an approach for studying multimodal HCI in the context of a computerized map. An implemented testbed allows us to conduct user studies and address issues toward understanding of hand gestures in a multimodal computer interface. Absence of an adequate gesture classification in HCI makes gesture interpretation difficult. We formalize a method for bootstrapping the interpretation process by a semantic classification of gesture primitives in HCI context. We distinguish two main categories of gesture classes based on their spatio-temporal deixis. Results of user studies revealed that gesture primitives, originally extracted from weather map narration, form patterns of co-occurrence with speech parts in association with their meaning in a visual display control system. The results of these studies indicated two levels of gesture meaning: individual stroke and motion complex. These findings define a direction in approaching interpretation in natural gesture-speech interfaces.

43 citations


Proceedings ArticleDOI
Rajeev Sharma1, J. Cai1, S. Chakravarthy1, I. Poddar1, Y. Sethi1 
26 Mar 2000
TL;DR: An HMM architecture for continuous gesture recognition framework and keyword spotting is presented and a statistical co-occurrence analysis of different gestures with a selected set of spoken keywords is conducted to explore the relation between gesture and speech.
Abstract: In order to incorporate naturalness in the design of human computer interfaces (HCI), it is desirable to develop recognition techniques capable of handling continuous natural gesture and speech inputs. Though many different researchers have reported high recognition rates for gesture recognition using hidden Markov models (HMM), the gestures used are mostly pre-defined and are bound with syntactical and grammatical constraints. But natural gestures do not string together in syntactical bindings. Moreover, strict classification of natural gestures is not feasible. We have examined hand gestures made in a very natural domain, that of a weather person narrating in front of a weather map. The gestures made by the weather person are embedded in a narration. This provides us with abundant data from an uncontrolled environment to study the interaction between speech and gesture in the context of a display. We hypothesize that this domain is very similar to that of a natural human-computer interface. We present an HMM architecture for continuous gesture recognition framework and keyword spotting. To explore the relation between gesture and speech, we conducted a statistical co-occurrence analysis of different gestures with a selected set of spoken keywords. We then demonstrate how this co-occurrence analysis can be exploited to improve the performance of continuous gesture recognition.

29 citations


01 Jan 2000
TL;DR: A framework for sensor-based motion planning of robotic manipulators using a topology representing network (TRN) and the representation of the Perceptual Control Manifold (PCM), a recently introduced concept for motion planning is presented.
Abstract: We present a framework for sensor-based motion planning of robotic manipulators using a topology representing network (TRN). Exploiting the perfectly topology preserving features of the network, the algorithm learns the representation of the Perceptual Control Manifold (PCM), a recently introduced concept for motion planning. This concept allows sensors to be integrated into robot motion planning. Besides a demonstration of the technical feasibility of motion planning through perfectly topology preserving maps the capabilities of this approach within an engineering framework, namely the implementation on a pneumatically driven robot arm (SoftArm), are demonstrated.

9 citations