scispace - formally typeset
Search or ask a question
Author

Inderpal Singh Mumick

Bio: Inderpal Singh Mumick is an academic researcher from Wilmington University. The author has contributed to research in topics: Voice browser & Wireless Markup Language. The author has an hindex of 5, co-authored 7 publications receiving 220 citations.

Papers
More filters
Patent
03 Jul 2002
TL;DR: In this paper, the authors propose a technique for synchronizing a visual browser and a voice browser, where the visual browser creates a historical record of events that have occurred during the navigation, and the voice browser uses this historical record to navigate the content in the same manner as occurred on the visual browsers.
Abstract: A technique for synchronizing a visual browser and a voice browser. A visual browser is used to navigate through visual content, such as WML pages. During the navigation, the visual browser creates a historical record of events that have occurred during the navigation. The voice browser uses this historical record to navigate the content in the same manner as occurred on the visual browser, thereby synchronizing to a state equivalent to that of the visual browser. The creation of the historical record may be performed by using a script to trap events, where the script contains code that records the trapped events. The synchronization technique may be used with a multi-modal application that permits the mode of input/output (I/O) to be changed between visual and voice browsers. When the mode is changed from visual to voice, the record of events captured by the visual browser is provided to the voice browser, thereby allowing the I/O mode to change seamlessly from visual to voice. Likewise, the voice browser captures events which may be provided to the visual browser when the I/O mode is changed from voice to visual.

119 citations

Patent
10 Apr 2002
TL;DR: In this article, mode swapping refers to the process of changing from one mode of communication (or one combination of modes) to another, which may be initiated by a user, by an application, or by another entity.
Abstract: Multi-modal applications are capable of interacting with a user in a voice mode, a visual mode, and/or other modes of communication. Mode swapping refers to the process of changing from one mode of communication (or one combination of modes) to another. Mode swapping may be initiated by a user, by an application, or by another entity. Mode swapping may be performed “adaptively,” in that an entity monitors the effectiveness of a communication mode and initiates a mode swap if communication is proceeding non-optimally in the current mode. Mode swapping may be performed by dropping a call to a telephone terminal in the currently-open mode and placing a call to that same device in another mode. In one embodiment, logic for mode swapping resides on a gateway proxy that mediates communication between an application and a telephone terminal.

61 citations

Patent
05 Aug 2002
TL;DR: In this article, a system for enabling multi-modal bookmarks comprises a bookmark repository (120) coupled to two browsers, where a user bookmarks content using a first browser, and a bookmark is stored in the bookmark repository; the user subsequently uses a second browser and requests that the bookmark be de-referenced to point the second browser to the bookmarked content.
Abstract: A system for enabling multi-modal bookmarks comprises a bookmark repository (120) coupled to two browsers. A user bookmarks content using a first browser, and a bookmark is stored in the bookmark repository (120). The user subsequently uses a second browser and requests that the bookmark be de-referenced to point the second browser to the bookmarked content. The second browser accesses the bookmark in the content repository and is directed to the bookmarked content. The first and second browsers may be visual and voice browsers (118) that render Wireless Markup Language (WML) and Voice Extensible Markup Language (VXML) content, respectively, wherein a mapping is provided that links a given piece of visual content to an equivalent piece of voice content.

16 citations

Patent
01 Oct 2002
TL;DR: In this article, a technique for allowing an automated participant in a telephone conversation, such as an Interactive Voice Response (IVR) system or a voice browser, to react to the placing of the conversation "on hold".
Abstract: A technique for allowing an automated participant in a telephone conversation, such as an Interactive Voice Response (IVR) system or a voice browser, to react to the placing of the conversation 'on hold'. When a conversation between an automated participant and a human participant is placed on hold, a signal is provided to the automated participant indicating that the voice circuit between the two participants has been broken. In response to the signal, the automated participant may take appropriate action, such as suspending its participation in the conversation, extending the time for which it awaits a response from the human participant, continuing the conversation in a 'visual' mode using the keypad and display on the human participant's handset, or some other appropriate action.

16 citations

Patent
09 Oct 2007
TL;DR: In this article, the authors propose a technique for synchronizing a visual browser and a voice browser, where the visual browser creates a historical record of events that have occurred during the navigation, and the voice browser uses this historical record to navigate the content in the same manner as occurred on the visual browsers.
Abstract: A technique for synchronizing a visual browser and a voice browser. A visual browser is used to navigate through visual content, such as WML pages. During the navigation, the visual browser creates a historical record of events that have occurred during the navigation. The voice browser uses this historical record to navigate the content in the same manner as occurred on the visual browser, thereby synchronizing to a state equivalent to that of the visual browser. The creation of the historical record may be performed by using a script to trap events, where the script contains code that records the trapped events. The synchronization technique may be used with a multi-modal application that permits the mode of input/output (I/O) to be changed between visual and voice browsers. When the mode is changed from visual to voice, the record of events captured by the visual browser is provided to the voice browser, thereby allowing the I/O mode to change seamlessly from visual to voice. Likewise, the voice browser captures events which may be provided to the visual browser when the I/O mode is changed from voice to visual.

5 citations


Cited by
More filters
PatentDOI
TL;DR: In this paper, a system for receiving speech and non-speech communications of natural language questions and commands, transcribing the speech and NN communications to textual messages, and executing the questions and/or commands is presented.
Abstract: Systems and methods are provided for receiving speech and non-speech communications of natural language questions and/or commands, transcribing the speech and non-speech communications to textual messages, and executing the questions and/or commands. The invention applies context, prior information, domain knowledge, and user specific profile data to achieve a natural environment for one or more users presenting questions or commands across multiple domains. The systems and methods creates, stores and uses extensive personal profile information for each user, thereby improving the reliability of determining the context of the speech and non-speech communications and presenting the expected results for a particular question or command.

1,164 citations

Patent
29 Aug 2006
TL;DR: In this article, a mobile system is provided that includes speech-based and non-speech-based interfaces for telematics applications that identify and uses context, prior information, domain knowledge, and user specific profile data to achieve a natural environment for users that submit requests and/or commands in multiple domains.
Abstract: A mobile system is provided that includes speech-based and non-speech-based interfaces for telematics applications. The mobile system identifies and uses context, prior information, domain knowledge, and user specific profile data to achieve a natural environment for users that submit requests and/or commands in multiple domains. The invention creates, stores and uses extensive personal profile information for each user, thereby improving the reliability of determining the context and presenting the expected results for a particular question or command. The invention may organize domain specific behavior and information into agents, that are distributable or updateable over a wide area network.

716 citations

Patent
11 Dec 2007
TL;DR: In this paper, a conversational, natural language voice user interface may provide an integrated voice navigation services environment, where the user can speak conversationally, using natural language, to issue queries, commands, or other requests relating to the navigation services provided in the environment.
Abstract: A conversational, natural language voice user interface may provide an integrated voice navigation services environment. The voice user interface may enable a user to make natural language requests relating to various navigation services, and further, may interact with the user in a cooperative, conversational dialogue to resolve the requests. Through dynamic awareness of context, available sources of information, domain knowledge, user behavior and preferences, and external systems and devices, among other things, the voice user interface may provide an integrated environment in which the user can speak conversationally, using natural language, to issue queries, commands, or other requests relating to the navigation services provided in the environment.

450 citations

Patent
04 Aug 2006
TL;DR: In this article, a conversational human-machine interface that includes conversational speech analyzer, a general cognitive model, an environmental model, and a personalized cognitive model to determine context, domain knowledge, and invoke prior information to interpret a spoken utterance or a received non-spoken message is presented.
Abstract: A system and method are provided for receiving speech and/or non-speech communications of natural language questions and/or commands and executing the questions and/or commands. The invention provides a conversational human-machine interface that includes a conversational speech analyzer, a general cognitive model, an environmental model, and a personalized cognitive model to determine context, domain knowledge, and invoke prior information to interpret a spoken utterance or a received non-spoken message. The system and method creates, stores and uses extensive personal profile information for each user, thereby improving the reliability of determining the context of the speech or non-speech communication and presenting the expected results for a particular question or command.

430 citations

Patent
16 Oct 2007
TL;DR: In this paper, a cooperative conversational voice user interface is presented, which builds upon short-term and long-term shared knowledge to generate one or more explicit and/or implicit hypotheses about an intent of a user utterance.
Abstract: A cooperative conversational voice user interface is provided. The cooperative conversational voice user interface may build upon short-term and long-term shared knowledge to generate one or more explicit and/or implicit hypotheses about an intent of a user utterance. The hypotheses may be ranked based on varying degrees of certainty, and an adaptive response may be generated for the user. Responses may be worded based on the degrees of certainty and to frame an appropriate domain for a subsequent utterance. In one implementation, misrecognitions may be tolerated, and conversational course may be corrected based on subsequent utterances and/or responses.

413 citations