scispace - formally typeset
Search or ask a question
Author

Gerald M. McCobb

Other affiliations: IBM
Bio: Gerald M. McCobb is an academic researcher from Nuance Communications. The author has contributed to research in topics: Grammar & Markup language. The author has an hindex of 19, co-authored 28 publications receiving 1473 citations. Previous affiliations of Gerald M. McCobb include IBM.

Papers
More filters
Patent
04 Feb 2008
TL;DR: In this article, a method is described for ordering recognition results produced by an automatic speech recognition (ASR) engine for a multimodal application implemented with a grammar of the multimodAL application in the ASR engine.
Abstract: A method is described for ordering recognition results produced by an automatic speech recognition ( ASR ) engine for a multimodal application implemented with a grammar of the multimodalapplication in the ASR engine, with the multimodal application operating in a multimodalbrowser on a multimodal device supporting multiple modes of interaction including a voice mode and one or more non-voice modes, the multimodal application 10 operativelycoupled to the ASR enginethrough a VoiceXML interpreter. The method includes: receiving, in the VoiceXML interpreter from the multimodal application, a voice utterance; determining, by the VoiceXML interpreter using the ASR engine, a plurality of recognition results in dependence upon the voice utterance and the grammar; determining, by the VoiceXML interpreter according to semantic interpretation scripts of the grammar, a 15 weight for each recognition result; and sorting, bythe VoiceXML interpreter, the plurality of recognition results in dependence upon the weight for each recognition result.

189 citations

Patent
08 Dec 2004
TL;DR: In this paper, a multimodal browser for rendering a multi-modal document on an end system defining a host can include a visual browser component for rendering visual content, if any, of the multimodi-al document, and a voice browser component, which can determine which of a plurality of speech processing configuration is used by the host in rendering the voice-based content.
Abstract: A multimodal browser for rendering a multimodal document on an end system defining a host can include a visual browser component for rendering visual content, if any, of the multimodal document, and a voice browser component for rendering voice-based content, if any, of the multimodal document. The voice browser component can determine which of a plurality of speech processing configuration is used by the host in rendering the voice-based content. The determination can be based upon the resources of the host running the application. The determination also can be based upon a processing instruction contained in the application.

178 citations

Patent
21 Oct 2005
TL;DR: In this article, a method of building a mixed-initiative grammar can include receiving one or more conjoin phrases, wherein each conjoin phrase is associated with a selected one of the plurality of directed dialog grammars, and receiving a user input specifying a selected grammar generation technique.
Abstract: A method of building a mixed-initiative grammar can include receiving one or more conjoin phrases, wherein each conjoin phrase is associated with a selected one of the plurality of directed dialog grammars, and receiving a user input specifying a selected grammar generation technique. The mixed-initiative grammar can be automatically generated, in accordance with the selected grammar generation technique, such that the mixed-initiative grammar specifies an allowable ordering of sets when interpreting a user spoken utterance and whether duplicative phrases are allowable within the user spoken utterance.

122 citations

Patent
20 May 2003
TL;DR: In this paper, a method for enhancing voice interactions within a portable multimodal computing device using visual messages is presented, where the message is a prompt for the speech input and/or a confirmation of the input.
Abstract: A method for enhancing voice interactions within a portable multimodal computing device using visual messages. A multimodal interface can be provided that includes an audio interface and a visual interface. A speech input can then be received and a voice recognition task can be performed upon at least a portion of the speech input. At least one message within the multimodal interface can be visually presented, wherein the message is a prompt for the speech input and/or a confirmation of the speech input.

99 citations

Patent
09 Aug 2005
TL;DR: In this paper, a computer-implemented method and system are provided for filling a graphic-based form field in response to a speech utterance, which includes generating a grammar corresponding to the form field, the grammar being based on a user profile and comprising a semantic interpretation string.
Abstract: A computer-implemented method and system are provided for filling a graphic-based form field in response to a speech utterance. The computer-implemented method includes generating a grammar corresponding to the form field, the grammar being based on a user profile and comprising a semantic interpretation string. The method further includes creating an auto-fill event based upon the at least one grammar and responsive to the speech utterance, the auto-fill event causing the filling of the form field with data corresponding to the user profile. The system includes a grammar-generating module for generating a grammar corresponding to the form field, the grammar being based on a user profile and comprising a semantic interpretation string. The system also includes an event module for creating an auto-fill event based upon the at least one grammar and responsive to the speech utterance, the event causing the filling of the form field with data corresponding to the user profile.

95 citations


Cited by
More filters
Patent
11 Jan 2011
TL;DR: In this article, an intelligent automated assistant system engages with the user in an integrated, conversational manner using natural language dialog, and invokes external services when appropriate to obtain information or perform various actions.
Abstract: An intelligent automated assistant system engages with the user in an integrated, conversational manner using natural language dialog, and invokes external services when appropriate to obtain information or perform various actions. The system can be implemented using any of a number of different platforms, such as the web, email, smartphone, and the like, or any combination thereof. In one embodiment, the system is based on sets of interrelated domains and tasks, and employs additional functionally powered by external services with which the system can interact.

1,462 citations

Patent
28 Sep 2012
TL;DR: In this article, a virtual assistant uses context information to supplement natural language or gestural input from a user, which helps to clarify the user's intent and reduce the number of candidate interpretations of user's input, and reduces the need for the user to provide excessive clarification input.
Abstract: A virtual assistant uses context information to supplement natural language or gestural input from a user. Context helps to clarify the user's intent and to reduce the number of candidate interpretations of the user's input, and reduces the need for the user to provide excessive clarification input. Context can include any available information that is usable by the assistant to supplement explicit user input to constrain an information-processing problem and/or to personalize results. Context can be used to constrain solutions during various phases of processing, including, for example, speech recognition, natural language processing, task flow processing, and dialog generation.

593 citations

Patent
08 Sep 2006
TL;DR: In this paper, a method for building an automated assistant includes interfacing a service-oriented architecture that includes a plurality of remote services to an active ontology, where the active ontologies includes at least one active processing element that models a domain.
Abstract: A method and apparatus are provided for building an intelligent automated assistant. Embodiments of the present invention rely on the concept of “active ontologies” (e.g., execution environments constructed in an ontology-like manner) to build and run applications for use by intelligent automated assistants. In one specific embodiment, a method for building an automated assistant includes interfacing a service-oriented architecture that includes a plurality of remote services to an active ontology, where the active ontology includes at least one active processing element that models a domain. At least one of the remote services is then registered for use in the domain.

389 citations

Patent
Aram Lindahl1
24 May 2012
TL;DR: In this paper, an electronic device may capture a voice command from a user and store contextual information about the state of the electronic device when the voice command is received, such as a desktop computer or a remote server.
Abstract: An electronic device may capture a voice command from a user. The electronic device may store contextual information about the state of the electronic device when the voice command is received. The electronic device may transmit the voice command and the contextual information to computing equipment such as a desktop computer or a remote server. The computing equipment may perform a speech recognition operation on the voice command and may process the contextual information. The computing equipment may respond to the voice command. The computing equipment may also transmit information to the electronic device that allows the electronic device to respond to the voice command.

385 citations

Patent
30 Sep 2011
TL;DR: In this article, the authors present a method for automatically determining whether a digital assistant application has been separately invoked by a user without regard to whether a user has separately invoked the application.
Abstract: The method includes automatically, without user input and without regard to whether a digital assistant application has been separately invoked by a user, determining that the electronic device is in a vehicle. In some implementations, determining that the electronic device is in a vehicle comprises detecting that the electronic device is in communication with the vehicle (e.g., via a wired or wireless communication techniques and/or protocols). The method also includes, responsive to the determining, invoking a listening mode of a virtual assistant implemented by the electronic device. In some implementations, the method also includes limiting the ability of a user to view visual output presented by the electronic device, provide typed input to the electronic device, and the like.

367 citations