About: VoiceXML is a(n) research topic. Over the lifetime, 457 publication(s) have been published within this topic receiving 6650 citation(s). The topic is also known as: VXML & Voice Extensible Markup Language.
26 Sep 2003-
Abstract: Call routing systems and methods are provided. A particular routing method comprises decoding a message based on an incoming call to determine whether a live agent is required. When a live agent is not required, a destination interactive voice response (IVR) application is identified, a determination of whether the destination IVR application is VoiceXML capable is made, and the incoming call is sent to the destination IVR application when the destination IVR application is VoiceXML capable. When the destination IVR application is not VoiceXML capable, a determination of whether the destination IVR application is capable of supporting an external data interface is made, and incoming call session data is routed to the destination IVR application when the destination IVR application is capable of supporting the external data interface. An audio file including one or more dual tone multi-frequency (DTMF) commands based on the incoming call session data is constructed and sent to the destination IVR application when the destination IVR application is not VoiceXML capable and is not capable of supporting an external data interface.
18 Apr 2001-
Abstract: A new application programming language is provided which is based on user interaction with any device which a user is employing to access any type of information. The new language is referred to herein as a “Conversational Markup Language (CML). In a preferred embodiment, CML is a high level XML based language for representing “dialogs” or “conversations” the user will have with any given computing device. For example, interaction may comprise, but is not limited to, visual based (text and graphical) user interaction and speech based user interaction. Such a language allows application authors to program applications using interaction-based elements referred to herein as “conversational gestures.” The present invention also provides for various embodiments of a multimodal browser capable of supporting the features of CML in accordance with various modality specific representations, e.g., HTML based graphical user interface (GUI) browser, VoiceXML based speech browser, etc.
23 Mar 2007-
Abstract: Methods, apparatus, and products are disclosed for supporting multi-lingual user interaction with a multimodal application, the application including a plurality of VoiceXML dialogs, each dialog characterized by a particular language, supporting multi-lingual user interaction implemented with a plurality of speech engines, each speech engine having a grammar and characterized by a language corresponding to one of the dialogs, with the application operating on a multimodal device supporting multiple modes of interaction including a voice mode and one or more non-voice modes, the application operatively coupled to the speech engines through a VoiceXML interpreter, the VoiceXML interpreter: receiving a voice utterance from a user; determining in parallel, using the speech engines, recognition results for each dialog in dependence upon the voice utterance and the grammar for each speech engine; administering the recognition results for the dialogs; and selecting a language for user interaction in dependence upon the administered recognition results.
04 Feb 2008-
Abstract: A method is described for ordering recognition results produced by an automatic speech recognition ( ASR ) engine for a multimodal application implemented with a grammar of the multimodalapplication in the ASR engine, with the multimodal application operating in a multimodalbrowser on a multimodal device supporting multiple modes of interaction including a voice mode and one or more non-voice modes, the multimodal application 10 operativelycoupled to the ASR enginethrough a VoiceXML interpreter. The method includes: receiving, in the VoiceXML interpreter from the multimodal application, a voice utterance; determining, by the VoiceXML interpreter using the ASR engine, a plurality of recognition results in dependence upon the voice utterance and the grammar; determining, by the VoiceXML interpreter according to semantic interpretation scripts of the grammar, a 15 weight for each recognition result; and sorting, bythe VoiceXML interpreter, the plurality of recognition results in dependence upon the weight for each recognition result.
03 Oct 2006-
Abstract: A multi-media telephone system capable of presenting multi-media content to users defined by structured script files incorporating markup tags, such as HTML, XML, SMIL or VoiceXML tags. The script files form “homepage” presentations displayed on the telephone stationset. The script files may be transmitted from called parties to calling parties at the time a telephone connection is first established, or transmitted at the request of either party thereafter, and are executed by a processor in the receiving stationset (such as a cellular telephone), providing both visual displays and audio output in an interactive interface which can be used to provide additional information or services from a menu of displayed or spoken option prompts.