会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 1. 发明申请
    • DYNAMICALLY DEFINING A VOICEXML GRAMMAR IN AN X+V PAGE OF A MULTIMODAL APPLICATION
    • 在多模式应用的X + V页中动态定义VOICEXML GRAMMAR
    • US20080195393A1
    • 2008-08-14
    • US11673675
    • 2007-02-12
    • Charles W. CrossHilary A. PikeLisa A. SeacatMarc T. White
    • Charles W. CrossHilary A. PikeLisa A. SeacatMarc T. White
    • G10L13/08
    • G10L15/193
    • Dynamically defining a VoiceXML grammar of a multimodal application, implemented with the multimodal application operating on a multimodal device supporting multiple modes of interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to a VoiceXML interpreter, and the method includes loading the X+V page by the multimodal application, from a web server into the multimodal device for execution, the X+V page including one or more VoiceXML grammars in one or more VoiceXML dialogs, including at least one in-line grammar that is declared but undefined; retrieving by the multimodal application a grammar definition for the in-line grammar from the web server without reloading the X+V page; and defining by the multimodal application the in-line grammar with the retrieved grammar definition before executing the VoiceXML dialog containing the in-line grammar.
    • 动态地定义多模式应用的VoiceXML语法,该多模式应用程序在多模式设备上运行,多模式设备支持包括语音模式和一种或多种非语音模式的多种交互模式,多模式应用可操作地耦合到VoiceXML解释器,并且 方法包括将多模式应用程序将X + V页面从Web服务器加载到多模式设备中以执行,X + V页面包括一个或多个VoiceXML对话框中的一个或多个VoiceXML语法,包括至少一个在线语法 这是宣布但未定义; 由多模式应用程序检索来自Web服务器的在线语法的语法定义,而不重新加载X + V页面; 并且在执行包含内联语法的VoiceXML对话之前,由多模式应用程序定义具有所检索的语法定义的在线语法。
    • 6. 发明申请
    • Document Session Replay for Multimodal Applications
    • 多模式应用程序的文档会话重放
    • US20080208587A1
    • 2008-08-28
    • US11678830
    • 2007-02-26
    • Shay Ben-DavidCharles W. CrossMarc T. White
    • Shay Ben-DavidCharles W. CrossMarc T. White
    • G10L11/00
    • G10L15/26G10L15/22
    • Methods, apparatus, and computer program products are described for document session replay for multimodal applications. including identifying, by a multimodal browser in dependence upon a log produced by a Form Interpretation Algorithm (‘FIA’) during a previous document session with a user, a speech prompt provided by a multimodal application in the previous document session; identifying, by a multimodal browser in replay mode in dependence upon the log, a response to the prompt provided by a user of the multimodal application in the previous document session; retrieving, by the multimodal browser in dependence upon the log, an X+V page of the multimodal application associated with the speech prompt and the response; rendering, by the multimodal browser, the visual elements of the retrieved X+V page; replaying, by the multimodal browser, the speech prompt; and replaying, by a multimodal browser, the response.
    • 描述用于多模式应用的文档会话重放的方法,装置和计算机程序产品。 包括通过多模式浏览器根据在与用户的先前文档会话期间由表单解释算法(“FIA”)产生的日志来识别由多模式应用在先前文档会话中提供的语音提示; 通过多模式浏览器根据日志在重放模式下识别由先前文档会话中的多模式应用的用户提供的提示的响应; 通过多模式浏览器根据日志检索与语音提示相关联的多模式应用的X + V页面和响应; 通过多模式浏览器呈现检索到的X + V页面的视觉元素; 由多模式浏览器重播演讲提示; 并通过多模式浏览器重播响应。
    • 7. 发明申请
    • Disambiguating A Speech Recognition Grammar In A Multimodal Application
    • 在多模式应用中消除语音识别语法
    • US20080208590A1
    • 2008-08-28
    • US11679274
    • 2007-02-27
    • Charles W. CrossMarc T. White
    • Charles W. CrossMarc T. White
    • G10L21/00
    • G10L15/22
    • Disambiguating a speech recognition grammar in a multimodal application, the multimodal application including voice activated hyperlinks, the voice activated hyperlinks voice enabled by a speech recognition grammar characterized by ambiguous terminal grammar elements, including maintaining by the multimodal browser a record of visibility of each voice activated hyperlink, the record of visibility including current visibility and past visibility on a display of the multimodal device of each voice activated hyperlink, the record of visibility further including an ordinal indication, for each voice activated hyperlink scrolled off display, of the sequence in which each such voice activated hyperlink was scrolled off display; recognizing by the multimodal browser speech from a user matching an ambiguous terminal element of the speech recognition grammar; selecting by the multimodal browser a voice activated hyperlink for activation, the selecting carried out in dependence upon the recognized speech and the record of visibility.
    • 在多模式应用中消除语音识别语法,多模式应用包括语音激活的超链接,由语义识别语法启用的语音激活的超链接语音,其特征在于模糊的终端语法元素,包括由多模式浏览器保持激活的每个语音的可见性记录 超链接,每个声音激活的超链接的多模式设备的显示器上的当前可见性和过去可视性的可见性记录,对于每个音频激活超链接滚动显示的进一步包括顺序指示的可见性记录,其中每个 这样的声音激活超链接被滚动显示; 通过用户匹配语音识别语法的模糊终端元素的多模式浏览器语音识别; 由多模式浏览器选择用于激活的语音激活超链接,根据识别的语音和可见性的记录进行选择。
    • 8. 发明申请
    • Speech-Enabled Content Navigation And Control Of A Distributed Multimodal Browser
    • 分布式多模态浏览器的语音启用内容导航和控制
    • US20080255851A1
    • 2008-10-16
    • US11734445
    • 2007-04-12
    • Soonthorn AtivanichayaphongCharles W. CrossGerald M. McCobb
    • Soonthorn AtivanichayaphongCharles W. CrossGerald M. McCobb
    • G10L21/00
    • G10L15/265G06F3/16G06F3/167G10L15/26
    • Speech-enabled content navigation and control of a distributed multimodal browser is disclosed, the browser providing an execution environment for a multimodal application, the browser including a graphical user agent (‘GUA’) and a voice user agent (‘VUA’), the GUA operating on a multimodal device, the VUA operating on a voice server, that includes: transmitting, by the GUA, a link message to the VUA, the link message specifying voice commands that control the browser and an event corresponding to each voice command; receiving, by the GUA, a voice utterance from a user, the voice utterance specifying a particular voice command; transmitting, by the GUA, the voice utterance to the VUA for speech recognition by the VUA; receiving, by the GUA, an event message from the VUA, the event message specifying a particular event corresponding to the particular voice command; and controlling, by the GUA, the browser in dependence upon the particular event.
    • 公开了一种分布式多模式浏览器的语音启用内容导航和控制,浏览器为多模式应用提供执行环境,浏览器包括图形用户代理(“GUA”)和语音用户代理(“VUA”), GUA在多模式设备上操作,VUA在语音服务器上操作,其包括:由GUA向VUA发送链接消息,指定控制浏览器的语音命令的链接消息和与每个语音命令相对应的事件; 由GUA接收来自用户的语音发音,指定特定语音命令的语音话语; 通过GUA向VUA发送语音识别语音识别语音; 由GUA接收来自VUA的事件消息,事件消息指定与特定语音命令对应的特定事件; 并由GUA根据特定事件控制浏览器。
    • 9. 发明申请
    • Configuring A Speech Engine For A Multimodal Application Based On Location
    • 配置基于位置的多模态应用的语音引擎
    • US20080208592A1
    • 2008-08-28
    • US11679297
    • 2007-02-27
    • Charles W. CrossIgor R. Jablokov
    • Charles W. CrossIgor R. Jablokov
    • G10L21/00
    • G10L15/24
    • Methods, apparatus, and products are disclosed for configuring a speech engine for a multimodal application based on location. The multimodal application operates on a multimodal device supporting multiple modes of user interaction with the multimodal application. The multimodal application is operatively coupled to a speech engine. Configuring a speech engine for a multimodal application based on location includes: receiving a location change notification in a location change monitor from a device location manager, the location change notification specifying a current location of the multimodal device; identifying, by the location change monitor, location-based configuration parameters for the speech engine in dependence upon the current location of the multimodal device, the location-based configuration parameters specifying a configuration for the speech engine at the current location; and updating, by the location change monitor, a current configuration for the speech engine according to the identified location-based configuration parameters.
    • 公开了基于位置配置用于多模式应用的语音引擎的方法,装置和产品。 多模式应用程序在支持多模式用户与多模态应用程序交互的多模式设备上运行。 多模式应用可操作地耦合到语音引擎。 基于位置为多模式应用配置语音引擎包括:从设备位置管理器在位置变化监视器中接收位置变化通知,所述位置变化通知指定多模态设备的当前位置; 根据所述多模式设备的当前位置,由所述位置变化监视器识别所述语音引擎的基于位置的配置参数,所述基于位置的配置参数指定所述语音引擎在当前位置的配置; 以及根据所识别的基于位置的配置参数,由所述位置变化监视器更新所述语音引擎的当前配置。
    • 10. 发明申请
    • Ordering Recognition Results Produced By An Automatic Speech Recognition Engine For A Multimodal Application
    • 由多模式应用程序自动语音识别引擎生成的订购识别结果
    • US20080208585A1
    • 2008-08-28
    • US11679284
    • 2007-02-27
    • Soonthorn AtivanichayaphongCharles W. CrossIgor R. JablokovGerald McCobb
    • Soonthorn AtivanichayaphongCharles W. CrossIgor R. JablokovGerald McCobb
    • G10L21/00
    • H04M3/4936G10L15/1815G10L15/19
    • Ordering recognition results produced by an automatic speech recognition (‘ASR’) engine for a multimodal application implemented with a grammar of the multimodal application in the ASR engine, with the multimodal application operating in a multimodal browser on a multimodal device supporting multiple modes of interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to the ASR engine through a VoiceXML interpreter, includes: receiving, in the VoiceXML interpreter from the multimodal application, a voice utterance; determining, by the VoiceXML interpreter using the ASR engine, a plurality of recognition results in dependence upon the voice utterance and the grammar; determining, by the VoiceXML interpreter according to semantic interpretation scripts of the grammar, a weight for each recognition result; and sorting, by the VoiceXML interpreter, the plurality of recognition results in dependence upon the weight for each recognition result.
    • 通过使用ASR引擎中的多模式应用程序的语法实现的多模式应用程序的自动语音识别(“ASR”)引擎进行的订购识别结果,多模式应用程序在支持多种交互模式的多模式设备的多模式浏览器中运行 包括语音模式和一个或多个非语音模式,通过VoiceXML解释器可操作地耦合到ASR引擎的多模式应用包括:在来自多模式应用的VoiceXML解释器中接收语音话语; 通过使用ASR引擎的VoiceXML解释器,根据语音发音和语法来确定多个识别结果; 通过VoiceXML解释器根据语法的语义解释脚本确定每个识别结果的权重; 以及由VoiceXML解释器根据每个识别结果的权重对多个识别结果进行排序。