会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 1. 发明申请
    • Document Session Replay for Multimodal Applications
    • 多模式应用程序的文档会话重放
    • US20080208587A1
    • 2008-08-28
    • US11678830
    • 2007-02-26
    • Shay Ben-DavidCharles W. CrossMarc T. White
    • Shay Ben-DavidCharles W. CrossMarc T. White
    • G10L11/00
    • G10L15/26G10L15/22
    • Methods, apparatus, and computer program products are described for document session replay for multimodal applications. including identifying, by a multimodal browser in dependence upon a log produced by a Form Interpretation Algorithm (‘FIA’) during a previous document session with a user, a speech prompt provided by a multimodal application in the previous document session; identifying, by a multimodal browser in replay mode in dependence upon the log, a response to the prompt provided by a user of the multimodal application in the previous document session; retrieving, by the multimodal browser in dependence upon the log, an X+V page of the multimodal application associated with the speech prompt and the response; rendering, by the multimodal browser, the visual elements of the retrieved X+V page; replaying, by the multimodal browser, the speech prompt; and replaying, by a multimodal browser, the response.
    • 描述用于多模式应用的文档会话重放的方法,装置和计算机程序产品。 包括通过多模式浏览器根据在与用户的先前文档会话期间由表单解释算法(“FIA”)产生的日志来识别由多模式应用在先前文档会话中提供的语音提示; 通过多模式浏览器根据日志在重放模式下识别由先前文档会话中的多模式应用的用户提供的提示的响应; 通过多模式浏览器根据日志检索与语音提示相关联的多模式应用的X + V页面和响应; 通过多模式浏览器呈现检索到的X + V页面的视觉元素; 由多模式浏览器重播演讲提示; 并通过多模式浏览器重播响应。
    • 2. 发明申请
    • DYNAMICALLY DEFINING A VOICEXML GRAMMAR IN AN X+V PAGE OF A MULTIMODAL APPLICATION
    • 在多模式应用的X + V页中动态定义VOICEXML GRAMMAR
    • US20080195393A1
    • 2008-08-14
    • US11673675
    • 2007-02-12
    • Charles W. CrossHilary A. PikeLisa A. SeacatMarc T. White
    • Charles W. CrossHilary A. PikeLisa A. SeacatMarc T. White
    • G10L13/08
    • G10L15/193
    • Dynamically defining a VoiceXML grammar of a multimodal application, implemented with the multimodal application operating on a multimodal device supporting multiple modes of interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to a VoiceXML interpreter, and the method includes loading the X+V page by the multimodal application, from a web server into the multimodal device for execution, the X+V page including one or more VoiceXML grammars in one or more VoiceXML dialogs, including at least one in-line grammar that is declared but undefined; retrieving by the multimodal application a grammar definition for the in-line grammar from the web server without reloading the X+V page; and defining by the multimodal application the in-line grammar with the retrieved grammar definition before executing the VoiceXML dialog containing the in-line grammar.
    • 动态地定义多模式应用的VoiceXML语法,该多模式应用程序在多模式设备上运行,多模式设备支持包括语音模式和一种或多种非语音模式的多种交互模式,多模式应用可操作地耦合到VoiceXML解释器,并且 方法包括将多模式应用程序将X + V页面从Web服务器加载到多模式设备中以执行,X + V页面包括一个或多个VoiceXML对话框中的一个或多个VoiceXML语法,包括至少一个在线语法 这是宣布但未定义; 由多模式应用程序检索来自Web服务器的在线语法的语法定义,而不重新加载X + V页面; 并且在执行包含内联语法的VoiceXML对话之前,由多模式应用程序定义具有所检索的语法定义的在线语法。
    • 3. 发明申请
    • Disambiguating A Speech Recognition Grammar In A Multimodal Application
    • 在多模式应用中消除语音识别语法
    • US20080208590A1
    • 2008-08-28
    • US11679274
    • 2007-02-27
    • Charles W. CrossMarc T. White
    • Charles W. CrossMarc T. White
    • G10L21/00
    • G10L15/22
    • Disambiguating a speech recognition grammar in a multimodal application, the multimodal application including voice activated hyperlinks, the voice activated hyperlinks voice enabled by a speech recognition grammar characterized by ambiguous terminal grammar elements, including maintaining by the multimodal browser a record of visibility of each voice activated hyperlink, the record of visibility including current visibility and past visibility on a display of the multimodal device of each voice activated hyperlink, the record of visibility further including an ordinal indication, for each voice activated hyperlink scrolled off display, of the sequence in which each such voice activated hyperlink was scrolled off display; recognizing by the multimodal browser speech from a user matching an ambiguous terminal element of the speech recognition grammar; selecting by the multimodal browser a voice activated hyperlink for activation, the selecting carried out in dependence upon the recognized speech and the record of visibility.
    • 在多模式应用中消除语音识别语法,多模式应用包括语音激活的超链接,由语义识别语法启用的语音激活的超链接语音,其特征在于模糊的终端语法元素,包括由多模式浏览器保持激活的每个语音的可见性记录 超链接,每个声音激活的超链接的多模式设备的显示器上的当前可见性和过去可视性的可见性记录,对于每个音频激活超链接滚动显示的进一步包括顺序指示的可见性记录,其中每个 这样的声音激活超链接被滚动显示; 通过用户匹配语音识别语法的模糊终端元素的多模式浏览器语音识别; 由多模式浏览器选择用于激活的语音激活超链接,根据识别的语音和可见性的记录进行选择。
    • 4. 发明申请
    • Speech-Enabled Content Navigation And Control Of A Distributed Multimodal Browser
    • 分布式多模态浏览器的语音启用内容导航和控制
    • US20080255851A1
    • 2008-10-16
    • US11734445
    • 2007-04-12
    • Soonthorn AtivanichayaphongCharles W. CrossGerald M. McCobb
    • Soonthorn AtivanichayaphongCharles W. CrossGerald M. McCobb
    • G10L21/00
    • G10L15/265G06F3/16G06F3/167G10L15/26
    • Speech-enabled content navigation and control of a distributed multimodal browser is disclosed, the browser providing an execution environment for a multimodal application, the browser including a graphical user agent (‘GUA’) and a voice user agent (‘VUA’), the GUA operating on a multimodal device, the VUA operating on a voice server, that includes: transmitting, by the GUA, a link message to the VUA, the link message specifying voice commands that control the browser and an event corresponding to each voice command; receiving, by the GUA, a voice utterance from a user, the voice utterance specifying a particular voice command; transmitting, by the GUA, the voice utterance to the VUA for speech recognition by the VUA; receiving, by the GUA, an event message from the VUA, the event message specifying a particular event corresponding to the particular voice command; and controlling, by the GUA, the browser in dependence upon the particular event.
    • 公开了一种分布式多模式浏览器的语音启用内容导航和控制,浏览器为多模式应用提供执行环境,浏览器包括图形用户代理(“GUA”)和语音用户代理(“VUA”), GUA在多模式设备上操作,VUA在语音服务器上操作,其包括:由GUA向VUA发送链接消息,指定控制浏览器的语音命令的链接消息和与每个语音命令相对应的事件; 由GUA接收来自用户的语音发音,指定特定语音命令的语音话语; 通过GUA向VUA发送语音识别语音识别语音; 由GUA接收来自VUA的事件消息,事件消息指定与特定语音命令对应的特定事件; 并由GUA根据特定事件控制浏览器。
    • 5. 发明申请
    • Configuring A Speech Engine For A Multimodal Application Based On Location
    • 配置基于位置的多模态应用的语音引擎
    • US20080208592A1
    • 2008-08-28
    • US11679297
    • 2007-02-27
    • Charles W. CrossIgor R. Jablokov
    • Charles W. CrossIgor R. Jablokov
    • G10L21/00
    • G10L15/24
    • Methods, apparatus, and products are disclosed for configuring a speech engine for a multimodal application based on location. The multimodal application operates on a multimodal device supporting multiple modes of user interaction with the multimodal application. The multimodal application is operatively coupled to a speech engine. Configuring a speech engine for a multimodal application based on location includes: receiving a location change notification in a location change monitor from a device location manager, the location change notification specifying a current location of the multimodal device; identifying, by the location change monitor, location-based configuration parameters for the speech engine in dependence upon the current location of the multimodal device, the location-based configuration parameters specifying a configuration for the speech engine at the current location; and updating, by the location change monitor, a current configuration for the speech engine according to the identified location-based configuration parameters.
    • 公开了基于位置配置用于多模式应用的语音引擎的方法,装置和产品。 多模式应用程序在支持多模式用户与多模态应用程序交互的多模式设备上运行。 多模式应用可操作地耦合到语音引擎。 基于位置为多模式应用配置语音引擎包括:从设备位置管理器在位置变化监视器中接收位置变化通知,所述位置变化通知指定多模态设备的当前位置; 根据所述多模式设备的当前位置,由所述位置变化监视器识别所述语音引擎的基于位置的配置参数,所述基于位置的配置参数指定所述语音引擎在当前位置的配置; 以及根据所识别的基于位置的配置参数,由所述位置变化监视器更新所述语音引擎的当前配置。
    • 6. 发明申请
    • Ordering Recognition Results Produced By An Automatic Speech Recognition Engine For A Multimodal Application
    • 由多模式应用程序自动语音识别引擎生成的订购识别结果
    • US20080208585A1
    • 2008-08-28
    • US11679284
    • 2007-02-27
    • Soonthorn AtivanichayaphongCharles W. CrossIgor R. JablokovGerald McCobb
    • Soonthorn AtivanichayaphongCharles W. CrossIgor R. JablokovGerald McCobb
    • G10L21/00
    • H04M3/4936G10L15/1815G10L15/19
    • Ordering recognition results produced by an automatic speech recognition (‘ASR’) engine for a multimodal application implemented with a grammar of the multimodal application in the ASR engine, with the multimodal application operating in a multimodal browser on a multimodal device supporting multiple modes of interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to the ASR engine through a VoiceXML interpreter, includes: receiving, in the VoiceXML interpreter from the multimodal application, a voice utterance; determining, by the VoiceXML interpreter using the ASR engine, a plurality of recognition results in dependence upon the voice utterance and the grammar; determining, by the VoiceXML interpreter according to semantic interpretation scripts of the grammar, a weight for each recognition result; and sorting, by the VoiceXML interpreter, the plurality of recognition results in dependence upon the weight for each recognition result.
    • 通过使用ASR引擎中的多模式应用程序的语法实现的多模式应用程序的自动语音识别(“ASR”)引擎进行的订购识别结果,多模式应用程序在支持多种交互模式的多模式设备的多模式浏览器中运行 包括语音模式和一个或多个非语音模式,通过VoiceXML解释器可操作地耦合到ASR引擎的多模式应用包括:在来自多模式应用的VoiceXML解释器中接收语音话语; 通过使用ASR引擎的VoiceXML解释器,根据语音发音和语法来确定多个识别结果; 通过VoiceXML解释器根据语法的语义解释脚本确定每个识别结果的权重; 以及由VoiceXML解释器根据每个识别结果的权重对多个识别结果进行排序。
    • 8. 发明授权
    • System for designing an application program to be independent of I/O
devices by utilizing application name, system name, and predetermined
hardware specific parameters of data objects
    • 通过利用应用程序名称,系统名称和数据对象的预定硬件特定参数来设计应用程序独立于I / O设备的系统
    • US5517635A
    • 1996-05-14
    • US991882
    • 1992-12-17
    • Charles W. CrossMark E. Trumbo
    • Charles W. CrossMark E. Trumbo
    • G06F3/00G06F3/038G06F9/44G06F5/00
    • G06F3/038G06F3/005G06F8/24
    • A system and a method for inputting/outputting device specific data that does not bind I/O device specific information to the application program until runtime is provided. This allows an application program developer to support a new I/O device without redesigning the application program to accommodate the new information and then compiling the new version of the application program. The system includes an Object Input/Output Manager (OIOM), a device dictionary, a plurality of I/O device modules, a plurality of I/O devices, and an I/O channel. The OIOM controls the overall flow of the system. The device dictionary and the I/O device modules contain I/O device specific information that is not known to the application program at the time it is compiled. When an application program needs to input or output device specific data, the user is presented with a list of I/O devices and selects one. After receiving the user's selection, the application program commands the OIOM to input or output the device specific data. The OIOM then gets the system name of the I/O device, loads the corresponding I/O device module, adds the I/O device specific information to the data object, and commands the I/O device module to input or output the data object. After the data object has been input or output, control passes back to the OIOM and finally to the application program.
    • 一种用于输入/输出设备特定数据的系统和方法,所述设备特定数据在提供运行时之前不将I / O设备特定信息绑定到应用程序。 这允许应用程序开发人员支持新的I / O设备,而无需重新设计应用程序来适应新信息,然后编译应用程序的新版本。 该系统包括对象输入/输出管理器(OIOM),设备字典,多个I / O设备模块,多个I / O设备和I / O通道。 监督厅控制系统的总体流量。 设备字典和I / O设备模块包含应用程序在编译时不知道的I / O设备特定信息。 当应用程序需要输入或输出设备特定数据时,会向用户显示I / O设备列表并选择一个。 在接收到用户的选择后,应用程序命令OIOM输入或输出设备特定的数据。 然后,OIOM获取I / O设备的系统名称,加载相应的I / O设备模块,将I / O设备特定信息添加到数据对象,并命令I / O设备模块输入或输出数据 目的。 数据对象输入或输出后,控制权返回到OIOM,最后到应用程序。
    • 10. 发明授权
    • Enabling voice selection of user preferences
    • 启用用户偏好设置的语音选择
    • US09083798B2
    • 2015-07-14
    • US11022464
    • 2004-12-22
    • Charles W. CrossYan Li
    • Charles W. CrossYan Li
    • G10L15/00G10L21/00H04M3/493G10L15/26G10L15/19
    • H04M3/4938G10L15/19G10L15/26H04M2201/40
    • A method, system and apparatus for voice enabling a user preference interface in a multimodal content browser. A method for voice enabling a user preference interface in a multimodal content browser can include matching voice input to a bound command in a speech grammar and invoking logic in the user preference interface consistent with the bound command in the speech grammar. The matching step can include comparing voice input to entries in a markup language specified speech grammar and locating the bound command in the specified speech grammar based upon the comparison. In this regard, the method further can include identifying a variable in the bound command, looking up the variable in a table, retrieving a corresponding parameter for the variable from the table, and replacing the variable with the corresponding parameter in the bound command.
    • 一种用于在多模式内容浏览器中启用用户偏好接口的语音的方法,系统和装置。 一种用于在多模式内容浏览器中启用用户偏好接口的语音的方法可以包括将语音输入与语音语法中的绑定命令相匹配,并且在用户偏好接口中调用与语音语法中的绑定命令一致的逻辑。 匹配步骤可以包括将语音输入与指定语言语法的标记语言中的条目进行比较,并且基于比较来定位指定语音语法中的绑定命令。 在这方面,该方法还可以包括标识绑定命令中的变量,查找表中的变量,从表中检索变量的相应参数,以及用该绑定命令中的相应参数替换该变量。