会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 53. 发明申请
    • Method and apparatus for improved speech recognition with supplementary information
    • 用于通过补充信息改进语音识别的方法和装置
    • US20050049860A1
    • 2005-03-03
    • US10652146
    • 2003-08-29
    • Jean-Claude JunquaRoland KuhnMatteo ContoliniRathinavelu Chengalvarayan
    • Jean-Claude JunquaRoland KuhnMatteo ContoliniRathinavelu Chengalvarayan
    • G10L15/08G10L15/10G10L15/22H04M1/27G10L15/00
    • H04M1/271G10L15/08G10L15/10G10L15/22
    • A method for improving recognition results of a speech recognizer uses supplementary information to confirm recognition results. A user inputs speech to a speech recognizer. The speech recognizer resides on a mobile device or on a server at a remote location. The speech recognizer determines a recognition result based on the input speech. A confidence measure is calculated for the recognition result. If the confidence measure is below a threshold, the user is prompted for supplementary data. The supplementary data is determined dynamically based on ambiguities between the input speech and the recognition result, wherein the supplementary data will distinguish the input speech over potential incorrect results. The supplementary data may be a subset of alphanumeric characters that comprise the input speech, or other data associated with a desired result, such as an area code or location. The user may provide the supplementary data verbally, or manually using a keypad, touchpad, touchscreen, or stylus pen.
    • 用于改善语音识别器的识别结果的方法使用补充信息来确认识别结果。 用户向语音识别器输入语音。 语音识别器驻留在移动设备或远程位置的服务器上。 语音识别器基于输入语音来确定识别结果。 计算识别结果的置信度量。 如果置信度量值低于阈值,则会提示用户提供补充数据。 基于输入语音和识别结果之间的模糊度来动态地确定补充数据,其中补充数据将通过潜在的不正确结果区分输入语音。 补充数据可以是组成输入语音的字母数字字符的子集,或与期望结果相关联的其他数据,例如区域代码或位置。 用户可以口头提供补充数据,或者使用键盘,触摸板,触摸屏或触控笔手动提供补充数据。
    • 54. 发明申请
    • Speech data mining for call center management
    • 语音数据挖掘用于呼叫中心管理
    • US20050010411A1
    • 2005-01-13
    • US10616006
    • 2003-07-09
    • Luca RigazioPatrick NguyenJean-Claude JunquaRobert Boman
    • Luca RigazioPatrick NguyenJean-Claude JunquaRobert Boman
    • G10L15/26G10L17/00G10L15/00
    • G10L15/26G10L17/00
    • A speech data mining system for use in generating a rich transcription having utility in call center management includes a speech differentiation module differentiating between speech of interacting speakers, and a speech recognition module improving automatic recognition of speech of one speaker based on interaction with another speaker employed as a reference speaker. A transcript generation module generates a rich transcript based on recognized speech of the speakers. Focused, interactive language models improve recognition of a customer on a low quality channel using context extracted from speech of a call center operator on a high quality channel with a speech model adapted to the operator. Mined speech data includes number of interaction turns, customer frustration phrases, operator polity, interruptions, and/or contexts extracted from speech recognition results, such as topics, complaints, solutions, and resolutions. Mined speech data is useful in call center and/or product or service quality management.
    • 用于产生在呼叫中心管理中具有效用的丰富录音的语音数据挖掘系统包括区分交互式扬声器的语音的语音区分模块和改善一个扬声器的语音的自动识别的语音识别模块, 作为参考发言人。 转录本生成模块基于扬声器的识别语音生成丰富的录音。 专注的交互式语言模型通过使用适合于操作员的语音模型,在高质量频道上从呼叫中心运营商的语音提取的上下文,改善对低质量信道上客户的识别。 挖掘的语音数据包括从诸如主题,投诉,解决方案和分辨率的语音识别结果中提取的交互轮廓数量,客户沮丧短语,运营商政治,中断和/或上下文。 挖掘的语音数据在呼叫中心和/或产品或服务质量管理中是有用的。
    • 57. 发明授权
    • Context-dependent acoustic models for medium and large vocabulary speech recognition with eigenvoice training
    • 用于具有本征语音训练的中大词汇语音识别的背景相关声学模型
    • US06571208B1
    • 2003-05-27
    • US09450392
    • 1999-11-29
    • Roland KuhnJean-Claude JunquaMatteo Contolini
    • Roland KuhnJean-Claude JunquaMatteo Contolini
    • G01L1700
    • G10L15/07
    • A reduced dimensionality eigenvoice analytical technique is used during training to develop context-dependent acoustic models for allophones. The eigenvoice technique is also used during run time upon the speech of a new speaker. The technique removes individual speaker idiosyncrasies, to produce more universally applicable and robust allophone models. In one embodiment the eigenvoice technique is used to identify the centroid of each speaker, which may then be “subtracted out” of the recognition equation. In another embodiment maximum likelihood estimation techniques are used to develop common decision tree frameworks that may be shared across all speakers when constructing the eigenvoice representation of speaker space.
    • 在训练期间使用减小的维度本征语音分析技术来开发用于异音素的上下文相关的声学模型。 特定语音技术在运行时也用于新演讲者的演讲。 该技术可以消除单个扬声器的特性,从而产生更普遍适用和强大的异音模型。 在一个实施例中,本征语音技术用于识别每个说话者的质心,然后可以将其“减去”识别方程。 在另一个实施例中,最大似然估计技术用于开发在构建扬声器空间的本征声表示时可以在所有扬声器之间共享的共同决策树框架。
    • 58. 发明授权
    • Automatic search of audio channels by matching viewer-spoken words against closed-caption/audio content for interactive television
    • 通过将观众口语与针对交互式电视的封闭字幕/音频内容相匹配来自动搜索音频频道
    • US06480819B1
    • 2002-11-12
    • US09258115
    • 1999-02-25
    • Robert BomanJean-Claude Junqua
    • Robert BomanJean-Claude Junqua
    • G06F1727
    • G10L15/26G10L15/1815
    • A method and apparatus is provided to enable a user watching and/or listening to a program to search for new information in the stream of a telecommunications data. The apparatus includes a voice recognition system that recognizes the user's request and causes a search to be performed in the long stream of data of at least one other telecommunication channel. The system includes a storage device for storing and processing the request. Upon recognition of the request, the incoming signal or signals are scanned for matches with the request. Upon finding the match between the request and the incoming signal, information related to the data is brought to the viewer's attention. This can be accomplished by either changing the viewer's station or by bringing in a split screen display forward into the display.
    • 提供了一种方法和装置,用于使用户能够观看和/或收听节目以搜索电信数据流中的新信息。 该装置包括语音识别系统,其识别用户的请求并且使得在至少另一个电信信道的长流数据中执行搜索。 该系统包括用于存储和处理该请求的存储装置。 一旦识别到请求,就会扫描输入信号或与该请求匹配的信号。 在找到请求和输入信号之间的匹配时,与数据相关的信息被引起观众的注意。 这可以通过改变观众的电台或将分屏显示向前推入显示器来实现。
    • 59. 发明授权
    • System for identifying and adapting a TV-user profile by means of speech technology
    • 通过语音技术识别和调整电视用户资料的系统
    • US06415257B1
    • 2002-07-02
    • US09383797
    • 1999-08-26
    • Jean-Claude JunquaRoland KuhnTony DavisYi ZhaoWeiying Li
    • Jean-Claude JunquaRoland KuhnTony DavisYi ZhaoWeiying Li
    • G10L1522
    • H04N21/4532G10L17/00H04N5/44543H04N21/42203H04N21/4394H04N21/4415H04N21/4667H04N21/482
    • Speech input supplied by the user is evaluated by the speaker verification/identification module, and based on the evaluation, parameters are retrieved from a user profile database. These parameters adapt the speech models of the speech recognizer and also supply the natural language parser with customized dialog grammars. The user's speech is then interpreted by the speech recognizer and natural language parser to determine the meaning of the user's spoken input in order to control the television tuner. The parser works in conjunction with a command module that mediates the dialog with the user, providing on-screen prompts or synthesized speech queries to elicit further input from the user when needed. The system integrates with an electronic program guide, so that the natural language parser is made aware of what programs are available when conducting the synthetic dialog with the user.
    • 由用户提供的语音输入由说话人验证/识别模块进行评估,并且基于评估,从用户简档数据库检索参数。 这些参数适应语音识别器的语音模型,并为自然语言解析器提供定制的对话语法。 用户的语音然后由语音识别器和自然语言解析器进行解释,以确定用户的口头输入的含义,以控制电视调谐器。 解析器与一个命令模块一起工作,该模块与用户中介对话,提供屏幕提示或合成语音查询,以便在需要时从用户中引出进一步的输入。 该系统与电子节目指南集成,使得自然语言解析器在与用户进行合成对话时了解哪些程序可用。