会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 83. 发明授权
    • Method and apparatus for automatically determining speaker characteristics for speech-directed advertising or other enhancement of speech-controlled devices or services
    • 用于自动确定语音导向广告的扬声器特性或语音控制设备或服务的其他增强的方法和装置
    • US08793127B2
    • 2014-07-29
    • US11930993
    • 2007-10-31
    • Harry PrintzVikas Gulati
    • Harry PrintzVikas Gulati
    • G10L17/00G10L15/16G10L15/18
    • G10L15/187G06F17/30861G06F17/30867G06Q30/02G10L15/02G10L15/142G10L15/18G10L15/22G10L17/26G10L2015/025
    • In addition to conveying primary information, human speech also conveys information concerning the speaker's gender, age, socioeconomic status, accent, language spoken, emotional state, or other personal characteristics, which is referred to as secondary information. Disclosed herein are both the means of automatic discovery and use of such secondary information to direct other aspects of the behavior of a controlled system. One embodiment of the invention comprises an improved method to determine, with high reliability, the gender of an adult speaker. A further embodiment of the invention comprises the use of this information to display a gender-appropriate advertisement to the user of an information retrieval system that uses a cell phone as the input and output device. The invention is not limited to gender and such secondary information can include, for example, any of information concerning the speaker's age, socioeconomic status, accent, language spoken, emotional state, or other personal characteristics.
    • 除了传达主要信息外,人类言语还传达有关演讲者的性别,年龄,社会经济状况,口音,口语,情绪状态或其他个人特征的信息,这被称为次要信息。 这里公开的是自动发现和使用这种二次信息以指导受控系统的行为的其他方面的手段。 本发明的一个实施例包括用于以高可靠性确定成人说话者的性别的改进方法。 本发明的另一实施例包括使用该信息向使用蜂窝电话的信息检索系统的用户显示针对性别的广告作为输入和输出设备。 本发明不限于性别,并且这种次要信息可以包括例如关于说话者的年龄,社会经济地位,口音,口语,情绪状态或其他个人特征的任何信息。
    • 85. 发明申请
    • REAL - TIME EMOTION TRACKING SYSTEM
    • 实时感应跟踪系统
    • US20140163960A1
    • 2014-06-12
    • US13712288
    • 2012-12-12
    • AT&T INTELLECTUAL PROPERTY I, L.P.
    • Dimitrios DIMITRIADISMazin E. GILBERTTaniya MISHRAHorst J. SCHROETER
    • G06F17/28
    • G10L25/63G10L17/04G10L17/26G10L25/48
    • Devices, systems, methods, media, and programs for detecting an emotional state change in an audio signal are provided. A plurality of segments of the audio signal is received, with the plurality of segments being sequential. Each segment of the plurality of segments is analyzed, and, for each segment, an emotional state and a confidence score of the emotional state are determined. The emotional state and the confidence score of each segment are sequentially analyzed, and a current emotional state of the audio signal is tracked throughout each of the plurality of segments. For each segment, it is determined whether the current emotional state of the audio signal changes to another emotional state based on the emotional state and the confidence score of the segment.
    • 提供了用于检测音频信号中的情绪状态改变的设备,系统,方法,媒体和程序。 接收音频信号的多个段,其中多个段是顺序的。 分析多个片段中的每个片段,并且针对每个片段,确定情感状态的情绪状态和置信评分。 顺序地分析每个片段的情绪状态和置信度得分,并且在多个片段中的每一个片段跟踪音频信号的当前情绪状态。 对于每个片段,基于片段的情绪状态和置信度分数确定音频信号的当前情绪状态是否改变到另一情感状态。
    • 87. 发明授权
    • Speaker characterization through speech analysis
    • 演讲者通过语音分析进行表征
    • US08682666B2
    • 2014-03-25
    • US13465487
    • 2012-05-07
    • Yoav DeganiYishai Zamir
    • Yoav DeganiYishai Zamir
    • G10L15/06G10L11/04G10L11/06G10L21/00
    • G10L15/06G10L15/1807G10L17/26
    • A computer implemented method, data processing system, apparatus and computer program product for determining current behavioral, psychological and speech styles characteristics of a speaker in a given situation and context, through analysis of current speech utterances of the speaker. The analysis calculates different prosodic parameters of the speech utterances, consisting of unique secondary derivatives of the primary pitch and amplitude speech parameters, and compares these parameters with pre-obtained reference speech data, indicative of various behavioral, psychological and speech styles characteristics. The method includes the formation of the classification speech parameters reference database, as well as the analysis of the speaker's speech utterances in order to determine the current behavioral, psychological and speech styles characteristics of the speaker in the given situation.
    • 一种计算机实现的方法,数据处理系统,装置和计算机程序产品,用于通过分析扬声器的当前语音话语来确定给定情况和语境中的扬声器的当前行为,心理和言语风格特征。 分析计算出语音话语的不同韵律参数,由主音阶和幅度语音参数的唯一二次导数组成,并将这些参数与预先获得的参考语音数据进行比较,表明各种行为,心理和言语风格特征。 该方法包括形成分类语音参数参考数据库,以及分析说话者的言语言,以便在给定情况下确定演讲者的当前行为,心理和言语风格特征。
    • 88. 发明授权
    • Sound event detecting module for a sound event recognition system and method thereof
    • 用于声音事件识别系统的声音事件检测模块及其方法
    • US08655655B2
    • 2014-02-18
    • US12981660
    • 2010-12-30
    • Yuh-Ching WangKuo-Yuan Li
    • Yuh-Ching WangKuo-Yuan Li
    • G10L15/08G10L25/03
    • G10L15/10G10L15/02G10L17/26G10L25/48
    • A sound event detecting module for detecting whether a sound event with characteristic of repeating is generated. A sound end recognizing unit recognizes ends of sounds according to a sound signal to generate sound sections and multiple sets of feature vectors of the sound sections correspondingly. A storage unit stores at least M sets of feature vectors. A similarity comparing unit compares the at least M sets of feature vectors with each other, and correspondingly generates a similarity score matrix, which stores similarity scores of any two of the sound sections of the at least M of the sound sections. A correlation arbitrating unit determines the number of sound sections with high correlations to each other according to the similarity score matrix. When the number is greater than one threshold value, the correlation arbitrating unit indicates that the sound event with the characteristic of repeating is generated.
    • 一种用于检测是否产生具有重复特征的声音事件的声音事件检测模块。 声音识别单元根据声音信号识别声音的结束,以相应地生成声音部分和声音部分的多组特征向量组。 存储单元至少存储M组特征向量。 相似度比较单元将至少M组特征向量彼此进行比较,并且相应地生成相似性评分矩阵,其存储声音部分的至少M个的声音部分中的任意两个的声音部分的相似性得分。 相关性仲裁单元根据相似性得分矩阵来确定彼此具有高相关性的声音段的数量。 当该数量大于一个阈值时,相关性仲裁单元指示产生具有重复特性的声音事件。
    • 89. 发明授权
    • Speech translation apparatus, method and program that generates insertion sentence explaining recognized emotion types
    • 语音翻译设备,方法和程序,生成插入句解释识别的情绪类型
    • US08635070B2
    • 2014-01-21
    • US13071549
    • 2011-03-25
    • Kazuo Sumita
    • Kazuo Sumita
    • G06F17/28G10L13/00G10L13/08
    • G06F17/2795G06F17/2854G06F17/289G10L17/26
    • According to one embodiment, a speech translation apparatus includes a receiving unit, a first recognition unit, a second recognition unit, a first generation unit, a translation unit, a second generation unit, a synthesis unit. The receiving unit is configured to receive a speech in a first language and convert to speech signal. The first recognition unit is configured to perform speech recognition and generate a transcription. The second recognition unit is configured to recognize which emotion type is included in the speech and generate emotion identification information including recognized emotion type(s). The first generation unit is configured to generate a filtered sentence. The translation unit is configured to generate a translation of the filtered sentence in the first language in a second language. The second generation unit is configured to generate an insertion sentence. The synthesis unit is configured to convert the filtered and the insertion sentences into speech signal.
    • 根据一个实施例,语音翻译装置包括接收单元,第一识别单元,第二识别单元,第一生成单元,翻译单元,第二生成单元,合成单元。 接收单元被配置为以第一语言接收语音并转换成语音信号。 第一识别单元被配置为执行语音识别并产生转录。 第二识别单元被配置为识别哪个情感类型被包括在语音中并且生成包括识别的情绪类型的情感识别信息。 第一代单元被配置为生成过滤句子。 翻译单元被配置为以第二语言生成第一语言的过滤句子的翻译。 第二代单元被配置为生成插入语句。 所述合成单元被配置为将所述滤波和插入语句转换为语音信号。
    • 90. 发明授权
    • Apparatus and method for automatic extraction of important events in audio signals
    • 自动提取音频信号中重要事件的装置和方法
    • US08635065B2
    • 2014-01-21
    • US10985446
    • 2004-11-10
    • Silke Goronzy-ThomaeThomas KempRalf KompeYin Hay LamKrzysztof MarasekRaquel Tato
    • Silke Goronzy-ThomaeThomas KempRalf KompeYin Hay LamKrzysztof MarasekRaquel Tato
    • G10L15/06G10L21/00G10L19/12G10L19/14G10L17/00
    • G10L25/00G10L15/00G10L17/26
    • The present invention discloses an apparatus for automatic extraction of important events in audio signals comprising: signal input means for supplying audio signals; audio signal fragmenting means for partitioning audio signals supplied by the signal input means into audio fragments of a predetermined length and for allocating a sequence of one or more audio fragments to a respective audio window; feature extracting means for analyzing acoustic characteristics of the audio signals comprised in the audio fragments and for analyzing acoustic characteristics of the audio signals comprised in the audio windows; and important event extraction means for extracting important events in audio signals supplied by the audio signal fragmenting means based on predetermined important event classifying rules depending on acoustic characteristics of the audio signals comprised in the audio fragments and on acoustic characteristics of the audio signals comprised in the audio windows, wherein each important event extracted by the important event extraction means comprises a discrete sequence of cohesive audio fragments corresponding to an important event included in the audio signals.
    • 本发明公开了一种用于自动提取音频信号中的重要事件的装置,包括:用于提供音频信号的信号输入装置; 用于将由信号输入装置提供的音频信号划分成预定长度的音频片段并用于将一个或多个音频片段的序列分配到相应音频窗口的音频信号分段装置; 特征提取装置,用于分析包含在音频片段中的音频信号的声学特性并分析包含在音频窗口中的音频信号的声学特性; 以及重要事件提取装置,用于根据包含在音频片段中的音频信号的声学特性以及包含在音频片段中的音频信号的声学特性,基于预定的重要事件分类规则,提取由音频信号分段装置提供的音频信号中的重要事件。 音频窗口,其中由重要事件提取装置提取的每个重要事件包括对应于包括在音频信号中的重要事件的粘性音频片段的离散序列。