会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 102. 发明授权
    • Statistical unit selection language models based on acoustic fingerprinting
    • 基于声指纹的统计单位选择语言模型
    • US09424835B2
    • 2016-08-23
    • US14850249
    • 2015-09-10
    • Google Inc.
    • Alexander GutkinJavier Gonzalvo FructuosoCyril Georges Luc Allauzen
    • G10L15/08G10L15/06G10L19/018G10L13/08
    • G10L15/063G10L13/08G10L19/018
    • Methods, systems, and apparatus, including computer programs encoded on computer storage media, for providing statistical unit selection language modeling based on acoustic fingerprinting. The methods, systems and apparatus include the actions of obtaining a unit database of acoustic units and, for each acoustic unit, linguistic data corresponding to the acoustic unit; obtaining stored data associating each acoustic unit with (i) a corresponding acoustic fingerprint and (ii) a probability of the linguistic data corresponding to the acoustic unit occurring in a text corpus; determining that the unit database of acoustic units has been updated to include one or more new acoustic units; for each new acoustic unit in the updated unit database: generating an acoustic fingerprint for the new acoustic unit; identifying an acoustic unit that (i) has an acoustic fingerprint that is indicated as similar to the fingerprint of the new acoustic unit, and (ii) has a stored associated probability.
    • 方法,系统和装置,包括在计算机存储介质上编码的计算机程序,用于提供基于声学指纹识别的统计单位选择语言建模。 方法,系统和装置包括获得单元数据库的动作,对于每个声学单元,对应于声学单元的语言数据; 获得将每个声学单元与(i)对应的声学指纹相关联的存储数据和(ii)与在文本语料库中发生的声学单元相对应的语言数据的概率; 确定声学单元的单元数据库已经被更新为包括一个或多个新的声学单元; 对于更新的单元数据库中的每个新的声学单元:为新的声学单元产生声学指纹; 识别(i)具有与新声学单元的指纹相似的声音指纹的声学单元,以及(ii)具有存储的相关概率。
    • 105. 发明申请
    • VEHICLE AND CONTROL METHOD THEREOF
    • 车辆及其控制方法
    • US20160111089A1
    • 2016-04-21
    • US14709139
    • 2015-05-11
    • HYUNDAI MOTOR COMPANY
    • Hyung Jin KIM
    • G10L15/22G10L15/28G10L15/06G10L15/00G10L15/10
    • G10L15/22G10L13/08G10L2015/223
    • A vehicle of recognizing received voice based on a language set in an external apparatus includes: a communication unit configured to receive text data stored in an external apparatus; a data converter configured to convert the received text data into voice data; a speech input unit configured to receive a speech from a user; a speech recognizer configured to recognize the received speech based on a language set in the external apparatus; and a controller configured to search for voice data corresponding to the recognized speech in the converted voice data, to generate a control command including the voice data found by the controller based on the recognized speech, and to transmit the control command to the external apparatus through the communication unit.
    • 基于外部设备中设置的语言识别接收到的语音的媒体包括:通信单元,被配置为接收存储在外部设备中的文本数据; 数据转换器,被配置为将所接收的文本数据转换成语音数据; 语音输入单元,被配置为从用户接收语音; 语音识别器,被配置为基于在外部设备中设置的语言来识别所接收的语音; 以及控制器,被配置为在转换的语音数据中搜索与识别的语音相对应的语音数据,以产生包括基于识别的语音由控制器发现的语音数据的控制命令,并且通过 通信单元。
    • 107. 发明授权
    • Content creation support apparatus, method and program
    • 内容创建支持设备,方法和程序
    • US09304987B2
    • 2016-04-05
    • US14301378
    • 2014-06-11
    • KABUSHIKI KAISHA TOSHIBA
    • Kosei FumeMasahiro Morita
    • G10L15/00G06F17/27G10L13/08G10L15/26G10L13/033
    • G06F17/2755G10L13/033G10L13/08G10L15/26
    • According to one embodiment, a content creation support apparatus includes a speech synthesis unit, a speech recognition unit, an extraction unit, a detection unit, a presentation unit and a selection unit. The speech synthesis unit performs a speech synthesis on a first text. The speech recognition unit performs a speech recognition on the synthesized speech to obtain a second text. The extraction unit extracts feature values by performing a morphological analysis on each of the first and second texts. The detection unit compares a first feature value of a first difference string and a second feature value of a second difference string. The presentation unit presents correction candidate(s) according to the second feature value. The selection unit selects one of the correction candidates in accordance with an instruction from a user.
    • 根据一个实施例,内容创建支持设备包括语音合成单元,语音识别单元,提取单元,检测单元,呈现单元和选择单元。 语音合成单元对第一文本执行语音合成。 语音识别单元对合成语音执行语音识别以获得第二文本。 提取单元通过对第一和第二文本中的每一个执行形态分析来提取特征值。 检测单元将第一差分字符串的第一特征值与第二差分字符串的第二特征值进行比较。 呈现单元根据第二特征值呈现校正候选。 选择单元根据来自用户的指令来选择一个校正候选。
    • 109. 发明授权
    • Speech recognition assisted evaluation on text-to-speech pronunciation issue detection
    • 语音识别辅助评估文本到语音发音问题检测
    • US09293129B2
    • 2016-03-22
    • US13785573
    • 2013-03-05
    • Microsoft Technology Licensing, LLC
    • Pei ZhaoBo YanLei HeZhe GengYiu-Ming Leung
    • G10L13/08
    • G10L13/086G10L13/08
    • Pronunciation issues for synthesized speech are automatically detected using human recordings as a reference within a Speech Recognition Assisted Evaluation (SRAE) framework including a Text-To-Speech flow and a Speech Recognition (SR) flow. A pronunciation issue detector evaluates results obtained at multiple levels of the TTS flow and the SR flow (e.g. phone, word, and signal level) by using the corresponding human recordings as the reference for the synthesized speech, and outputs possible pronunciation issues. A signal level may be used to determine similarities/differences between the recordings and the TTS output. A model level checker may provide results to the pronunciation issue detector to check the similarities of the TTS and the SR phone set including mapping relations. Results from a comparison of the SR output and the recordings may also be evaluation by the pronunciation issue detector. The pronunciation issue detector outputs a list that lists potential pronunciation issue candidates.
    • 在语音识别辅助评估(SRAE)框架内使用人类录音作为参考,自动检测合成语音的发音问题,包括文本到语音流和语音识别(SR)流。 发音问题检测器通过使用相应的人类记录作为合成语音的参考来评估在多个级别的TTS流和SR流(例如电话,字和信号电平)上获得的结果,并输出可能的发音问题。 可以使用信号电平来确定记录和TTS输出之间的相似/差异。 模型级检查器可以向发音问题检测器提供结果,以检查包括映射关系的TTS和SR电话机的相似性。 通过比较SR输出和记录的结果也可以由发音问题检测器进行评估。 发音问题检测器输出列出潜在发音问题候选人的列表。
    • 110. 发明授权
    • Methods and apparatus for predicting prosody in speech synthesis
    • 用于预测语音合成中的韵律的方法和装置
    • US09286886B2
    • 2016-03-15
    • US13012740
    • 2011-01-24
    • Stephen MinnisAndrew P. Breen
    • Stephen MinnisAndrew P. Breen
    • G10L13/08G10L13/10
    • G10L13/10G10L13/08
    • Techniques for predicting prosody in speech synthesis may make use of a data set of example text fragments with corresponding aligned spoken audio. To predict prosody for synthesizing an input text, the input text may be compared with the data set of example text fragments to select a best matching sequence of one or more example text fragments, each example text fragment in the sequence being paired with a portion of the input text. The selected example text fragment sequence may be aligned with the input text, e.g., at the word level, such that prosody may be extracted from the audio aligned with the example text fragments, and the extracted prosody may be applied to the synthesis of the input text using the alignment between the input text and the example text fragments.
    • 用于预测语音合成中的韵律的技术可以利用具有对应的口头音频的示例文本片段的数据集。 为了预测合成输入文本的韵律,可以将输入文本与示例文本片段的数据集进行比较,以选择一个或多个示例文本片段的最佳匹配序列,每个示例中的文本片段与一部分 输入文本。 所选择的示例文本片段序列可以与输入文本(例如,在字级别)对齐,使得可以从与示例文本片段对齐的音频中提取韵律,并且所提取的韵律可以应用于输入的合成 文本使用输入文本和示例文本片段之间的对齐。