会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 1. 发明授权
    • Online distorted speech estimation within an unscented transformation framework
    • 一个无限转换框架内的在线扭曲语音估计
    • US08731916B2
    • 2014-05-20
    • US12948935
    • 2010-11-18
    • Deng LiJinyu LiDong YuYifan Gong
    • Deng LiJinyu LiDong YuYifan Gong
    • G10L21/02
    • G10L19/005G10L15/20
    • Noise and channel distortion parameters in the vectorized logarithmic or the cepstral domain for an utterance may be estimated, and subsequently the distorted speech parameters in the same domain may be updated using an unscented transformation framework during online automatic speech recognition. An utterance, including speech generated from a transmission source for delivery to a receiver, may be received by a computing device. The computing device may execute instructions for applying the unscented transformation framework to speech feature vectors, representative of the speech, in order to estimate, in a sequential or online manner, static noise and channel distortion parameters and dynamic noise distortion parameters in the unscented transformation framework. The static and dynamic parameters for the distorted speech in the utterance may then be updated from clean speech parameters and the noise and channel distortion parameters using non-linear mapping.
    • 可以估计用于话语的向量化对数或倒频域中的噪声和信道失真参数,并且随后可以在在线自动语音识别期间使用无密码变换框架来更新相同域中的失真语音参数。 包括从发送源产生的用于传送到接收机的语音的话语可以被计算设备接收。 计算设备可以执行用于将无声变换框架应用于代表语音的语音特征向量的指令,以便以顺序或在线方式估计无密度变换框架中的静态噪声和信道失真参数以及动态噪声失真参数 。 然后可以使用非线性映射从干净的语音参数和噪声和信道失真参数中更新话音中失真语音的静态和动态参数。
    • 2. 发明申请
    • ONLINE DISTORTED SPEECH ESTIMATION WITHIN AN UNSCENTED TRANSFORMATION FRAMEWORK
    • 在一个未经规定的转换框架内的在线失真的语音估计
    • US20120130710A1
    • 2012-05-24
    • US12948935
    • 2010-11-18
    • Deng LiJinyu LiDong YuYifan Gong
    • Deng LiJinyu LiDong YuYifan Gong
    • G10L15/00
    • G10L19/005G10L15/20
    • Noise and channel distortion parameters in the vectorized logarithmic or the cepstral domain for an utterance may be estimated, and subsequently the distorted speech parameters in the same domain may be updated using an unscented transformation framework during online automatic speech recognition. An utterance, including speech generated from a transmission source for delivery to a receiver, may be received by a computing device. The computing device may execute instructions for applying the unscented transformation framework to speech feature vectors, representative of the speech, in order to estimate, in a sequential or online manner, static noise and channel distortion parameters and dynamic noise distortion parameters in the unscented transformation framework. The static and dynamic parameters for the distorted speech in the utterance may then be updated from clean speech parameters and the noise and channel distortion parameters using non-linear mapping.
    • 可以估计用于话语的向量化对数或倒频域中的噪声和信道失真参数,并且随后可以在在线自动语音识别期间使用无密码变换框架来更新相同域中的失真语音参数。 包括从发送源产生的用于传送到接收机的语音的话语可以被计算设备接收。 计算设备可以执行用于将无声变换框架应用于代表语音的语音特征向量的指令,以便以顺序或在线方式估计无密度变换框架中的静态噪声和信道失真参数以及动态噪声失真参数 。 然后可以使用非线性映射从干净的语音参数和噪声和信道失真参数中更新话音中失真语音的静态和动态参数。
    • 4. 发明授权
    • Phase sensitive model adaptation for noisy speech recognition
    • 嘈杂语音识别的相敏模型适应
    • US08214215B2
    • 2012-07-03
    • US12236530
    • 2008-09-24
    • Jinyu LiLi DengDong YuYifan GongAlejandro Acero
    • Jinyu LiLi DengDong YuYifan GongAlejandro Acero
    • G10L15/14
    • G10L15/065G10L15/20
    • A speech recognition system described herein includes a receiver component that receives a distorted speech utterance. The speech recognition also includes an updater component that is in communication with a first model and a second model, wherein the updater component automatically updates parameters of the second model based at least in part upon joint estimates of additive and convolutive distortions output by the first model, wherein the joint estimates of additive and convolutive distortions are estimates of distortions based on a phase-sensitive model in the speech utterance received by the receiver component. Further, distortions other than additive and convolutive distortions, including other stationary and nonstationary sources, can also be estimated used to update the parameters of the second model.
    • 本文描述的语音识别系统包括接收失真的语音话语的接收机组件。 所述语音识别还包括与第一模型和第二模型通信的更新器组件,其中所述更新器组件至少部分地基于由所述第一模型输出的加法和卷积失真的联合估计来自动更新所述第二模型的参数 其中,加法和卷积失真的联合估计是基于由接收器部件接收的语音发声中的相敏模型的失真估计。 此外,还可以估计用于更新第二模型参数的除加法和卷积失真之外的失真,包括其他静止和非平稳源。
    • 5. 发明申请
    • PHASE SENSITIVE MODEL ADAPTATION FOR NOISY SPEECH RECOGNITION
    • 语音识别的相敏感模型适应
    • US20100076758A1
    • 2010-03-25
    • US12236530
    • 2008-09-24
    • Jinyu LiLi DengDong YuYifan GongAlejandro Acero
    • Jinyu LiLi DengDong YuYifan GongAlejandro Acero
    • G10L15/20G10L15/14
    • G10L15/065G10L15/20
    • A speech recognition system described herein includes a receiver component that receives a distorted speech utterance. The speech recognition also includes an updater component that is in communication with a first model and a second model, wherein the updater component automatically updates parameters of the second model based at least in part upon joint estimates of additive and convolutive distortions output by the first model, wherein the joint estimates of additive and convolutive distortions are estimates of distortions based on a phase-sensitive model in the speech utterance received by the receiver component. Further, distortions other than additive and convolutive distortions, including other stationary and nonstationary sources, can also be estimated used to update the parameters of the second model.
    • 本文描述的语音识别系统包括接收失真的语音话语的接收机组件。 所述语音识别还包括与第一模型和第二模型通信的更新器组件,其中所述更新器组件至少部分地基于由所述第一模型输出的加法和卷积失真的联合估计来自动更新所述第二模型的参数 其中,加法和卷积失真的联合估计是基于由接收器部件接收的语音发声中的相敏模型的失真估计。 此外,还可以估计用于更新第二模型参数的除加法和卷积失真之外的失真,包括其他静止和非平稳源。
    • 9. 发明授权
    • Model training for automatic speech recognition from imperfect transcription data
    • 从不完美的转录数据自动语音识别的模型训练
    • US09280969B2
    • 2016-03-08
    • US12482142
    • 2009-06-10
    • Jinyu LiYifan GongChaojun LiuKaisheng Yao
    • Jinyu LiYifan GongChaojun LiuKaisheng Yao
    • G10L15/00G10L15/06G10L15/065
    • G10L15/063G10L15/065
    • Techniques and systems for training an acoustic model are described. In an embodiment, a technique for training an acoustic model includes dividing a corpus of training data that includes transcription errors into N parts, and on each part, decoding an utterance with an incremental acoustic model and an incremental language model to produce a decoded transcription. The technique may further include inserting silence between a pair of words into the decoded transcription and aligning an original transcription corresponding to the utterance with the decoded transcription according to time for each part. The technique may further include selecting a segment from the utterance having at least Q contiguous matching aligned words, and training the incremental acoustic model with the selected segment. The trained incremental acoustic model may then be used on a subsequent part of the training data. Other embodiments are described and claimed.
    • 描述了用于训练声学模型的技术和系统。 在一个实施例中,用于训练声学模型的技术包括将包括转录错误的训练数据的语料库划分成N个部分,并且在每个部分上,用增量声学模型和增量语言模型解码语音以产生解码的转录。 该技术可以进一步包括将一对单词之间的沉默插入解码的转录中,并根据每个部分的时间将与发音对应的原始转录与解码的转录对准。 该技术可以进一步包括从具有至少Q个连续匹配对齐字的话语中选择一段,以及使用所选择的段来训练增量声学模型。 然后可以在训练数据的后续部分上使用经过训练的增量声学模型。 描述和要求保护其他实施例。