会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 1. 发明授权
    • Face feature analysis for automatic lipreading and character animation
    • 面部特征分析,用于自动修剪和角色动画
    • US6028960A
    • 2000-02-22
    • US716959
    • 1996-09-20
    • Hans Peter GrafEric David Petajan
    • Hans Peter GrafEric David Petajan
    • G06K9/00G06K9/46
    • G06K9/00268
    • A face feature analysis which begins by generating multiple face feature candidates, e.g., eyes and nose positions, using an isolated frame face analysis. Then, a nostril tracking window is defined around a nose candidate and tests are applied to the pixels therein based on percentages of skin color area pixels and nostril area pixels to determine whether the nose candidate represents an actual nose. Once actual nostrils are identified, size, separation and contiguity of the actual nostrils is determined by projecting the nostril pixels within the nostril tracking window. A mouth window is defined around the mouth region and mouth detail analysis is then applied to the pixels within the mouth window to identify inner mouth and teeth pixels and therefrom generate an inner mouth contour. The nostril position and inner mouth contour are used to generate a synthetic model head. A direct comparison is made between the inner mouth contour generated and that of a synthetic model head and the synthetic model head is adjusted accordingly. Vector quantization algorithms may be used to develop a codebook of face model parameters to improve processing efficiency. The face feature analysis is suitable regardless of noise, illumination variations, head tilt, scale variations and nostril shape.
    • 一种面部特征分析,其通过使用孤立的框架面分析来生成多个面部特征候选物,例如眼睛和鼻子位置。 然后,围绕鼻子候选者定义鼻孔跟踪窗口,并且基于皮肤颜色面积像素和鼻孔区域像素的百分比将测试应用于其中的像素,以确定鼻子候选者是否表示实际的鼻子。 一旦确定了实际鼻孔,就可以通过在鼻孔跟踪窗口内投射鼻孔像素来确定实际鼻孔的尺寸,分离和连续性。 嘴口围绕嘴部区域定义,然后将口腔细节分析应用于口腔内的像素,以识别内部嘴和牙齿像素,从而产生内部嘴部轮廓。 鼻孔位置和内嘴轮廓用于产生合成模型头。 直接比较所生成的内口轮廓和合成模型头的内口轮廓,并相应地调整合成模型头。 矢量量化算法可用于开发面部模型参数的码本,以提高处理效率。 面部特征分析适用于噪声,照明变化,头部倾斜,尺度变化和鼻孔形状。
    • 2. 发明授权
    • Systems and methods for encoding and decoding video streams
    • 视频流编码和解码的系统和方法
    • US08638846B1
    • 2014-01-28
    • US10601495
    • 2003-06-23
    • Eric CosattoHans Peter GrafJoern Ostermann
    • Eric CosattoHans Peter GrafJoern Ostermann
    • H04N7/12
    • H04N19/573G06T9/001H04N19/21H04N19/23H04N19/55
    • Systems and methods for encoding/decoding a video stream. Animated talking heads are coded using partial offline encoding, multiple video streams, and multiple reference frames. The content of a face animation video that is known beforehand is encoded offline and the remaining content is encoded online and included in the video stream. To reduce bit rate, a server can stream multiple video sequences to the client and the video sequences are stored in the client's frame store. The server sends instructions to play a particular video sequence instead of streaming the particular video sequence. Multiple video streams can also be streamed to the client. Positional data and blending data are also sent to properly position one video stream relative to another video stream and to blend one video stream into another video stream.
    • 用于对视频流进行编码/解码的系统和方法。 动画通话头使用部分离线编码,多个视频流和多个参考帧进行编码。 预先知道的面部动画视频的内容被离线编码,并且剩余的内容被在线编码并包括在视频流中。 为了降低比特率,服务器可以将多个视频序列流式传输到客户端,视频序列存储在客户端的帧存储器中。 服务器发送播放特定视频序列的指令,而不是流式传输特定的视频序列。 多个视频流也可以流式传输到客户端。 还发送位置数据和混合数据以相对于另一个视频流适当地定位一个视频流,并将一个视频流混合到另一视频流中。
    • 5. 发明授权
    • System and method for sending multi-media messages using emoticons
    • 使用表情发送多媒体消息的系统和方法
    • US07921013B1
    • 2011-04-05
    • US11214666
    • 2005-08-30
    • Joern OstermannMehmet Reha CivanlarEric CosattoHans Peter GrafYann Andre LeCun
    • Joern OstermannMehmet Reha CivanlarEric CosattoHans Peter GrafYann Andre LeCun
    • G10L13/00G10L13/06G10L21/00
    • G06Q10/107G06F17/241G10L13/00
    • A system and method of providing sender-customization of multi-media messages through the use of emoticons is disclosed. The sender inserts the emoticons into a text message. As an animated face audibly delivers the text, emoticons associated with the message are started a predetermined period of time or number of words prior to the position of the emoticon in the message text and completed a predetermined length of time or number of words following the location of the emoticon. The sender may insert emoticons through the use of emoticon buttons that are icons available for choosing. Upon sender selections of an emoticon, an icon representing the emoticon is inserted into the text at the position of the cursor. Once an emoticon is chosen, the sender may also choose the amplitude for the emoticon and increased or decreased amplitude will be displayed in the icon inserted into the message text.
    • 公开了通过使用表情符号来提供多媒体消息的发送者定制的系统和方法。 发件人将表情符号插入文本消息。 作为动画面部听觉地传送文本,与消息相关联的表情符号在消息文本中表情符号的位置之前的预定时间段或字数开始,并且完成预定时间长度或位置之后的字数 的表情符号。 发件人可以通过使用可用于选择的图标的表情符号按钮来插入表情符号。 在表情符号的发送者选择之后,表示表情符号的图标被插入到光标位置的文本中。 一旦选择了表情符号,发送者也可以选择表情符号的幅度,增加或减小的幅度将显示在插入消息文本的图标中。
    • 8. 发明授权
    • System and method of controlling sound in a multi-media communication application
    • 多媒体通信应用中的声音控制系统及方法
    • US06963839B1
    • 2005-11-08
    • US09999526
    • 2001-11-02
    • Joern OstermannMehmet Reha CivanlarHans Peter GrafThomas M. Isaacson
    • Joern OstermannMehmet Reha CivanlarHans Peter GrafThomas M. Isaacson
    • G06F17/28G10L13/00G10L13/08G10L21/00
    • G10L13/08
    • A method for customizing a voice in a multi-media message created by a sender for a recipient is disclosed. The multi-media message comprises a text message from the sender to be delivered by an animated entity. The method comprises presenting an option to the sender to insert voice emoticons into the text message associated with parameters of a voice used by the animated entity to deliver the text message. The message is then delivered wherein the voice of the animated entity is modified throughout the message according to the voice emoticons. The voice emoticons may relate to features such as voice stress, volume, pauses, emotion, yelling, or whispering. After the sender inserts various voice emoticons into the text of the message, the animated entity delivers the multi-media message giving effect to each voice emoticon in the text. A volume or intensity of the voice emoticons may be given effect by repeating the tags. In this case, delivering the multi-media message further comprises delivering the multi-media message at a variable level associated with the number of times a respective voice emoticon is repeated. In this manner, the sender may control the presentation of the message to increase the overall effectiveness of the multi-media message.
    • 公开了一种用于定制由发送方为接收者创建的多媒体消息中的语音的方法。 多媒体消息包括来自发送者的要被动画实体递送的文本消息。 该方法包括向发送者呈现选项以将语音表情符号插入与由动画实体使用的语音的参数相关联的文本消息中以递送文本消息。 然后传递消息,其中根据语音表情符号在整个消息中修改动画实体的语音。 语音表情符号可能与语音压力,音量,停顿,情感,吼叫或耳语等功能有关。 在发送者将各种语音表情符号插入消息的文本之后,动画实体递送多媒体消息,使文本中的每个语音表情符合效果。 语音表情符号的音量或强度可以通过重复标签来实现。 在这种情况下,传送多媒体消息还包括以与重复相应的语音表情符号的次数相关联的可变级别递送多媒体消息。 以这种方式,发送者可以控制消息的呈现以增加多媒体消息的整体有效性。
    • 10. 发明授权
    • System and method for triphone-based unit selection for visual speech synthesis
    • 用于视觉语音合成的基于耳机的单元选择的系统和方法
    • US09583098B1
    • 2017-02-28
    • US11924025
    • 2007-10-25
    • Eric CosattoHans Peter GrafFu Jie Huang
    • Eric CosattoHans Peter GrafFu Jie Huang
    • G10L15/26G06T15/70G10L15/08G10L13/07H04N19/00
    • G10L15/08G10L13/07G10L15/26G10L2021/105H04N19/00
    • A system and method for generating a video sequence having mouth movements synchronized with speech sounds are disclosed. The system utilizes a database of n-phones as the smallest selectable unit, wherein n is larger than 1 and preferably 3. The system calculates a target cost for each candidate n-phone for a target frame using a phonetic distance, coarticulation parameter, and speech rate. For each n-phone in a target sequence, the system searches for candidate n-phones that are visually similar according to the target cost. The system samples each candidate n-phone to get a same number of frames as in the target sequence and builds a video frame lattice of candidate video frames. The system assigns a joint cost to each pair of adjacent frames and searches the video frame lattice to construct the video sequence by finding the optimal path through the lattice according to the minimum of the sum of the target cost and the joint cost over the sequence.
    • 公开了一种用于产生具有与语音的同步的口部动作的视频序列的系统和方法。 该系统利用n电话的数据库作为最小的可选单元,其中n大于1,并且优选地为3.系统使用语音距离,协调参数和目标帧来计算目标帧的每个候选n电话的目标成本 言语速度 对于目标序列中的每个n电话,系统根据目标成本搜索视觉上类似的候选n电话。 系统对每个候选n电话进行采样,以获得与目标序列相同数量的帧,并建立候选视频帧的视频帧格点。 系统为每对相邻帧分配联合成本,并通过根据目标成本和序列中的联合成本的总和的最小值找到通过网格的最优路径来搜索视频帧格以构建视频序列。