会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 1. 发明授权
    • Automatic video summarization using a measure of shot importance and a frame-packing method
    • 自动视频摘要使用拍摄重要性的度量和帧打包方法
    • US06535639B1
    • 2003-03-18
    • US09267529
    • 1999-03-12
    • Shingo UchihachiJonathan T. FooteLynn Wilcox
    • Shingo UchihachiJonathan T. FooteLynn Wilcox
    • G06K936
    • G06K9/00711G06F17/30799G06F17/30843G06K9/00751G11B27/031G11B27/28Y10S707/99933Y10S707/99945
    • A measure of importance is calculated for segmented parts of a video. The segmented parts are determined by segmenting the video into component shots and then merging by iteration the component shots based on similarity or other factors. Segmentation may also be determined by clustering frames of the video, and creating segments from the same cluster ID. The measure of importance is calculated based on a normalized weight of each segment and on length and rarity of each shot/segmented part. The importance measure may be utilized to generate a video summary by selecting the most important segments and generating representative frames for the selected segments. A thresholding process is applied to the importance score to provide a predetermined number or an appropriate number generated on the fly of shots or segments to be represented by frames. The representative frames are then packed into the video summary. The sizes of the frames to be packed are predetermined by their importance measure and adjusted according to space availability. Packing based on a grid and an exhaustive search of frame combinations to fill each row in the grid. A cost algorithm and a space-filling rule are utilized to determine the best fit of frames. The video summary may be presented on either a paper interface referencing or a web page linking the frames of the summary to points of the video.
    • 为视频的分段部分计算重要度量度。 分割部分通过将视频分割成分量拍摄确定,然后通过迭代合并基于相似性或其他因素的分量拍摄。 分割也可以通过对视频的帧进行聚类来确定,并且从相同的集群ID创建分段。 重要性的度量是基于每个段的归一化权重和每个镜头/分割部分的长度和稀有度来计算的。 可以通过选择最重要的段并为所选择的段产生代表性帧来利用重要度量来产生视频摘要。 将阈值处理应用于重要性得分以提供在由帧表示的拍摄或片段的飞行中产生的预定数量或适当数量。 然后将代表性的框架包装在视频摘要中。 要包装的帧的大小由其重要性度量预先确定,并根据空间可用性进行调整。 基于网格的包装和对帧组合的详尽搜索来填充网格中的每一行。 使用成本算法和空格填充规则来确定帧的最佳拟合。 视频摘要可以呈现在纸张界面引用或将摘要的帧链接到视频的点的网页上。
    • 2. 发明授权
    • Methods and apparatuses for interactive similarity searching, retrieval and browsing of video
    • 视频互动相似搜索,检索和浏览的方法和装置
    • US07246314B2
    • 2007-07-17
    • US10859832
    • 2004-06-03
    • Jonathan T. FooteAndreas GirgensohnLynn Wilcox
    • Jonathan T. FooteAndreas GirgensohnLynn Wilcox
    • G06F15/00G06F14/00
    • G06K9/00758G06F17/30814G06F17/30825G06F17/3084
    • Methods for interactive selecting video queries consisting of training images from a video for a video similarity search and for displaying the results of the similarity search are disclosed. The user selects a time interval in the video as a query definition of training images for training an image class statistical model. Time intervals can be as short as one frame or consist of disjoint segments or shots. A statistical model of the image class defined by the training images is calculated on-the-fly from feature vectors extracted from transforms of the training images. For each frame in the video, a feature vector is extracted from the transform of the frame, and a similarity measure is calculated using the feature vector and the image class statistical model. The similarity measure is derived from the likelihood of a Gaussian model producing the frame. The similarity is then presented graphically, which allows the time structure of the video to be visualized and browsed. Similarity can be rapidly calculated for other video files as well, which enables content-based retrieval by example. A content-aware video browser featuring interactive similarity measurement is presented. A method for selecting training segments involves mouse click-and-drag operations over a time bar representing the duration of the video; similarity results are displayed as shades in the time bar. Another method involves selecting periodic frames of the video as endpoints for the training segment.
    • 公开了用于交互式选择由用于视频相似性搜索的视频的训练图像组成的视频查询和用于显示相似性搜索的结果的方法。 用户选择视频中的时间间隔作为用于训练图像类统计模型的训练图像的查询定义。 时间间隔可以短到一帧,或者由不相交的片段或镜头组成。 从训练图像变换中提取的特征向量,计算由训练图像定义的图像类别的统计模型。 对于视频中的每个帧,从帧的变换中提取特征向量,并且使用特征向量和图像类统计模型来计算相似度度量。 相似性度量是从产生帧的高斯模型的可能性得出的。 然后以图形方式呈现相似性,这允许视频的时间结构可视化和浏览。 也可以为其他视频文件快速计算相似度,从而实现基于内容的检索。 介绍了具有交互式相似度测量功能的内容感知视频浏览器。 用于选择训练段的方法涉及通过表示视频持续时间的时间条来进行鼠标点击和拖动操作; 相似度结果在时间栏中显示为阴影。 另一种方法是选择视频的周期帧作为训练段的端点。
    • 3. 发明授权
    • Methods and apparatuses for interactive similarity searching, retrieval, and browsing of video
    • 视频互动相似检索,检索和浏览的方法和装置
    • US06774917B1
    • 2004-08-10
    • US09266558
    • 1999-03-11
    • Jonathan T. FooteAndreas GirgensohnLynn Wilcox
    • Jonathan T. FooteAndreas GirgensohnLynn Wilcox
    • G06F1500
    • G06K9/00758G06F17/30814G06F17/30825G06F17/3084
    • Method for interactive selecting video consisting of training images from a video for a video similarity search and for displaying the results of the similarity search are disclosed. The user selects a time interval in the video as a query definition of training images for training an image class statistical model. Time intervals can be as short as one frame or consist of disjoint segments or shots. A statistical model of the image class defined by the training images is calculated on-the-fly from feature vectors extracted from transforms of the training images. For each frame in the video, a feature vector is extracted from the transform of the frame, and a similarity measure is calculated using the feature vector and the image class statistical model. The similarity measure is derived from the likelihood of a Gaussian model producing the frame. The similarity is then presented graphically, which allows the time structure of the video to be visualized and browsed. Similarity can be rapidly calculated for other video files as well, which enables content-based retrieval by example. A content-aware video browser featuring interactive similarity measurement is presented. A method for selecting training segments involves mouse click-and-drag operations over a time bar representing the duration of the video; similarity results are displayed as shades in the time bar. Another method involves selecting periodic frames of the video as endpoints for the training segment.
    • 公开了一种用于交互式选择视频组合的视频相似性搜索的视频的训练图像和用于显示相似性搜索的结果的方法。 用户选择视频中的时间间隔作为用于训练图像类统计模型的训练图像的查询定义。 时间间隔可以短到一帧,或者由不相交的片段或镜头组成。 从训练图像变换中提取的特征向量,计算由训练图像定义的图像类别的统计模型。 对于视频中的每个帧,从帧的变换中提取特征向量,并且使用特征向量和图像类统计模型来计算相似度度量。 相似性度量是从产生帧的高斯模型的可能性得出的。 然后以图形方式呈现相似性,这允许视频的时间结构可视化和浏览。 也可以为其他视频文件快速计算相似度,从而实现基于内容的检索。 介绍了具有交互式相似度测量功能的内容感知视频浏览器。 用于选择训练段的方法涉及通过表示视频持续时间的时间条来进行鼠标点击和拖动操作; 相似度结果在时间栏中显示为阴影。 另一种方法是选择视频的周期帧作为训练段的端点。
    • 4. 发明授权
    • Methods and apparatuses for segmenting an audio-visual recording using image similarity searching and audio speaker recognition
    • 用于使用图像相似性搜索和音频扬声器识别分割视听记录的方法和装置
    • US06404925B1
    • 2002-06-11
    • US09266561
    • 1999-03-11
    • Jonathan T. FooteLynn Wilcox
    • Jonathan T. FooteLynn Wilcox
    • G06K962
    • G06K9/00758G06F17/30746G06F17/30787G10L17/00G11B27/28Y10S707/99931Y10S707/99933
    • Methods for segmenting audio-video recording of meetings containing slide presentations by one or more speakers are described. These segments serve as indexes into the recorded meeting. If an agenda is provided for the meeting, these segments can be labeled using information from the agenda. The system automatically detects intervals of video that correspond to presentation slides. Under the assumption that only one person is speaking during an interval when slides are displayed in the video, possible speaker intervals are extracted from the audio soundtrack by finding these regions. Since the same speaker may talk across multiple slide intervals, the acoustic data from these intervals is clustered to yield an estimate of the number of distinct speakers and their order. Clustering the audio data from these intervals yields an estimate of the number of different speakers and their order. Merged clustered audio intervals corresponding to a single speaker are then used as training data for a speaker segmentation system. Using speaker identification techniques, the full video is then segmented into individual presentations based on the extent of each presenter's speech. The speaker identification system optionally includes the construction of a hidden Markov model trained on the audio data from each slide interval. A Viterbi assignment then segments the audio according to speaker.
    • 描述了由一个或多个扬声器分割包含幻灯片呈现的会议音频视频记录的方法。 这些段作为记录会议的索引。 如果为会议提供议程,则可以使用来自议程的信息来标记这些细分。 系统自动检测与演示幻灯片相对应的视频间隔。 假设在视频中显示幻灯片的间隔期间只有一个人正在说话,通过查找这些区域,可以从音频音轨提取可能的扬声器间隔。 由于相同的说话者可以在多个幻灯片间隔中进行交谈,所以将来自这些间隔的声学数据进行聚类,以产生不同扬声器数量及其顺序的估计。 从这些间隔聚集音频数据产生不同扬声器数量及其顺序的估计。 然后将对应于单个扬声器的合并的群集音频间隔用作用于讲话者分割系统的训练数据。 使用扬声器识别技术,根据每位演讲者的讲话范围,将完整的视频分割成单独的演示文稿。 扬声器识别系统可选地包括针对来自每个幻灯片间隔的音频数据训练的隐马尔可夫模型的构造。 维特比分配然后根据扬声器分割音频。
    • 6. 发明授权
    • Methods and apparatuses for video segmentation, classification, and retrieval using image class statistical models
    • 使用图像类统计模型进行视频分割,分类和检索的方法和装置
    • US06751354B2
    • 2004-06-15
    • US09266637
    • 1999-03-11
    • Jonathan T. FooteLynn WilcoxAndreas Girgensohn
    • Jonathan T. FooteLynn WilcoxAndreas Girgensohn
    • G06K962
    • G06K9/00758G06K9/6277G06K9/6297
    • Techniques for classifying video frames using statistical models of transform coefficients are disclosed. After optionally being decimated in time and space, image frames are transformed using a discrete cosine transform or Hadamard transform. The methods disclosed model image composition and operate on grayscale images. The resulting transform matrices are reduced using truncation, principal component analysis, or linear discriminant analysis to produce feature vectors. Feature vectors of training images for image classes are used to compute image class statistical models. Once image class statistical models are derived, individual frames are classified by the maximum likelihood resulting from the image class statistical models. Thus, the probabilities that a feature vector derived from a frame would be produced from each of the image class statistical models are computed. The frame is classified into the image class corresponding to the image class statistical model which produced the highest probability for the feature vector derived from the frame. Optionally, frame sequence information is taken into account by applying a hidden Markov model to represent image class transitions from the previous frame to the current frame. After computing all class probabilities for all frames in the video or sequence of frames using the image class statistical models and the image class transition probabilities, the final class is selected as having the maximum likelihood. Previous frames are selected in reverse order based upon their likelihood given determined current states.
    • 公开了使用变换系数的统计模型对视频帧进行分类的技术。 在可选地在时间和空间中抽取后,使用离散余弦变换或Hadamard变换来转换图像帧。 该方法公开了模型图像组合,并对灰度图像进行操作。 所得到的变换矩阵使用截断,主成分分析或线性判别分析来减少以产生特征向量。 用于图像类的训练图像的特征向量用于计算图像类的统计模型。 一旦导出了图像类统计模型,则通过图像类统计模型产生的最大似然分类各个帧。 因此,计算从每个图像类统计模型产生从帧导出的特征向量的概率。 该帧被分类为对应于从帧产生的特征向量产生最高概率的图像类统计模型的图像类别。 可选地,通过应用隐马尔科夫模型来表示从先前帧到当前帧的图像类转换来考虑帧序列信息。 在使用图像类统计模型和图像类转换概率计算帧的视频或序列中的所有帧的所有类概率之后,选择最终类具有最大似然。 根据给定确定的当前状态的可能性,以相反的顺序选择先前的帧。
    • 8. 发明授权
    • Unusual event detection via collaborative video mining
    • 通过协作视频挖掘异常事件检测
    • US08009193B2
    • 2011-08-30
    • US11446893
    • 2006-06-05
    • Hanning ZhouDon KimberLynn Wilcox
    • Hanning ZhouDon KimberLynn Wilcox
    • H04N7/12
    • G06K9/00771G06K9/6218
    • Embodiments of the present invention describe a collaborative framework for mining of surveillance videos to detect abnormal events, which introduces a two-stage training process to alleviate the high false alarm problem. In the first stage, unsupervised clustering is performed on the segments of the video streams and a set of abnormal events are combined with user feedback to generate a clean training set. In the second stage, the clean training set is used to train a more precise model for the analysis of normal events and the motion detection results from multiple cameras can be cross validated and combined. This description is not intended to be a complete description of, or limit the scope of, the invention. Other features, aspects, and objects of the invention can be obtained from a review of the specification, the figures, and the claims.
    • 本发明的实施例描述了用于检测异常事件的监视视频挖掘的协同框架,其引入了两阶段训练过程以减轻高错误警报问题。 在第一阶段,对视频流的片段执行无监督的聚类,并且将一组异常事件与用户反馈相结合以产生干净的训练集。 在第二阶段,清洁训练集用于训练更准确的模型,用于分析正常事件,并且可以交叉验证并组合来自多个摄像机的运动检测结果。 本说明书不是对本发明的完整描述或限制本发明的范围。 本发明的其它特征,方面和目的可以通过对说明书,附图和权利要求的评述来获得。
    • 9. 发明授权
    • Methods and interfaces for event timeline and logs of video streams
    • 事件时间线和视频流日志的方法和接口
    • US07996771B2
    • 2011-08-09
    • US11324971
    • 2006-01-03
    • Andreas GirgensohnFrank M. ShipmanLynn Wilcox
    • Andreas GirgensohnFrank M. ShipmanLynn Wilcox
    • G06F3/00
    • G06F17/3079G06F17/30802G06F17/30843G08B13/19682
    • Techniques for generating timelines and event logs from one or more fixed-position cameras based on the identification of activity in the video are presented. Various embodiments of the invention include an assessment of the importance of the activity, the creation of a timeline identifying events of interest, and interaction techniques for seeing more details of an event or alternate views of the video. In one embodiment, motion detection is used to determine activity in one or more synchronized video streams. In another embodiment, events are determined based on periods of activity and assigned importance assessments based on the activity, important locations in the video streams, and events from other sensors. In different embodiments, the interface consists of a timeline, event log, and map.
    • 提出了基于视频中的活动识别从一个或多个固定位置摄像机生成时间线和事件日志的技术。 本发明的各种实施例包括评估活动的重要性,创建识别感兴趣事件的时间线以及用于查看视频的事件或替代视图的更多细节的交互技术。 在一个实施例中,运动检测用于确定一个或多个同步视频流中的活动。 在另一个实施例中,基于活动的周期和基于活动的重要性评估,视频流中的重要位置以及来自其他传感器的事件来确定事件。 在不同的实施例中,接口由时间线,事件日志和地图组成。