Story segmentation in news videos using visual and text cues

Authors

    Authors

    Y. Zhai; A. Yilmaz;M. Shah

    Comments

    Authors: contact us about adding a copy of your work at STARS@ucf.edu

    Keywords

    Computer Science, Theory & Methods; Imaging Science & Photographic; Technology

    Abstract

    In this paper, we present a framework for segmenting the news programs into different story topics. The proposed method utilizes both visual and text information of the video. We represent the news video by a Shot Connectivity Graph (SCG), where the nodes in the graph represent the shots in the video, and the edges between nodes represent the transitions between shots. The cycles in the graph correspond to the story segments in the news program. We first detect the cycles in the graph by finding the anchor persons in the video. This provides us with the coarse segmentation of the news video. The initial segmentation is later refined by the detections of the weather and sporting news, and the merging of similar stories. For the weather detection, the global color information of the images and the motion of the shots are considered. We have used the text obtained from automatic speech recognition (ASR) for detecting the potential sporting shots to form the sport stories. Adjacent stories with similar semantic meanings are further merged based on the visual and text similarities. The proposed framework has been tested on a widely used data set provided by NIST, which contains the ground truth of the story boundaries, and competitive evaluation results have been obtained.

    Journal Title

    Image and Video Retrieval, Proceedings

    Volume

    3568

    Publication Date

    1-1-2005

    Document Type

    Article

    Language

    English

    First Page

    92

    Last Page

    102

    WOS Identifier

    WOS:000231148700013

    ISSN

    0302-9743; 3-540-27858-3

    Share

    COinS