Automatic Story Segmentation for TV News Video Using Multiple Modalities
Author(s) -
Émilie Dumont,
Georges Quénot
Publication year - 2012
Publication title -
international journal of digital multimedia broadcasting
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.164
H-Index - 17
eISSN - 1687-7586
pISSN - 1687-7578
DOI - 10.1155/2012/732514
Subject(s) - computer science , segmentation , context (archaeology) , task (project management) , process (computing) , artificial intelligence , modality (human–computer interaction) , annotation , modalities , video retrieval , information retrieval , pattern recognition (psychology) , paleontology , social science , management , sociology , economics , biology , operating system
While video content is often stored in rather large files or broadcasted in continuous streams, users are often interested in retrieving only a particular passage on a topic of interest to them. It is, therefore, necessary to split video documents or streams into shorter segments corresponding to appropriate retrieval units. We propose here a method for the automatic segmentation of TV news videos into stories. A-multiple-descriptor based segmentation approach is proposed. The selected multimodal features are complementary and give good insights about story boundaries. Once extracted, these features are expanded with a local temporal context and combined by an early fusion process. The story boundaries are then predicted using machine learning techniques. We investigate the system by experiments conducted using TRECVID 2003 data and protocol of the story boundary detection task, and we show that the proposed approach outperforms the state-of-the-art methods while requiring a very small amount of manual annotation.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom