Unsupervised scene detection and commentator building using multi-modal chains

Gert-Jan Poulisse, Georgios Patsis, Marie-Francine Moens, Borko Furht (Editor)

Research output: Contribution to journalArticlepeer-review

3 Citations (Scopus)

Abstract

This paper presents a novel unsupervised method for identifying the semantic
structure in long semi-structured video streams. We identify chains, i.e., local clusters of
repeated features from both the video stream and audio transcripts. Each chain serves as an
indicator that the temporal interval it demarcates is part of the same semantic event. By
layering all the chains over each other, dense regions emerge from the overlapping chains,
from which we can identify the semantic structure of the video. We present two clustering
strategies that accomplish this task, and compare them against a baseline Scene Transition
Graph approach. We then develop a commentator that provides a semantic labeling of the
resultant video segmentation.
Original languageEnglish
Pages (from-to)159-175
Number of pages16
JournalMultimedia Tools and Applications
Volume70
Issue number1
Publication statusPublished - 1 May 2014

Bibliographical note

Borko Furht

Keywords

  • Semantic event detection
  • Feature extraction
  • . Multi-modal scene segmentation
  • Video summarization

Fingerprint

Dive into the research topics of 'Unsupervised scene detection and commentator building using multi-modal chains'. Together they form a unique fingerprint.

Cite this