MULTIMODAL INFORMATION FUSION AND TEMPORAL INTEGRATION FOR VIOLENCE DETECTION IN MOVIES

Cédric Penet 1, 2 Claire-Hélène Demarty 1 Guillaume Gravier 2 Patrick Gros 2
2 TEXMEX - Multimedia content-based indexing
IRISA - Institut de Recherche en Informatique et Systèmes Aléatoires, Inria Rennes – Bretagne Atlantique
Abstract : This paper presents a violent shots detection system that studies several methods for introducing temporal and multimodal information in the framework. It also investigates different kinds of Bayesian network structure learning algorithms for modelling these problems. The system is trained and tested using the MediaEval 2011 Affect Task corpus, which comprises of 15 Hollywood movies. It is experimentally shown that both multimodality and temporality add interesting information into the system. Moreover, the analysis of the links between the variables of the resulting graphs yields important observations about the quality of the structure learning algorithms. Overall, our best system achieved 50% false alarms and 3% missed detection, which is among the best submissions in the MediaEval campaign.
Document type :
Conference papers
Complete list of metadatas

Cited literature [8 references]  Display  Hide  Download

https://hal.inria.fr/hal-00671016
Contributor : Cédric Penet <>
Submitted on : Thursday, February 16, 2012 - 2:42:38 PM
Last modification on : Friday, November 16, 2018 - 1:22:33 AM
Long-term archiving on : Thursday, June 14, 2012 - 4:35:39 PM

File

CedricPenet_MultimodalInformat...
Files produced by the author(s)

Identifiers

  • HAL Id : hal-00671016, version 1

Citation

Cédric Penet, Claire-Hélène Demarty, Guillaume Gravier, Patrick Gros. MULTIMODAL INFORMATION FUSION AND TEMPORAL INTEGRATION FOR VIOLENCE DETECTION IN MOVIES. ICASSP - 37th International Conference on Acoustics, Speech, and Signal Processing (2012), Mar 2012, Kyoto, Japan. ⟨hal-00671016⟩

Share

Metrics

Record views

1019

Files downloads

499