A scalable framework for joint clustering and synchronizing multi-camera videos - Inria - Institut national de recherche en sciences et technologies du numérique Access content directly
Conference Papers Year : 2013

A scalable framework for joint clustering and synchronizing multi-camera videos

Ashish Bagri
  • Function : Author
  • PersonId : 946520
Franck Thudor
  • Function : Author
  • PersonId : 946521
Alexey Ozerov
  • Function : Author
  • PersonId : 930358
Pierre Hellier
  • Function : Author
  • PersonId : 946522

Abstract

This paper describes a method to cluster and synchronize large scale audio-video sequences recorded by multiple users during an event. The proposed method is designed to jointly cluster audio content and synchronize sequences in each cluster to create a multi-view presentation of the event. The method is roughly based on cross-correlation of local audio features. In this paper, three main contributions are presented to obtain a scalable and accurate framework. First, a salient representation of features is used to reduce the computation complexity while maintaining high performance. Second, an intermediate clustering step is introduced to limit the number of comparisons required. Third, a voting approach is proposed to avoid tuning thresholds for cross-correlation. This framework was tested on 164 YouTube concert videos and results demonstrated the efficiency of the method with a correct clustering of 98.8% of the sequences.
Fichier principal
Vignette du fichier
Bagri_et_al_EUSIPCO_2013.pdf (151.5 Ko) Télécharger le fichier
Origin : Files produced by the author(s)

Dates and versions

hal-00870381 , version 1 (07-10-2013)

Identifiers

  • HAL Id : hal-00870381 , version 1

Cite

Ashish Bagri, Franck Thudor, Alexey Ozerov, Pierre Hellier. A scalable framework for joint clustering and synchronizing multi-camera videos. 21st European Signal Processing Conference (EUSIPCO 2013), Sep 2013, Marrakech, Morocco. ⟨hal-00870381⟩
87 View
147 Download

Share

Gmail Facebook X LinkedIn More