One-Step Time-Dependent Future Video Frame Prediction with a Convolutional Encoder-Decoder Neural Network - Archive ouverte HAL Access content directly
Conference Papers Year : 2016

One-Step Time-Dependent Future Video Frame Prediction with a Convolutional Encoder-Decoder Neural Network

(1) , (2) , (1) , (1) , (2)
1
2

Abstract

There is an inherent need for machines to have a notion of how entities within their environment behave and to anticipate changes in the near future. In this work, we focus on anticipating future appearance, given the current frame of a video. Typical methods are used either to predict the next frame of a video or to predict future optical flow or trajectories based on a single video frame. This work presents an experiment on stretching the ability of CNNs to anticipate appearance at an arbitrarily given near future time, by conditioning our predicted video frames on a continuous time variable. We show that CNNs can learn an intrinsic representation of typical appearance changes over time and successfully generate realistic predictions in one step-at a deliberate time difference in the near future. The method is evaluated on the KTH human actions dataset and compared to a baseline consisting of an analogous CNN architecture that is not time-aware.
Fichier principal
Vignette du fichier
Vukotic_NCCV_2016.pdf (1.36 Mo) Télécharger le fichier
Origin : Files produced by the author(s)
Loading...

Dates and versions

hal-01467064 , version 1 (14-02-2017)

Identifiers

  • HAL Id : hal-01467064 , version 1

Cite

Vedran Vukotić, Silvia-Laura Pintea, Christian Raymond, Guillaume Gravier, Jan van Gemert. One-Step Time-Dependent Future Video Frame Prediction with a Convolutional Encoder-Decoder Neural Network. Netherlands Conference on Computer Vision (NCCV), Dec 2016, Lunteren, Netherlands. ⟨hal-01467064⟩
818 View
784 Download

Share

Gmail Facebook Twitter LinkedIn More