scispace - formally typeset
Open AccessProceedings ArticleDOI

Video Frame Synthesis Using Deep Voxel Flow

Reads0
Chats0
TLDR
Deep voxel flow as mentioned in this paper combines the advantages of optical flow and neural network-based methods by training a deep network that learns to synthesize video frames by flowing pixel values from existing ones, which can be applied at any video resolution.
Abstract
We address the problem of synthesizing new video frames in an existing video, either in-between existing frames (interpolation), or subsequent to them (extrapolation). This problem is challenging because video appearance and motion can be highly complex. Traditional optical-flow-based solutions often fail where flow estimation is challenging, while newer neural-network-based methods that hallucinate pixel values directly often produce blurry results. We combine the advantages of these two methods by training a deep network that learns to synthesize video frames by flowing pixel values from existing ones, which we call deep voxel flow. Our method requires no human supervision, and any video can be used as training data by dropping, and then learning to predict, existing frames. The technique is efficient, and can be applied at any video resolution. We demonstrate that our method produces results that both quantitatively and qualitatively improve upon the state-of-the-art.

read more

Citations
More filters
Proceedings ArticleDOI

Scale-Recurrent Network for Deep Image Deblurring

TL;DR: A Scale-recurrent Network (SRN-DeblurNet) is proposed and shown to produce better quality results than state-of-the-arts, both quantitatively and qualitatively in single image deblurring.
Proceedings ArticleDOI

Super SloMo: High Quality Estimation of Multiple Intermediate Frames for Video Interpolation

TL;DR: In this paper, an end-to-end convolutional neural network is proposed for variable-length multi-frame video interpolation, where the motion interpretation and occlusion reasoning are jointly modeled.
Proceedings ArticleDOI

Video Frame Interpolation via Adaptive Separable Convolution

TL;DR: In this article, a deep fully convolutional neural network is proposed to estimate pairs of 1D kernels for all pixels simultaneously, which allows for the incorporation of perceptual loss to train the network to produce visually pleasing frames.
Journal ArticleDOI

Video Enhancement with Task-Oriented Flow

TL;DR: Task-Oriented Flow (TOFlow) as mentioned in this paper is a self-supervised, task-specific representation for low-level video processing, which is trained in a supervised manner.
Proceedings ArticleDOI

Detail-Revealing Deep Video Super-Resolution

TL;DR: In this article, a sub-pixel motion compensation (SPMC) layer is proposed to fuse multiple frames to reveal image details, which can generate visually and quantitatively high quality results without the need of parameter tuning.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Journal ArticleDOI

Image quality assessment: from error visibility to structural similarity

TL;DR: In this article, a structural similarity index is proposed for image quality assessment based on the degradation of structural information, which can be applied to both subjective ratings and objective methods on a database of images compressed with JPEG and JPEG2000.
Journal ArticleDOI

Generative Adversarial Nets

TL;DR: A new framework for estimating generative models via an adversarial process, in which two models are simultaneously train: a generative model G that captures the data distribution and a discriminative model D that estimates the probability that a sample came from the training data rather than G.
Proceedings Article

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Related Papers (5)