scispace - formally typeset
Proceedings ArticleDOI

Real-Time Neural Style Transfer for Videos

TLDR
This work proposes a hybrid loss to capitalize on the content information of input frames, the style information of a given style image, and the temporal information of consecutive frames to calculate the temporal loss during the training stage.
Abstract
Recent research endeavors have shown the potential of using feed-forward convolutional neural networks to accomplish fast style transfer for images. In this work, we take one step further to explore the possibility of exploiting a feed-forward network to perform style transfer for videos and simultaneously maintain temporal consistency among stylized video frames. Our feed-forward network is trained by enforcing the outputs of consecutive frames to be both well stylized and temporally consistent. More specifically, a hybrid loss is proposed to capitalize on the content information of input frames, the style information of a given style image, and the temporal information of consecutive frames. To calculate the temporal loss during the training stage, a novel two-frame synergic training mechanism is proposed. Compared with directly applying an existing image style transfer method to videos, our proposed method employs the trained network to yield temporally consistent stylized videos which are much more visually pleasant. In contrast to the prior video style transfer method which relies on time-consuming optimization on the fly, our method runs in real time while generating competitive visual results.

read more

Content maybe subject to copyright    Report

Citations
More filters
Posted Content

Video-to-Video Synthesis

TL;DR: In this article, a video-to-video synthesis approach under the generative adversarial learning framework is proposed, which achieves high-resolution, photorealistic, temporally coherent video results on a diverse set of input formats.
Posted Content

Neural Style Transfer: A Review

TL;DR: A comprehensive overview of the current progress in NST can be found in this paper, where the authors present several evaluation methods and compare different NST algorithms both qualitatively and quantitatively, concluding with a discussion of various applications of NST and open problems for future research.
Journal ArticleDOI

Neural Style Transfer: A Review

TL;DR: A taxonomy of current algorithms in the field of NST is proposed and several evaluation methods are presented and compared to compare different NST algorithms both qualitatively and quantitatively.
Book ChapterDOI

Learning Blind Video Temporal Consistency

TL;DR: An efficient approach based on a deep recurrent network for enforcing temporal consistency in a video that can handle multiple and unseen tasks, including but not limited to artistic style transfer, enhancement, colorization, image-to-image translation and intrinsic image decomposition.
Journal ArticleDOI

Consistent video depth estimation

TL;DR: An algorithm for reconstructing dense, geometrically consistent depth for all pixels in a monocular video by using a learning-based prior, i.e., a convolutional neural network trained for single-image depth estimation.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Proceedings ArticleDOI

ImageNet: A large-scale hierarchical image database

TL;DR: A new database called “ImageNet” is introduced, a large-scale ontology of images built upon the backbone of the WordNet structure, much larger in scale and diversity and much more accurate than the current image datasets.
Proceedings ArticleDOI

Fully convolutional networks for semantic segmentation

TL;DR: The key insight is to build “fully convolutional” networks that take input of arbitrary size and produce correspondingly-sized output with efficient inference and learning.