scispace - formally typeset
Journal ArticleDOI

Unsupervised Deep Image Fusion With Structure Tensor Representations

Reads0
Chats0
TLDR
Deep Image Fusion Network (DIF-Net) as discussed by the authors proposes an unsupervised loss function using the structure tensor representation of the multi-channel image contrasts, which is minimized by a stochastic deep learning solver with large-scale examples.
Abstract
Convolutional neural networks (CNNs) have facilitated substantial progress on various problems in computer vision and image processing. However, applying them to image fusion has remained challenging due to the lack of the labelled data for supervised learning. This paper introduces a deep image fusion network (DIF-Net), an unsupervised deep learning framework for image fusion. The DIF-Net parameterizes the entire processes of image fusion, comprising of feature extraction, feature fusion, and image reconstruction, using a CNN. The purpose of DIF-Net is to generate an output image which has an identical contrast to high-dimensional input images. To realize this, we propose an unsupervised loss function using the structure tensor representation of the multi-channel image contrasts. Different from traditional fusion methods that involve time-consuming optimization or iterative procedures to obtain the results, our loss function is minimized by a stochastic deep learning solver with large-scale examples. Consequently, the proposed method can produce fused images that preserve source image details through a single forward network trained without reference ground-truth labels. The proposed method has broad applicability to various image fusion problems, including multi-spectral, multi-focus, and multi-exposure image fusions. Quantitative and qualitative evaluations show that the proposed technique outperforms existing state-of-the-art approaches for various applications.

read more

Citations
More filters
Journal ArticleDOI

Image fusion meets deep learning: A survey and perspective

TL;DR: In this paper, a comprehensive review and analysis of latest deep learning methods in different image fusion scenarios is provided, and the evaluation for some representative methods in specific fusion tasks are performed qualitatively and quantitatively.
Journal ArticleDOI

Multi-focus image fusion: A Survey of the state of the art

TL;DR: A comprehensive overview of existing multi-focus image fusion methods is presented and a new taxonomy is introduced to classify existing methods into four main categories: transformdomain methods, spatial domain methods, methods combining transform domain and spatial domain, and deep learning methods.
Journal ArticleDOI

SESF-Fuse: an unsupervised deep model for multi-focus image fusion

TL;DR: A novel unsupervised deep learning model is proposed to address multi-focus image fusion problem and analyzes sharp appearance in deep feature instead of original image to achieve state-of-art fusion performance.
Journal ArticleDOI

Deep Coupled Feedback Network for Joint Exposure Fusion and Image Super-Resolution

TL;DR: Zhang et al. as discussed by the authors proposed a deep Coupled Feedback Network (CF-Net) to achieve multi-exposure image fusion (MEF) and image super-resolution (SR) simultaneously.
Journal ArticleDOI

Benchmarking and comparing multi-exposure image fusion algorithms

TL;DR: A benchmark for multi-exposure image fusion (MEFB) is proposed which consists of a test set of 100 image pairs, a code library of 16 algorithms, 20 evaluation metrics, 1600 fused images and a software toolkit and it is expected that MEFB will serve as an effective platform for researchers to compare performances and investigate MEF algorithms.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal ArticleDOI

Generative Adversarial Nets

TL;DR: A new framework for estimating generative models via an adversarial process, in which two models are simultaneously train: a generative model G that captures the data distribution and a discriminative model D that estimates the probability that a sample came from the training data rather than G.
Proceedings Article

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Book ChapterDOI

Microsoft COCO: Common Objects in Context

TL;DR: A new dataset with the goal of advancing the state-of-the-art in object recognition by placing the question of object recognition in the context of the broader question of scene understanding by gathering images of complex everyday scenes containing common objects in their natural context.
Related Papers (5)