scispace - formally typeset
Search or ask a question
Proceedings ArticleDOI

Image fusion using multivariate and multidimensional EMD

TL;DR: It is demonstrated that while multidimensional extensions, by design, may seem more appropriate for tasks related to image processing, the proposed multivariate extension outperforms these in image fusion applications owing to its mode-alignment property for IMFs.
Abstract: We present a novel methodology for the fusion of multiple (two or more) images using the multivariate extension of empirical mode decomposition (MEMD). Empirical mode decomposition (EMD) is a data-driven method which decomposes input data into its intrinsic oscillatory modes, known as intrinsic mode functions (IMFs), without making a priori assumptions regarding the data. We show that the multivariate and multidimensional extensions of EMD are suitable for image fusion purposes. We further demonstrate that while multidimensional extensions, by design, may seem more appropriate for tasks related to image processing, the proposed multivariate extension outperforms these in image fusion applications owing to its mode-alignment property for IMFs. Case studies involving multi-focus image fusion and pan-sharpening of multi-spectral images are presented to demonstrate the effectiveness of the proposed method.

Summary (2 min read)

1. INTRODUCTION

  • The authors propose a hybrid (multi-scale and pixel-level) and data-driven scheme for image fusion based on multivariate extensions of empirical mode decomposition (MEMD) algorithm [6] .
  • The authors also compare their results with the standard bi-dimensional EMD [7] based fusion approach.
  • The EMD based fusion methods are employed since they are fully data adaptive, enable fusion of intrinsic scales at local level, and allow fusion of matched spatial frequency content between input images.
  • Standard multiscale methods (based on Fourier and wavelet transform) employ static filter banks and predefined basis functions which hinder the fusion of matched spatial frequency content between input images.
  • In both cases, the fusion results obtained from the proposed scheme outperforms the results obtained by BDEMD both qualitatively and quantitatively.

2. EMD AND ITS MULTIVARIATE AND MULTIDIMENSIONAL EXTENSIONS

  • The recursive sifting algorithm operates by defining the upper and lower envelopes of an input signal by interpolating its extrema.
  • The local mean m(k) is then estimated by averaging these envelopes, which is subsequently subtracted from the input signal x(k) to obtain the fast oscillating signal d(k) = x(k) − m(k).
  • The sifting process stops when d(k) has inadequate extrema.

2.2. Bi-dimensional EMD (BDEMD)

  • Bi-dimensional EMD [7] is a generic extension of EMD for images.
  • Various algorithms for computing BDEMD decomposition exist which mainly differ in the way the extrema are interpolated to obtain upper and lower envelopes.
  • Radial basis functions (tensor product) or B-splines are commonly used methods for interpolation [7] , whereas the method by Linderhed [8] uses thin-plate splines for the interpolation of the extrema.

2.3. Multivariate EMD (MEMD)

  • Multivariate EMD (MEMD) algorithm extends the functionality of EMD to signals containing multiple channels [6] .
  • The rationale behind the MEMD is to separate inherent rotations (rather than oscillations) within a signal.
  • This is achieved by estimating the local mean of a multivariate signal in multidimensional spaces where the signal resides.
  • For multivariate signals, however, the concept of extrema cannot be defined in clear terms and therefore envelopes cannot be obtained as a trivial extension of univariate case.
  • Note that the MEMD produced diagonally dominant correlograms of IMFs as compared to the BDEMD, proving that the same-indexed IMFs generated from MEMD are highly correlated, a major requirement in most fusion applications.

4. MEMD-AND BDEMD-BASED IMAGE FUSION

  • The block diagram of the proposed multivariate EMD based fusion algorithm is shown in Fig. 2 .
  • The BDEMD based fusion algorithm operates similarly to the MEMD algorithm illustrated above.
  • Note that owing to the empirical nature of the EMD algorithm, typically different number of IMFs are obtained for multiple input images resulting in mismatched IMFs, thus hindering the fusion process.

5. CASE STUDY 1: MULTI-FOCUS IMAGE FUSION

  • Fig. 3 shows a subset of input multi-focus images and the fused images obtained from the two methods; only three out of seven input images are shown due to the space restrictions.
  • Note from Figs. 3(a-c ) that each input image has some specific objects within focus: Image 1, for instance, focuses on the nearest objects such as the coin, whereas the Image 2 and Image 3 focus on the middle and the farthest objects respec- shows pie-charts highlighting the number of cases in terms of percentage where each method performed best.
  • It can be observed that for E and SF measures, the MEMD yielded superior results for approximately all input data sets, whereas for QABF , the MEMD produced better results for 83% of the input data sets.

6. CASE STUDY 2: PAN-SHARPENING

  • The authors next performed experiments for Pan-sharpening of multispectral (MS) images using the existing BDEMD-based fusion and the proposed MEMD-based fusion algorithms.
  • The ground truth for both data sets was also available (not shown in Fig. 5 due to space restrictions) and was used for the quantitative analysis of the fusion results.
  • The details added by the fusion process were extracted by subtracting the original intensity plane I from Î, which were then separately added to the B, G, R, and NIR components of the MS image to obtain the Pan-sharpened MS image.
  • The authors employed the following set of performance metrics for this purpose: i) Relative dimensionless global error in synthesis (ER, ideally 0), ii) Spectral Angle Mapper (S, ideally 0), and iii) Quaternion Index (Q4, ideally 100%) [2] .
  • Please observe that in both data sets the proposed scheme performed better than AWT and BDEMD fusion methods for all performance metrics with an exception to S value for the Tolouse image where AWT performed better.

7. CONCLUSIONS

  • The authors have presented a method for the fusion of multiple images using multivariate empirical mode decomposition (MEMD) algorithm.
  • The superiority of the method has been demonstrated on a large data set for two applications: i) multi-focus fusion, and ii) pan-sharpening of multi-spectral images.
  • In addition to the qualitative analysis, the authors have also employed a wide range of quantitative performance measures to compare the fusion results obtained from the two approaches.

Did you find this useful? Give us your feedback

Content maybe subject to copyright    Report

IMAGE FUSION USING MULTIVARIATE AND MULTIDIMENSIONAL EMD
Naveed ur Rehman
1
M. Murtaza Khan
2
Ishaq Sohaib
1
M. Jehanzaib
1
Shoaib Ehsan
3
Klaus McDonald-Maier
3
1
Department of Electrical Engineering, COMSATS Institute of Information Technology, Park Road, Islamabad, Pakistan
2
School of Electrical Engineering and Computer Science, National University of Sciences and Technology, Islamabad, Pakistan
3
School of Computer Science and Electronic Engineering, University of Essex, Wivenhoe Park, Colchester CO4 3SQ, UK
ABSTRACT
We present a novel methodology for the fusion of multiple
(two or more) images using the multivariate extension of
empirical mode decomposition (MEMD). Empirical mode
decomposition (EMD) is a data-driven method which decom-
poses input data into its intrinsic oscillatory modes, known
as intrinsic mode functions (IMFs), without making a priori
assumptions regarding the data. We show that the multivari-
ate and multidimensional extensions of EMD are suitable
for image fusion purposes. We further demonstrate that
while multidimensional extensions, by design, may seem
more appropriate for tasks related to image processing, the
proposed multivariate extension outperforms these in image
fusion applications owing to its mode-alignment property
for IMFs. Case studies involving multi-focus image fusion
and pan-sharpening of multi-spectral images are presented to
demonstrate the effectiveness of the proposed method.
Index Terms Empirical mode decomposition (EMD),
Multivariate EMD (MEMD), Bidimensional EMD, Multi-
focus image fusion, Pan-sharpening.
1. INTRODUCTION
Image fusion is the process of combining multiple images to
produce a single image which carries more information than
any of the images used for blending [1]. Fusion techniques
are useful for cases where the limitations of optical sensors
and imaging conditions make it difficult to view multiple ob-
jects clearly in a single image. In such cases, multiple images
are obtained with each containing partial information about
a scene. Multi-focus and Multi-exposure images are two ex-
ample classes in which the relevant objects may be obscured
as a result of either being out of focus or not being prop-
erly exposed to the light source. These sets of images can be
merged to present the complete information in a single image
via multi-focus or multi-exposure image fusion.
Similarly, in remote sensing applications, we often require
both high spatial and spectral information in a single image
which is not physically possible to obtain via available sen-
sors. In such cases, fusion of high spatial resolution panchro-
matic image (PAN) and low spatial resolution (but high spec-
tral resolution) multispectral (MS) images is performed to ob-
tain the desired high spatial and spectral resolution MS image.
This process is also called Pan-sharpening [2].
Image fusion methods may be characterized as pixel-level
fusion, multi-scale fusion and hybrid fusion techniques. The
key steps of pixel level fusion include: i) generation of a quan-
titative map of information content for each image; ii) com-
parison of information content at pixel level; iii) assigning
weights to individual pixels (or a set of pixels) based on in-
formation content; iv) and weighted recombination to obtain
fused image. The advantages of such class of methods in-
clude their low computational cost and simplicity, while the
main disadvantage is their susceptibility to noise. Multi-scale
techniques, on the other hand, operate by first decomposing
input images in terms of their frequency components which
are then combined to obtain a single fused image. Here, the
main steps include: i) converting input images into transform
domain coefficients; ii) assigning weights to the coefficients
based on information content; iii) selecting the relevant coeffi-
cients; and iv) taking the inverse transform. Typical examples
are the methods based on Gaussian pyramids, Fast Fourier
Transform (FFT), Discrete Cosine Transform (DCT) [3], and
Discrete Wavelet Transform (DWT) [4].
We propose a hybrid (multi-scale and pixel-level) and
data-driven scheme for image fusion based on multivariate
extensions of empirical mode decomposition (MEMD) al-
gorithm [6]. We also compare our results with the standard
bi-dimensional EMD (BDEMD) [7] based fusion approach.
The EMD based fusion methods are employed since they
are fully data adaptive, enable fusion of intrinsic scales at
local level, and allow fusion of matched spatial frequency
content between input images. Standard multiscale methods
(based on Fourier and wavelet transform) employ static filter
banks and predefined basis functions which hinder the fusion
of matched spatial frequency content between input images.
We demonstrate the potential of the proposed scheme in two
application scenarios: a) multi-focus image fusion; and b)
pan-sharpening of MS images. In both cases, the fusion re-
sults obtained from the proposed scheme outperforms the
results obtained by BDEMD both qualitatively and quantita-
tively.

2. EMD AND ITS MULTIVARIATE AND
MULTIDIMENSIONAL EXTENSIONS
2.1. Standard EMD
Empirical mode decomposition (EMD) [5] is a data-driven
method which decomposes an arbitrary signal x(k) into a set
of multiple oscillatory components called the intrinsic mode
functions (IMFs) via an iterative process known as sifting al-
gorithm [6]. The IMFs represent the intrinsic temporal modes
(scales) that are present in the input data which when added
together reproduce the input x(k), as shown in eq. (1) below:
x(k) =
M
X
m=1
c
m
(k) + r(k) (1)
The residual r(k) does not contain any oscillations and repre-
sents a trend within the signal.
The recursive sifting algorithm operates by defining the
upper and lower envelopes of an input signal by interpolat-
ing its extrema. The local mean m(k) is then estimated by
averaging these envelopes, which is subsequently subtracted
from the input signal x(k) to obtain the fast oscillating sig-
nal d(k) = x(k) m(k). Next, d(k) is checked for an IMF
condition; if it is not satisfied, the process is repeated until
the condition for IMF is satisfied and we obtain an IMF. The
sifting process stops when d(k) has inadequate extrema.
2.2. Bi-dimensional EMD (BDEMD)
Bi-dimensional EMD (BDEMD) [7] is a generic extension
of EMD for images. Various algorithms for computing
BDEMD decomposition exist which mainly differ in the way
the extrema are interpolated to obtain upper and lower en-
velopes. Radial basis functions (tensor product) or B-splines
are commonly used methods for interpolation [7], whereas
the method by Linderhed [8] uses thin-plate splines for the
interpolation of the extrema.
2.3. Multivariate EMD (MEMD)
Multivariate EMD (MEMD) algorithm extends the function-
ality of EMD to signals containing multiple channels [6]. The
rationale behind the MEMD is to separate inherent rotations
(rather than oscillations) within a signal. This is achieved by
estimating the local mean of a multivariate signal in multidi-
mensional spaces where the signal resides. For multivariate
signals, however, the concept of extrema cannot be defined
in clear terms and therefore envelopes cannot be obtained as
a trivial extension of univariate case. To address this issue,
MEMD operates by projecting an input multivariate signal in
V uniformly spaced directions on a unit p-sphere; the extrema
of the so projected signals are then interpolated to obtain mul-
tiple envelopes which are subsequently averaged to obtain the
local mean.
3. MEMD VS BDEMD: MODE ALIGNMENT
Fig. 1 shows correlations of normalized IMFs for two multi-
focus images (Fig. 3(a) and Fig. 3(c)), obtained using the
IMF index
IMF index
1 2 3 4 5 6
1
2
3
4
5
6
IMF index
IMF index
2 4 6 8
2
4
6
8
0.2
0.4
0.6
0.8
Fig. 1: Cross-correlation of normalised IMFs for multi-focus
images (left) BDEMD (right) MEMD.
1
ˆ
I
MEMD
I
1
P
I
1
1
1
2
I
2
P
I
2
1
I
2
2
I
M
P
I
M
1
I
M
2
Pixel
Fusion
Pixel
Fusion
Pixel
Fusion
M
I
ˆ
2
ˆ
I
I
ˆ
Fused Image
Input Images
Fig. 2: The proposed scheme illustrating the local fusion of
P arbitrary images to yield a single fused image
ˆ
I using the
MEMD algorithm.
BDEMD (left) and MEMD (right). Note that the MEMD
produced diagonally dominant correlograms of IMFs as com-
pared to the BDEMD, proving that the same-indexed IMFs
generated from MEMD are highly correlated, a major require-
ment in most fusion applications. This mode alignment prop-
erty of MEMD is a result of direct processing of input images
within MEMD, whereas the lack of it in BDEMD is due to
the fact that it processes multiple input images separately.
4. MEMD- AND BDEMD-BASED IMAGE FUSION
The proposed algorithm based on MEMD operates by first
converting the P input images into a vector form by concate-
nating their rows/columns. The resulting vectors are then put
together to form a multivariate signal containing P number of
data channels. MEMD is next applied to the resulting signal
yielding M number of IMFs for each channel; let us denote
the mth IMF of the pth channel (input image) by I
p
m
(a, b),
where m = 1 . . . M; n = 1 . . . P ; and a and b represent
the spatial coordinates. To perform fusion at the pixel level,

(a) Image 1 (b) Image 3 (c) Image 7 (d) BDEMD (e) MEMD
Fig. 3: Multi-focus image fusion results for data set 1.
sub-images I
p
m
(a, b) are divided into small windows of size
N × N and their variances ξ
p
m
(a, b) are computed. Next, the
sub-images I
p
m
(a, b) are assigned local weights, W F
p
m
(a, b),
based on ξ
p
m
(a, b) by using the following relation:
W F
p
m
(a, b) =
ξ
p
m
(a, b)
P
P
p=1
ξ
p
m
(a, b)
(2)
This means that the IMFs exhibiting greater variance are as-
signed higher weights W F
p
m
(a, b) than those exhibiting lower
variances, thereby maximising their contribution to the fused
image. To obtain the mth IMF of the fused image, the
IMFs of all P input images {I
p
m
(a, b)}
P
p=1
are multiplied by
their respective weight factors {W F
p
m
(a, b)}
P
p=1
and added
together to obtain:
ˆ
I
m
(a, b) =
P
X
p=1
W F
p
m
(a, b) × I
p
m
(a, b) (3)
This procedure is repeated for all M IMFs to obtain a set of
fused IMFs {
ˆ
I
m
(a, b)}
M
m=1
which are added together to yield
the fused image
ˆ
I.
ˆ
I(a, b) =
M
X
m=1
ˆ
I
m
(a, b) (4)
The block diagram of the proposed multivariate EMD based
fusion algorithm is shown in Fig. 2.
The BDEMD based fusion algorithm operates similarly to
the MEMD algorithm illustrated above. The only difference
is that instead of a single operation of MEMD on all input
images, BDEMD is applied separately on P input images to
obtain M
p
IMFs, where p = 1 . . . M. Note that owing to
the empirical nature of the EMD algorithm, typically differ-
ent number of IMFs are obtained for multiple input images
resulting in mismatched IMFs, thus hindering the fusion pro-
cess.
5. CASE STUDY 1: MULTI-FOCUS IMAGE FUSION
As a case study, we first performed fusion of multi-focus im-
ages using MEMD and compared the results with those ob-
tained from BDEMD based fusion approach. For this pur-
pose, multiple images of 30 different scenes were used; seven
images were taken of each scene with different parts of the
scene out-of-focus in each image.
For quantitative evaluation of the fusion results, we
have employed Entropy (E) [11], objective image fusion
(QABF ) [10] and the spatial frequency (SF ) [9] perfor-
mance measures. Collectively, these measures served to
quantify the spatial distortion as well as the information
present in the fused image. The parameters used in MEMD
algorithm were: number of direction vectors V = 8 and the
square window length of N = 9. The standard BDEMD
algorithm given in [7] was implemented for fusion purposes.
Fig. 3 shows a subset of input multi-focus images and the
1 5 10 15 20 25 30
0
20
40
60
80
Entropy score
100%
1 5 10 15 20 25 30
0
2
4
6
SF score
100%
1 5 10 15 20 25 30
0
0.2
0.4
0.6
0.8
Data set index
QABF score
83%
17%
Fig. 4: Quantitative comparison of the proposed fusion
schemes on 30 input multi-focus images. (Left column)
Bar graphs of the values of quantitative measures, including
%Entropy, %SF, and QABF shown respectively from top to
bottom, obtained for the MEMD- (dark green), and BDEMD-
(yellow) based fusion methods. (Right column) Pie charts
of the quantitative measures highlighting the relative perfor-
mance of the MEMD (dark green), and BDEMD (yellow).
fused images obtained from the two methods; only three out
of seven input images are shown due to the space restrictions.
Note from Figs. 3(a-c) that each input image has some spe-
cific objects within focus: Image 1, for instance, focuses on
the nearest objects such as the coin, whereas the Image 2 and
Image 3 focus on the middle and the farthest objects respec-

(a) MS (b) PAN (c) AWT (d) BDEMD (e) MEMD
Fig. 5: Fusion results for Pleiades Tolouse Image.
tively. The proposed MEMD-based fusion algorithm resulted
in a single output image, shown in Fig. 3(d), which has all
objects within focus. Similarly, the BDEMD-based fused
image also yielded an improved image, though not as sharp
as the one obtained via MEMD: Please observe the improved
sharpness of the word ‘DUTCH’ written on the key chain in
the MEMD-fused image. Similar trend was observed in most
of the 30 input multi-focus data sets used in our experiments,
with MEMD outperforming BDEMD-based fusion. This is
evident from Fig. 4 (left) which shows the bar graphs of the
values of the quantitative performance measures (E, SF , and
QABF ) for all data sets. To complement that, Fig. 4 (right)
shows pie-charts highlighting the number of cases in terms
of percentage where each method performed best. It can be
observed that for E and SF measures, the MEMD yielded
superior results for approximately all input data sets, whereas
for QABF , the MEMD produced better results for 83% of
the input data sets.
6. CASE STUDY 2: PAN-SHARPENING
We next performed experiments for Pan-sharpening of mul-
tispectral (MS) images using the existing BDEMD-based
fusion and the proposed MEMD-based fusion algorithms.
Their performance was compared against a multiscale Pan-
sharpening algorithm called `atrous wavelet transform (AWT)
method [12]. The simulated Pleiades data set consisting of
1024 × 1024 pixels of i) Strasbourg and ii) Tolouse was used
in our experiments. All the input MS images contained four
bands i.e. blue (B), green (G), red (R) and near-infrared
(NIR). The ground truth for both data sets was also available
(not shown in Fig. 5 due to space restrictions) and was used
for the quantitative analysis of the fusion results.
Pan-sharpening of MS images via MEMD and BDEMD
was performed as follows: the intensity plane I of the in-
put MS image was first obtained by averaging the bands of
the MS image. The so obtained intensity image I was then
fused with the high resolution panchromatic image to obtain
the ‘detailed’ intensity image
ˆ
I. The details added by the fu-
sion process were extracted by subtracting the original inten-
sity plane I from
ˆ
I, which were then separately added to the
B, G, R, and NIR components of the MS image to obtain the
Pan-sharpened MS image.
Figs. 5(a-b) show the source MS and PAN images of
Strasbourg city taken from the Pleiades sensor. The fused im-
ages obtained using AWT, BDEMD and MEMD algorithms
are shown in Figs. 5(c-e) respectively. It can be noticed that
the spectral performance of the images obtained from the pro-
posed MEMD-based method, shown in Fig. 5(c), matches the
result of the state-of-the-art AWT technique. The proposed
method, however, showed much improved spatial perfor-
mance as compared to both AWT and BDEMD based fusion
methods.
Table 1: Quantitative Results of Pan-sharpening
Pleiades Strasbourgh Pleiades Tolouse
AWT BD.. MEMD AWT BD.. MEMD
S 2.91 4.96 2.90 4.70 5.23 5.01
ER 4.247 6.52 2.981 5.74 5.72 3.52
Q4 89.31 54.36 93.89 94.2 59.22 94.98
The improved performance of MEMD can be further val-
idated by the quantitative results of pan-sharpening on both
data sets. We employed the following set of performance
metrics for this purpose: i) Relative dimensionless global er-
ror in synthesis (ER, ideally 0), ii) Spectral Angle Mapper
(S, ideally 0), and iii) Quaternion Index (Q4, ideally 100%)
[2]. The results of the quantitative analysis are presented in
Table I with the best value for each quality measure are la-
beled in bold. Please observe that in both data sets the pro-
posed scheme performed better than AWT and BDEMD fu-
sion methods for all performance metrics with an exception to
S value for the Tolouse image where AWT performed better.
Superiority of MEMD over BDEMD can be attributed to data
adaptive and local nature of its decomposition which mani-
fested in improved spatial performance in both case studies.
7. CONCLUSIONS
We have presented a method for the fusion of multiple images
using multivariate empirical mode decomposition (MEMD)
algorithm. The superiority of the method has been demon-
strated on a large data set for two applications: i) multi-focus
fusion, and ii) pan-sharpening of multi-spectral images. In
addition to the qualitative analysis, we have also employed a
wide range of quantitative performance measures to compare
the fusion results obtained from the two approaches.

8. REFERENCES
[1] T. Stathaki, “Image Fusion: Algorithms and Applications, Aca-
demic Press, 2008.
[2] L. Alparone, L. Wald, J. Chanussot, C. Thomas, P. Gamba and
L. M. Bruce, “Comparison of Pansharpening Algorithms: Out-
come of the 2006 GRS-S Data-Fusion Contest”, IEEE Transac-
tions on Geoscience and Remote Sensing, vol. 45, no. 10, pp.
3012–3021, 2007.
[3] J. Tang, A contrast based image fusion technique in the DCT
domain”, Digital Signal Processing, vol. 14, no. 3, pp. 218–226,
2004.
[4] G. Pajares and J. M. Cruz, A wavelet-based image fusion tuto-
rial”, Pattern Recognition, vol. 37, no. 9, pp. 1855–1872, 2004.
[5] N. E. Huang, Z. Shen, S. R. Long, M. L. Wu, H. H. Shih, Z.
Quanan, N. C. Yen, C. C. Tung, and H. H. Liu, “The empirical
mode decomposition and the Hilbert spectrum for nonlinear and
non-stationary time series analysis”, Proceedings of the Royal
Society A, vol. 454, pp. 903–995, 1998.
[6] N. Rehman and D. P. Mandic, “Multivariate Empirical Mode
Decomposition”, Proceedings of the Royal Society of Engineer-
ing A, vol. 466, no. 2117, pp. 1291–1302, 2009.
[7] J. Nunes, Y. Bouaoune, E. Delechelle, O. Niang, and Ph. Bunel,
“Image analysis by bi-dimensional empirical mode decompo-
sition”, Image and Visual Computing, vol. 21, no. 12, pp.
10191026, 2003.
[8] A. Linderhed, “Image EMD: A new tool for image processing”,
Advances in Adaptive Data Analysis, vol. 1, no. 2, pp. 265294,
2009.
[9] Y. Zheng, E. A. Essock, B. C. Hansen and A. M. Haun, A new
metric based on extended spatial frequency and its application
to DWT based fusion algorithms”, Information Fusion vol. 8,
no. 2, pp. 177–192, 2005.
[10] C. S. Xydeas and V. Petrovic, “Objective image fusion perfor-
mance measure”, IEEE Electronics Letters, vol. 36, no. 4, pp.
308–309, 2000.
[11] C. Schmid, and R. Mohr, and C. Bauckhage, “Evaluation of
Interest Point Detectors”, International Journal of Computer Vi-
sion, vol. 37, no. 2, pp. 151-172, 2000.
[12] J. Nunez, X. Otazu, O. Aldrich, A. Prades, V. Pala, R. Arbiol,
”Multiresolution-based image fusion with additive wavelet de-
composition”, IEEE Transactions on Geoscience and Remote
Sensing, vol.37, no.3, pp. 1204-1211, 1999.
View publication statsView publication stats
Citations
More filters
Journal ArticleDOI
TL;DR: It is shown that MEMD overcomes the limitations of standard EMD and yields improved spatial and spectral performance in the context of pansharpening of MS images.
Abstract: We propose a novel class of schemes for the pansharpening of multispectral (MS) images using a multivariate empirical mode decomposition (MEMD) algorithm. MEMD is an extension of the empirical mode decomposition (EMD) algorithm, which enables the decomposition of multivariate data into its intrinsic oscillatory scales. The ability of MEMD to process multichannel data directly by performing data-driven, local, and multiscale analysis makes it a perfect match for pansharpening applications, a task for which standard univariate EMD is ill-equipped due to the nonuniqueness, mode-mixing, and mode-misalignment issues. We show that MEMD overcomes the limitations of standard EMD and yields improved spatial and spectral performance in the context of pansharpening of MS images. The potential of the proposed schemes is further demonstrated through comparative analysis against a number of standard pansharpening algorithms on both simulated Pleiades and real-world IKONOS data sets.

38 citations


Cites background from "Image fusion using multivariate and..."

  • ...That ability of MEMD to cater for more than two input images has been highlighted for multiexposure and multifocus images in [23]....

    [...]

Journal ArticleDOI
TL;DR: The proposed MNCMD is capable of handling time-varying signal efficiently in an elegant variational optimization framework and can extract an optimal set of multivariate modes and their corresponding instantaneous frequencies without requiring more user-defined parameters than the original NCMD.

26 citations

Proceedings ArticleDOI
01 Mar 2022
TL;DR: In this article , the authors reviewed the application of AI techniques in motor FDD in recent years and divided the process of feature extraction and fault classification into two main steps, and presented the characteristics of all techniques reviewed are summarized.
Abstract: The motor drive system plays a significant role in the safety of electric vehicles as a bridge for power transmission. Meanwhile, to enhance the efficiency and stability of the drive system, more and more studies based on AI technology are devoted to the fault detection and diagnosis (FDD) of the motor drive system. This article reviews the application of AI techniques in motor FDD in recent years. AI-based FDD is divided into two main steps: feature extraction and fault classification. The application of different signal processing methods in feature extraction is discussed. In particular, the application of traditional machine learning and deep learning algorithms for fault classification is presented in detail. In addition, the characteristics of all techniques reviewed are summarized. Finally, the latest developments, research gaps, and future challenges in fault monitoring and diagnosis of motor faults are discussed.

19 citations

Journal ArticleDOI
TL;DR: A novel bidimensional MEMD is proposed which directly projects a bidimensional multivariate signal, which is composed of multiple images, on the unit hypersphere in the $n$ -dimensional space and the mean surface is estimated by interpolating the multivariate scatter data so as to extract common spatio-temporal scales across multiple images.
Abstract: Empirical mode decomposition (EMD) is a fully data-driven technique designed for multi-scale decomposition of signals into their natural scale components, called intrinsic mode functions (IMFs). When EMD is directly applied to perform fusion of multivariate data from multiple and heterogeneous sources, the problem of uniqueness, that is, different numbers of decomposition levels for different sources, is likely to occur, due to the empirical nature of EMD. Although the multivariate EMD (MEMD) has been proposed for temporal data, which employs real-valued projections along multiple directions on a unit hypersphere in the $n$ -dimensional space to calculate the envelope and the local mean of multivariate signals, in order to guarantee the uniqueness of the scales, its direct usefulness in 2D multi-scale image fusion is still limited, due to its inability to maintain the spatial information. To address this issue, we propose a novel bidimensional MEMD (BMEMD) which directly projects a bidimensional multivariate signal, which is composed of multiple images, on the unit hypersphere in the $n$ -dimensional space. This is achieved via real-valued surface projections and the mean surface is estimated by interpolating the multivariate scatter data so as to extract common spatio-temporal scales across multiple images. Case studies involving texture analysis and multi-focus image fusion are presented to demonstrate the effectiveness of the proposed method.

15 citations


Cites background or methods from "Image fusion using multivariate and..."

  • ...Although the simulations in [23], [24] illustrate the potentiality of MEMD in fusion of multi-focus images and pan-...

    [...]

  • ...However, a prerequisite to implement these EMD extensions in 2D image processing is to vectorize multiple images by concatenating their columns/rows to form a one-dimensional multivariate signal [21]–[23]....

    [...]

Journal ArticleDOI
TL;DR: This paper presents an efficient implementation of multivariate empirical mode decomposition (MEMD) algorithm, a multivariate extension of EMD algorithm, and compared the optimized implementation of MEMD, using GPU, with the MATLAB implementation for hexa-variate andhexa-deca-Variate data sets, and observed that the GPU-based optimized implementation results in approximately 6% performance improvements in terms of time consumption.
Abstract: This paper presents an efficient implementation of multivariate empirical mode decomposition (MEMD) algorithm, a multivariate extension of EMD algorithm. Analogous to EMD, MEMD decomposes a multivariate signal into its intrinsic mode functions using joint rotational mode. The algorithm is computationally intensive because of its recursive nature and any increase in input data size results in a non-linear increase in its execution time. Therefore, it is extremely time-consuming to obtain a decomposition of signal, such as EEG into its intrinsic modes using MEMD. As the interest in applying MEMD algorithm in various domains is increasing, there is a need to develop an optimized implementation of the algorithm, since it requires repeated execution of the same operations and computationally extensive interpolations on each projected vector. This can be done using GPGPU, because it has the power to process similar function on different blocks of data. We have compared the optimized implementation of MEMD, using GPU, with the MATLAB implementation for hexa-variate and hexa-deca-variate data sets, and observed that the GPU-based optimized implementation results in approximately $6\times \sim 16\times $ performance improvements in terms of time consumption.

8 citations


Additional excerpts

  • ...Though MEMD algorithm is designed for time varying signals; however, it can be used for image processing applications [29], [30]....

    [...]

References
More filters
Journal ArticleDOI
TL;DR: In this paper, a new method for analysing nonlinear and nonstationary data has been developed, which is the key part of the method is the empirical mode decomposition method with which any complicated data set can be decoded.
Abstract: A new method for analysing nonlinear and non-stationary data has been developed. The key part of the method is the empirical mode decomposition method with which any complicated data set can be dec...

18,956 citations

Journal ArticleDOI
TL;DR: Two evaluation criteria for interest points' repeatability rate and information content are introduced and different interest point detectors are compared using these two criteria.
Abstract: Many different low-level feature detectors exist and it is widely agreed that the evaluation of detectors is important. In this paper we introduce two evaluation criteria for interest points' repeatability rate and information content. Repeatability rate evaluates the geometric stability under different transformations. Information content measures the distinctiveness of features. Different interest point detectors are compared using these two criteria. We determine which detector gives the best results and show that it satisfies the criteria well.

1,690 citations


"Image fusion using multivariate and..." refers background in this paper

  • ...(1) below: x(k) = M∑ m=1 cm(k) + r(k) (1) The residual r(k) does not contain any oscillations and represents a trend within the signal....

    [...]

Journal ArticleDOI
TL;DR: Experimental results clearly indicate that this metric reflects the quality of visual information obtained from the fusion of input images and can be used to compare the performance of different image fusion algorithms.
Abstract: A measure for objectively assessing the pixel level fusion performance is defined. The proposed metric reflects the quality of visual information obtained from the fusion of input images and can be used to compare the performance of different image fusion algorithms. Experimental results clearly indicate that this metric is perceptually meaningful.

1,446 citations


"Image fusion using multivariate and..." refers background in this paper

  • ...(1) below: x(k) = M∑ m=1 cm(k) + r(k) (1) The residual r(k) does not contain any oscillations and represents a trend within the signal....

    [...]

Journal ArticleDOI
TL;DR: This tutorial performs a synthesis between the multiscale-decomposition-based image approach, the ARSIS concept, and a multisensor scheme based on wavelet decomposition, i.e. a multiresolution image fusion approach.

1,187 citations


"Image fusion using multivariate and..." refers methods in this paper

  • ...Typical examples are the methods based on Gaussian pyramids, Fast Fourier Transform (FFT), Discrete Cosine Transform (DCT) [3], and Discrete Wavelet Transform (DWT) [4]....

    [...]

Journal ArticleDOI
TL;DR: The authors developed a technique, based on multiresolution wavelet decomposition, for the merging and data fusion of high-resolution panchromatic and multispectral images which is clearly better than the IHS and LHS mergers in preserving both spectral and spatial information.
Abstract: The standard data fusion methods may not be satisfactory to merge a high-resolution panchromatic image and a low-resolution multispectral image because they can distort the spectral characteristics of the multispectral data. The authors developed a technique, based on multiresolution wavelet decomposition, for the merging and data fusion of such images. The method presented consists of adding the wavelet coefficients of the high-resolution image to the multispectral (low-resolution) data. They have studied several possibilities concluding that the method which produces the best results consists in adding the high order coefficients of the wavelet transform of the panchromatic image to the intensity component (defined as L=(R+G+B)/3) of the multispectral image. The method is, thus, an improvement on standard intensity-hue-saturation (IHS or LHS) mergers. They used the "a trous" algorithm which allows the use of a dyadic wavelet to merge nondyadic data in a simple and efficient scheme. They used the method to merge SPOT and LANDSAT/sup TM/ images. The technique presented is clearly better than the IHS and LHS mergers in preserving both spectral and spatial information.

1,151 citations


"Image fusion using multivariate and..." refers background in this paper

  • ...The sifting process stops when d(k) has inadequate extrema....

    [...]

Frequently Asked Questions (16)
Q1. What is the common method for determining the upper and lower envelopes of an input signal?

Radial basis functions (tensor product) or B-splines are commonly used methods for interpolation [7], whereas the method by Linderhed [8] uses thin-plate splines for the interpolation of the extrema. 

The authors present a novel methodology for the fusion of multiple ( two or more ) images using the multivariate extension of empirical mode decomposition ( MEMD ). The authors show that the multivariate and multidimensional extensions of EMD are suitable for image fusion purposes. The authors further demonstrate that while multidimensional extensions, by design, may seem more appropriate for tasks related to image processing, the proposed multivariate extension outperforms these in image fusion applications owing to its mode-alignment property for IMFs. 

Empirical mode decomposition (EMD) [5] is a data-driven method which decomposes an arbitrary signal x(k) into a set of multiple oscillatory components called the intrinsic mode functions (IMFs) via an iterative process known as sifting algorithm [6]. 

The recursive sifting algorithm operates by defining the upper and lower envelopes of an input signal by interpolating its extrema. 

For quantitative evaluation of the fusion results, the authors have employed Entropy (E) [11], objective image fusion (QABF ) [10] and the spatial frequency (SF ) [9] performance measures. 

For this purpose, multiple images of 30 different scenes were used; seven images were taken of each scene with different parts of the scene out-of-focus in each image. 

The authors employed the following set of performance metrics for this purpose: i) Relative dimensionless global error in synthesis (ER, ideally 0), ii) Spectral Angle Mapper (S, ideally 0), and iii) Quaternion Index (Q4, ideally 100%) [2]. 

It can be observed that for E and SF measures, the MEMD yielded superior results for approximately all input data sets, whereas for QABF , the MEMD produced better results for 83% of the input data sets. 

Superiority of MEMD over BDEMD can be attributed to data adaptive and local nature of its decomposition which manifested in improved spatial performance in both case studies. 

The authors next performed experiments for Pan-sharpening of multispectral (MS) images using the existing BDEMD-based fusion and the proposed MEMD-based fusion algorithms. 

To address this issue, MEMD operates by projecting an input multivariate signal in V uniformly spaced directions on a unit p-sphere; the extrema of the so projected signals are then interpolated to obtain multiple envelopes which are subsequently averaged to obtain the local mean. 

Please observe that in both data sets the proposed scheme performed better than AWT and BDEMD fusion methods for all performance metrics with an exception to S value for the Tolouse image where AWT performed better. 

The IMFs represent the intrinsic temporal modes (scales) that are present in the input data which when added together reproduce the input x(k), as shown in eq. (1) below:x(k) = M∑ m=1 cm(k) + r(k) (1)The residual r(k) does not contain any oscillations and represents a trend within the signal. 

The superiority of the method has been demonstrated on a large data set for two applications: i) multi-focus fusion, and ii) pan-sharpening of multi-spectral images. 

This mode alignment property of MEMD is a result of direct processing of input images within MEMD, whereas the lack of it in BDEMD is due to the fact that it processes multiple input images separately. 

The details added by the fusion process were extracted by subtracting the original intensity plane The authorfrom Î , which were then separately added to the B, G, R, and NIR components of the MS image to obtain the Pan-sharpened MS image.