scispace - formally typeset
Search or ask a question
Journal ArticleDOI

The Nonsubsampled Contourlet Transform: Theory, Design, and Applications

01 Oct 2006-IEEE Transactions on Image Processing (IEEE TRANSACTION ON IMAGE PROCESSING)-Vol. 15, Iss: 10, pp 3089-3101
TL;DR: This paper proposes a design framework based on the mapping approach, that allows for a fast implementation based on a lifting or ladder structure, and only uses one-dimensional filtering in some cases.
Abstract: In this paper, we develop the nonsubsampled contourlet transform (NSCT) and study its applications. The construction proposed in this paper is based on a nonsubsampled pyramid structure and nonsubsampled directional filter banks. The result is a flexible multiscale, multidirection, and shift-invariant image decomposition that can be efficiently implemented via the a trous algorithm. At the core of the proposed scheme is the nonseparable two-channel nonsubsampled filter bank (NSFB). We exploit the less stringent design condition of the NSFB to design filters that lead to a NSCT with better frequency selectivity and regularity when compared to the contourlet transform. We propose a design framework based on the mapping approach, that allows for a fast implementation based on a lifting or ladder structure, and only uses one-dimensional filtering in some cases. In addition, our design ensures that the corresponding frame elements are regular, symmetric, and the frame is close to a tight one. We assess the performance of the NSCT in image denoising and enhancement applications. In both applications the NSCT compares favorably to other existing methods in the literature

Summary (1 min read)

Introduction

  • In this paper the authors study the nonsubsampled contourlet transform.
  • The authors show how zeroes can be imposed in the filters so that the iterated structure produces regular basis functions.
  • The authors apply the constructed transform in image noise removal where the results obtained are comparable to the state-of-the art, being superior in some cases.

2.1. The Nonsubsampled Pyramid

  • Such expansion is similar to the 1-D̀a trouswavelet expansion [6] and has a redundancy ofJ + 1 whenJ is the number of decomposition stages.
  • The proposed structure is thus different from the tensor producta tròus algorithm.

2.2. The Nonsubsampled Directional Filter Bank

  • The directional filter bank [5] is constructed by combining critically sampled fan filter banks and pre/post re-sampling operations.
  • The NSCT is obtained by carefully combining the 2-D nonsubsampled pyramid and the nonsubsampled DFB [7].
  • The resulting filtering structure approximates the ideal partition of the frequency plane displayed in Figure 1.
  • If the authors relax the tightness constraint then the design becomes more flexible.
  • The authors restrict their attention to zero phase FIR designs only.

3.1. Zeros of the mapped filters

  • For the pyramid case, high order zeros atω1 = π andω2 = π result in smoothness of the scaling function and wavelet associated with the iterated filter bank.
  • The authors point out that for the approximation of polynomial surfaces, point zeros at(±π,±π) would suffice.
  • Point zeros alone do not guarantee a “reasonable” frequency response of the pyramid filters.
  • The following proposition characterizes the mapping function that generates such zeros.
  • Proposition 1 Let G(z) be an-th order polynomial withn ≥ 1 and roots{zi}ni=1 where eachzi has multiplicityni.

3.2. Implementation through lifting.

  • Thus, one can construct NS filter banks starting with theLazynonsubsampled filters given byG0(z) = K, H1(z) = 1−K and then adding update and predict operators.
  • Moreover, givenH0(z) andH1(z), the factorization is easily computed.
  • Assume without loss of generality that the degree of the highpass prototype filterH(1D)1 (x) is smaller than the degree ofH(1D)0 (x).
  • The nonsubsampled pyramid filter bank is almost tight.

3.3. Fan Filter Design

  • Similar to the pyramid case, due to lack of factorization tools the authors resort to non-tight solutions obtained through mapping.
  • Thus, the methodology is similar to the pyramid case, the distinction being the mapping function.
  • Following the work in [2] the authors choose the threshold locally using Ti,j,n = σ2nij σi,j,n whereσi,j,n denotes the variance of then-th coefficient at thei-th direction in thej-th scale andσ2nij the noise variance in the corresponding subband.
  • Further results and comparisons can be found in [7].

Did you find this useful? Give us your feedback

Content maybe subject to copyright    Report

NONSUBSAMPLED CONTOURLET TRANSFORM: FILTER DESIGN AND APPLICATIONS
IN DENOISING
Arthur L. da Cunha, J. Zhou and Minh N. Do
University of Illinois at Urbana-Champaign
Department of Electrical and Computer Engineering
Coordinated Science Laboratory, Urbana, IL 61801
Email:{cunhada,jzhou2,minhdo}@uiuc.edu
ABSTRACT
In this paper we study the nonsubsampled contourlet transform.
We address the corresponding filter design problem using the Mc-
Clellan transformation. We show how zeroes can be imposed in
the filters so that the iterated structure produces regular basis func-
tions. The proposed design framework yields filters that can be
implemented efficiently through a lifting factorization. We apply
the constructed transform in image noise removal where the results
obtained are comparable to the state-of-the art, being superior in
some cases.
1. INTRODUCTION
A number of image processing tasks are efficiently carried out in
a domain other than the pixel domain, often by means of an in-
vertible linear transformation. This linear transformation can be
redundant or not, depending on whether the set of basis functions
is linear independent. By allowing redundancy, it is possible to
enrich the set of basis functions so that the representation is more
efficient in capturing some signal behavior. Imaging applications
such as edge detection, contour detection, denoising and image
restoration can greatly benefit from redundant representations.
In the context of multiscale expansions implemented with fil-
ter banks, dropping the basis requirement offers the possibility of
an expansion that is shift-invariant, a crucial property in a number
of applications. For instance, in image denoising via thresholding
in the wavelet domain, the lack of shift-invariance causes pseudo-
Gibbs phenomena around singularities [1]. As a result, most of
the state-of-the-art wavelet denoising routines (see e.g. [2], [3])
use an expansion with less shift sensitivity than the standard max-
imally decimated wavelet decomposition.
In addition to shift-invariance, an efficient image representa-
tion has to account for the geometry pervasive in natural scenes.
The contourlet transform [4] is a directional multiscale transform
that is constructed by combining the Laplacian pyramid (LP) and
the directional filter bank (DFB) [5]. Due to downsamplers and up-
samplers present in both the LP and DFB, the contourlet transform
is not shift-invariant.
The non-subsampled contourlet transform (NSCT) is obtained
by coupling a nonsubsampled pyramid structure with the nonsub-
sampled DFB (Section 2). In this paper we address the filter design
Thanks to CAPES and NSF for funding.
problem of the NSCT (Section 3) and show its effectiveness in im-
age denoising (Section 4).
(π, π)
(π, π)
ω
1
ω
2
Fig. 1. The idealized frequency partitioning obtained with the
NSCT.
2. THE NONSUBSAMPLED CONTOURLET
TRANSFORM
The idea behind a fully shift invariant multiscale directional ex-
pansion similar to contourlets is to obtain the frequency partition-
ing of Figure 1 without resorting to critically sampled structures
that have periodically time-varying units such as downsamplers
and upsamplers. The NSCT construction can thus be divided into
two parts: (1) A nonsubsampled pyramid structure which ensures
the multiscale property and (2) A nonsubsampled DFB structure
which gives directionality. Next we describe each part in detail.
2.1. The Nonsubsampled Pyramid
The shift sensitivity of the LP can be remedied by replacing it with
a 2-channel nonsubsampled 2-D filter bank structure. Such expan-
sion is similar to the 1-D
`
a trous wavelet expansion [6] and has
a redundancy of J + 1 when J is the number of decomposition
stages. The ideal frequency support of the low-pass filter at the
j-th stage is the region [
π
2
j
,
π
2
j
] × [
π
2
j
,
π
2
j
]. Accordingly, the
support of the high-pass filter is the complement of the low-pass
support region on the [
π
2
j+1
,
π
2
j+1
]×[
π
2
j1
,
π
2
j1
] square. The
proposed structure is thus different from the tensor product a tr
`
ous
algorithm. It has J + 1 redundancy. By contrast, the 2-D
`
a trous
algorithm has 3J + 1 redundancy.

H
0
(z)
H
1
(z)
G
0
(z)
G
1
(z)
(a)
F
0
(z)
F
1
(z)
E
0
(z)
E
1
(z)
(b)
Fig. 2. Two kinds of desired responses. (a) The pyramid desired
response. (b) The fan desired response.
2.2. The Nonsubsampled Directional Filter Bank
The directional filter bank [5] is constructed by combining criti-
cally sampled fan filter banks and pre/post re-sampling operations.
The result is a tree-structured filter bank which splits the frequency
plane into directional wedges.
A fully shift-invariant directional expansion is obtained by sim-
ply switching off the downsamplers and upsamplers in the DFB
equivalent filter bank. Due to multirate identities, this is equiva-
lent to switching off each of the downsamplers in the tree struc-
ture, while still keeping the re-sampling operations that can be ab-
sorbed by the filters. This results in a tree structure composed of
two-channel nonsubsampled filter banks.
The NSCT is obtained by carefully combining the 2-D non-
subsampled pyramid and the nonsubsampled DFB (NSDFB) [7].
The resulting filtering structure approximates the ideal partition
of the frequency plane displayed in Figure 1. It must be noted
that, different from the contourlet expansion, the NSCT has a re-
dundancy given by R =
P
J
j=0
2
l
j
where 2
l
j
is the number of
directions at scale j.
3. FILTER DESIGN
The filter design problem in the proposed NSCT construction can
be split into two related parts: (1) design of the pyramid filters
in the NS pyramid and (2) design of the nonsubsampled fan fil-
ter bank which constitute the basic building block of the NSDFB.
Figure 2 illustrates these nonsubsampled filter banks and their ide-
alized frequency responses. The goal is to design a set of filters
that satisfy the Bezout identity
H
0
(z)G
0
(z) + H
1
(z)G
1
(z) = 1, (1)
and, in addition, approximate the ideal frequency responses (Fig.
2.) A nonsubsampled filter bank underlies a frame expansion in
2
(Z
2
) and the frame is tight whenever G
i
(z) = H
i
(z
1
), i =
0, 1 [8]. Tight FIR designs often require spectral factorization [8]
which is hard in 2-D. If we relax the tightness constraint then the
design becomes more flexible. In addition, non-tight FIR designs
can be linear-phase, thus avoiding phase distortion and allowing
for symmetric extension at the boundaries.
A set of 2-D filters satisfying (1) can be constructed with the
aid of the McClellan transformation. The design can be summa-
rized in the following steps:
1. Construct a set of 1-D polynomials {H
(1D)
0
, H
(1D)
1
, G
(1D)
0
,
G
(1D)
1
} that satisfy the Bezout identity.
2. Construct a mapping function f(z) so that the 2-D filters
H
(1D)
0
(f(z)), H
(1D)
1
(f(z)), G
(1D)
0
(f(z)), G
(1D)
1
(f(z)) sat-
isfy the Bezout identity, have desired frequency response
and satisfy the “zeros” condition (see below).
The 1-D filters are called prototype filters. We restrict our
attention to zero phase FIR designs only. In this case we write
f(z) =
˜
f(x, y) with x = (z
1
+ z
1
1
)/2 and y = (z
2
+ z
1
2
)/2.
3.1. Zeros of the mapped filters
It is desirable to have zero moments in the mapped filters. For
the pyramid case, high order zeros at ω
1
= π and ω
2
= π result
in smoothness of the scaling function and wavelet associated with
the iterated filter bank. We point out that for the approximation
of polynomial surfaces, point zeros at (±π, ±π) would suffice.
However, point zeros alone do not guarantee a “reasonable” fre-
quency response of the pyramid filters. Furthermore, to obtain a
high degree of smoothness, a high number of point zeros is often
needed. By contrast, a small number of line zeros at ω
1
= π and
ω
2
= π yield a fairly good degree of smoothness. The following
proposition characterizes the mapping function that generates such
zeros.
Proposition 1 Let G(z) be a n-th order polynomial with n 1
and roots {z
i
}
n
i=1
where each z
i
has multiplicity n
i
. Suppose we
want a mapping function f(x, y) such that
G (f(x, y)) = (x c)
N
x
(y d)
N
y
r(x, y) (2)
where r (x, y) is a bivariate polynomial. Then G(f(x, y)) has the
form in (2) if and only if f(x, y) takes the form
f(x, y) = z
j
+ (x c)
N
x
(y d)
N
y
r
f
(x, y) (3)
for some root z
j
and with r
f
(x, y) a bivariate polynomial and
N
x
, N
y
such that N
x
n
i
N
x
and N
y
n
i
N
y
.
Proof: See [7].
3.2. Implementation through lifting.
If a set of filters {H
0
(z), H
1
(z), G
0
(z), G
1
(z)} satisfy (1), then
a different solution can be obtained by setting G
0
(z) = G
0
(z) +
r(z) and H
1
(z) = H
1
(z) s(z) in which case the Bezout’s rela-
tion is satisfied provided that
H
0
(z)r(z) = G
1
(z)s(z) (4)
Consequently, we see that r(z) is an FIR function that has G
1
(z)
as a factor so that r(z) = G
1
(z)f(z) and similarly we have s(z) =
H
0
(z)g(z). Substituting in (4) we conclude that g(z) = f(z). We
then have the following:
Proposition 2 Let H
0
(z), H
1
(z), G
0
(z), G
1
(z) be FIR 2-D fil-
ters satisfying the Bezout relation (1). Then, the new filters defined
as
G
0
(z) = G
0
(z) G
1
(z)P (z), H
1
(z) = H
1
(z) + H
0
(z)P (z)
(5)

or
H
0
(z) = H
0
(z) + H
1
(z)U(z), G
1
(z) = G
1
(z) + G
0
(z)U(z)
(6)
where P (z) and U (z) are 2-D FIR filters also satisfy Bezout’s
relation.
Proposition 2 characterizes the NSFB in terms of lifting steps in
the same way as in the critically sampled filter bank. In this con-
text, (5) is a predict operation whereas (6) is the update operation.
Thus, one can construct NS filter banks starting with the Lazy non-
subsampled filters given by G
0
(z) = K, H
1
(z) = 1K and then
adding update and predict operators. Figure 3 shows the resulting
ladder structure with one predict and one update step.
In the 1-D case, one can show the above factorization is com-
plete, as a consequence of the Euclidean algorithm for Laurent
polynomials. Moreover, given H
0
(z) and H
1
(z), the factorization
is easily computed. For the 2-D case, computing a factorization
is hard as a Euclidean algorithm is absent. When the filters are
+
+
++
+
x[n] x[n]
U
1 K
K
UPP
Fig. 3. Lifting structure for the nonsubsampled filter bank.
designed via mapping then a factorization can be obtained from
the 1-D filters. Assume without loss of generality that the degree
of the highpass prototype filter H
(1D)
1
(x) is smaller than the de-
gree of H
(1D)
0
(x). Since there are synthesis filters G
(1D)
0
(x) and
G
(1D)
1
(x) such that Bezout’s identity is satisfied, it follows that
gcd{H
(1D)
0
, H
(1D)
1
} = 1. The Euclidean algorithm then allows us
to write [9]
H
(1D)
0
(x)
H
(1D)
1
(x)
!
=
N
Y
i=0
1 0
p
i
(x) 1
1 q
i
(x)
0 1
1
0
for q
i
, p
i
polynomials. As a result we can obtain a 2-D lifting fac-
torization by replacing x with the mapping function f(x, y). In
general, the implementation using predict/update stages halves the
number of multiplications/additions of the direct form. The com-
plexity can be reduced further if the mapping filter is separable.
The next example illustrates the main points of this section.
Example 1 We construct the prototypes to each have one zero at
x = 1 and minimize the mean square distance of their coeffi-
cients. This ensures that the underlying frame in the NSFB is close
to being tight. We get
H
(1D)
0
(x) =
1
2
(x + 1)
2 + (1
2)x
and
G
(1D)
0
(x) =
1
2
(x + 1)
2 + (4 3
2)x + (2
2 3)x
2
.
To guarantee the response of the high-pass filters we impose
H
(1D)
1
(x) = H
(1D)
0
(x) and G
(1D)
1
(x) = G
(1D)
0
(x). The lifting
-2
0
2
-2
0
2
0
0.25
0.5
0.75
1
-2
0
2
(a) H
0
(z)
-2
0
2
-2
0
2
0
0.25
0.5
0.75
1
-2
0
2
(b) H
1
(z)
-2
0
2
-2
0
2
0
0.25
0.5
0.75
1
-2
0
2
(c) G
0
(z)
-2
0
2
-2
0
2
0
0.25
0.5
0.75
1
-2
0
2
(d) G
1
(z)
Fig. 4. Design example 1 with maximally flat filters. The nonsubsampled
pyramid filter bank is almost tight.
factorization of the prototype filters is given by
H
(1D)
0
(x)
H
(1D)
1
(x)
!
=
1 αx
0 1
1 0
βx 1
1 γx
0 1
K
1
K
2
(7)
with α = γ = 1
2, β =
1
2
, K
1
= K
2
=
1
2
. This imple-
mentation uses 5 multiplies/sample whereas the direct one yields
10 multiplies/sample. Following Proposition 1, we set f (x, y) =
1 + 2P
2,4
(x)P
2,4
(y) using the maximally flat polynomials:
P
N,L
(x) := (1 + x)
N
L1N
X
l=0
N + l 1
l
!
2
Nl
(1 x)
l
.
From Proposition 1, each of the resulting highpass filters has a
4-th order zero at ω
1
= π and ω
2
= π . The sizes of the filters
H
0
(z) and G
0
(z) are 13 × 13 and 19 × 19 respectively. Figure 4
shows the obtained frequency responses.
-2
0
2
-2
0
2
0
0.25
0.5
0.75
1
-2
0
2
(a) F
0
(z)
-2
0
2
-2
0
2
0
0.25
0.5
0.75
1
-2
0
2
(b) E
0
(z)
Fig. 5. Fan filters designed with prototype filters of Example 1 and dia-
mond maximally flat mapping filters.
3.3. Fan Filter Design
Similar to the pyramid case, due to lack of factorization tools we
resort to non-tight solutions obtained through mapping. Thus, the
methodology is similar to the pyramid case, the distinction being

“Lena” PSNR (dB)
σ Noisy SI-AdaptShr [2] BivShrink [3] NSCT
10 28.13 - 35.34 35.38
15 24.65 33.37 33.67 33.67
20 22.13 32.08 32.40 32.43
25 20.17 31.13 31.40 31.45
30 18.63 - 30.54 30.64
“Barbara” PSNR (dB)
σ Noisy SI-AdaptShr [2] BivShrink [3] NSCT
10 28.17 - 33.35 33.98
15 24.65 31.11 31.31 31.94
20 22.15 29.49 29.80 30.51
25 20.22 28.30 28.61 29.40
30 18.63 - 27.65 28.49
Table 1. Denoising results for various soft-threshold estimators.
the mapping function. A useful family of mapping functions in this
context is the maximally flat diamond one. The fan shaped map-
ping can be obtained from the diamond one by a simple change
of variables. The maximally flat diamond mapping polynomial is
obtained by imposing flatness at the points (x, y) = (1, 1) and
(x, y) = (0, 0). For instance, with the same prototypes of Ex-
ample 1, we use a maximally flat mapping to obtain the fan fil-
ters F
0
(z) and E
0
(z) shown in Figure 5. The other filters F
1
(z),
E
1
(z) are modulated versions of F
0
(z), E
0
(z). The sizes of
F
0
(z) and E
0
(z) are 21 × 21 and 31 × 31 respectively.
4. DENOISING EXPERIMENTS
In other to illustrate the potential of the NSCT we attempt to re-
move additive Gaussian Noise (AWGN) of images by means of a
threshold estimator. We perform soft thresholding independently
in each high pass filter output of the NSCT. Following the work in
[2] we choose the threshold locally using
T
i,j,n
=
σ
2
n
ij
σ
i,j,n
where σ
i,j,n
denotes the variance of the n-th coefficient at the i-th
direction in the j-th scale and σ
2
n
ij
the noise variance in the cor-
responding subband. We estimate the variances locally using the
neighboring coefficients and a maximum likelihood estimator. The
noise variance in each subband is inferred using a Monte-Carlo
technique where the variances are computed for a few normal-
ized AWGN images and then averaged to stabilize the results. To
benchmark the performance of the NSCT we have used the Bi-
variate shrinkage estimator of [3] which uses a redundant complex
wavelet transform and the context adaptive thresholding estimator
of [2] which uses the nonsubsampled wavelet transform (NSWT).
Table 1 shows the PSNR results obtained for the various methods.
The NSCT performs comparably to the BivShrink estimator for
the “Lena” image and is superior for the Barbara image. Figure 6
show the denoised “Lena” image obtained with the NSCT and the
NSWT using the same denoising routine. Notice that the NSCT
exhibits a much better reconstruction of edge features thus attest-
ing the effectiveness of the proposed transform. Further results
and comparisons can be found in [7].
(a) (b)
Fig. 6. Image Denoising with the NSCT. (a) Denoised ”Lena” im-
age using the NSWT, PSNR = 31.92dB. (b) Denoised with NSCT,
PSNR=32.43dB.
5. CONCLUSION
In this paper we have addressed the filter design problem of the
NSCT and applied the construction in image denoising. The de-
sign methodology proposed also allows for a fast implementation
through lifting factorization which halves the number of opera-
tions when compared to the direct form. The proposed transform
structure with the new designed filters has proven very efficient in
noise removal surpassing the undecimated discrete wavelet trans-
form and being comparable (sometimes superior) to other state-
ofthe- art denoising methods.
6. REFERENCES
[1] R. R. Coifman and D. L. Donoho, “Translation invariant de-noising,
in Wavelets and Statistics, A. Antoniadis and G. Oppenheim, Eds.
New York: Springer-Verlag, 1995, pp. 125–150.
[2] S. G. Chang, B. Yu, and M. Vetterli, “Spatially adaptive wavelet
thresholding with context modeling for image denoising, IEEE Trans.
Img. Proc., vol. 9, no. 9, pp. 1522–1531, September 2000.
[3] L. Sendur and I. W. Selesnick, “Bivariate shrinkage with local variance
estimation, IEEE Signal Processing Letters, vol. 9, no. 12, pp. 438–
441, December 2002.
[4] M. N. Do and M. Vetterli, “The contourlet transform: An efficient
directional multiresolution image representation, IEEE Trans. Img.
Processing, to appear, 2005.
[5] R. H. Bamberger and M. J. T. Smith, A filter bank for the directional
decomposition of images: Theory and design, IEEE Trans. Signal
Processing, vol. 40, no. 4, pp. 882–893, April 1992.
[6] M. J. Shensa, “The discrete wavelet transform: Wedding the
`
a trous
and Mallat algorithms. IEEE Trans. Signal Proc., vol. 40, no. 10, pp.
2464–2482, October 1992.
[7] A. L. Cunha, J. Zhou, and M. N. Do, “The nonsubsampled contourlet
transform: Theory, design, and applications, IEEE Trans. Img. Proc.,
submitted, 2005.
[8] Z. Cvetkovic and M. Vetterli, “Oversampled filter banks, IEEE Trans-
actions on Signal Processing, vol. 46, no. 5, pp. 1245–1255, May
1998.
[9] R. E. Blahut, Fast Algorithms for Digital Signal Processing. Addison-
Wesley, 1985.
Citations
More filters
Journal ArticleDOI
TL;DR: In this paper, the authors proposed a deep convolutional neural network (CNN)-based algorithm for solving ill-posed inverse problems, which combines multiresolution decomposition and residual learning in order to learn to remove these artifacts while preserving image structure.
Abstract: In this paper, we propose a novel deep convolutional neural network (CNN)-based algorithm for solving ill-posed inverse problems. Regularized iterative algorithms have emerged as the standard approach to ill-posed inverse problems in the past few decades. These methods produce excellent results, but can be challenging to deploy in practice due to factors including the high computational cost of the forward and adjoint operators and the difficulty of hyperparameter selection. The starting point of this paper is the observation that unrolled iterative methods have the form of a CNN (filtering followed by pointwise non-linearity) when the normal operator ( $H^{*}H$ , where $H^{*}$ is the adjoint of the forward imaging operator, $H$ ) of the forward model is a convolution. Based on this observation, we propose using direct inversion followed by a CNN to solve normal-convolutional inverse problems. The direct inversion encapsulates the physical model of the system, but leads to artifacts when the problem is ill posed; the CNN combines multiresolution decomposition and residual learning in order to learn to remove these artifacts while preserving image structure. We demonstrate the performance of the proposed network in sparse-view reconstruction (down to 50 views) on parallel beam X-ray computed tomography in synthetic phantoms as well as in real experimental sinograms. The proposed network outperforms total variation-regularized iterative reconstruction for the more realistic phantoms and requires less than a second to reconstruct a $512\times 512$ image on the GPU.

1,757 citations

Journal ArticleDOI
22 Apr 2010
TL;DR: This paper surveys the various options such training has to offer, up to the most recent contributions and structures of the MOD, the K-SVD, the Generalized PCA and others.
Abstract: Sparse and redundant representation modeling of data assumes an ability to describe signals as linear combinations of a few atoms from a pre-specified dictionary. As such, the choice of the dictionary that sparsifies the signals is crucial for the success of this model. In general, the choice of a proper dictionary can be done using one of two ways: i) building a sparsifying dictionary based on a mathematical model of the data, or ii) learning a dictionary to perform best on a training set. In this paper we describe the evolution of these two paradigms. As manifestations of the first approach, we cover topics such as wavelets, wavelet packets, contourlets, and curvelets, all aiming to exploit 1-D and 2-D mathematical models for constructing effective dictionaries for signals and images. Dictionary learning takes a different route, attaching the dictionary to a set of examples it is supposed to serve. From the seminal work of Field and Olshausen, through the MOD, the K-SVD, the Generalized PCA and others, this paper surveys the various options such training has to offer, up to the most recent contributions and structures.

1,345 citations

Journal ArticleDOI
TL;DR: The numerical experiments presented in this paper demonstrate that the discrete shearlet transform is very competitive in denoising applications both in terms of performance and computational efficiency.

972 citations

Journal ArticleDOI
Jiayi Ma1, Yong Ma1, Chang Li1
TL;DR: This survey comprehensively survey the existing methods and applications for the fusion of infrared and visible images, which can serve as a reference for researchers inrared and visible image fusion and related fields.

849 citations

Journal ArticleDOI
TL;DR: A novel fusion algorithm, named Gradient Transfer Fusion (GTF), based on gradient transfer and total variation (TV) minimization is proposed, which can keep both the thermal radiation and the appearance information in the source images.

729 citations

References
More filters
Book
01 May 1992
TL;DR: This paper presents a meta-analyses of the wavelet transforms of Coxeter’s inequality and its applications to multiresolutional analysis and orthonormal bases.
Abstract: Introduction Preliminaries and notation The what, why, and how of wavelets The continuous wavelet transform Discrete wavelet transforms: Frames Time-frequency density and orthonormal bases Orthonormal bases of wavelets and multiresolutional analysis Orthonormal bases of compactly supported wavelets More about the regularity of compactly supported wavelets Symmetry for compactly supported wavelet bases Characterization of functional spaces by means of wavelets Generalizations and tricks for orthonormal wavelet bases References Indexes.

16,073 citations


Additional excerpts

  • ...D filtering operations only....

    [...]

Journal ArticleDOI
TL;DR: In this article, the regularity of compactly supported wavelets and symmetry of wavelet bases are discussed. But the authors focus on the orthonormal bases of wavelets, rather than the continuous wavelet transform.
Abstract: Introduction Preliminaries and notation The what, why, and how of wavelets The continuous wavelet transform Discrete wavelet transforms: Frames Time-frequency density and orthonormal bases Orthonormal bases of wavelets and multiresolutional analysis Orthonormal bases of compactly supported wavelets More about the regularity of compactly supported wavelets Symmetry for compactly supported wavelet bases Characterization of functional spaces by means of wavelets Generalizations and tricks for orthonormal wavelet bases References Indexes.

14,157 citations

Journal ArticleDOI
TL;DR: A technique for image encoding in which local operators of many scales but identical shape serve as the basis functions, which tends to enhance salient image features and is well suited for many image analysis tasks as well as for image compression.
Abstract: We describe a technique for image encoding in which local operators of many scales but identical shape serve as the basis functions. The representation differs from established techniques in that the code elements are localized in spatial frequency as well as in space. Pixel-to-pixel correlations are first removed by subtracting a lowpass filtered copy of the image from the image itself. The result is a net data compression since the difference, or error, image has low variance and entropy, and the low-pass filtered image may represented at reduced sample density. Further data compression is achieved by quantizing the difference image. These steps are then repeated to compress the low-pass image. Iteration of the process at appropriately expanded scales generates a pyramid data structure. The encoding process is equivalent to sampling the image with Laplacian operators of many scales. Thus, the code tends to enhance salient image features. A further advantage of the present code is that it is well suited for many image analysis tasks as well as for image compression. Fast algorithms are described for coding and decoding.

6,975 citations


"The Nonsubsampled Contourlet Transf..." refers methods in this paper

  • ...The contourlet transform [14] is a multidirectional and multiscale transform that is constructed by combining the Laplacian pyramid [16], [17] with the directional filter bank (DFB) proposed in [18]....

    [...]

Book
01 Jul 1992
TL;DR: In this paper, a review of Discrete-Time Multi-Input Multi-Output (DIMO) and Linear Phase Perfect Reconstruction (QLP) QMF banks is presented.
Abstract: 1. Introduction 2. Review of Discrete-Time Systems 3. Review of Digital Filters 4. Fundamentals of Multirate Systems 5. Maximally Decimated Filter Banks 6. Paraunitary Perfect Reconstruction Filter Banks 7. Linear Phase Perfect Reconstruction QMF Banks 8. Cosine Modulated Filter Banks 9. Finite Word Length Effects 10. Multirate Filter Bank Theory and Related Topics 11. The Wavelet Transform and Relation to Multirate Filter Banks 12. Multidimensional Multirate Systems 13. Review of Discrete-Time Multi-Input Multi-Output LTI Systems 14. Paraunitary and Lossless Systems Appendices Bibliography Index

4,757 citations


"The Nonsubsampled Contourlet Transf..." refers methods in this paper

  • ...For example, image compression and denoising are efficiently done in the wavelet transform domain [1], [ 2 ]....

    [...]

Journal ArticleDOI
TL;DR: A "true" two-dimensional transform that can capture the intrinsic geometrical structure that is key in visual information is pursued and it is shown that with parabolic scaling and sufficient directional vanishing moments, contourlets achieve the optimal approximation rate for piecewise smooth functions with discontinuities along twice continuously differentiable curves.
Abstract: The limitations of commonly used separable extensions of one-dimensional transforms, such as the Fourier and wavelet transforms, in capturing the geometry of image edges are well known. In this paper, we pursue a "true" two-dimensional transform that can capture the intrinsic geometrical structure that is key in visual information. The main challenge in exploring geometry in images comes from the discrete nature of the data. Thus, unlike other approaches, such as curvelets, that first develop a transform in the continuous domain and then discretize for sampled data, our approach starts with a discrete-domain construction and then studies its convergence to an expansion in the continuous domain. Specifically, we construct a discrete-domain multiresolution and multidirection expansion using nonseparable filter banks, in much the same way that wavelets were derived from filter banks. This construction results in a flexible multiresolution, local, and directional image expansion using contour segments, and, thus, it is named the contourlet transform. The discrete contourlet transform has a fast iterated filter bank algorithm that requires an order N operations for N-pixel images. Furthermore, we establish a precise link between the developed filter bank and the associated continuous-domain contourlet expansion via a directional multiresolution analysis framework. We show that with parabolic scaling and sufficient directional vanishing moments, contourlets achieve the optimal approximation rate for piecewise smooth functions with discontinuities along twice continuously differentiable curves. Finally, we show some numerical experiments demonstrating the potential of contourlets in several image processing applications.

3,948 citations


"The Nonsubsampled Contourlet Transf..." refers background in this paper

  • ...These include adaptive schemes such as wedgelets [9], [10] and bandelets [11], and non-adaptive ones such as curvelets [12 ] and contourlets [13]....

    [...]

  • ...In particular, it can satisfy the a nisotropic scaling law — a key property in establishing the expansion nonlinear approxim ation behavior [12], [13]....

    [...]

Frequently Asked Questions (12)
Q1. What are the contributions mentioned in the paper "Nonsubsampled contourlet transform: filter design and applications in denoising" ?

In this paper the authors study the nonsubsampled contourlet transform. The authors show how zeroes can be imposed in the filters so that the iterated structure produces regular basis functions. 

The design methodology proposed also allows for a fast implementation through lifting factorization which halves the number of operations when compared to the direct form. 

The proposed transform structure with the new designed filters has proven very efficient in noise removal surpassing the undecimated discrete wavelet transform and being comparable (sometimes superior) to other stateofthe- art denoising methods. 

The directional filter bank [5] is constructed by combining critically sampled fan filter banks and pre/post re-sampling operations. 

thesupport of the high-pass filter is the complement of the low-pass support region on the [− π2j+1 , π 2j+1 ]× [− π 2j−1 , π 2j−1 ] square. 

The maximally flat diamond mapping polynomial is obtained by imposing flatness at the points (x, y) = (1, 1) and (x, y) = (0, 0). 

The liftingfactorization of the prototype filters is given by H(1D) 0 (x) H (1D) 1 (x)! =1 αx 0 11 0βx 11 γx 0 1K1 K2(7)with α = γ = 1 − √2, β = 1√ 2 , K1 = K2 = 1√ 2 . 

Suppose the authors want a mapping function f(x, y) such thatG (f(x, y)) = (x − c)Nx (y − d)Ny r(x, y) (2)where r(x, y) is a bivariate polynomial. 

Due to multirate identities, this is equivalent to switching off each of the downsamplers in the tree structure, while still keeping the re-sampling operations that can be absorbed by the filters. 

Following Proposition 1, the authors set f(x, y) = −1 + 2P2,4(x)P2,4(y) using the maximally flat polynomials:PN,L(x) := (1 + x) N L−1−NX l=0 N + l − 1 l! 2−N−l(1 − x)l. 

Following the work in [2] the authors choose the threshold locally usingTi,j,n = σ2nij σi,j,nwhere σi,j,n denotes the variance of the n-th coefficient at the i-th direction in the j-th scale and σ2nij the noise variance in the corresponding subband. 

The noise variance in each subband is inferred using a Monte-Carlo technique where the variances are computed for a few normalized AWGN images and then averaged to stabilize the results.