scispace - formally typeset
Open AccessJournal ArticleDOI

Direct kernel biased discriminant analysis: a new content-based image retrieval relevance feedback algorithm

Reads0
Chats0
TLDR
Experimental results are reported on a real-world image collection to demonstrate that the proposed methods outperform the traditional kernel BDA (KBDA) and the support vector machine (SVM) based RF algorithms.
Abstract
In recent years, a variety of relevance feedback (RF) schemes have been developed to improve the performance of content-based image retrieval (CBIR). Given user feedback information, the key to a RF scheme is how to select a subset of image features to construct a suitable dissimilarity measure. Among various RF schemes, biased discriminant analysis (BDA) based RF is one of the most promising. It is based on the observation that all positive samples are alike, while in general each negative sample is negative in its own way. However, to use BDA, the small sample size (SSS) problem is a big challenge, as users tend to give a small number of feedback samples. To explore solutions to this issue, this paper proposes a direct kernel BDA (DKBDA), which is less sensitive to SSS. An incremental DKBDA (IDKBDA) is also developed to speed up the analysis. Experimental results are reported on a real-world image collection to demonstrate that the proposed methods outperform the traditional kernel BDA (KBDA) and the support vector machine (SVM) based RF algorithms

read more

Content maybe subject to copyright    Report

BIROn - Birkbeck Institutional Research Online
Tao, D. and Tang, X. and Li, Xuelong and Rui, Y. (2006) Direct kernel
biased discriminant analysis: a new content-based image retrieval relevance
feedback algorithm. IEEE Transactions on Multimedia 8 (4), pp. 716-727.
ISSN 1520-9210.
Downloaded from:
Usage Guidelines:
Please refer to usage guidelines at or alternatively
contact lib-eprints@bbk.ac.uk.

Birkbeck ePrints: an open access repository of the
research output of Birkbeck College
http://eprints.bbk.ac.uk
Tao, Dacheng; Tang, Xiaoou; Li, Xuelong and Rui,
Yong (2006). Direct kernel biased discriminant
analysis: a new content-based image retrieval
relevance feedback algorithm. IEEE Transactions
on Multimedia 8 (4) 716-727.
Copyright © 2006 IEEE. Reprinted from IEEE Transactions on Multimedia
(ISSN 1520-9210). This material is posted here with permission of the IEEE.
Such permission of the IEEE does not in any way imply IEEE endorsement of
any of Birkbeck’s products or services. Internal or personal use of this
material is permitted. However, permission to reprint/republish this material for
advertising or promotional purposes or for creating new collective works for
resale or redistribution must be obtained from the IEEE by writing to pubs-
permissions@ieee.org.
By choosing to view this document, you agree to all provisions of the
copyright laws protecting it.
Citation for this copy:
Tao, Dacheng; Tang, Xiaoou; Li, Xuelong and Rui, Yong (2006). Direct kernel
biased discriminant analysis: a new content-based image retrieval relevance
feedback algorithm. London: Birkbeck ePrints. Available at:
http://eprints.bbk.ac.uk/archive/00000452
Citation as published:
Tao, Dacheng; Tang, Xiaoou; Li, Xuelong and Rui, Yong (2006). Direct kernel
biased discriminant analysis: a new content-based image retrieval relevance
feedback algorithm. IEEE Transactions on Multimedia 8 (4) 716-727.
http://eprints.bbk.ac.uk
Contact Birkbeck ePrints at lib-eprints@bbk.ac.uk

716 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 8, NO. 4, AUGUST 2006
Direct Kernel Biased Discriminant Analysis:
A New Content-Based Image Retrieval
Relevance Feedback Algorithm
Dacheng Tao, Student Member, IEEE, Xiaoou Tang, Senior Member, IEEE, Xuelong Li, Member, IEEE, and
Yong Rui, Senior Member, IEEE
Abstract—In recent years, a variety of relevance feedback (RF)
schemes have been developed to improve the performance of con-
tent-based image retrieval (CBIR). Given user feedback informa-
tion, the key to a RF scheme is how to select a subset of image fea-
tures to construct a suitable dissimilarity measure. Among various
RF schemes, biased discriminant analysis (BDA) based RF is one of
the most promising. It is based on the observation that all positive
samples are alike, while in general each negative sample is negative
in its own way. However, to use BDA, the small sample size (SSS)
problem is a big challenge, as users tend to give a small number
of feedback samples. To explore solutions to this issue, this paper
proposes a direct kernel BDA (DKBDA), which is less sensitive to
SSS. An incremental DKBDA (IDKBDA) is also developed to speed
up the analysis. Experimental results are reported on a real-world
image collection to demonstrate that the proposed methods outper-
form the traditional kernel BDA (KBDA) and the support vector
machine (SVM) based RF algorithms.
Index Terms—Biased discriminant analysis (BDA), content-
based image retrieval (CBIR), direct kernel biased discriminant
analysis (DKBDA), incremental direct kernel biased discriminant
analysis (IDKBDA), kernel biased discriminant analysis (KBDA),
relevance feedback (RF).
I. INTRODUCTION
W
ITH the explosive growth in image records and the rapid
increase of computer power, retrieving images from a
large-scale image database becomes one of the most active re-
search fields [1], [2]. To give all images text annotations manu-
ally is tedious and impractical and to automatically annotate an
image is beyond current technology.
Content-based image retrieval (CBIR) is a technique to re-
trieve images semantically relevant to the user’s query from an
image database. It is based on automatically extracted visual
features from an image, such as color [3], [4], [10]–[12], texture
[5]–[10], [12], and shape [11]–[13]. However, the gap between
these low-level visual features and high-level semantic mean-
ings usually leads to poor performance.
Manuscript received October 31, 2004; revised May 31, 2005. The work of
D. Tao and X. Tang was supported by grants from the Research Grants Council
of the Hong Kong (SAR). The associate editor coordinating the review of this
manuscript and approving it for publication was Dr. Jiebo Luo.
D. Tao and X. Li are with the School of Computer Science and Information
Systems, Birkbeck, University of London, London WC1E 7HX, U.K. (e-mail:
dacheng@dcs.bbk.ac.uk; xuelong@dcs.bbk.ac.uk).
X. Tang is with the Department of Information Engineering, The Chinese
University of Hong Kong, Shatin, N.T., Hong Kong (e-mail: xtang@ie.cuhk.
edu.hk).
Y. Rui is with the Microsoft Research, Redmond, WA 98052 USA (e–mail:
yongrui@microsoft.com).
Digital Object Identifier 10.1109/TMM.2005.861375
Relevance feedback (RF) is a way to bridge this gap and to
scale the performance in CBIR systems [14]–[17]. RF focuses
on the interactions between the user and the search engine by
letting the user label semantically positive or negative samples.
RF is different from the traditional classification problem be-
cause the user is not likely to label a large number of images.
As a result, small sample learning methods, where the number
of the training samples is much smaller than the dimension of
the descriptive features, are important in CBIR RF. Discrimi-
nant analysis [18]–[26] and the support vector machine (SVM)
method [27]–[31] are two small sample learning methods used
in recent years to obtain
state-of-the-art performances.
Discriminant analysis [18] is one of the most popular solu-
tions for the small sample learning problem. In the last 20 years,
Fisher linear discriminant analysis (LDA) has been successfully
used in face recognition [19]–[23], [26]. LDA was first used
in CBIR for feature selection and extracts the most discrimi-
nant subset feature for image retrieval. The remaining images
in the database were then projected onto the subspace and fi-
nally, some similarity or dissimilarity measures were used to
sort these images. However, with LDA all negative feedbacks
are deemed equivalent, and this is a severe limitation of the
method because all positive examples are alike and each neg-
ative example is negative in its own way. With this observation,
biased discriminant analysis (BDA) [24], [25] was developed
by Zhou and Huang to scale the performance of CBIR and ob-
tained a more satisfactory result. In the BDA model, the nega-
tive feedbacks are required to stay away from the center of the
positive feedbacks. Motivated by the kernel trick successfully
used in pattern recognition [32], Zhou et al. also generalized
the BDA to the kernel feature space as the kernel biased dis-
criminant analysis (KBDA). KBDA performs much better than
BDA [24], [25]. Just like LDA, BDA and KBDA also lead to
the small sample size (SSS) problem [33] because the number
of the sample is much smaller than the dimension of the repre-
sentative features of images. Traditionally, the SSS problem is
solved by the regularization method [33], [24], [25].
However, the regularization method to solve the SSS problem
is not a good choice for LDA, as is pointed out by many papers
on face recognition [20]–[23], [26]. We aim to significantly im-
prove the performance of CBIR RF and utilize the direct idea
to the BDA algorithm in the kernel feature space. We name the
approach as the direct kernel BDA (DKBDA) [34]. DKBDA is
motivated by a) direct LDA (DLDA) [23], [26], which has been
successfully applied to face recognition; (b) unlike face recogni-
tion, image retrieval deals with diverse images, so the nonlinear
1520-9210/$20.00 © 2006 IEEE

TAO et al.: DIRECT KERNEL BIASED DISCRIMINANT ANALYSIS 717
properties of image features should be considered because of
the success of kernel algorithms in pattern recognition.
The DKBDA algorithm can be regarded as an enhanced
KBDA. According to the kernel trick idea, the original input
space is rst nonlinearly mapped to an arbitrarily high dimen-
sion feature space, in which the distribution of the images
patterns is linearized. Then, the DLDA idea [23], [26] is used to
obtain a set of optimal discriminant basis vectors in the kernel
feature space. The BDA criterion is modied as in Liu
et al.
[20], so that a robust result can be gained.
The following section describes the related previous work:
BDA, KBDA, and DLDA; DKBDA is then proposed in Sec-
tion III; in Section IV, an image retrieval system is introduced;
in Section V, a large number of experiments validate the effec-
tiveness and efciency of DKBDA on a large real world image
database; possible future work is briey described in Section VI;
nally, Section VII draws conclusions. Detailed deduction of
DKBDA is given in Appendix A; Appendix B provides full de-
duction of the incremental DKBDA (IDKBDA).
II. P
REVIOUS WORK
In this section, previous work including Direct Linear Dis-
criminant Analysis (DLDA), Biased Discriminant Analysis
(BDA), and Kernel Biased Discriminant Analysis (KBDA) are
introduced.
A. Direct Linear Discriminant Analysis (DLDA)
Before describing DLDA [23], we rst describe linear dis-
criminant analysis (LDA) [18].
LDA tries to nd the best discriminating subspace for dif-
ferent classes. It is spanned by a set of vectors
, which aims
at maximizing the ratio between
and , the within-class
scatter matrix and the between-class scatter matrix:
(1)
Assume the training set contains
individual classes and each
class
has samples. Then and are dened as
(2)
where
. is the mean vector
of the total training set.
is the mean
vector for the individual class
. is the sample belongs to
class
. Therefore, can be computed from the eigenvectors
of
.Given equals 2, LDA changes to Fisher discrimi-
nant analysis (FDA); otherwise, multiple discriminant analysis
(MDA).
LDA has the SSS problem when the number of the training
samples is smaller than the dimension of the low-level visual
features, which is almost always true for CBIR RF.
Yu et al. [23] propose a DLDA method. It accepts high-
dimensional data as input, and optimizes Fishers criterion
directly without any feature extraction or dimension reduction
steps. So, it takes advantage of all the information within and
outside of the null space of
. In this approach, is rst
diagonalized, then the null space of
is removed:
(3)
where
comprises eigenvectors and comprises the corre-
sponding nonzero eigenvalues of
. is transformed to
(4)
where
is diagonalized by eigenanalysis:
(5)
The LDA transformation matrix is dened as
(6)
In DLDA, the null space of
is removed, and the discrimi-
nant vectors are restricted in the subspace spanned by class cen-
ters. It is assumed that the null space of
contains no discrim-
inative information at all.
B. Biased Discriminant Analysis (BDA)
Zhou et al. [24], [25] developed BDA, which denes the
-class classification problem. This means there is an unknown
number of classes but the user is only interested in one class.
BDA tries to nd the subspace to discriminate the positive
samples (the only class of concern to the user) and negative sam-
ples (unknown number of classes). It is spanned by a set of vec-
tors
maximizing the ratio between the biased matrix and
the positive covariance matrix
:
(7)
Assume the training set contains
positive and negative
samples.
and can be dened as (8):
(8)
where
denotes the positive samples, denotes the negative
samples, and
is the mean vector of the
positive samples.
can be computed from the eigenvectors
of
. Firstly, BDA minimizes the variance of the positive
samples. Then it maximizes the distance between the two cen-
ters of the positive feedbacks and all negative feedbacks.
C. Kernel Biased Discriminant Analysis (KBDA)
The data is in a nonlinear space, in which the kernel method is
successfullyused.Therefore,BDAisgeneralized toitskernelver-
sion, named as KBDA. To obtain the nonlinear generalization, the
linear input space is mapped to a nonlinear kernel feature space:
(9)
(10)

718 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 8, NO. 4, AUGUST 2006
The data is mapped from into a po-
tentially much higher dimensional feature space
. Now, given
a learning problem, one can consider the BDA in
instead of
. In other words, the idea behind KBDA is to perform the
BDA in the feature space
instead of the input space .
Let
and be the positive within-class scatter and the
negative scatter with respect to positive centroid matrices in the
feature space
. They can be respectively expressed as follows:
(11)
(12)
where
is the centroid of positive
samples,
is the positive samples number, and is the neg-
ative samples number. KBDA determines a set of optimal dis-
criminant basis vectors
, which, according to
eigenvectors of
, can be obtained to solve the following
eigenvalue problem:
(13)
The dimension of the feature space
is arbitrarily high, and
possibly innite. Fortunately, there is no need to use the exact
to calculate , because the kernel method can be utilized
to avoid mapping the feature point from the linear input space
to a nonlinear kernel feature space. This mapping is based on
replacing the dot product with a kernel function in
.
In KBDA based RF, the number of feedback samples is much
smaller than the dimension of the low-level visual feature. This
leads to a degenerated
, i.e., the SSS problem or the matrix
singular problem. Zhou et al. [24], [25] solve the SSS problem
by the regularized version
and , which adds small quanti-
ties to the diagonal of the scatter matrices. However, this is not
an optimal solution and sometimes it may lead to an ill-posed
problem, which limits the performance of their method.
III. D
IRECT KERNEL BIASED DISCRIMINANT ANALYSIS
(DKBDA) AND ITS INCREMENTAL VERSION
The regularization method to solve the SSS problem is not a
good choice for LDA, as is pointed out by many papers on face
recognition [20][23], and [26]. We aim to signicantly improve
the performance of CBIR RF and utilize the direct idea to the
BDA algorithm in the kernel feature space. This direct method
is proposed based on all positive examples are alike and each
negative example is negative in its own way [24], [25]. We name
the approach as the direct kernel BDA (DKBDA).
DKBDA is motivated by a) the fact that direct LDA (DLDA)
[23], [26], recently developed for face recognition, has made
some advances; and b) unlike face recognition, image retrieval
deals with diverse images, so the nonlinear properties of image
features should be considered because of the success of kernel
algorithms in pattern recognition.
DKBDA can be regarded as an enhanced KBDA. According
to the kernel trick idea, the original input space is rst nonlin-
early mapped to an arbitrarily high dimension feature space,
in which the distribution of the images patterns is linearized.
Then, the DLDA idea [23], [26] is used to obtain a set of op-
timal discriminant basis vectors in the kernel feature space. The
BDA criterion is modied as in Liu et al. [20], so that a robust
result can be gained. First of all, the kernel matrix
is intro-
duced:
(14)
where
where stands for positive feedback samples, and is the
number of positive feedback samples;
stands for negative
feedback samples, and
is the number of negative feedback
samples.
is the kernel function. Some typical kernel func-
tions can be employed, such as Polynomial, Gaussian, or Sig-
moid based kernel functions.
DKBDA begins from the analysis of the negative scatter with
respect to positive centroid matrix (11). Since the dimension of
could be arbitrarily innitive, it is impossible to calculate
directly and implement eigen analysis with .
Fortunately, this can be avoided through the following analysis:
(15)
The dimension of
is the number of negative RF sam-
ples. The next problem is to obtain the matrix:
(16)
, , , and
should then be calculated. The detailed deductions can be seen
from Appendix A and the results are given by the following
formulations:
(17)
(18)
(19)
where
is an by 1 column vector (all terms equal to 1).

Citations
More filters
Book ChapterDOI

Statistical Pattern Recognition

TL;DR: This chapter introduces the subject of statistical pattern recognition (SPR) by considering how features are defined and emphasizes that the nearest neighbor algorithm achieves error rates comparable with those of an ideal Bayes’ classifier.
Journal ArticleDOI

A survey of graph edit distance

TL;DR: The research advance of G ED is surveyed in order to provide a review of the existing literatures and offer some insights into the studies of GED.
Journal ArticleDOI

A comprehensive review of current local features for computer vision

TL;DR: This review focuses on detectors and local descriptors, which are widely utilized in a large number of applications, e.g., object categorization, image retrieval, robust matching, and robot localization.
Journal ArticleDOI

Gabor-Based Region Covariance Matrices for Face Recognition

TL;DR: A new method for human face recognition by utilizing Gabor-based region covariance matrices as face descriptors is presented, using both pixel locations and Gabor coefficients to form the covariant matrices.
Journal ArticleDOI

Marginal Fisher Analysis and Its Variants for Human Gait Recognition and Content- Based Image Retrieval

TL;DR: Extensions of the recently proposed marginal Fisher analysis (MFA) to address dimensionality reduction problems for human gait recognition and content-based image retrieval and deal with the relevance feedback problem by extending MFA to marginal biased analysis.
References
More filters
Book

The Nature of Statistical Learning Theory

TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Book

Matrix Analysis

TL;DR: In this article, the authors present results of both classic and recent matrix analyses using canonical forms as a unifying theme, and demonstrate their importance in a variety of applications, such as linear algebra and matrix theory.
Journal ArticleDOI

Content-based image retrieval at the end of the early years

TL;DR: The working conditions of content-based retrieval: patterns of use, types of pictures, the role of semantics, and the sensory gap are discussed, as well as aspects of system engineering: databases, system architecture, and evaluation.
Journal ArticleDOI

Color indexing

TL;DR: In this paper, color histograms of multicolored objects provide a robust, efficient cue for indexing into a large database of models, and they can differentiate among a large number of objects.
Journal ArticleDOI

Texture features for browsing and retrieval of image data

TL;DR: Comparisons with other multiresolution texture features using the Brodatz texture database indicate that the Gabor features provide the best pattern retrieval accuracy.
Related Papers (5)
Frequently Asked Questions (10)
Q1. What are the contributions in "Direct kernel biased discriminant analysis: a new content-based image retrieval relevance feedback algorithm" ?

To explore solutions to this issue, this paper proposes a direct kernel BDA ( DKBDA ), which is less sensitive to SSS. 

In the future, the authors plan to generalize the tuning method to select the parameters of kernel-based algorithms. For RF in CBIR, the training size of the training set is small, so the leave-one-out method to tune the parameters can be used. 

DKBDA is motivated by a) the fact that direct LDA (DLDA) [23], [26], recently developed for face recognition, has made some advances; and b) unlike face recognition, image retrieval deals with diverse images, so the nonlinear properties of imagefeatures should be considered because of the success of kernel algorithms in pattern recognition. 

In the case of the top ten results, after four iterations, the precision of the proposed DBDA is already higher than 90% while seven iterations are required for the BDA algorithm and more than nine iterations for the SVM RF. 

Generally in a CBIR RF system images are represented by the three main features: color [3], [4], and [10]–[12], texture [5]–[10], [12], and shape [11]–[13]. 

the user provides feedback by clicking on the “thumb up” or “thumb down” button according to his/her judgment of the relevance of the sorted images. 

Zhou et al. [24], [25] solve the SSS problem by the regularized version and , which adds small quantities to the diagonal of the scatter matrices. 

LDA has the SSS problem when the number of the training samples is smaller than the dimension of the low-level visual features, which is almost always true for CBIR RF. 

In their experiments, the computer does the relevance feedback iterations automatically without mislabeled samples using the 80 concept groups described previously. 

IDKBDA is proved to be of approximately the same capabilities as DKBDA, but it can speed up the DKBDA remarkably by saving about 20% of the running time (9 and 11 h for all the 300 queries and nine iterations for each query).