scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Case Retrieval in Medical Databases by Fusing Heterogeneous Information

01 Jan 2011-IEEE Transactions on Medical Imaging (IEEE Trans Med Imaging)-Vol. 30, Iss: 1, pp 108-118
TL;DR: In this paper, a content-based heterogeneous information retrieval framework was proposed to retrieve possibly incomplete documents, consisting of several images and semantic information, from a database; more complex data types such as videos can also be included in the framework.
Abstract: A novel content-based heterogeneous information retrieval framework, particularly well suited to browse medical databases and support new generation computer aided diagnosis (CADx) systems, is presented in this paper. It was designed to retrieve possibly incomplete documents, consisting of several images and semantic information, from a database; more complex data types such as videos can also be included in the framework. The proposed retrieval method relies on image processing, in order to characterize each individual image in a document by their digital content, and information fusion. Once the available images in a query document are characterized, a degree of match, between the query document and each reference document stored in the database, is defined for each attribute (an image feature or a metadata). A Bayesian network is used to recover missing information if need be. Finally, two novel information fusion methods are proposed to combine these degrees of match, in order to rank the reference documents by decreasing relevance for the query. In the first method, the degrees of match are fused by the Bayesian network itself. In the second method, they are fused by the Dezert-Smarandache theory: the second approach lets us model our confidence in each source of information (i.e., each attribute) and take it into account in the fusion process for a better retrieval performance. The proposed methods were applied to two heterogeneous medical databases, a diabetic retinopathy database and a mammography screening database, for computer aided diagnosis. Precisions at five of 0.809 ± 0.158 and 0.821 ± 0.177, respectively, were obtained for these two databases, which is very promising.

Summary (4 min read)

Introduction

  • The knowledge of medical experts is a mixture of textbook knowledge and experience through real life clinical cases, so the assumption that analogous problems have similar solutions makes sense to them.
  • Medical CBR systems are intended to be used as follows: should a physician be doubtful about his/her diagnosis, he/she can send the available data about the patient to the system; the system selects and displays the most similar documents, along with their associated medical interpretations, which may help him/her confirm or invalidate his/her diagnosis by analogy.
  • The first fusion operator is incorporated in the Bayesian network: the computation of the degree of match, with respect to a given attribute, relies on the design of conditional probabilities relating this attribute to the overall degree of match.
  • The main advantage of the proposed approach, over standard feature selection / feature classification approaches, is that a retrieval model is trained separately for each attribute.

A. Description of Bayesian Networks

  • A Bayesian network [22] is a probabilistic graphical model that represents a set of variables and their probabilistic dependencies.
  • It is a directed acyclic graph whose nodes represent variables, and whose edges encode conditional independencies between the variables.
  • Because a Bayesian network can completely model the variables and their relationships, it can be used to answer queries about them.
  • Typically, it is used to estimate unknown probabilities for a subset of variables when other variables (the evidence variables) are observed.
  • By transforming the network into a cycle-free hypergraph, and performing inference in this hypergraph, Lauritzen and Spiegelhalter proposed an exact inference algorithm with relatively low complexity [23]; this algorithm was used in the proposed system.

B. Learning a Bayesian Network from Data

  • A Bayesian network is defined by a structure and the conditional probability of each node given its parents in that structure (or its prior probability if it does not have any parent).
  • C. Including Images in a Bayesian Network Contextual information are included as usual in a Bayesian network: a variable with a finite set of states, one for each possible attribute value, is defined for each field.
  • The maximum likelihood estimators of the wavelet coefficient distribution in each subband are used as a signature.
  • In order to define several states for an “image variable”, similar images are clustered with an unsupervised classification algorithm, thanks to the image signatures and the associated distance measure above, also known as 2) Signature Clustering.
  • When each sample has been assigned a class label, mutual information between clusters and class labels can be used to determine the optimal number of clusters K̂ [29] (see equation (7)).

D. System Design

  • The design of the system is described hereafter and illustrated in Fig.
  • So, an intermediate network is built from data, using Cheng’s algorithm (see section II-B).
  • In their experiments, the query document xq belongs to the test dataset and x belongs to the training dataset.
  • E. Computing the Conditional Probabilities Pq(Ai = aij |Q) To compute Pq(Ai = aij |Q), the authors first estimate Pq(Q|Ai = aij): the probability that a reference document x, with full membership to the state aij of attribute Ai, is relevant.
  • The authors define the semantic similarity between documents x and xq , with respect to Ai, as follows: ∑ j ∑ k αij(x)Sijkαik(xq) (10) where Sikl is the correlation between two states aik and ail of Ai, regarding the class of the documents at these states.

F. Retrieval Process

  • The different reference documents in the database are then processed sequentially.
  • The reference documents in the database are then ranked in decreasing order of the computed posterior probability Pq(Q|x).

A. Description of the Dezert-Smarandache Theory

  • The Dezert-Smarandache Theory (DSmT) of plausible and paradoxical reasoning, proposed in recent years [17], lets us combine any types of independent sources of information represented in term of belief functions.
  • DSmT is mainly focused on the fusion of uncertain, highly conflicting and imprecise sources of evidence.
  • The belief mass functions mi must be first specified by the user for each source of information, i = 1..M (mi functions used in their system are described below, in paragraph III-C).
  • Another difference between DST and DSmT comes from the underlying rules of combinations.
  • Pignistic probability [32], a possible compromise, is used instead (see below, in paragraph III-D); other probabilistic transformations are available [33].

C. System Design

  • The sensitivity (resp. the specificity) represents the degree of confidence in a positive (resp. negative) answer to test Ti; mi(Q ∪ Q̄) is assigned the degree of uncertainty.
  • The sensitivity of Ti, for a given threshold τi, is defined as the percentage of pairs of training documents (y1, y2) from the same class such that Ti(y1, y2) is true.
  • Similarly, the specificity of Ti is defined as the percentage of pairs of training documents (z1, z2) from different classes such that Ti(z1, z2) is false.
  • As τi increases, sensitivity increases and specificity decreases.

D. Retrieval Process

  • If the ith attribute of xq is available, the degree of match dmi(x, xq) is computed according to αij(x) (see equation 13) and the belief masses are computed according to test Ti(x, xq).
  • The sources available for xq are then fused.
  • Once the sources available for xq are fused by the proposed rule of combination, the pignistic probability betP (Q) is computed following equation 16.
  • The reference documents are then ranked in decreasing order of betP (Q).

IV. APPLICATION TO MEDICAL IMAGE DATABASES

  • The proposed method has been applied to CADx on two heterogeneous databases.
  • First, it has been applied to diabetic retinopathy severity assessment on a dataset (DRD) built at the Inserm U650 laboratory, in collaboration with ophthalmologists of Brest University Hospital.
  • Then, it has been applied to breast cancer screening on a public access database (DDSM).

A. Diabetic Retinopathy Database (DRD)

  • The diabetic retinopathy database contains retinal images of diabetic patients, with associated anonymized information on the pathology.
  • The latter case, namely diabetic retinopathy, can lead to blindness.
  • Patients records consist of at most 10 images per eye (see Fig. 6) and 13 contextual attributes; 12.1% of these images and 40.5% of these contextual attribute values are missing.
  • The disease severity level, according to ICDRS classification [35], was assessed by a single expert for all 67 patients: because of intra-observer variability, the reference standard is imperfect.

B. Digital Database for Screening Mammography (DDSM)

  • The DDSM project [36], involving the Massachusetts General Hospital, the University of South Florida and the Sandia National laboratories, has built a mammographic image database for research on breast cancer screening.
  • Each patient file has been graded by a physician.
  • Patients are then classified in three groups: normal, benign and cancer.
  • The reference standard is also affected by intra- and interobserver variability in this dataset.

C. Objective of the System

  • For each query document, the authors want to retrieve the most similar reference documents in a given database.
  • The average precision at K measures how good a fusion method is at combining feature-specific distance measures into a semantically meaningful distance measure.

D. Patient File Features

  • In those databases, each patient file consists of both digital images and contextual information.
  • Contextual attributes (13 in DRD, 2 in DDSM) are processed as-is in the CBR system.
  • A possible solution is to segment these images and extract domain specific information (such as the number of lesions); for DRD, the number of automatically detected microaneurysms (the most frequent lesion of diabetic retinopathy) [37] is used.
  • This kind of approach requires expert knowledge and a robust segmentation of images, which is not always possible because of acquisition variability.
  • So, an additional solution to characterize images by their digital content, without segmenting images, is proposed: a feature vector is extracted from the wavelet decomposition of the image [25].

E. Training and Test Sets

  • Each subset Vi, i = 1..5, is used in turn as test set while the remaining four subsets are used for training the retrieval system.
  • Note that the test set is completely independent from the training process.

F. Results

  • Precisions obtained with each fusion method are reported in table III.
  • For both databases, at K = 5, the average precision is greater than 0.8; it means that, on average, more than 80% of the selected documents are relevant for a query.
  • To assess the performance of the proposed fusion framework, independently of the underlying image signatures (described in II-C1), it was compared to an early fusion [6] and a late fusion method [8] based on the same image signatures.
  • The results the authors obtained for these methods are summarized in table III.

V. DISCUSSION AND CONCLUSIONS

  • The authors introduced two methods to include image series and their signatures, with contextual information, in a CBR system.
  • The first method uses a Bayesian network to model the relationships between attributes.
  • The wavelet transform based signature, presented in section II-C, can be applied to any n-dimensional digital signal, using its n-dimensional wavelet transform (n = 1 for sound, n = 3 for video, etc) [39].
  • On DRD for instance, a precision at five of 0.6 can be reached by providing less than 30% of the attributes (see Fig. 8): with this precision, the majority of the retrieved documents (3 out of 5) belong to the right class.
  • Table III shows that the difference, in terms of retrieval performance, between single image retrieval [25] and heterogeneous document retrieval, comes from the combination of image features extracted from several images, more than the inclusion of contextual attributes.

Did you find this useful? Give us your feedback

Figures (12)

Content maybe subject to copyright    Report

HAL Id: hal-00704248
https://hal.archives-ouvertes.fr/hal-00704248
Submitted on 5 Jun 2012
HAL is a multi-disciplinary open access
archive for the deposit and dissemination of sci-
entic research documents, whether they are pub-
lished or not. The documents may come from
teaching and research institutions in France or
abroad, or from public or private research centers.
L’archive ouverte pluridisciplinaire HAL, est
destinée au dépôt et à la diusion de documents
scientiques de niveau recherche, publiés ou non,
émanant des établissements d’enseignement et de
recherche français ou étrangers, des laboratoires
publics ou privés.
Case retrieval in medical databases by fusing
heterogeneous information.
Gwénolé Quellec, Mathieu Lamard, Guy Cazuguel, Christian Roux, Béatrice
Cochener
To cite this version:
Gwénolé Quellec, Mathieu Lamard, Guy Cazuguel, Christian Roux, Béatrice Cochener. Case retrieval
in medical databases by fusing heterogeneous information.. IEEE Transactions on Medical Imaging,
Institute of Electrical and Electronics Engineers, 2011, 30 (1), pp.108-18. �10.1109/TMI.2010.2063711�.
�hal-00704248�

IEEE TRANSACTIONS ON MEDICAL IMAGING 1
Case Retrieval in Medical Databases by Fusing
Heterogeneous Information
Gw
´
enol
´
e Quellec, Mathieu Lamard, Guy Cazuguel, Member, IEEE, Christian Roux, Fellow Member, IEEE and
B
´
eatrice Cochener
Abstract—A novel content-based heterogeneous information
retrieval framework, particularly well suited to browse med-
ical databases and support new generation Computer Aided
Diagnosis (CADx) systems, is presented in this paper. It was
designed to retrieve possibly incomplete documents, consisting
of several images and semantic information, from a database;
more complex data types such as videos can also be included in
the framework. The proposed retrieval method relies on image
processing, in order to characterize each individual image in a
document by their digital content, and information fusion. Once
the available images in a query document are characterized,
a degree of match, between the query document and each
reference document stored in the database, is defined for each
attribute (an image feature or a metadata). A Bayesian network
is used to recover missing information if need be. Finally, two
novel information fusion methods are proposed to combine these
degrees of match, in order to rank the reference documents
by decreasing relevance for the query. In the first method, the
degrees of match are fused by the Bayesian network itself. In
the second method, they are fused by the Dezert-Smarandache
theory: the second approach lets us model our confidence in each
source of information (i.e. each attribute) and take it into account
in the fusion process for a better retrieval performance. The
proposed methods were applied to two heterogeneous medical
databases, a diabetic retinopathy database and a mammography
screening database, for computer aided diagnosis. Precisions at
five of 0.809±0.158 and 0.821±0.177, respectively, were obtained
for these two databases, which is very promising.
Index Terms—Medical databases, Heterogeneous information
retrieval, Information fusion, Diabetic retinopathy, Mammogra-
phy
I. INTRODUCTION
T
WO main tasks in Computer Aided Diagnosis (CADx)
using medical images are extraction of relevant informa-
tion from images and combination of the extracted features
with other sources of information to automatically or semi-
automatically generate a reliable diagnosis. One promising
Copyright (c) 2010 IEEE. Personal use of this material is permitted.
However, permission to use this material for any other purposes must be
obtained from the IEEE by sending a request to pubs-permissions@ieee.org.
G. Quellec, G. Cazuguel, and C. Roux are with the INSTITUT TELE-
COM/TELECOM Bretagne, Dpt ITI, Brest, F-29200 France, and also with
the Institut National de la Sant
´
e et de la Recherche M
´
edicale (INSERM),
U650, Brest, F-29200 France (e-mail: gwenole.quellec@telecom-bretagne.eu;
guy.cazuguel@telecom-bretagne.eu; christian.roux@telecom-bretagne.eu).
M. Lamard is with the University of Bretagne Occidentale, Brest, F-
29200 France, and also with the Institut National de la Sant
´
e et de
la Recherche M
´
edicale (INSERM), U650, Brest, F-29200 France (e-mail:
mathieu.lamard@univ-brest.fr).
B. Cochener is with the Centre Hospitalier Universitaire de Brest, Service
d’Ophtalmologie, Brest, F-29200 France, also with the University of Bretagne
Occidentale, Brest, F-29200 France, and also with the Institut National de la
Sant
´
e et de la Recherche M
´
edicale (INSERM), U650, Brest, F-29200 France
(e-mail: Beatrice.Cochener-lamard@chu-brest.fr)
way to achieve the second goal is to take advantage of
the growing number of digital medical databases either for
heterogeneous data mining, i.e. for extracting new knowledge,
or for heterogeneous information retrieval, i.e. for finding
similar heterogeneous medical records (e.g. consisting of
digital images and metadata). This paper presents a generic
solution to use digital medical databases for heterogeneous
information retrieval, and solve CADx problems using Case-
Based Reasoning (CBR) [
1].
CBR was introduced in the early 1980s as a new decision
support tool. It relies on the idea that analogous problems have
similar solutions. In CBR, interpreting a new situation revolves
around the retrieval of relevant documents in a case database.
The knowledge of medical experts is a mixture of textbook
knowledge and experience through real life clinical cases, so
the assumption that analogous problems have similar solutions
makes sense to them. This is the reason why there is a growing
interest in CBR for the development of medical decision
support systems [2]. Medical CBR systems are intended to
be used as follows: should a physician be doubtful about
his/her diagnosis, he/she can send the available data about
the patient to the system; the system selects and displays the
most similar documents, along with their associated medical
interpretations, which may help him/her confirm or invalidate
his/her diagnosis by analogy. Therefore, the purpose of such
a system is not to replace physicians’ diagnosis, but rather to
aid their diagnosis. Medical documents often consist of digital
information such as images and symbolic information such as
clinical annotations. In the case of Diabetic Retinopathy, for
instance, physicians analyze heterogeneous series of images
together with contextual information such as the age, sex and
medical history of the patient. Moreover, medical information
is sometimes incomplete and uncertain, two problems that
require a particular attention. As a consequence, original
CBR systems, designed to process simple documents such
as homogeneous and comprehensive attribute vectors, are
clearly unsuited to complex CADx applications. On one hand,
some CBR systems have been designed to manage symbolic
information [3]. On the other hand, some others, based on
Content-Based Image Retrieval [
4], have been designed to
manage digital images [
5]. However, few attempts have been
made to merge the two kinds of approaches. We consider in
this paper a larger class of problems: CBR in heterogeneous
databases.
To retrieve heterogeneous information, some simple ap-
proaches, based on early fusion (i.e. attributes are fused in
feature space) [6], [7] or late fusion (i.e. attributes are fused
in semantic space) [8], [9], [10] have been presented in the

IEEE TRANSACTIONS ON MEDICAL IMAGING 2
literature. A few application-specific approaches [
11], [12],
[
13], [14], [15], as well as a generic retrieval system, based
on dissimilarity spaces and relevance feedback [
16], have also
been presented. We introduce in this paper a novel generic
approach that does not require relevance feedback from the
user. The proposed system is able to manage incomplete
information and the aggregation of heterogeneous attributes:
symbolic and multidimensional digital information (we focus
on digital images, but the same principle can be applied to
any n-dimensional signals). The proposed approach is based
on a Bayesian network and the Dezert-Smarandache theory
(DSmT) [
17]. Bayesian networks have been used previously
in retrieval systems, either for keyword based retrieval [
18],
[
19] or for content-based image or video retrieval [20], [21].
The Dezert-Smarandache theory is more and more widely used
in remote sensing applications [17], however, to our knowl-
edge, this is its first medical application. In our approach, a
Bayesian network is used to model the relationships between
the different attributes (the extracted features of each digital
image and each contextual information field): we associate
each attribute with a variable in the Bayesian network. It lets us
compare incomplete documents: the Bayesian network is used
to estimate the probability of unknown variables (associated
with missing attributes) knowing the value of other variables
(associated with available attributes). Information coming from
each attribute is then used to derive an estimation of the degree
of match between a query document and a reference document
in the database. Then, these estimations are fused; two fusion
operators are introduced in this paper for this purpose. The
first fusion operator is incorporated in the Bayesian network:
the computation of the degree of match, with respect to a
given attribute, relies on the design of conditional probabilities
relating this attribute to the overall degree of match. An
evolution of this fusion operator that models our confidence in
each source of information (i.e. each attribute) is introduced. It
is based on the Dezert-Smarandache theory. In order to model
our confidence in each source of information, within this
second fusion operator, an uncertainty component is included
in the belief mass function characterizing the evidence coming
from this source of information.
The main advantage of the proposed approach, over standard
feature selection / feature classification approaches, is that a
retrieval model is trained separately for each attribute. This
is useful to process incomplete documents: in the proposed
approach, we simply combine the models associated with all
available attributes; as a comparison, a standard classifier relies
on feature combinations, and therefore may become invalid
when input feature vectors are incomplete. Also, because each
attribute is processed separately, the curse of dimensionality
is avoided. Therefore, it is not necessary to select the most
relevant features: instead, we simply weight each feature by a
confidence measure.
The paper is organized as follows. Section
II presents
the proposed Bayesian network based retrieval. Section III
presents the Bayesian network and Dezert-Smarandache theory
based retrieval. These methods are applied in section IV to
CADx in two heterogeneous databases: a diabetic retinopa-
thy database and a mammography database. We end with a
discussion and a conclusion in section
V.
II. BAYESIAN NETWORK BASED RETRIEVAL
A. Description of Bayesian Networks
A Bayesian network [
22] is a probabilistic graphical model
that represents a set of variables and their probabilistic depen-
dencies. It is a directed acyclic graph whose nodes represent
variables, and whose edges encode conditional independencies
between the variables. Examples of Bayesian networks are
given in Fig.
1.
(a) (b) (c)
Fig. 1. Examples of Bayesian Networks. Fig. (a) shows a chain. Fig. (b)
shows a polytree, i.e. a network in which there is at most one (undirected)
path between two nodes. Fig. (c) shows a network containing a cycle: <
A, D, E, C, A >.
In the example of Fig.
1 (b), the edge from the parent
node A to its child node D indicates that variable A has a
direct influence on variable D. Each edge in the graph is
associated with a conditional probability matrix expressing
the probability of a child variable given one of its parent
variables. For instance, if A = { a
0
, a
1
} and D = {d
0
, d
1
, d
2
},
then A D is assigned the following (3 × 2) conditional
probability matrix P (D|A):
P (D|A) =
P (D = d
0
|A = a
0
) P (D = d
0
|A = a
1
)
P (D = d
1
|A = a
0
) P (D = d
1
|A = a
1
)
P (D = d
2
|A = a
0
) P (D = d
2
|A = a
1
)
(1)
A directed acyclic graph is a Bayesian Network relative
to a set of variables {X
1
, . . . , X
n
} if the joint distribution
P (X
1
, . . . , X
n
) can be expressed as in equation
2:
P (X
1
, . . . , X
n
) =
n
i=1
P (X
i
|parents(X
i
)) (2)
where parents(X) is the set of nodes such that Y X is in
the graph Y parents(X). Because a Bayesian network
can completely model the variables and their relationships,
it can be used to answer queries about them. Typically, it
is used to estimate unknown probabilities for a subset of
variables when other variables (the evidence variables) are
observed. This process of computing the posterior distribution
of variables, given evidence, is called probabilistic inference.
In Bayesian networks containing cycles, exact inference is
a NP-hard problem. Approximate inference algorithms have
been proposed, but their accuracies depend on the network’s
structure; therefore, they are not general. By transforming the
network into a cycle-free hypergraph, and performing infer-
ence in this hypergraph, Lauritzen and Spiegelhalter proposed
an exact inference algorithm with relatively low complexity
[
23]; this algorithm was used in the proposed system.

IEEE TRANSACTIONS ON MEDICAL IMAGING 3
B. Learning a Bayesian Network from Data
A Bayesian network is defined by a structure and the
conditional probability of each node given its parents in that
structure (or its prior probability if it does not have any parent).
These parameters can be learned automatically from data.
Defining the structure consists in finding pairs of nodes (X, Y )
directly dependent, i.e. such that:
X and Y are not independent (P (X, Y ) 6= P (X)P (Y ))
There is no node set Z such that X and Y are indepen-
dent given Z (P (X, Y |Z) 6= P (X|Z)P (Y |Z))
Independence and conditional independence can be assessed
by mutual information (see equation 3) and conditional mutual
information (see equation
4), respectively.
I(X, Y ) =
x,y
P (x, y) log
P (x, y)
P (x)P (y)
(3)
I(X, Y |Z) =
x,y,z
P (x, y, z) log
P (x, y|z)
P (x|z)P (y|z)
(4)
Two nodes are independent (resp. conditionally independent)
if mutual information (resp. conditional mutual information)
is smaller than a given threshold ǫ, 0 ǫ < 1. Ideally, ǫ
should be equal to 0. However, in the presence of noise, some
meaningless edges (links) can appear. These edges can also
unnecessarily increase the computation time. To avoid this, in
this study, ǫ was chosen in advance to be equal to 0.1. This
number is independent of dataset cardinality [
24].
The structure of the Bayesian network, as well as edge
orientation, was obtained by Cheng’s algorithm [24]. This
algorithm was chosen for its complexity: complexity is poly-
nomial in the number of variables, as opposed to exponential
in competing algorithms.
C. Including Images in a Bayesian Network
Contextual information are included as usual in a Bayesian
network: a variable with a finite set of states, one for each
possible attribute value, is defined for each field.
To include images in a Bayesian network, we first define a
variable for each image in the sequence. For each “image
variable”, we follow the usual steps of Content-Based Image
Retrieval (CBIR) [
4]: 1) building a signature for each image
(i.e. extracting a feature vector summarizing their digital
content), and 2) defining a distance measure between two
signatures (see section
II-C1). Thus, measuring the distance
between two images comes down to measuring the distance
between two signatures. Similarly, in a Bayesian network,
defining states for an “image variable” comes down to defining
states for the signature of the corresponding images. To
this aim, similar image signatures are clustered, as described
below, and each cluster is associated with a state. Thanks to
this process, image signatures can be included in a Bayesian
network like any other variable.
1) Image Signature and Distance Measure: in previous
works on CBIR, we proposed to extract a signature for
images from their wavelet transform [
25]. These signatures
model the distribution of the wavelet coefficients in each
subband of the decomposition; as a consequence they provide
a multiscale description of images. To characterize the wavelet
coefficient distribution in a given subband, Wouwer’s work
was applied [
26]: Wouwer has shown that this distribution can
be modeled by a generalized Gaussian function. The maximum
likelihood estimators of the wavelet coefficient distribution
in each subband are used as a signature. These estimators
can be computed directly from wavelet-based compressed
images (such as JPEG-2000 compressed images), which can
be useful when a large number of images has to be processed.
A simplified version of Do’s generalized Gaussian parameter
estimation method [
27], [25] is proposed in appendix A to
reduce computation times. Any wavelet basis can be used to
decompose images. However, the effectiveness of the extracted
signatures largely depends on the choice of this basis. For
this reason, we proposed to search for an optimal wavelet
basis [
25] within the lifting scheme framework, which is
implemented in the compression standards. To compare two
signatures, Do proposed the use of the Kullback-Leibler di-
vergence between wavelet coefficient distributions P and Q
in two subbands [27]:
D(P ||Q) =
R
p(x) log
p(x)
q(x)
dx (5)
where p and q are the densities of P and Q, respectively.
A symmetric version of the Kullback-Leibler divergence was
used, since clustering algorithms require (symmetric) distance
measures:
1
2
(D(P ||Q) + D(Q||P )) (6)
Finally, the distance between two images is defined as a
weighted sum of these distances over the subbands, noted
W SD; weights are tuned by a genetic algorithm to maximize
retrieval performance on the training set [
25]. The ability to
select a weight vector and a wavelet basis makes this image
representation highly tunable. We have shown in previous
works the superiority of the proposed image signature, in
terms of retrieval performance, over several well-known image
signatures [
25].
2) Signature Clustering: in order to define several states
for an “image variable”, similar images are clustered with
an unsupervised classification algorithm, thanks to the image
signatures and the associated distance measure above. Any
algorithm can be used, provided that the distance measure
can be specified. We chose the well-known Fuzzy C-Means
algorithm (FCM) [
28] and replaced the Euclidean distance by
W SD described above. In this algorithm, each document is
assigned to each cluster k = 1..K with a fuzzy membership
u
k
, 0 u
k
1, such that
K
k=1
u
k
= 1, which can
be interpreted as a probability. Finding the right number of
clusters is generally a difficult problem. However, when each
sample has been assigned a class label, mutual information
between clusters and class labels can be used to determine the
optimal number of clusters
ˆ
K [
29] (see equation (7)).
ˆ
K = argmax
K
C
c=1
K
k=1
P (c, k) log
C+K
P (c, k)
P (c)P (k)
(7)
where c = 1..C are the class labels, P (c, k) is the joint proba-
bility distribution function of the class and cluster labels, P (c)

IEEE TRANSACTIONS ON MEDICAL IMAGING 4
and P (k) are the marginal probability distribution functions.
Other continuous variables can be discretized similarly: the
age of a person, one-dimensional signals, videos, etc.
D. System Design
QUERY - case in the testing setOFFLINE - on the training set
Learn the probabilistic
relationships
between variables
(section II.B).
Compute the correlation
between two
states of a variable
(section II-E2).
Intermediate network
(Fig 3(a))
Correlations
Compute the probabilistic
relationships between the
variables and the query node
(section II-E).
Add a query node Q to
the intermediate network
(section II-D).
Query-specific network
(Fig 3(b))
case x in the training set
Probabilistic inference on
the query-specific network
using x as evidence
(sections II-A, II-F)
Cases in the training set
ranked in decreasing
order of P(Q|x)
Fig. 2. Bayesian Network based Retrieval. Solid-lined arrows mean “leads
to” or “is followed by” and dashed-lined arrows mean “is used by”.
Let x
q
be a query document and M be the number of
attributes.
Definition
: A document x is said to be relevant for x
q
if x
and x
q
belong to the same class.
To assess the relevance of each reference document in a
database for x
q
, we define a Bayesian network with the
following variables:
a set of variables {A
i
, i = 1..M }, where A
i
represents
the i
th
attribute of x,
a Boolean variable Q = x is relevant for x
q
(
¯
Q = x
is not relevant for x
q
”).
The design of the system is described hereafter and illustrated
in Fig.
2. To build the network, the first step is to learn the
different relationships between the attributes {A
i
, i = 1..M}.
So, an intermediate network is built from data, using Cheng’s
algorithm (see section
II-B). In that purpose, the studied
database is divided into a training dataset and a test dataset.
Cheng’s algorithm is applied to the training dataset. In our
experiments, the query document x
q
belongs to the test dataset
and x belongs to the training dataset. To build this Bayesian
network, a finite number of states a
ij
is defined for each
variable A
i
, i = 1..M . To learn the relationships between these
variables, we use the membership degree of any document y
in the training dataset to each state a
ij
of each variable A
i
,
noted α
ij
(y). If A
i
is a nominal variable, α
ij
(y) is boolean;
for instance, if y is a male then α
sex
′′
,male
′′
(y) = 1 and
α
sex
′′
,female
′′
(y) = 0. If A
i
is a continuous variable (such
as an image-based feature), α
ik
(y) is the fuzzy membership
of y to each cluster k = 1..K (see section
II-C2). An example
of intermediate network is given in Fig. 3 (a).
(a) Intermediate network
(b) Query-specific network
Fig. 3. Retrieval Bayesian Network (built for the database presented in
section
IV-A). In the example of Fig. (b), attributes A
1
, ..., A
6
, A
8
, A
10
,
A
13
, A
14
, A
15
, A
17
, A
18
, A
22
, A
23
are available for the query document
x
q
, so the associated nodes are then connected to node Q.
Q is then integrated in the network. For retrieval, the
attributes of x are observable evidences for Q, as a con-
sequence the associated variables should be descendants of
Q. In the retrieval network, the probabilistic dependences
between Q and each variable A
i
depend on x
q
. In fact, x
q
specifies which attributes should be found in the retrieved
documents in order to meet the user’s needs. So, when the
i
th
attribute of x
q
is available, we connect the two nodes Q
and A
i
and we estimate the associated conditional probability
matrix P
q
(A
i
= a
ij
|Q) according to x
q
(see Fig.
3 (b)).
The index q denotes that the probability depends on x
q
. A
query-specific network is obtained: its structure depends on
which attributes are available for the query document and the
conditional probability matrices depend on the value taken for
these available attributes by the query document. This network
is used to assess the relevance of any reference document for
x
q
.
E. Computing the Conditional Probabilities P
q
(A
i
= a
ij
|Q)
To compute P
q
(A
i
= a
ij
|Q), we first estimate P
q
(Q|A
i
=
a
ij
): the probability that a reference document x, with full
membership to the state a
ij
of attribute A
i
, is relevant.
P
q
(A
i
= a
ij
|Q) can then be computed thanks to Bayes’
theorem (see equation (8)). The prior probability
P
q
(
Q
)
is
required; it can be estimated by the probability that two
documents belong to the same class, i.e. the probability that
both documents belong to class 1 or that both documents
belong to class 2, etc., hence equation 9:
P (A|B) =
P (B|A)P (A)
P (B)
(8)
P
q
(Q) =
C
c=1
(P (c))
2
(9)

Citations
More filters
Journal ArticleDOI
TL;DR: This paper presents a review of state-of-the-art medical CBIR approaches in five main categories: two-dimensional image retrieval, retrieval of images with three or more dimensions, the use of nonimage data to enhance the retrieval, multimodality image retrieved, and retrieval from diverse datasets.
Abstract: Medical imaging is fundamental to modern healthcare, and its widespread use has resulted in the creation of image databases, as well as picture archiving and communication systems. These repositories now contain images from a diverse range of modalities, multidimensional (three-dimensional or time-varying) images, as well as co-aligned multimodality images. These image collections offer the opportunity for evidence-based diagnosis, teaching, and research; for these applications, there is a requirement for appropriate methods to search the collections for images that have characteristics similar to the case(s) of interest. Content-based image retrieval (CBIR) is an image search technique that complements the conventional text-based retrieval of images by using visual features, such as color, texture, and shape, as search criteria. Medical CBIR is an established field of study that is beginning to realize promise when applied to multidimensional and multimodality medical data. In this paper, we present a review of state-of-the-art medical CBIR approaches in five main categories: two-dimensional image retrieval, retrieval of images with three or more dimensions, the use of nonimage data to enhance the retrieval, multimodality image retrieval, and retrieval from diverse datasets. We use these categories as a framework for discussing the state of the art, focusing on the characteristics and modalities of the information used during medical image retrieval.

194 citations

Journal ArticleDOI
TL;DR: In this paper, the authors present a list of recommendations focusing on the creation of large repositories of test data created by international consortia, easily accessible via moderated Web sites, including multicenter annotations by multiple experts, specific to clinical tasks.
Abstract: This paper concerns the validation of automatic retinal image analysis (ARIA) algorithms. For reasons of space and consistency, we concentrate on the validation of algorithms processing color fundus camera images, currently the largest section of the ARIA literature. We sketch the context (imaging instruments and target tasks) of ARIA validation, summarizing the main image analysis and validation techniques. We then present a list of recommendations focusing on the creation of large repositories of test data created by international consortia, easily accessible via moderated Web sites, including multicenter annotations by multiple experts, specific to clinical tasks, and capable of running submitted software automatically on the data stored, with clear and widely agreed-on performance criteria, to provide a fair comparison.

166 citations

Journal ArticleDOI
TL;DR: This survey will directly help researchers understand the research developments of MSIF under RST and provide state-of-the-art understanding in specialized literature, as well as clarify the approaches and application of MSif in RST research community.
Abstract: Multi-Source Information Fusion (MSIF) is a comprehensive and interdisciplinary subject, and is referred to as, multi-sensor information fusion which was originated in the 1970s. Nowadays, the types and updates of data are becoming more multifarious and frequent, which bring new challenges for information fusion to deal with the multi-source data. Consequently, the construction of MSIF models suitable for different scenarios and the application of different fusion technologies are the core problems that need to be solved urgently. Rough set theory (RST) provides a computing paradigm for uncertain data modeling and reasoning, especially for classification issues with noisy, inaccurate or incomplete data. Furthermore, due to the rapid development of MSIF in recent years, the methodologies of learning under RST are becoming increasingly mature and systematic, unveiling a framework which has not been mentioned in the literature. In order to better clarify the approaches and application of MSIF in RST research community, this paper reviews the existing models and technologies from the perspectives of MSIF model (i.e., homogeneous and heterogeneous MSIF model), multi-view rough sets information fusion model (i.e., multi-granulation, multi-scale and multi-view decisions information fusion models), parallel computing information fusion model, incremental learning fusion technology and cluster ensembles fusion technology. Finally, RST based MSIF related research directions and challenges are also covered and discussed. By providing state-of-the-art understanding in specialized literature, this survey will directly help researchers understand the research developments of MSIF under RST.

105 citations

Journal ArticleDOI
TL;DR: Results show that the proposed approach can be implemented in real practice for analysing noisy radiography images, which have many useful medical applications such as diagnosis of diseases related to lung, breast, musculoskeletal or pediatric studies.
Abstract: This paper introduces a three-step framework for classifying multiclass radiography images The first step utilizes a de-noising technique based on wavelet transform (WT) and the statistical Kolmogorov Smirnov (KS) test to remove noise and insignificant features of the images An unsupervised deep belief network (DBN) is designed for learning the unlabelled features in the second step Although small-scale DBNs have demonstrated significant potential, the computational cost of training the restricted Boltzmann machine is a major issue when scaling to large networks Moreover, noise in radiography images can cause a significant corruption of information that hinders the performance of DBNs The combination of WT and KS test in the first step helps improve performance of DBNs Discriminative feature subsets obtained in the first two steps serve as inputs into classifiers in the third step for evaluations Five frequently used classifiers including naive Bayes, radial basis function network, random forest, sequential minimal optimization, and support vector machine and four different case studies are implemented for experiments using the Image Retrieval in Medical Application data set The experimental results show that the three-step framework has significantly reduced computational cost and yielded a great performance for multiclass radiography image classification Along with effective applications in image processing in other fields published in the literature, deep learning network in this paper has again demonstrated its robustness in handling a complex set of medical images This implies that the proposed approach can be implemented in real practice for analysing noisy radiography images, which have many useful medical applications such as diagnosis of diseases related to lung, breast, musculoskeletal or pediatric studies

101 citations

Journal ArticleDOI
TL;DR: A new multimodal medical image retrieval approach based on the recent advances in the statistical graphic model and deep learning is developed, which first investigates a new extended probabilistic Latent Semantic Analysis model to integrate the visual and textual information from medical images to bridge the semantic gap.
Abstract: Medical imaging is becoming a vital component of war on cancer. Tremendous amounts of medical image data are captured and recorded in a digital format during cancer care and cancer research. Facing such an unprecedented volume of image data with heterogeneous image modalities, it is necessary to develop effective and efficient content-based medical image retrieval systems for cancer clinical practice and research. While substantial progress has been made in different areas of content-based image retrieval (CBIR) research, direct applications of existing CBIR techniques to the medical images produced unsatisfactory results, because of the unique characteristics of medical images. In this paper, we develop a new multimodal medical image retrieval approach based on the recent advances in the statistical graphic model and deep learning. Specifically, we first investigate a new extended probabilistic Latent Semantic Analysis model to integrate the visual and textual information from medical images to bridge the semantic gap. We then develop a new deep Boltzmann machine-based multimodal learning model to learn the joint density model from multimodal information in order to derive the missing modality. Experimental results with large volume of real-world medical images have shown that our new approach is a promising solution for the next-generation medical imaging indexing and retrieval system.

60 citations

References
More filters
Book
01 Jan 1988
TL;DR: Probabilistic Reasoning in Intelligent Systems as mentioned in this paper is a complete and accessible account of the theoretical foundations and computational methods that underlie plausible reasoning under uncertainty, and provides a coherent explication of probability as a language for reasoning with partial belief.
Abstract: From the Publisher: Probabilistic Reasoning in Intelligent Systems is a complete andaccessible account of the theoretical foundations and computational methods that underlie plausible reasoning under uncertainty. The author provides a coherent explication of probability as a language for reasoning with partial belief and offers a unifying perspective on other AI approaches to uncertainty, such as the Dempster-Shafer formalism, truth maintenance systems, and nonmonotonic logic. The author distinguishes syntactic and semantic approaches to uncertainty—and offers techniques, based on belief networks, that provide a mechanism for making semantics-based systems operational. Specifically, network-propagation techniques serve as a mechanism for combining the theoretical coherence of probability theory with modern demands of reasoning-systems technology: modular declarative inputs, conceptually meaningful inferences, and parallel distributed computation. Application areas include diagnosis, forecasting, image interpretation, multi-sensor fusion, decision support systems, plan recognition, planning, speech recognition—in short, almost every task requiring that conclusions be drawn from uncertain clues and incomplete information. Probabilistic Reasoning in Intelligent Systems will be of special interest to scholars and researchers in AI, decision theory, statistics, logic, philosophy, cognitive psychology, and the management sciences. Professionals in the areas of knowledge-based systems, operations research, engineering, and statistics will find theoretical and computational tools of immediate practical use. The book can also be used as an excellent text for graduate-level courses in AI, operations research, or applied probability.

15,671 citations

Book
01 Jan 1976
TL;DR: This book develops an alternative to the additive set functions and the rule of conditioning of the Bayesian theory: set functions that need only be what Choquet called "monotone of order of infinity." and Dempster's rule for combining such set functions.
Abstract: Both in science and in practical affairs we reason by combining facts only inconclusively supported by evidence. Building on an abstract understanding of this process of combination, this book constructs a new theory of epistemic probability. The theory draws on the work of A. P. Dempster but diverges from Depster's viewpoint by identifying his "lower probabilities" as epistemic probabilities and taking his rule for combining "upper and lower probabilities" as fundamental. The book opens with a critique of the well-known Bayesian theory of epistemic probability. It then proceeds to develop an alternative to the additive set functions and the rule of conditioning of the Bayesian theory: set functions that need only be what Choquet called "monotone of order of infinity." and Dempster's rule for combining such set functions. This rule, together with the idea of "weights of evidence," leads to both an extensive new theory and a better understanding of the Bayesian theory. The book concludes with a brief treatment of statistical inference and a discussion of the limitations of epistemic probability. Appendices contain mathematical proofs, which are relatively elementary and seldom depend on mathematics more advanced that the binomial theorem.

14,565 citations

Journal ArticleDOI
TL;DR: The working conditions of content-based retrieval: patterns of use, types of pictures, the role of semantics, and the sensory gap are discussed, as well as aspects of system engineering: databases, system architecture, and evaluation.
Abstract: Presents a review of 200 references in content-based image retrieval. The paper starts with discussing the working conditions of content-based retrieval: patterns of use, types of pictures, the role of semantics, and the sensory gap. Subsequent sections discuss computational steps for image retrieval systems. Step one of the review is image processing for retrieval sorted by color, texture, and local geometry. Features for retrieval are discussed next, sorted by: accumulative and global features, salient points, object and shape features, signs, and structural combinations thereof. Similarity of pictures and objects in pictures is reviewed for each of the feature types, in close connection to the types and means of feedback the user of the systems is capable of giving by interaction. We briefly discuss aspects of system engineering: databases, system architecture, and evaluation. In the concluding section, we present our view on: the driving force of the field, the heritage from computer vision, the influence on computer vision, the role of similarity and of interaction, the need for databases, the problem of evaluation, and the role of the semantic gap.

6,447 citations

Journal ArticleDOI
TL;DR: An overview of the foundational issues related to case-based reasoning is given, some of the leading methodological approaches within the field are described, and the current state of the field is exemplified through pointers to some systems.
Abstract: Case-based reasoning is a recent approach to problem solving and learning that has got a lot of attention over the last few years. Originating in the US, the basic idea and underlying theories have spread to other continents, and we are now within a period of highly active research in case-based reasoning in Europe, as well. This paper gives an overview of the foundational issues related to case-based reasoning, describes some of the leading methodological approaches within the field, and exemplifies the current state through pointers to some systems. Initially, a general framework is defined, to which the subsequent descriptions and discussions will refer. The framework is influenced by recent methodologies for knowledge level descriptions of intelligent systems. The methods for case retrieval, reuse, solution testing, and learning are summarized, and their actual realization is discussed in the light of a few example systems that represent different CBR approaches. We also discuss the role of case-based methods as one type of reasoning and learning method within an integrated system architecture.

5,750 citations