scispace - formally typeset
Open AccessBook ChapterDOI

3D invariants with high robustness to local deformations for automated pollen recognition

Reads0
Chats0
TLDR
A new technique for the extraction of features from 3D volumetric data sets based on group integration is presented, which is robust to local arbitrary deformations and nonlinear gray value changes, but is still sensitive to fine structures.
Abstract
We present a new technique for the extraction of features from 3D volumetric data sets based on group integration. The features are invariant to translation, rotation and global radial deformations. They are robust to local arbitrary deformations and nonlinear gray value changes, but are still sensitive to fine structures. On a data set of 389 confocally scanned pollen from 26 species we get a precision/recall of 99.2% with a simple 1NN classifier. On volumetric transmitted light data sets of about 180,000 airborne particles, containing about 22,700 pollen grains from 33 species, recorded with a low-cost optic in a fully automated online pollen monitor the mean precision for allergenic pollen is 98.5% (recall: 86.5%) and for the other pollen 97.5% (recall: 83.4%).

read more

Content maybe subject to copyright    Report

3D Invariants with High Robustness to Local
Deformations for Automated Pollen Recognition
Olaf Ronneberger, Qing Wang, and Hans Burkhardt
Albert-Ludwigs-Universit¨at Freiburg, Institut ur Informatik, Lehrstuhl ur
Mustererkennung und Bildverarbeitung, Georges-K¨ohler-Allee Geb. 052,
79110 Freiburg, Deutschland
{ronneber,qwang,burkhardt}@informatik.uni-freiburg.de
Abstract. We present a new technique for the extraction of features
from 3D volumetric data sets based on group integration. The features
are invariant to translation, rotation and global radial deformations.
They are robust to local arbitrary deformations and nonlinear gray value
changes, but are still sensitive to fine structures. On a data set of 389 con-
focally scanned pollen from 26 species we get a precision/recall of 99.2%
with a simple 1NN classifier. On volumetric transmitted light data sets of
about 180,000 airborne particles, containing about 22,700 pollen grains
from 33 species, recorded with a low-cost optic in a fully automated
online pollen monitor the mean precision for allergenic pollen is 98.5%
(recall: 86.5%) and for the other pollen 97.5% (recall: 83.4%).
1 Introduction
Nearly all worldwide pollen forecasts are still based on manual counting of pollen
in air samples under the microscope. Within the BMBF-founded project “OM-
NIBUSS” a first demonstrator of a fully automated online pollen monitor was
developed, that integrates the collection, preparation and microscopic analysis
of air samples. Due to commercial interests, no details of the developed pattern
recognition algorithms were published within the last three years. This is the
first time that we show how this machine works behind the scenes.
Challenges in pollen recognition. Due to the great intra class variability and
only verysubtle inter-class differences, automated pollen recognition is a very chal-
lenging but still largely unsolved problem. As most pollen grains are nearly spher-
ical and the subtle differences are mainly found near the surface, a pollen expert
needs the full 3D information (usually by “focussing through” the transparent
pollen grain). An additional difficulty is that pollen grains are often agglomerated
and that the air samples containlots of other airborne particles. For a reliable mea-
surement of high allergenic pollen (e.g. Artemisia. A few such pollen grains per m
3
of air can already cause allergic reactions) the avoidance of false positives is one
of the most important requirements for a fully automated system.
State of the art. Almost all published articles concerning pollen recognition
deal with very low numbers of pollen grains from only a few species and use
F.A. Hamprecht, C. Schn¨orr, and B. ahne (Eds.): DAGM 2007, LNCS 4713, pp. 425–435, 2007.
c
Springer-Verlag Berlin Heidelberg 2007

426 O. Ronneberger, Q. Wang, and H. Burkhardt
manually prepared pure pollen samples, e.g. [1]. Only [4] used a data set from
real air samples containing a reasonable number of pollen grains (3686) from
27 species. But even on a reduced data set containing only 8 species and dust
particles, the recall was only 64,9% with a precision of 30%.
Main Contribution. In this paper we describe the extension of the Haar-
integration framework [9,6,7,8] (further denoted as “HI framework”) to global
and local deformations. This is achieved by creating synthetic channels con-
taining the segmentation borders and employing special parameterized kernel
functions. Due to the sparsity of non-zero-values in the synthetic channels the
resulting integral features are highly localized in the real space, while the frame-
work automatically guarantees the desired invariance properties.
For efficient computation of these integrals we make use of the sparsity of
the data in the synthetic channels and use a Fourier or spherical harmonics
(“SH”) series expansion (for the desired rotation invariance) to compute multiple
features at the same time.
a) volume rendering of
confocal data set
b) horizontal and vertical
cuts of confocal data set
c) horizontal and vertical cuts
of transmitted light data set
Fig. 1. 3D recordings of Betula pollen grains. In transmitted light microscopy the
recording properties in z-direction (the direction of the optical axis) are significantly
different from those in the xy-direction, because the effects of diffraction, refraction
and absorption depend on the direction of the transmitted light. Furthermore there
is a significant loss of information in z-direction due to the low-pass property of the
optical transfer function.
2MaterialandMethods
Data Sets. To demonstrate the generality of the proposed invariants and com-
pare them to earlier results, we use two different pollen data sets in this article.
Both contain 3D volumetric recordings of pollen grains.
The “confocal data set” contains 389 pollen grains from 26 German pollen
taxa, recorded with a confocal laser scanning microscope (fig 1a,b). For further
details on this data set refer to [6].
The “pollen monitor data set” contains about 180,000 airborne particles in-
cluding about 22,700 pollen grains from air samples that were collected, prepared

3D Invariants with High Robustness to Local Deformations 427
and recorded with transmitted light microscopy from the online pollen monitor
from March to September 2006 in Freiburg and Z¨urich (fig. 1c). All 180,000
particles were manually labeled by pollen experts.
Segmentation. To find the 3D surface of the pollen grains in the confocal data
set, we use the graph cut algorithm described in [2]. The original data were first
scaled down. The edge costs to source and sink were modeled by a Gaussian
distribution relative to the mean and minimum gray value. We added voxel-to-
voxel edges to the 124 neighborhood, where the weight was a Gaussian of the
gray differences. The resulting binary mask was then smoothly scaled up to the
original size.
The first step in processing the pollen monitor data set is the detection of
circular objects with voxel-wise vector based gray-scale invariants, similar to
those in [8]. For each detected circular object the precise border in the sharpest
layer is searched: As parts of the object border are often missing or not clear, we
use snakes to find a smooth and complete border. To avoid the common problem
of snakes being attracted to undesired edges (if plain gradient magnitude is used
asforcefield),wetakethestepsdepictedinfig2.
a) sharpest layer b) found edges c) weighted edges d) final snake
1. Applying modified Canny edge
detection.
As pollen grains have a nearly
round shape, the edges that are
approximately perpendicular to
the radial direction are more rele-
vant. We replace the gradient with
its radial component in the orig-
inal Canny edge detection algo-
rithm.
2. Model-based weighting of the
edges.
The curvatures and relative loca-
tions of the edges are analyzed
and each edge is given a different
weight. Some edges are even elim-
inated. As a result, a much clearer
weighted edge image is obtained.
3. Employing snakes to find the
final border.
The initial contour is chosen to be
the circle found in the detection
step. The external force field is the
so-called “gradient vector flow”
[10] computed from the weighted
edge image
Fig. 2. Segmentation of transmitted light microscopic images
2.1 Construction of Invariants
For the construction of invariants we use the combination of a normalization
and Haar-integration [9,6,7,8](see eq. (1)) over a transformation group con-
taining rotations and deformations (Haar-integration has nothing to do with
Haar wavelets). In contrast to the very general approach in [6], we now use the

428 O. Ronneberger, Q. Wang, and H. Burkhardt
object center and the outer border found in the segmentation step to extract
more distinctive features describing certain regions of the object.
T [f](X):=
G
f(gX)dg
G : transformation group
g : one element of the transformation group
dg : Haar measure
f : nonlinear kernel function
X : n-dim, multi-channel data set
(1)
Invariance to translations. Invariance to translations is achieved by moving
the center of mass of the segmentation mask to the origin. The final features are
quite insensitive to errors in this normalization step, because they are computed
“far” away from this center and only the direction to it is used.
Invariance to rotation. Invariance to rotation around the object center is
achieved by integration over the rotation group. In the confocal data set we can
model a 3D rotation of a real-world object by a 3D rotation of the recorded volu-
metric data set (see fig. 1b). In contrast to this, the transmitted light microscopic
image stacks from the pollen monitor data set show very different characteristics
in xy- and z-direction, (see fig. 1c). A rotation around the x- or y-axis of the
real-world object results in so different gray value distributions, that it is more
reasonable to model only the rotation around the z-axis, resulting in a planar
rotation invariance.
Invariance to global Deformations and Robustness to local Deforma-
tions. The deformation model consists of two parts. The global deformations
are modeled by a simple shift in radial direction e
r
, which depends only on the
angular coordinates (see figure 3a). For full 3D-rotations described in spherical
coordinates x =(x
r
,x
ϕ
,x
ϑ
)thismodelis
x
= x + γ
γ
γ(x)withγ
γ
γ(x)=γ(x
ϕ
,x
ϑ
) · e
r
(x
ϕ
,x
ϑ
) . (2)
For rotations around the z-axis described in cylindrical coordinates x=(x
r
,x
ϕ
,x
z
)
we get
x
= x + γ
γ
γ(x)withγ
γ
γ(x)=γ(x
ϕ
) · e
r
(x
ϕ
) . (3)
Please note, that this deformation is well defined only for r>γ(ϕ), which is
no problem in the present application, because the features are computed “far”
away from the center.
The smaller local deformations are described by an arbitrary displacement
field D(x) such that
x
= x + D(x)(4)
(see fig. 3b). For the later partial Haar-integration [3] over all possible realizations
of this displacement field, it is sufficient to know only the probability for the
occurrence of a certain relative displacement r within this field as
p
D(x + d) D(x)=r
= p
d
(r; d) x, d IR
3
, (5)

3D Invariants with High Robustness to Local Deformations 429
a) Global deformation model (radial) b) Local deformation model (arbitrary)
Fig. 3. Possible realizations of the deformation models
where we select p
d
(r; d) to be a rotationally symmetric Gaussian distribution
with a standard deviation σ = d·σ
d
.
While we achieve full invariance to radial deformations by full Haar-integration
we can only reach robustness to local deformations by partial Haar-integration.
But this non-invariance in the second case is exactly the desired behavior. In com-
bination with appropriate kernel functions this results in a continuous mapping of
objects (with weak or strong local deformations) into the feature space.
The kernel functions. Instead of selecting a certain fixed number of kernel
functions, we introduce parameterized kernel functions here. Embedded into the
HI framework, each new combination of kernel parameters results in a new in-
variant feature. For multiple kernel parameters, we now have a multidimensional
invariant feature array describing the object.
Robustness to gray value transformations. To become robust to gray value trans-
formations the information is split into gradient direction (which is very robust
even under nonlinear gray value transformations) and gradient magnitude. This
was already successfully applied to the HI framework in [8] and to confocal pollen
data sets in [5].
Synthetic channels with segmentation results. To feed the segmentation informa-
tion into the HI framework we simply render the surface (confocal data set) or
the contour of the sharpest layer (transmitted light data set) as delta-peaks into
a new channel S and extend the kernel-function with two additional points that
sense the gray value in this channel. The only condition for this technique is
that the computation of the synthetic channel and the action of transformation
group can be exchanged without the result being changed (i.e., we must get the
same result if we first extract the surface and then rotate and deform the volume
and vice versa).
Resulting kernel function. To achieve the requested properties we construct 4-
point kernels, where 2 points of the kernel a
1
and a
2
sense the segmentation

Citations
More filters
Proceedings ArticleDOI

Fast and robust segmentation of spherical particles in volumetric data sets from brightfield microscopy

TL;DR: The system is based on a voting procedure that finds the centers and radii of the particles and a subsequent precise segmentation with an active contour approach to meet the demands of an online pollenmonitor for high speed and low memory consumption.
Proceedings ArticleDOI

Fast computation of 3D spherical Fourier harmonic descriptors - a complete orthonormal basis for a rotational invariant representation of three-dimensional objects

TL;DR: Having created an orthonormal basis on the ball with all the gentle properties known from the spherical harmonics theory and Fourier theory, this paper is able to compute efficiently a multi-scale representation of 3D objects that leads to highly discriminative rotation-invariant features, which will be called spherical Fourier harmonic descriptors (SFHD).

3D invariants for automated pollen recognition

TL;DR: The development of 3D invariants for the recognition of biological structures is described, based on the Haar-integration-framework of Schulz-Mirbach (1995b), which allows for a fast and reliable detection of spherical objects in cluttered environments.
Journal ArticleDOI

A Local Feature Descriptor Based on SIFT for 3D Pollen Image Recognition

TL;DR: A local feature extraction method based on SIFT that can effectively describe the pollen image and is robust to the rotation, translation and scaling of the image is presented.
Book ChapterDOI

Analysis of Relevant Features for Pollen Classification

TL;DR: A set of relevant image-based features for the recognition of top allergenic pollen taxa is proposed and demonstrated to overcome the intra-class variance and inter-class similarity in a SVM classification scheme with a performance comparable to the state of the art procedures.
References
More filters
Journal ArticleDOI

An experimental comparison of min-cut/max- flow algorithms for energy minimization in vision

TL;DR: This paper compares the running times of several standard algorithms, as well as a new algorithm that is recently developed that works several times faster than any of the other methods, making near real-time performance possible.
Journal ArticleDOI

Snakes, shapes, and gradient vector flow

TL;DR: This paper presents a new external force for active contours, which is computed as a diffusion of the gradient vectors of a gray-level or binary edge map derived from the image, and has a large capture range and is able to move snakes into boundary concavities.
Book ChapterDOI

An Experimental Comparison of Min-cut/Max-flow Algorithms for Energy Minimization in Vision

TL;DR: The goal of this paper is to provide an experimental comparison of the efficiency of min-cut/max flow algorithms for applications in vision, comparing the running times of several standard algorithms, as well as a new algorithm that is recently developed.
Journal ArticleDOI

Automatic recognition of biological particles in microscopic images

TL;DR: This system was developed to classify 12 categories of particles found in human urine; it achieves a 93.2% correct classification rate in this application and this performance is considered good.
Book ChapterDOI

Invariant Features for Gray Scale Images

TL;DR: This paper considers image rotations and translations and presents algorithms for constructing invariant features and develops algorithms for recognizing several objects in a single scene without the necessity to segment the image beforehand.
Related Papers (5)
Frequently Asked Questions (14)
Q1. What are the contributions in "3d invariants with high robustness to local deformations for automated pollen recognition" ?

The authors present a new technique for the extraction of features from 3D volumetric data sets based on group integration. 

Due to the sparsity of non-zero-values in the synthetic channels the resulting integral features are highly localized in the real space, while the framework automatically guarantees the desired invariance properties. 

A few such pollen grains per m3 of air can already cause allergic reactions) the avoidance of false positives is one of the most important requirements for a fully automated system. 

While the authors achieve full invariance to radial deformations by full Haar-integration the authors can only reach robustness to local deformations by partial Haar-integration. 

The first step in processing the pollen monitor data set is the detection of circular objects with voxel-wise vector based gray-scale invariants, similar to those in [8]. 

From the training set only the “clean” (not agglomerated, not contaminated) pollen and the “non-pollen” particles from a few samples were used to train the support vector machine (SVM) using the RBF-kernel (radial basis function) and the one-vs-rest multi-class approach. 

For the application on the pollen monitor data set (rotational invariance only around the z-axis), q is split into a radial distance qr to the segmentation border and the z-distance to the central plane qz. 

The best sampling of the parameter space of the kernel functions (corresponding to the inner class deformations of the objects), was found by cross validation on the training data set, resulting in Nqr ×Nqz ×Nc×n = 31×11×16×16 = 87296 “structural” features (using kernel function k1) and 8 “shape” features (usingkernel function k2). 

Within the BMBF-founded project “OMNIBUSS” a first demonstrator of a fully automated online pollen monitor was developed, that integrates the collection, preparation and microscopic analysis of air samples. 

This is achieved by creating synthetic channels containing the segmentation borders and employing special parameterized kernel functions. 

The “pollen monitor data set” contains about 180,000 airborne particles including about 22,700 pollen grains from air samples that were collected, preparedand recorded with transmitted light microscopy from the online pollen monitor from March to September 2006 in Freiburg and Zürich (fig. 1c). 

the group of arbitrary deformations GD and the group of rotations GR the final Haar integral becomes:T = ∫GR∫Gγ∫GDf ( gRgγgDS, gRgγgDX ) p(D) dgD dgγ dgR , (8)where p(D) is the probability for the occurrence of the local displacement field D. The transformation of the data set is described by (gX)(x) =: X(x′), wherex′ = Rx︸︷︷︸ rotation + γ(Rx)︸ ︷︷ ︸ global deformation+ 

For 3D rotations this framework uses a spherical-harmonics series expansion, and for planar rotations around the z-axis it is simplified to a Fourier series expansion. 

As most pollen grains are nearly spherical and the subtle differences are mainly found near the surface, a pollen expert needs the full 3D information (usually by “focussing through” the transparent pollen grain).