scispace - formally typeset
Open AccessJournal ArticleDOI

A paraperspective factorization method for shape and motion recovery

TLDR
This work has shown that the paraperspective factorization method can be applied to a much wider range of motion scenarios, including image sequences containing motion toward the camera and aerial image sequences of terrain taken from a low-altitude airplane.
Abstract
The factorization method, first developed by Tomasi and Kanade (1992), recovers both the shape of an object and its motion from a sequence of images, using many images and tracking many feature points to obtain highly redundant feature position information. The method robustly processes the feature trajectory information using singular value decomposition (SVD), taking advantage of the linear algebraic properties of orthographic projection. However, an orthographic formulation limits the range of motions the method can accommodate. Paraperspective projection, first introduced by Ohta et al. (1981), is a projection model that closely approximates perspective projection by modeling several effects not modeled under orthographic projection, while retaining linear algebraic properties. Our paraperspective factorization method can be applied to a much wider range of motion scenarios, including image sequences containing motion toward the camera and aerial image sequences of terrain taken from a low-altitude airplane.

read more

Content maybe subject to copyright    Report

206 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 19, NO. 3, MARCH 1997
A Paraperspective Factorization Method for
Shape and Motion Recovery
Conrad J. Poelman and Takeo Kanade,
Fellow, IEEE
Abstract
—The factorization method, first developed by Tomasi and Kanade, recovers both the shape of an object and its motion
from a sequence of images, using many images and tracking many feature points to obtain highly redundant feature position
information. The method robustly processes the feature trajectory information using singular value decomposition (SVD), taking
advantage of the linear algebraic properties of orthographic projection. However, an orthographic formulation limits the range of
motions the method can accommodate. Paraperspective projection, first introduced by Ohta, is a projection model that closely
approximates perspective projection by modeling several effects not modeled under orthographic projection, while retaining linear
algebraic properties. Our paraperspective factorization method can be applied to a much wider range of motion scenarios, including
image sequences containing motion toward the camera and aerial image sequences of terrain taken from a low-altitude airplane.
Index Terms
—Motion analysis, shape recovery, factorization method, three-dimensional vision, image sequence analysis, singular
value decomposition.
——————————
——————————
1I
NTRODUCTION
ECOVERING
the geometry of a scene and the motion of
the camera from a stream of images is an important
task in a variety of applications, including navigation, ro-
botic manipulation, and aerial cartography. While this is
possible in principle, traditional methods have failed to
produce reliable results in many situations [2].
Tomasi and Kanade [13], [14] developed a robust and ef-
ficient method for accurately recovering the shape and mo-
tion of an object from a sequence of images, called the fac-
torization method. It achieves its accuracy and robustness by
applying a well-understood numerical computation, the
singular value decomposition (SVD), to a large number of
images and feature points, and by directly computing
shape without computing the depth as an intermediate
step. The method was tested on a variety of real and syn-
thetic images, and was shown to perform well even for
distant objects, where traditional triangulation-based ap-
proaches tend to perform poorly.
The Tomasi-Kanade factorization method, however, as-
sumed an orthographic projection model. The applicability of
the method is therefore limited to image sequences created
from certain types of camera motions. The orthographic
model contains no notion of the distance from the camera to
the object. As a result, shape reconstruction from image se-
quences containing large translations toward or away from
the camera often produces deformed object shapes, as the
method tries to explain the size differences in the images by
creating size differences in the object. The method also sup-
plies no estimation of translation along the camera’s optical
axis, which limits its usefulness for certain tasks.
There exist several perspective approximations which
capture more of the effects of perspective projection while
remaining linear. Scaled orthographic projection, sometimes
referred to as “weak perspective” [5], accounts for the scaling
effect of an object as it moves towards and away from the
camera. Paraperspective projection, first introduced by Ohta
[6] and named by Aloimonos [1], accounts for the scaling
effect as well as the different angle from which an object is
viewed as it moves in a direction parallel to the image plane.
In this paper, we present a factorization method based
on the paraperspective projection model. The paraperspec-
tive factorization method is still fast, and robust with re-
spect to noise. It can be applied to a wider realm of situa-
tions than the original factorization method, such as se-
quences containing significant depth translation or con-
taining objects close to the camera, and can be used in ap-
plications where it is important to recover the distance to
the object in each image, such as navigation.
We begin by describing our camera and world reference
frames and introduce the mathematical notation that we use.
We review the original factorization method as defined in
[13], presenting it in a slightly different manner in order to
make its relation to the paraperspective method more appar-
ent. We then present our paraperspective factorization
method, followed by a description of a perspective refine-
ment step. We conclude with the results of several experi-
ments which demonstrate the practicality of our system.
2P
ROBLEM
D
ESCRIPTION
In a shape-from-motion problem, we are given a sequence
of F images taken from a camera that is moving relative to
an object. Assume for the time being that we locate P
prominent feature points in the first image, and track these
0162-8828/97/$10.00 © 1997 IEEE
————————————————
C.J. Poelman is with the Satellite Assessment Center (WSAT), USAF
Phillips Laboratory, Albuquerque, NM 87117-5776.
E-mail: poelmanc@plk.af.mil.
T. Kanade is with the School of Computer Science, Carnegie Mellon Uni-
versity, 5000 Forbes Avenue, Pittsburgh, PA 15213-3890.
E-mail: tk@cs.cmu.edu.
M
anuscript received June 15, 1994; revised Jan. 10, 1996. Recommended for accep-
tance by S. Peleg.
For information on obtaining reprints of this article, please send e-mail to:
transpami@computer.org, and reference IEEECS Log Number P97001.
R

POELMAN AND KANADE: A PARAPERSPECTIVE FACTORIZATION METHOD FOR SHAPE AND MOTION RECOVERY 207
points from each image to the next, recording the coordi-
nates uv
fp fp
,
ej
of each point p in each image f. Each feature
point p that we track corresponds to a single world point,
located at position s
p
in some fixed world coordinate sys-
tem. Each image f was taken at some camera orientation,
which we describe by the orthonormal unit vectors i
f
, j
f
, and
k
f
, where i
f
and j
f
correspond to the x and y axes of the cam-
era’s image plane, and k
f
points along the camera’s line of
sight. We describe the position of the camera in each frame f
by the vector t
f
indicating the camera’s focal point. This
formulation is illustrated in Fig. 1.
Fig. 1. Coordinate system.
The result of the feature tracker is a set of P feature point
coordinates
uv
fp fp
,
ej
for each of the F frames of the image
sequence. From this information, our goal is to estimate the
shape of the object as
$
s
p
for each object point, and the mo-
tion of the camera as
$
i
f
,
$
j
f
,
$
k
f
, and
$
t
f
for each frame in the
sequence.
3T
HE
O
RTHOGRAPHIC
F
ACTORIZATION
M
ETHOD
This section presents a summary of the orthographic factori-
zation method developed by Tomasi and Kanade. A more
detailed description of the method can be found in [13].
3.1 Orthographic Projection
The orthographic projection model assumes that rays are
projected from an object point along the direction parallel
to the camera’s optical axis, so that they strike the image
plane orthogonally, as illustrated in Fig. 2. A point p whose
location is s
p
will be observed in frame f at image coordi-
nates
uv
fp fp
,
ej
, where
uv
fp f
p
ffpf
p
f
=◊ - = -
ist jst
ej ej
(1)
These equations can be rewritten as
uxvy
fp f
p
ffpf
p
f
=◊+ =+
ms ns
(2)
where
xy
fff fff
=- =- ti tj
ej ej
(3)
mi nj
ff ff
==
(4)
Fig. 2. Orthographic projection in two dimensions. Dotted lines indicate
perspective projection.
3.2 Decomposition
All of the feature point coordinates
uv
fp fp
,
ej
are entered in a
2
FP
¥ measurement matrix W.
W
uu
uu
vv
vv
P
FFP
P
FFP
=
L
N
M
M
M
M
M
M
O
Q
P
P
P
P
P
P
11 1
1
11 1
1
K
KKK
K
K
KKK
K
(5)
Each column of the measurement matrix contains the ob-
servations for a single point, while each row contains the
observed u-coordinates or v-coordinates for a single frame.
Equation (2) for all points and frames can now be combined
into the single matrix equation
WMST=+11K (6)
where M is the
23F¥
motion matrix whose rows are the m
f
and n
f
vectors, S is the 3
¥ P
shape matrix whose columns
are the s
p
vectors, and T is the 2 1F ¥ translation vector
whose elements are the x
f
and y
f
.
Up to this point, Tomasi and Kanade placed no restric-
tions on the location of the world origin, except that it be
stationary with respect to the object. Without loss of gener-
ality, they position the world origin at the center of mass of
the object, denoted by c, so that
cs
==
=
Â
1
0
1
P
p
p
P
(7)
Because the sum of any row of S is zero, the sum of any
row i of W is
PT
i
. This enables them to compute the ith
element of the translation vector T directly from W, simply
by averaging the ith row of the measurement matrix. The
translation is the subtracted from W, leaving a “registered”
measurement matrix
WWT
*
=-
11K. Because W* is the
product of a 2 3
F ¥
motion matrix M and a 3
¥ P
shape

208 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 19, NO. 3, MARCH 1997
matrix S, its rank is at most three. When noise is present in
the input, the W* will not be exactly of rank three, so the
Tomasi-Kanade factorization method uses the SVD to find
the best rank three approximation to W*, factoring it into
the product
WMS
*
$
$
= (8)
3.3 Normalization
The decomposition of (8) is only determined up to a linear
transformation. Any non-singular 3 3
¥
matrix A and its
inverse could be inserted between
$
M
and
$
S
, and their
product would still equal W*. Thus the actual motion and
shape are given by
MMASAS
==
-
$
$
1
(9)
with the appropriate
33¥
invertible matrix A selected. The
correct A can be determined using the fact that the rows of
the motion matrix M (which are the m
f
and n
f
vectors) repre-
sent the camera axes, and therefore they must be of a certain
form. Since i
f
and j
f
are unit vectors, we see from (4) that
mn
ff
22
11
==
(10)
and because they are orthogonal,
mn
ff
◊=0
(11)
Equations (10) and (11) give us 3F equations which we call
the metric constraints. Using these constraints, we solve for
the 3 3
¥
matrix A which, when multiplied by
$
M
, produces
the motion matrix M that best satisfies these constraints.
Once the matrix A has been found, the shape and motion
are computed from (9).
4T
HE
P
ARAPERSPECTIVE
F
ACTORIZATION
M
ETHOD
The Tomasi-Kanade factorization method was shown to be
computationally inexpensive and highly accurate, but its
use of an orthographic projection assumption limited the
method’s applicability. For example, the method does not
produce accurate results when there is significant transla-
tion along the camera’s optical axis, because orthography
does not account for the fact that an object appears larger
when it is closer to the camera. We must model this and
other perspective effects in order to successfully recover
shape and motion in a wider range of situations. We choose
an approximation to perspective projection known as
paraperspective projection, which was introduced by Ohta
et al. [6] in order to solve a shape from texture problem.
Although the paraperspective projection equations are
more complex than those for orthography, their basic form
is the same, enabling us develop a method analogous to
that developed by Tomasi and Kanade.
4.1 Paraperspective Projection
Paraperspective projection closely approximates perspec-
tive projection by modeling both the scaling effect (closer
objects appear larger than distant ones) and the position
effect (objects in the periphery of the image are viewed
from a different angle than those near the center of projec-
tion [1]) while retaining the linear properties of ortho-
graphic projection. Paraperspective projection is related to,
but distinct from, the affine camera model, as described in
Appendix A. The paraperspective projection of an object
onto an image, illustrated in Fig. 3, involves two steps.
1) An object point is projected along the direction of the
line connecting the focal point of the camera to the
object’s center of mass, onto a hypothetical image
plane parallel to the real image plane and passing
through the object’s center of mass.
2) The point is then projected onto the real image plane
using perspective projection. Because the hypothetical
plane is parallel to the real image plane, this is
equivalent to simply scaling the point coordinates by
the ratio of the camera focal length and the distance
between the two planes.
1
In general, the projection of a point p along direction r, onto
the plane with normal n and distance from the origin d, is
given by the equation
¢
=-
◊-
pp
pn
rn
r
d
(12)
In frame f, each object point s
p
is projected along the direc-
tion
ct
-
f
(which is the direction from the camera’s focal
point to the object’s center of mass) onto the plane defined
by normal k
f
and distance from the origin
ck
f
. The result
¢
s
fp
of this projection is
¢
=-
◊-
-◊
-ss
sk ck
ct k
ct
fp
p
p
ff
ff
f
ejej
ej
ej
(13)
The perspective projection of this point onto the image
plane is given by subtracting t
f
from
¢
s
fp
to give the position
of the point in the camera’s coordinate system, and then
scaling the result by the ratio of the camera’s focal length l
to the depth to the object’s center of mass z
f
. Adjusting for
the aspect ratio a and projection center oo
xy
,
ej
yields the
coordinates of the projection in the image plane,
u
l
z
o
v
la
z
o
z
fp
f
f
fp f
x
fp
f
f
fp f
y
fff
=
¢
-+
=
¢
-+
=-
i
st
j
st
ct k
ej
ej
ej
where
(14)
Substituting (13) into (14) and simplifying gives the general
paraperspective equations for
u
fp
and
v
fp
1. The scaled orthographic projection model (also known as “weak per-
spective”) is similar to paraperspective projection, except that the direction
of the initial projection in Step 1 is parallel to the camera’s optical axis
rather than parallel to the line connecting the object’s center of mass to the
camera’s focal point. This model captures the scaling effect of perspective
projection, but not the position effect, as explained in Appendix B.

POELMAN AND KANADE: A PARAPERSPECTIVE FACTORIZATION METHOD FOR SHAPE AND MOTION RECOVERY 209
u
zz
o
v
la
zz
o
fp
f
f
ff
f
f
p
ff
x
fp
f
f
ff
f
f
p
ff
y
=
-
◊-
L
N
M
M
M
O
Q
P
P
P
◊-+-
R
S
|
T
|
U
V
|
W
|
+
=
-
◊-
L
N
M
M
M
O
Q
P
P
P
◊-+-
R
S
|
T
|
U
V
|
W
|
+
1
i
ict
ksccti
j
jct
kscctj
ej
ejej
ej
ejej
(15)
We simplify these equations by assuming unit focal length,
unit aspect ratio, and (0, 0) center of projection. This re-
quires that the image coordinates
uv
fp fp
,
ej
be adjusted to
account for these camera parameters before commencing
shape and motion recovery.
Fig. 3. Paraperspective projection in two dimensions. Dotted lines indi-
cate perspective projection. Æ indicates parallel lines.
In [3] the factorization approach is extended to handle
multiple objects moving separately, which requires each
object to be projected based on its own mass center. How-
ever, since this paper addresses the single object case, we
can further simplify our equations by placing the world
origin at the object’s center of mass so that by definition
cs
==
=
Â
1
0
1
P
p
p
P
(16)
This reduces (15) to
u
zz
v
zz
fp
f
f
ff
f
f
p
ff
fp
f
f
ff
f
f
p
ff
=+
L
N
M
O
Q
P
◊-
R
S
|
T
|
U
V
|
W
|
=+
L
N
M
O
Q
P
◊-
R
S
|
T
|
U
V
|
W
|
1
1
i
it
ks ti
j
jt
ks tj
ej
ej
(17)
These equations can be rewritten as
uxvy
fp f
p
ffpf
p
f
=◊+ =+
ms ns
(18)
where
z
fff
=-
tk
(19)
x
z
y
z
f
ff
f
f
ff
f
=-
=-
ti tj
(20)
m
ik
n
jk
f
fff
f
f
fff
f
x
z
y
z
=
-
=
-
(21)
Notice that (18) has a form identical to its counterpart for
orthographic projection, (2), although the corresponding
definitions of
x
f
,
y
f
,
m
f
, and
n
f
differ. This enables us to
perform the basic decomposition of the matrix in the same
manner that Tomasi and Kanade did for orthographic
projection.
4.2 Paraperspective Decomposition
We can combine (18), for all points p from 1 to P, and all
frames f from 1 to F, into the single matrix equation
uu
uu
vv
vv
x
x
y
y
P
FFP
P
FFP
F
F
P
F
F
11 1
1
11 1
1
1
1
1
1
1
11
K
KKK
K
K
KKK
K
K
K
K
K
K
K
L
N
M
M
M
M
M
M
O
Q
P
P
P
P
P
P
=
L
N
M
M
M
M
M
M
O
Q
P
P
P
P
P
P
+
L
N
M
M
M
M
M
M
O
Q
P
P
P
P
P
P
m
m
n
n
ss (22)
or in short
WMST=+
11K (23)
where W is the
2FP¥
measurement matrix, M is the
23
F¥
motion matrix, S is the 3
¥ P
shape matrix, and T is
the 2 1
F ¥
translation vector.
Using (16) and (18), we can write
u x Px Px
v y Py Py
fp
p
P
f
p
f
p
P
f
p
ff
p
P
fp
p
P
f
p
f
p
P
f
p
ff
p
P
=◊+=+=
=◊+=+=
== =
== =
ÂÂ Â
ÂÂ Â
11 1
11 1
ms m s
ns n s
ej
ej
(24)
Therefore we can compute
x
f
and
y
f
, which are the ele-
ments of the translation vector T, immediately from the
image data as
x
P
uy
P
v
ffp
p
P
ffp
p
P
==
==
ÂÂ
11
11
(25)
Once we know the translation vector T, we subtract it from
W, giving the registered measurement matrix
WWT MS
*
=- =11
K
(26)
Since W* is the product of two matrices each of rank at most
three, W* has rank at most three, just as it did in the ortho-
graphic projection case. If there is noise present, the rank of
W* will not be exactly three, but by computing the SVD of

210 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 19, NO. 3, MARCH 1997
W* and only retaining the largest three singular values, we
can factor it into
WMS
*
$
$
=
(27)
where
$
M
is a 2 3F ¥ matrix and
$
S
is a 3 ¥ P matrix. Using
the SVD to perform this factorization guarantees that the
product
$
$
MS
is the best possible rank three approximation
to W*, in the sense that it minimizes the sum of squares dif-
ference between corresponding elements of W* and
$
$
MS
.
4.3 Paraperspective Normalization
Just as in the orthographic case, the decomposition of W*
into the product of
$
M
and
$
S
by (27) is only determined up
to a linear transformation matrix A. Again, we determine
this matrix A by observing that the rows of the motion ma-
trix M (the
m
f
and
n
f
vectors) must be of a certain form.
Taking advantage of the fact that
i
f
,
j
f
, and
k
f
are unit
vectors, from (21) we observe that
mn
f
f
f
f
f
f
x
z
y
z
2
2
2
2
2
2
11
=
+
=
+
(28)
We know the values of
x
f
and
y
f
from our initial registra-
tion step, but we do not know the value of the depth
z
f
.
Thus we cannot impose individual constraints on the mag-
nitudes of
m
f
and
n
f
as was done in the orthographic fac-
torization method. However, we can adopt the following
constraint on the magnitudes of
m
f
and
n
f
mn
f
f
f
ff
xyz
2
2
2
22
11
1
+
=
+
=
F
H
G
G
I
K
J
J
(29)
In the case of orthographic projection, one constraint on
m
f
and
n
f
was that they each have unit magnitude, as re-
quired by (10). In the above paraperspective case, we sim-
ply require that their magnitudes be in a certain ratio.
There is also a constraint on the angle relationship of
m
f
and
n
f
. From (21), and the knowledge that
i
f
,
j
f
, and
k
f
are orthogonal unit vectors,
mn
ikjk
ff
fff
f
fff
f
ff
f
x
z
y
z
xy
z
◊=
-
-
=
2
(30)
The problem with this constraint is that, again,
z
f
is un-
known. We could use either of the two values given in (29)
for 1
2
/
z
f
, but in the presence of noisy input data the two
will not be exactly equal, so we use the average of the two
quantities. We choose the arithmetic mean over the geomet-
ric mean or some other measure in order to keep the solu-
tion of these constraints linear. Thus our second constraint
becomes
mn
mn
ff ff
f
f
f
f
xy
xy
◊=
+
+
+
F
H
G
G
G
I
K
J
J
J
1
2
11
2
2
2
2
(31)
This is the paraperspective version of the orthographic con-
straint given by (11), which required that the dot product of
m
f
and
n
f
be zero.
Equations (29) and (31) are homogeneous constraints,
which could be trivially satisfied by the solution
"==f
ff
mn
0
, or M = 0. To avoid this solution, we im-
pose the additional constraint
m
1
1
=
(32)
This does not effect the final solution except by a scaling
factor.
Equations (29), (31), and (32) give us 2F + 1 equations,
which are the paraperspective version of the metric con-
straints. We compute the 3 3
¥
matrix A such that
MMA
=
$
best satisfies these metric constraints in the least sum-of-
squares error sense. This is a simple problem because the
constraints are linear in the six unique elements of the
symmetric
33¥
matrix
QAA
T
=
. We use the metric con-
straints to compute Q, compute its Jacobi Transformation
QLL
T
=L
, where L is the diagonal eigenvalue matrix, and
as long as Q is positive definite,
AL
T
=L
12/
ej
. A non-
positive-definite Q indicates that unmodeled distortion has
overwhelmed the third singular value of the measurement
matrix, due possibly to noise, perspective effects, insuffi-
cient rotational motion, a planar object shape, or a combi-
nation of these effects.
4.4 Paraperspective Motion Recovery
Once the matrix A has been determined, we compute the
shape matrix
SAS=
-
1
$
and the motion matrix
MMA=
$
.
For each frame f, we now need to recover the camera ori-
entation vectors
$
i
f
,
$
j
f
, and
$
k
f
from the vectors
m
f
and
n
f
,
which are the rows of the matrix M. From (21) we see that
$
$
$
$
imkjnk
ffffffffff
zx zy=+ =+
(33)
From this and the knowledge that
$
i
f
,
$
j
f
, and
$
k
f
must be
orthonormal, we determine that
$$
$$$
$
$
$
$
ij m k n k k
imk
jnk
f f f f ff ff ff f
fffff
fffff
zx zy
zx
zy
¥= + ¥ + =
=+=
=+=
ejej
1
1
(34)
Again, we do not know a value for
z
f
, but using the rela-
tions specified in (29) and the additional knowledge that
$
k
f
=
1, (34) can be reduced to
GH
ff f
$
k= (35)
where
GHx
y
f
ff
f
f
ff
f
=
¥
L
N
M
M
M
M
O
Q
P
P
P
P
=-
-
L
N
M
M
M
O
Q
P
P
P
~~
~
~
mn
m
n
ej
1
(36)

Citations
More filters

Multiple View Geometry in Computer Vision.

TL;DR: This book is referred to read because it is an inspiring book to give you more chance to get experiences and also thoughts and it will show the best book collections and completed collections.
Book

Computer Vision: Algorithms and Applications

TL;DR: Computer Vision: Algorithms and Applications explores the variety of techniques commonly used to analyze and interpret images and takes a scientific approach to basic vision problems, formulating physical models of the imaging process before inverting them to produce descriptions of a scene.
Journal ArticleDOI

A Multibody Factorization Method for Independently Moving Objects

TL;DR: A new method for separating and recovering the motion and shape of multiple independently moving objects in a sequence of images by introducing a mathematical construct of object shapes, called the shape interaction matrix, which is invariant to both the object motions and the selection of coordinate systems.
Book ChapterDOI

A Factorization Based Algorithm for Multi-Image Projective Structure and Motion

TL;DR: A practical method for the recovery of projective shape and motion from multiple images of a scene by the factorization of a matrix containing the images of all points in all views, using only fundamental matrices and epipoles estimated from the image data.
Journal ArticleDOI

Using Unmanned Aerial Vehicles (UAV) for High-Resolution Reconstruction of Topography: The Structure from Motion Approach on Coastal Environments

TL;DR: The UAV-based approach to Structure from Motion approach to low-altitude aerial imageries collected by Unmanned Aerial Vehicle was demonstrated to be a straightforward one and accuracy of the vertical dataset was comparable with results obtained by TLS technology.
References
More filters
Proceedings Article

An iterative image registration technique with an application to stereo vision

TL;DR: In this paper, the spatial intensity gradient of the images is used to find a good match using a type of Newton-Raphson iteration, which can be generalized to handle rotation, scaling and shearing.
Book

Numerical Recipes in C: The Art of Scientific Computing

TL;DR: Numerical Recipes: The Art of Scientific Computing as discussed by the authors is a complete text and reference book on scientific computing with over 100 new routines (now well over 300 in all), plus upgraded versions of many of the original routines, with many new topics presented at the same accessible level.

Numerical Recipes in FORTRAN - The Art of Scientific Computing - Second Edition

TL;DR: This paper presents a list of recommended recipes for making CDRom decks and some examples of how these recipes can be modified to suit theommelier's needs.
Related Papers (5)
Frequently Asked Questions (14)
Q1. What contributions have the authors mentioned in the paper "A paraperspective factorization method for shape and motion recovery" ?

The method robustly processes the feature trajectory information using singular value decomposition ( SVD ), taking advantage of the linear algebraic properties of orthographic projection. 

Generally about six steps were required for convergence of a single point or frame refinement, so a complete refinement step requires 6P inversions of 3 3¥ matrices and 6F inversions of 6 6¥ matrices. 

It achieves its accuracy and robustness by applying a well-understood numerical computation, the singular value decomposition (SVD), to a large number of images and feature points, and by directly computing shape without computing the depth as an intermediate step. 

While holding the shape constant, the minimization with respect to the motion variables can be performed independently for each frame. 

Due to the bumpy motion of the plane and the instability of the hand-held camera, features often moved by as much as 30 pixels from one image to the next. 

The authors perform the individual minimizations, fitting six motion variables to P equations or fitting three shape variables to 2F equations, using the Levenberg-Marquardt method [8]. 

In image sequences in which the object being viewed translates significantly toward or away from the camera or across the camera’s field of view, the paraperspective factorization method performs significantly better than the orthographic method. 

The principle that the measurement matrix has rank three, as put forth by Tomasi and Kanade in [14], was dependent on the use of an orthographic projection model. 

A common drawback of iterative methods on complex non-linear error surfaces is that the final result can be highly dependent on the initial value. 

The C implementation of the paraperspective factorization method required about 20-24 seconds to solve a system of 60 frames and 60 points on a Sun 4/65, with most of this time spent computing the singular value decomposition of the measurement matrix. 

The orthographic projection model assumes that rays are projected from an object point along the direction parallel to the camera’s optical axis, so that they strike the image plane orthogonally, as illustrated in Fig. 

When noise is present in the input, the W* will not be exactly of rank three, so the Tomasi-Kanade factorization method uses the SVD to find the best rank three approximation to W*, factoring it into the productW MS* $ $= (8)The decomposition of (8) is only determined up to a linear transformation. 

The shape recovered by the orthographic factorization method was rather deformed (see Fig. 8) and the recovered motion incorrect, because the method could notaccount for the scaling and position effects which are prominent in the sequence. 

The scaled orthographic factorization method is very similar to theparaperspective factorization method; the metric constraints for the method are m nf f 2 2 = , m nf f◊ = 0 , and m1 1= .