scispace - formally typeset
Search or ask a question
Proceedings ArticleDOI

Combining central and peripheral vision for reactive robot navigation

23 Jun 1999-Vol. 2, pp 2646-2651
TL;DR: A new method for vision-based, reactive robot navigation that enables a robot to move in the middle of the free space by exploiting both central and peripheral vision and is computationally efficient.
Abstract: In this paper we present a new method for vision-based, reactive robot navigation that enables a robot to move in the middle of the free space by exploiting both central and peripheral vision. The robot employs a forward-looking camera for central vision and two side-looking cameras for sensing the periphery of its visual field. The developed method combines the information acquired by this trinocular vision system and produces low-level motor commands that keep the robot in the middle of the free space. The approach follows the purposive vision paradigm in the sense that vision is not studied in isolation but in the context of the behaviors that the system is engaged as well as the environment and the robot's motor capabilities. It is demonstrated that by taking into account these issues, vision processing can be drastically simplified still giving rise to quite complex behaviors. The proposed method does not make strict assumptions about the environment, requires very low level information to be extracted from the images, produces a robust robot behavior and is computationally efficient. Results obtained by bath simulations and from a prototype on-line implementation demonstrate the effectiveness of the method.

Summary (2 min read)

1 Introduction

  • The term navigation refers to the capability of a system to move autonomously in its environment by using its own sensors.
  • The more specific term visual navigation is used for the process of motion control based on the analysis of data gathered by visual sensors.
  • The topic of visual navigation is *This research has been carried out during the first author's 1996-97 appointment to CVAPINADAIKTH and was funded under the VIRGO research network (EC Contract No ERBFMRX-CT96-0049) of the TMR Programme.

2 The behavior, the environment and the body

  • In this work the authors assume a robot that can translate in the forward direction and rotate (pan) around its vertical axis (Fig. 1 ).
  • The term reactive is used to express lack of a particular destination that could be set by using maps of the environment, landmark recognition etc.
  • The behavior is based [5] on velocity information computed at the left and right eyes of the bee.
  • Second, it turns out that control is facilitated when the cameras are slanted.
  • If the robot is in the middle of the free space, it perceives equal distances from the walls only if its pose is parallel to the walls.

3 Method description

  • In Eq. ( 13), .F is a quantity that can be directly computed from functions of normal flow that have been extracted from the central and peripheral cameras.
  • Function F is equal to zero when the left and right cameras are in equal distances from world points and takes positive or negative values depending on whether the right camera is farther or closer from obstacles compared to the left camera.
  • Therefore, the computable quantity T can be used to control the rotational velocity by keeping the quantity F as close to zero as possible, achieving this way the desired behavior.
  • In their robot setup the authors avoid this special case.

4 Implementation issues -Experimental results

  • In their present algorithm, the authors did not allow for individual motions of the cameras (eye movements).
  • This is roughly tantamount to the assumption of approximately known FOE while exhibiting this behavior.
  • The authors did not calibrate the head so that the central camera pointed exactly in the forward motion translation direction.
  • In fact, the authors noticed that in many of the successful navigation experiments the optical axis of the central camera was 5-10 degrees off the xs # 0.

5 Conclusions

  • The method does not make strict assumptions about the environment, it requires very low level information to be extracted from the images, it produces a robust robot behavior and it is computationally efficient.
  • Results obtained by both simulations and from a prototype on-line implementation demonstrate the effectiveness of the method.
  • Peripheral vision seems to be very useful for achieving certain behaviors and its combination with central vision seems natural and appears to be powerful.

Did you find this useful? Give us your feedback

Content maybe subject to copyright    Report

Combining Central and Peripheral Vision
for Reactive Robot Navigation*
Antonis
A.
&gyros Fredrik Bergholm
Computer Vision and Robotics Lab.
Computational Vision and Active Perception Lab.
ICSFORTH
Heraklion
-
Crete
-
Greece
argyros @ics.forth.gr
Abstract
In
this paper; we present a new method for vision-based,
reactive robot navigation that enables a robot
to
move in
the middle of the free space
by
exploiting both central and
peripheral vision. The robot employs a forward-looking
camera for central vision and
two
side-looking cameras for
sensing the periphery of its visual jield. The developed
method combines the information acquired by this trinocu-
lar vision system and produces low-level motor commands
that keep the robot in the middle of the free space. The ap-
proach follows the purposive vision paradigm in the sense
that vision is not studied in isolation but in the context of the
behaviors that the system is engaged as well as the environ-
ment and the robot’s motor capabilities. It is demonstrated
that by taking into account these issues, vision processing
can be drastically simpl$ed, still giving rise to quite com-
plex behaviors. The proposed method does not make strict
assumptions about the environment, requires very low level
information to be extracted from the images, produces a ro-
bust robot behaviorand is computationally eficient. Results
obtained by both simulations and from a prototype on-line
implementation demonstrate the effectiveness of the method.
1
Introduction
The term navigation refers to the capability of a system
to move autonomously in its environment by using its own
sensors. The more specific term visual navigation is used for
the process
of
motion control based on the analysis of data
gathered by visual sensors. The topic of visual navigation is
*This
research has been carried out during the first author’s 1996-97
appointment to CVAPINADAIKTH and was funded under the VIRGO
research network (EC Contract
No
ERBFMRX-CT96-0049) of the TMR
Programme.
NADNKTH
Stockholm
-
Sweden
fredrikb
@
nada.kth.se
of
particular importance mainly because
of
the rich percep-
tual input provided by vision. Moreover, navigation that is
based on other types
of
sensors,
in
contrast to vision, often
requires modification of the environment (e.g. insertion of
emitters) which imposes constraints on the application of
such methods in unknown environments.
The problem of visual navigation has been traditionally
treated without taking very much into account the environ-
ment
of
the robot, its body and the characteristics
of
the de-
sired behavior. Typically, monocular
or
stereoscopic visual
systems are assumed and the effort is focused on construct-
ing
a
general representation of the environment that may
thereafter support the solution of any vision-related prob-
lem. During the last decade, a new vision paradigm has
attracted the interest
of
the computational vision research
community. According to this paradigm, called active and
purposive vision
[I],
vision is more readily understood in
the context
of
the behaviors
in
which the system is engaged.
Consequently, vision attemtps to explore the aspects
of
the
world that are important
for
the system at a given point
in time, instead of aiming at a general representation
of
the environment which, besides being extremely difficult to
extract,
it
is probably not needed either. The interest in
purposive vision is largely motivated by the fact that all bi-
ological vision systems are highly active and purposive
[2].
The purposiveness
of
visual processes enables the formula-
tion and the solution of simpler problems that have a relative
small number of possible solutions and can be treated
in
a
qualitative manner
[3].
In this paper, we describe a new method for visual robot
navigation based on the principles of purposive vision. By
employing a forward-looking camera for central vision and
two side-looking cameras for sensing the periphery
of
the
visual field, reactive robot navigation has been achieved.
The developed method combines the information acquired
by this trinocular vision system to produce low-level
mo-
tor commands that keep the robot in the middle of the free
646
0-7695-0149-4/99
$10.00
0
1999
IEEE
In Proc. of the Comp. Vision and Pat. Recogn. Conf., (CVPR’99), Vol. 2, Ft. Collins, Colorado, Jun. 21-25, 1999.

Figure
1.
Top-down view of the robot geome-
try. The placement of the two peripheral cam-
eras is also shown.
space. The aim of this work is to investigate how the de-
sign of a visual system can assist robots with specific bodies
and motor capabilities
in
exhibiting particular behaviors. It
is demonstrated that considering the behavior and the mo-
tor capabilities of a robot when designing its visual system,
leads to theoretically simpler and computationally more ef-
ficient solutions.
The rest of the paper is organized as follows. Section 2
describes the requirements that the target behavior poses to
the design of the robot’s visual system. Section
3
presents is-
sues related to the motion information that can be computed
by each camera as well as how this information is processed
and used to drive the robot. Section
4
presents results from
simulations of the method as well as implementation issues
and results obtained by an on-line implementation on a real
robotic platform. Finally, in section 5, the main conclusions
of this work are summarized and future research plans are
described.
2
The behavior, the environment and the
body
The study of the behaviors that should be exhibited by
an observer, its environment as well as the specifics of the
observer, provides valuable hints on how the sensors should
be placed in order to facilitate the implementation of a par-
ticular behavior. In this work we assume a robot that can
translate in the forward direction and rotate (pan) around
its vertical axis (Fig.
1).
We aim at developing a vision
based reactive navigation capability that enables the robot
to navigate in flat-floor indoor environments (long corridors,
narrow passages, rooms), avoiding collisions with walls and
obstacles. The term reactive is used to express lack of a
particular destination that could be set by using maps of
Figure
2.
The KTH head with
two
extra cam-
eras mounted on it for implementing periph-
eral vision.
the environment, landmark recognition etc. Free space is
defined based on the motor capabilities of the robot: the
robot moves on a plane and, therefore, all
3D
structures
that do not belong to this plane can be potentially harmful
if the robot crashes on them. Since the robot is about to
“live” in indoor environments, it is expected to be able to
handle situations where long corridors and narrow passages
are encountered. It can be shown that difficulties arise when
only central vision is used (i.e. a camera or a fixating stereo
configuration at the direction of translation). Consider for
example a camera with a field of view of
30
degrees that
is placed in a a 2-meters wide corridor with its optical axis
parallel to the walls. The camera can only see walls that
are approximately
4
meters ahead and it is therefore quite
difficult to maneuver accurately. On the other hand, the use
of cameras with wide field of view
[4]
give rise to depth
dependent geometric distortions that are difficult to correct.
In order to implement this behavior,
it
appears quite natural
to exploit the information provided by peripheral vision, i.e.
visual information at large angles with respect to the direc-
tion of forward translational motion (see for example the
configuration in Fig.
2).
By using such a camera configura-
tion, the robot is able to perceive walls and obstacles that are
immediately close to it. Moreover, the target behavior may
be implemented by indirectly comparing crude structure in-
formation acquired by the left and right peripheral cameras
instead of computing precise structure information. This ap-
proach is motivated by experiments that study the behavior
of honeybees [5]. In these experiments, bees were trained
to navigate along corridors towards a source of food. The
bees were observed to navigate in the middle of the corridor.
The eyes of the bees are pointing laterally (at about
180
degrees). The behavior is based [5] on velocity information
computed at the left and right eyes of the bee. In simple
terms, if a non-rotating (no panningkilting) bee is
in
the
center of the corridor, it perceives the world as ‘‘leaving’’ its
optical field with the same velocity in both eyes, while if
the bee is closer to one of the sides of the corridor, it per-
ceives it as moving faster. For a non-rotating observer, the
difference in the observed velocities depends only on depth.
647

a
Figure
3.
(a), (b) If a robot with lateral pe-
ripheral cameras is in the middle of the free
space, it perceives equal distances from left
and right walls independently of pose. (c) and
(d) If the peripheral cameras are slanted, the
distances at left and right are equal only
if
the
robot’s pose is parallel to the walls.
Therefore, balancing the flow, balances the distances to the
left and the right of the observer. Santos-Victor et al.
[6]
proposed the divergent stereo approach in order to exploit
this finding in robots. They exploit visual information that
is captured by two cameras with optical axes of opposite
ori-
entation that are mounted perpendicularly to the direction
of forward translation. Our research differs to the approach
in
[6]
in
several ways. First, peripheral cameras are not
placed
in
opposite directions because decisions on forward
motion should not be influenced by “past” structure infor-
mation. Proximity calculations based on data at
90
degrees
angle to the motion direction are sort of obsolete for most
reactive schemes and navigation situations. Second,
it
turns
out that control is facilitated when the cameras are slanted.
See for example Figs. 3(a) and 3(b) where the cameras
are placed laterally on the robot body. The robot perceives
equal distances at its left and right side, independently of
pose. However, the situation is different in cases 3(c) and
3(d) where the cameras are slanted towards the direction of
translation. If the robot is in the middle of the free space,
it perceives equal distances from the walls only if its pose
is parallel to the walls. Therefore, in this case, flow balanc-
ing fixes also the pose of the robot. Last, but not least, we
study the effects of the observer’s rotational motion in the
flow computed by the two peripheral cameras. A moving
robot is not only translating but also rotating and this rota-
tion affects the computed flow. We also show how central
vision (i.e. visual information acquired in the direction of
the translation) can be used along with peripheral vision in
order to simplify the problems to be solved.
3
Method description
Consider an arbitrary 3D reference coordinate system
(RCS). Consider also a 3D camera coordinate system (CCS)
that is positioned at the optical center (nodal point) of a
pinhole camera. Assume that the center
of
the RCS remains
fixed at coordinates
(X,
,
Y,,
2,)
with respect to the CICS.
If the RCS moves with 3D translational velocity
(U,
V,
W)
and 3D rotational velocity
(a,
p,
y),
the equations relating
the 2D velocity
(U,
v)
of an image point
p(x, y)
to the 3D
motion parameters of the projected 3D point
P(X,
Y,
2:)
are
[71:
-Uf
+
XW
-
axY,
+
P(XX,
+
Zsf)
+
yYsf
-
Z
U=
+
a--@
-+f
+yy
xy
f
(; )
+
a
-+f
-p---yx
(?
)
7
where
f
denotes the focal length of the camera.
The projection of the optical flow
(U,
v)
along the inten-
sity gradient direction (i.e. the perpendicular to the edge at
that point) is also known as normal flow. The normal flow
is less informative than optical flow but can be computed
robustly and efficiently from image sequences by just using
differentiation techniques. Moreover, in contrast to the com-
putation of optical flow, no environmental assumptions such
as smoothness are required for normal flow computation.
For the above reasons, the proposed method for reactive
robot navigation relies on the computation of the normal
flow field. Let
(nz,
n,)
be the unit vector in the gradient
direction. The magnitude
uM
of
the normal flow vector is
given by:
uM
=
nzu+nyw
(2)
By substituting Eqs.
(1)
in Eq. (2) we obtain:
+
a)
v
-
az,
+
yx,
-
nyf
(
+
(ynz
-xny)y
(3)
In
our
robot setup (see Fig
l),
we set the RCS on the robot’s
body
so
that the
Y
axis coincides with the robot’s rotational
axis and the
Z
axis is parallel to the robot’s translational
motion.
As
it has already been discussed, we assume that
the robot is capable of translating with velocity
S
and rotate
(pan) with velocity
p.
We set the nodal point of the right
peripheral camera
so
that the center of the RCS is at coor-
dinates
(-X,,
0,
-ZS).
Similarly we set the nodal points
648

of the left peripheral camera and of the central camera
so
that the center of the RCS is at coordinates (Xs,
0,
-Zs)
and
(0, 0, 0),
respectively. The translational velocity
S
of the robot produces translational velocities
(-U,
0,
W),
(U,
0,
W) and
(0,
0,
S)
for the right, left and central cam-
eras, respectively. For all three cameras, the rotational ve-
locity of the robot produces a rotational velocity
(0,
P,
0)
at
the camera coordinate system.
By taking into account the above considerations for each
of the left
(L),
right
(R),
and central
(C)
cameras, Eq. (3)
gives:
UCM
=
bxf
where
ZL,
ZR
and
Zc
represent the depth of the
3D
points
perceived by the left, right and central cameras, respectively.
By selecting normal flow vectors
for
which it holds that
xn,
+
yn,
=
O',
we obtain:
and
and and by selecting normal flow vectors for which it holds
that
(nx,
nY)
=
(0,l)
(the vertical normalflows), we obtain:
The left sides of Eqs.
(7),
(8)
and
(9)
can be computed
from images because they employ normal flow values, point
coordinates and gradient directions. The right sides of Eqs.
(7)
and
(9)
employ functions of depth and can get an even
simpler form by noting that the function acquired by the
central camera (Eq.
(8)),
gives the rotation. Thus, Eqs.
(7)
and
(9)
after derotation, become:
'The selected
normal flow
vectors
are
those that are tangent
to
circles
centered at the
image
origin
and
Thus, central vision can be used to derotate the flow fields
produced at the peripheral cameras. Having exploited this
observation, it turns out that:
Equation (12) can be rewritten in a simpler form as follows:
In Eq. (13),
.F
is
a quantity that can be directly computed
from functions of normal flow that have been extracted from
the central and peripheral cameras.
C
is an unknown con-
stant (of known sign) that depends on the characteristics of
the body of the observer as well as its constant translational
velocity. Function
F
is equal to zero when the left and right
cameras are in equal distances from world points and takes
positive
or
negative values depending on whether the right
camera is farther
or
closer from obstacles compared to the
left camera. Therefore, the computable quantity
T
can be
used to control the rotational velocity by keeping the quan-
tity
F
as close to zero as possible, achieving this way the
desired behavior. Note that
C
is equal to zero if the nodal
points of the left, right and central cameras are collinear be-
cause in this case
U2
=
W. In our robot setup we avoid
this special case.
4
Implementation issues
-
Experimental re-
sults
An experimental evaluation of the proposed method has
been based on both simulation results as well as on results
obtained by an on-line implementation of the method on
a real robotic platform. Simulations have been based on
the KHEPERA simulator
[8],
which has been modified to
simulate the central and peripheral cameras of the robot. The
aim of the simulation experiments was to test the control
law used to drive the robot. Thus the function
.F
of Eq.
(1
3) has been simulated and the robot was set to navigate in
various environments. Several experiments were conducted.
Figure
4
shows a sample run. Thin dark lines represent
the walls of the corridor-like environment. The thick dark
line is the trace of the robot. It can be observed that the
robot started at the bottom-right end of the environment and,
after reaching the end of the corridor has started moving
649

Figure
4.
A
run
of
the simulated robot.
backwards. Moreover, the robot moves in a smooth path
among the various obstacles of the environment.
One of the most interesting results of simulations was the
difference in the behavior of the simulated robot depending
on whether the peripheral cameras were laterally placed or
slanted with respect to the direction of translation. ,It turns
out that slanted cameras result in smooth robot paths, while
the laterally placed ones produce snake-like robot motion
patterns.
The simulation experiments are, of course, not adequate
for testing the performance of the method when real vision
processes are employed. For this reason, an on-line imple-
mentation of the method has been realized.
The
platform
used was a LABMATE ROBUTER on which “Charlie”, the
KTH active vision head has been mounted. Two extra cam-
eras were mounted on “Charlie” implementing peripheral
vision
(Fig.
2).
Only one
of
the central cameras was used
to
implement central vision. Note that the central camera
is
not placed at coordinates
(0,
0,O)
with respect to the RCS as
it has been theoretically assumed. The process of selecting
normal flow vectors at specific directions leads to cancelling
of motion components. It turns out that Eq.
(8)
still holds if
In our implementation, a
SUN
Ultra Sparc was respon-
sible for peripheral vision processing and a PENTIUM pro-
cessor running LINUX was responsible for central vision
processing. The distributed processing as well as the inter-
process communication was based on the TCX communi-
cations library
[9].
Various navigation scenarios have been
tested in which the robot successfully managed to perform
maneuvers in narrow passages. In Figs.
5
and
6,
we present
snapshots from two different navigation sessions.
In our present algorithm, we did not allow for individual
motions of the cameras (eye movements). This is roughly
tantamount to the assumption of approximately known FOE
while exhibiting this behavior. However, we did not cali-
brate the head
so
that the central camera pointed exactly
in
the forward motion translation direction. In fact, we noticed
that in many of the successful navigation experiments the
optical axis of the central camera was
5-10
degrees off the
xs
#
0.
Figure
5.
Snapshots
of
a navigation session
(left
to right, top to bottom).
650

Citations
More filters
Journal ArticleDOI
TL;DR: An algorithm for visual obstacle avoidance of autonomous mobile robot by balancing the amount of left and right side flow to avoid obstacles is developed, this technique allows robot navigation without any collision with obstacles.
Abstract: In this paper we try to develop an algorithm for visual obstacle avoidance of autonomous mobile robot. The input of the algorithm is an image sequence grabbed by an embedded camera on the B21r robot in motion. Then, the optical flow information is extracted from the image sequence in order to be used in the navigation algorithm. The optical flow provides very important information about the robot environment, like: the obstacles disposition, the robot heading, the time to collision and the depth. The strategy consists in balancing the amount of left and right side flow to avoid obstacles, this technique allows robot navigation without any collision with obstacles. The robustness of the algorithm will be showed by some

144 citations

Journal ArticleDOI
TL;DR: The present work considers the development of a wheelchair for people with special needs, which is capable of navigating semi-autonomously within its workspace and takes into account the non-holonomic kinematic constraints of the wheelchair, in order to guarantee certain desired features of the closed-loop system.
Abstract: The present work considers the development of a wheelchair for people with special needs, which is capable of navigating semi-autonomously within its workspace. This system is expected to prove useful to people with impaired mobility and limited fine motor control of the upper extremities. Among the implemented behaviors of this robotic system are the avoidance of obstacles, the motion in the middle of the free space and the following of a moving target specified by the user (e.g., a person walking in front of the wheelchair). The wheelchair is equipped with sonars, which are used for distance measurement in preselected critical directions, and with a panoramic camera with a 360 degree field of view, which is used for following a moving target. After suitably processing the color sequence of the panoramic images using the color histogram of the desired target, the orientation of the target with respect to the wheelchair is determined, while its distance is determined by the sonars. The motion control laws developed for the system use the sensory data and take into account the non-holonomic kinematic constraints of the wheelchair, in order to guarantee certain desired features of the closed-loop system, such as stability. Moreover, they are as simplified as possible to minimize implementation requirements. An experimental prototype has been developed at ICS–FORTH, based on a commercially-available wheelchair. The sensors, the computing power and the electronics needed for the implementation of the navigation behaviors and of the user interfaces (touch screen, voice commands) were developed as add-on modules and integrated with the wheelchair.

72 citations


Cites methods from "Combining central and peripheral vi..."

  • ...The method presented in [14] is based on the method proposed in [1] and it is inspired by experiments on the navigational capabilities of bees [3, 11]....

    [...]

Journal ArticleDOI
TL;DR: This work investigates the deployment of a submersible platform with stereovision imaging capability for three-dimensional mapping of benthic habitats and other sea-floor structures over local areas through the fusion of various visual cues, including motion and stereo within a Kalman filter.
Abstract: We investigate the deployment of a submersible platform with stereovision imaging capability for three-dimensional (3D) mapping of benthic habitats and other sea-floor structures over local areas. A complete framework is studied, comprising: 1) suitable trajectories to be executed for data collection; 2) data processing for positioning and trajectory followed by online frame-to-frame and frame-to-mosaic registration of images, as well as recursive global realignment of positions along the path; and 3) 3D mapping by the fusion of various visual cues, including motion and stereo within a Kalman filter. The computational requirements of the system are evaluated, formalizing how processing may be achieved in real time. The proposed scenario is simulated for testing with known ground truth to assess the system performance, to quantify various errors, and to identify how performance may be improved. Experiments with underwater images are also presented to verify the performance of various components and the overall scheme.

67 citations


Cites background or methods from "Combining central and peripheral vi..."

  • ...Stereovision systems have been extensively studied for various terrestrial and planetary applications, including object reconstruction in industrial robotics systems [14], [24]; indoor and outdoor mobile robotic navigation [ 2 ], [12], [60], [62], [74]; autonomous vehicles for highway and offroad navigation [6], [11], [55]; and planetary rovers [26], [38], [50], [59]....

    [...]

  • ...Stereo imaging is a classic technique that has been utilized commonly for 3-D reconstruction in industrial robotics manufacturing systems [14], [19], [24], [67], [69], as well as for obstacle avoidance and 3-D mapping in the deployment of autonomous terrestrial and space mobile platforms [ 2 ], [12], [26], [38], [50], [59], [60], [62], [74]....

    [...]

Journal ArticleDOI
TL;DR: The proposed method relies on the extraction of primitive visual information from appropriately selected areas of a panoramic visual field and its direct use in the control law and the resulting control laws were shown to possess the required stability properties.
Abstract: A reactive robotic centering behavior based on panoramic vision is presented. It is inspired by the way insects exploit visual information in analogous navigation tasks. By employing a panoramic camera, the development of the centering behavior is simplified both from a theoretical and from an implementation point of view. The proposed method relies on the extraction of primitive visual information from appropriately selected areas of a panoramic visual field and its direct use in the control law. Experimental results from an implementation of this method on a robotic platform demonstrate a centering behavior which can be achieved in real-time and with high accuracy. The proposed technique circumvents the need to address complex problems of 3D structure estimation and the resulting control laws were shown to possess the required stability properties.

59 citations


Cites methods from "Combining central and peripheral vi..."

  • ...Since the computation of the quantity L 1,2 involves only the horizontal component of the optical flow, normal flow in selected edge directions (e.g., vertical edges) could have been used instead of optical flow, as in [9]....

    [...]

  • ...This problem is corrected in [9], where a trinocular camera system is employed....

    [...]

Journal ArticleDOI
TL;DR: A bibliography of nearly 1700 references related to computer vision and image analysis, arranged by subject matter is presented, including computational techniques; feature detection and segmentation; image and scene analysis; and motion.

39 citations

References
More filters
01 Jan 1995
TL;DR: It is claimed that the state of computer architecture has been a strong influence on models of thought in Artificial Intelligence over the last thirty years.
Abstract: Computers and Thought are the two categories that together define Artificial Intelligence as a discipline. It is generally accepted that work in Artificial Intelligence over the last thirty years has had a strong influence on aspects of computer architectures. In this paper we also make the converse claim; that the state of computer architecture has been a strong influence on our models of thought. The Von Neumann model of computation has lead Artificial Intelligence in particular directions. Intelligence in biological systems is completely different. Recent work in behavior-based Artificial Intelligence has produced new models of intelligence that are much closer in spirit to biological systems. The non-Von Neumann computational models they use share many characteristics with biological computation.

1,796 citations


"Combining central and peripheral vi..." refers background in this paper

  • ...The interest in purposive vision is largely motivated by the fact that all biological vision systems are highly active and purposive [2]....

    [...]

Book ChapterDOI
24 Aug 1991
TL;DR: In this article, the authors make the converse claim that the state of computer architecture has been a strong influence on our models of thought, and they use non-Von Neumann computational models they use share many characteristics with biological computation.
Abstract: Computers and Thought are the two categories that together define Artificial Intelligence as a discipline. It is generally accepted that work in Artificial Intelligence over the last thirty years has had a strong influence on aspects of computer architectures. In this paper we also make the converse claim; that the state of computer architecture has been a strong influence on our models of thought. The Von Neumann model of computation has lead Artificial Intelligence in particular directions. Intelligence in biological systems is completely different. Recent work in behavior-based Artificial Intelligence has produced new models of intelligence that are much closer in spirit to biological systems. The non-Von Neumann computational models they use share many characteristics with biological computation.

1,537 citations

Journal ArticleDOI
TL;DR: It is shown that from a monocular view of a rigid, textured, curved surface it is possible, in principle, to determine the gradient of the surface at any point, and the motion of the eye relative to it, from the velocity field of the changing retinal image, and its first and second spatial derivatives.
Abstract: It is shown that from a monocular view of a rigid, textured, curved surface it is possible, in principle, to determine the gradient of the surface at any point, and the motion of the eye relative to it, from the velocity field of the changing retinal image, and its first and second spatial derivatives. The relevant equations are redundant, thus providing a test of the rigidity assumption. They involve, among other observable quantities, the components of shear of the retinal velocity field, suggesting that the visual system may possess specialized channels for computing these components.

1,468 citations


"Combining central and peripheral vi..." refers background in this paper

  • ...If the RCS moves with 3D translational velocity (U; V;W ) and 3D rotational velocity ( ; ; ), the equations relating the 2D velocity (u; v) of an image point p(x; y) to the 3D motion parameters of the projected 3D point P (X;Y; Z) are [7]: u = Uf + xW xYs + (xXs + Zsf) + Ysf Z + xy f x2 f + f + y...

    [...]

Book
12 Jan 1993
TL;DR: Sometimes, reading is very boring and it will take long time starting from getting the book and start reading, but in modern era, you can take the developing technology by utilizing the internet and search for the book that is needed.
Abstract: Sometimes, reading is very boring and it will take long time starting from getting the book and start reading. However, in modern era, you can take the developing technology by utilizing the internet. By internet, you can visit this page and start to search for the book that is needed. Wondering this active vision is the one that you need, you can go for downloading. Have you understood how to get it?

1,425 citations

Proceedings ArticleDOI
16 Jun 1990
TL;DR: The traditional view of the problem of computer vision as a recovery problem is questioned, and the paradigm of purposive-qualitative vision is offered as an alternative and the design of the Medusa of CVL is described.
Abstract: The traditional view of the problem of computer vision as a recovery problem is questioned, and the paradigm of purposive-qualitative vision is offered as an alternative. This paradigm considers vision as a general recognition problem (recognition of objects, patterns or situations). To demonstrate the usefulness of the framework, the design of the Medusa of CVL is described. It is noted that this machine can perform complex visual tasks without reconstructing the world. If it is provided with intentions, knowledge of the environment, and planning capabilities, it can perform highly sophisticated navigational tasks. It is explained why the traditional structure from motion problem cannot be solved in some cases and why there is reason to be pessimistic about the optimal performance of a structure from motion module. New directions for future research on this problem in the recovery paradigm, e.g., research on stability or robustness, are suggested. >

338 citations


"Combining central and peripheral vi..." refers background in this paper

  • ...The purposiveness of visual processes enables the formulation and the solution of simpler problems that have a relative small number of possible solutions and can be treated in a qualitative manner [3]....

    [...]