scispace - formally typeset
Proceedings ArticleDOI

Learning the Point Gathering Task Using Shared Value Function In Mobile Robots

22 Jan 2009-pp 9-13

TL;DR: The algorithm was analyzed through simulations for up to 5 robots and the analysis indicates that communication helped robots perform significantly faster than when they acted independently - measured using the path-length of the first robot to reach the goal as the metric.

AbstractManaging the actions of several agents to perform tasks which require coordination and cooperation pose significant research challenges. One such challenge is to synchronize the agents’ view of the system to help them take the ‘right’ actions. In this paper, we propose an algorithm called MRCPG (Mobile Robot Coordination Point Gathering Algorithm) for coordinating the actions of a team of mobile robots. The aim is to gather these robots at a particular location in a 2-dimensional plane which is determined during execution. The robots are randomly deployed in the plane and they achieve the goal by communicating periodically. In addition, we impose a Reinforcement Learning framework and the robots learn a Shared Value Function (SVF) based on scalar rewards received. The SVF is used to select the best possible action in each state until at least one robot reaches the goal. Then a Reach-distance heuristic is used to direct the remaining robots to the goal. The algorithm was analyzed through simulations for up to 5 robots and the analysis indicates that communication helped robots perform significantly faster than when they acted independently - measured using the path-length of the first robot to reach the goal as the metric. We also observed that increasing team size enhances the effect of communication and hastens task completion.

...read more


Citations
More filters
Book ChapterDOI
17 Dec 2015
TL;DR: This work presents a novel approach to solving the distribution of the environmental feedback signal among learning agents in Multiagent Reinforcement Learning (MARL).
Abstract: In Multiagent Reinforcement Learning (MARL), a single scalar reinforcement signal is the sole reliable feedback that members of a team of learning agents can receive from the environment around them. Hence, the distribution of the environmental feedback signal among learning agents, also known as the “Multiagent Credit Assignment” (MCA), is among the most challenging problems in MARL.

References
More filters
Journal ArticleDOI
01 Dec 1998
TL;DR: New reactive behaviors that implement formations in multirobot teams are presented and evaluated and demonstrate the value of various types of formations in autonomous, human-led and communications-restricted applications, and their appropriateness in different types of task environments.
Abstract: New reactive behaviors that implement formations in multirobot teams are presented and evaluated. The formation behaviors are integrated with other navigational behaviors to enable a robotic team to reach navigational goals, avoid hazards and simultaneously remain in formation. The behaviors are implemented in simulation, on robots in the laboratory and aboard DARPA's HMMWV-based unmanned ground vehicles. The technique has been integrated with the autonomous robot architecture (AuRA) and the UGV Demo II architecture. The results demonstrate the value of various types of formations in autonomous, human-led and communications-restricted applications, and their appropriateness in different types of task environments.

2,819 citations

Journal ArticleDOI
TL;DR: A variant of the potential field method is used to produce the appropriate velocity and steering commands for the robot and demonstrates the feasibility of this approach.
Abstract: Motor schemas serve as the basic unit of behavior specifica tion for the navigation of a mobile robot. They are multiple concurrent processes that operate in conjunction with asso ciated perceptual...

1,058 citations

Journal ArticleDOI
01 Dec 1999
TL;DR: The results of computer simulation under a more realistic model give convincing indication that the algorithm, if implemented on physical robots, will be robust against sensor and control error.
Abstract: We present a distributed algorithm for converging autonomous mobile robots with limited visibility toward a single point. Each robot is an omnidirectional mobile processor that repeatedly: 1) observes the relative positions of those robots that are visible; 2) computes its new position based on the observation using the given algorithm; 3) moves to that position. The robots' visibility is limited so that two robots can see each other if and only if they are within distance V of each other and there are no other robots between them. Our algorithm is memoryless in the sense that the next position of a robot is determined entirely from the positions of the robots that it can see at that moment. The correctness of the algorithm is proved formally under an abstract model of the robot system in which: 1) each robot is represented by a point that does not obstruct the view of other robots; 2) the robots' motion is instantaneous; 3) there are no sensor and control error; 4) the issue of collision is ignored. The results of computer simulation under a more realistic model give convincing indication that the algorithm, if implemented on physical robots, will be robust against sensor and control error.

558 citations

Journal ArticleDOI
TL;DR: This paper presents a protocol that allows anonymous oblivious robots with limited visibility to gather in the same location in finite time, provided they have orientation (i.e., agreement on a coordinate system), indicating that, with respect to gathering, orientation is at least as powerful as instantaneous movements.
Abstract: In this paper we study the problem of gathering a collection of identical oblivious mobile robots in the same location of the plane. Previous investigations have focused mostly on the unlimited visibility setting, where each robot can always see all the others regardless of their distance.In the more difficult and realistic setting where the robots have limited visibility, the existing algorithmic results are only for convergence (towards a common point, without ever reaching it) and only for semi-synchronous environments, where robots' movements are assumed to be performed instantaneously.In contrast, we study this problem in a totally asynchronous setting, where robots' actions, computations, and movements require a finite but otherwise unpredictable amount of time. We present a protocol that allows anonymous oblivious robots with limited visibility to gather in the same location in finite time, provided they have orientation (i.e., agreement on a coordinate system).Our result indicates that, with respect to gathering, orientation is at least as powerful as instantaneous movements.

395 citations

Journal ArticleDOI
TL;DR: A sufficient condition for the stability of a desired formation pattern for a fleet of robots each equipped with the navigation strategy based on nearest neighbor tracking is developed and simple navigation strategies for robots moving in formation are derived.
Abstract: The problem of deriving navigation strategies for a fleet of autonomous mobile robots moving in formation is considered. Here, each robot is represented by a particle with a spherical effective spatial domain and a specified cone of visibility. The global motion of each robot in the world space is described by the equations of motion of the robot's center of mass. First, methods for formation generation are discussed. Then, simple navigation strategies for robots moving in formation are derived. A sufficient condition for the stability of a desired formation pattern for a fleet of robots each equipped with the navigation strategy based on nearest neighbor tracking is developed. The dynamic behavior of robot fleets consisting of three or more robots moving in formation in a plane is studied by means of computer simulation.

393 citations