scispace - formally typeset
Search or ask a question
Proceedings ArticleDOI

Reinforcement learning control of robot manipulators in uncertain environments

10 Feb 2009-pp 1-6
TL;DR: This work investigates here the robust tracking performance of reinforcement learning control of manipulators, subjected to parameter variations and extraneous disturbances, and shows the importance of fuzzy Q-learning control.
Abstract: Considerable attention has been given to the design of stable controllers for robot manipulators, in the presence of uncertainties. We investigate here the robust tracking performance of reinforcement learning control of manipulators, subjected to parameter variations and extraneous disturbances. Robustness properties in terms of average error, absolute maximum errors and absolute maximum control efforts, have been compared for reinforcement learning systems using various parameterized function approximators, such as fuzzy, neural network, decision tree, and support vector machine. Simulation results show the importance of fuzzy Q-learning control. Further improvements in this control approach through dynamic fuzzy Q-learning have also been highlighted.
Citations
More filters
Journal ArticleDOI
TL;DR: An example of an implementation of a novel model-free Q-learning based discrete optimal adaptive controller for a humanoid robot arm that uses a novel adaptive dynamic programming (ADP) reinforcement learning (RL) approach to develop an optimal policy on-line.

172 citations


Cites background from "Reinforcement learning control of r..."

  • ...Shah and Gopal (2009) have presented reinforcement learning control for robot manipulators in uncertain environments....

    [...]

Journal ArticleDOI
TL;DR: In this paper, a model-free predictive control (MFPC) framework is proposed to take care of both the issues of conventional MPC and the excessive computational burden associated with the control optimization.

28 citations

Journal ArticleDOI
27 Sep 2017-Robotics
TL;DR: The authors review and discuss the learning control in robotic manipulators and some issues in learning control for robotic manipulator and are able to give a general guideline for future research.

11 citations


Cites background from "Reinforcement learning control of r..."

  • ...In [74], the tracking performance of reinforcement learning control for a two link robotic mechanism that has parameter variations and external disturbances is studied....

    [...]

Proceedings ArticleDOI
13 Oct 2010
TL;DR: This paper introduces a preliminary design of deicing robot, which travels on transmission lines and automatically remove ices, and implements a graphical simulation environment and uses it to evaluate the Q-learning based line-grasping control policy study algorithm.
Abstract: Ice coating in power networks could result in power-tower collapse and power interruption. This paper introduces a preliminary design of deicing robot, which travels on transmission lines and automatically remove ices. Inevitably, the deicing robot will encounter some obstacles. To cross an obstacle, the deicing robot needs to control its arms to grasp transmission line over the obstacle. In this paper, Q-learning, one of reinforcement learning algorithms is applied to study the line-grasping control strategies for deicing robot. We implement a graphical simulation environment and use it to evaluate the Q-learning based line-grasping control policy study algorithm. Simulation results show that the proposed algorithm is promising for the line-grasping control of deicing robot.

4 citations


Cites background from "Reinforcement learning control of r..."

  • ...And some papers gave the comparison among them [13, 14]....

    [...]

Proceedings ArticleDOI
05 Jul 2009
TL;DR: This paper proposes a method for hybridization of model-based approach with RL, which is the right solution for such control problems and shows superiority in terms of robustness of the controller to parameter variations in the plant.
Abstract: Reinforcement learning (RL) is a popular learning paradigm to adaptive learning control of nonlinear systems, and is able to work without an explicit model. However, learning from scratch, i.e., without any a priori knowledge, is a daunting undertaking, which results in long training time and instability of learning process with large continuous state space. For physical systems, one must consider that the design of controller is very rarely a tabula rasa: some approximate mathematical model of the system is always available. In this paper, our focus is on control applications wherein the system to be controlled is a physical system. We can always obtain at least an approximate mathematical model of the plant to be controlled. We propose a method for hybridization of model-based approach with RL, which is the right solution for such control problems. The superiority of proposed hybrid approach has been established through simulation experiments on a cart-pole balance bench mark problem, comparing it with model-free RL system. We have used fuzzy inference system for function approximation; it can deal with continuous action space in Q-learning. Comparison with other function approximators has shown its superiority in terms of robustness of the controller to parameter variations in the plant.

2 citations

References
More filters
Book
Vladimir Vapnik1
01 Jan 1995
TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Abstract: Setting of the learning problem consistency of learning processes bounds on the rate of convergence of learning processes controlling the generalization ability of learning processes constructing learning algorithms what is important in learning theory?.

40,147 citations


"Reinforcement learning control of r..." refers methods in this paper

  • ...SVM Q-learning SVM is a new universal learning machine in the framework of structural risk minimization (SRM) [13]....

    [...]

  • ...D. SVM Q-learning SVM is a new universal learning machine in the framework of structural risk minimization (SRM) [13]....

    [...]

  • ...SVM uses a kernel function that satisfies Mercer’s condition [13], to map the input data into a highdimensional feature space, and then construct a linear optimal separating hyper plane in that space....

    [...]

  • ...SRM has better generalization ability and is superior to the traditional empirical risk minimization (ERM) principle....

    [...]

  • ...The membership function parameters used in this paper are same as in [13]....

    [...]

Book
01 Jan 1988
TL;DR: This book provides a clear and simple account of the key ideas and algorithms of reinforcement learning, which ranges from the history of the field's intellectual foundations to the most recent developments and applications.
Abstract: Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives when interacting with a complex, uncertain environment. In Reinforcement Learning, Richard Sutton and Andrew Barto provide a clear and simple account of the key ideas and algorithms of reinforcement learning. Their discussion ranges from the history of the field's intellectual foundations to the most recent developments and applications. The only necessary mathematical background is familiarity with elementary concepts of probability. The book is divided into three parts. Part I defines the reinforcement learning problem in terms of Markov decision processes. Part II provides basic solution methods: dynamic programming, Monte Carlo methods, and temporal-difference learning. Part III presents a unified view of the solution methods and incorporates artificial neural networks, eligibility traces, and planning; the two final chapters present case studies and consider the future of reinforcement learning.

37,989 citations


"Reinforcement learning control of r..." refers methods in this paper

  • ...In order to explore the set of possible actions and acquire experience through the RL signals, actions are selected using an exploration/exploitation policy (EEP) [11]....

    [...]

  • ...If †iu , the action selected in rule iR is ε -greedy iu (where ε -greedy is a function implementing the EEP strategy), while *iu is the maximizing action, i.e., *( , ) max ( , )i b mq i u q i b≤= , then Q-value for the inferred action ku is † 1 1 ( , ) ( ) ( , ) ( ) N Nk k k k i i i i i Q x u x q i u xα α = = = ∑ ∑ , and value of state kx is: 1 1 ( ) ( ) ( , ) ( ) N Nk k k i i i i i V x x q i u xα α∗ = = = ∑ ∑ ....

    [...]

  • ...This information is used to calculate temporal difference (TD) [11] approximation error as: 1 ( ) ( , ) k k k n k Q c V x Q x u γ + ∆ = + − and q parameter values are updated as...

    [...]

Journal ArticleDOI
TL;DR: A least squares version for support vector machine (SVM) classifiers that follows from solving a set of linear equations, instead of quadratic programming for classical SVM's.
Abstract: In this letter we discuss a least squares version for support vector machine (SVM) classifiers. Due to equality type constraints in the formulation, the solution follows from solving a set of linear equations, instead of quadratic programming for classical SVM‘s. The approach is illustrated on a two-spiral benchmark classification problem.

8,811 citations

01 Jan 1989

4,916 citations


"Reinforcement learning control of r..." refers background or methods in this paper

  • ...One of the most popular RL approaches is the Q-learning [3]....

    [...]

  • ...It is an adaptation of Watkin’s Q-learning [3] for FIS, where both the actions and Q-functions are inferred from fuzzy rules....

    [...]

Journal ArticleDOI
TL;DR: Reinforcement learning methods are presented as a computationally simple, direct approach to the adaptive optimal control of nonlinear systems.
Abstract: Neural network reinforcement learning methods are described and considered as a direct approach to adaptive optimal control of nonlinear systems. These methods have their roots in studies of animal learning and in early learning control work. An emerging deeper understanding of these methods is summarized that is obtained by viewing them as a synthesis of dynamic programming and stochastic approximation methods. The focus is on Q-learning systems, which maintain estimates of utilities for all state-action pairs and make use of these estimates to select actions. The use of hybrid direct/indirect methods is briefly discussed. >

437 citations


"Reinforcement learning control of r..." refers background in this paper

  • ...RL is a computationally simple, direct approach to the adaptive optimal control of nonlinear systems [2]....

    [...]