scispace - formally typeset
Search or ask a question

Showing papers by "Robert Babuska published in 2020"


Journal ArticleDOI
01 Sep 2020
TL;DR: This paper proposes to employ symbolic regression (SR) to construct parsimonious process models described by analytic equations and demonstrates on a real pendulum system that the analytic model found enables a RL controller to successfully perform the swing-up task, based on a model constructed from only 100 data samples.
Abstract: Developing mathematical models of dynamic systems is central to many disciplines of engineering and science. Models facilitate simulations, analysis of the system’s behavior, decision making and design of automatic control algorithms. Even inherently model-free control techniques such as reinforcement learning (RL) have been shown to benefit from the use of models, typically learned online. Any model construction method must address the tradeoff between the accuracy of the model and its complexity, which is difficult to strike. In this paper, we propose to employ symbolic regression (SR) to construct parsimonious process models described by analytic equations. We have equipped our method with two different state-of-the-art SR algorithms which automatically search for equations that fit the measured data: Single Node Genetic Programming (SNGP) and Multi-Gene Genetic Programming (MGGP). In addition to the standard problem formulation in the state-space domain, we show how the method can also be applied to input–output models of the NARX (nonlinear autoregressive with exogenous input) type. We present the approach on three simulated examples with up to 14-dimensional state space: an inverted pendulum, a mobile robot, and a bipedal walking robot. A comparison with deep neural networks and local linear regression shows that SR in most cases outperforms these commonly used alternative methods. We demonstrate on a real pendulum system that the analytic model found enables a RL controller to successfully perform the swing-up task, based on a model constructed from only 100 data samples.

16 citations


Proceedings ArticleDOI
25 Jun 2020
TL;DR: In this article, the authors propose a multi-objective symbolic regression approach that is driven by both the training data and the prior knowledge of the properties the desired model should manifest.
Abstract: In symbolic regression, the search for analytic models is typically driven purely by the prediction error observed on the training data samples. However, when the data samples do not sufficiently cover the input space, the prediction error does not provide sufficient guidance toward desired models. Standard symbolic regression techniques then yield models that are partially incorrect, for instance, in terms of their steady-state characteristics or local behavior. If these properties were considered already during the search process, more accurate and relevant models could be produced. We propose a multi-objective symbolic regression approach that is driven by both the training data and the prior knowledge of the properties the desired model should manifest. The properties given in the form of formal constraints are internally represented by a set of discrete data samples on which candidate models are exactly checked. The proposed approach was experimentally evaluated on three test problems with results clearly demonstrating its capability to evolve realistic models that fit the training data well while complying with the prior knowledge of the desired model characteristics at the same time. It outperforms standard symbolic regression by several orders of magnitude in terms of the mean squared deviation from a reference model.

14 citations


Posted Content
TL;DR: A novel approach is presented that enables a direct deployment of the trained policy on real robots using a new powerful simulator capable of domain randomization and a tailored reward scheme fine-tuned on images collected from real-world environments.
Abstract: Visual navigation is essential for many applications in robotics, from manipulation, through mobile robotics to automated driving. Deep reinforcement learning (DRL) provides an elegant map-free approach integrating image processing, localization, and planning in one module, which can be trained and therefore optimized for a given environment. However, to date, DRL-based visual navigation was validated exclusively in simulation, where the simulator provides information that is not available in the real world, e.g., the robot's position or image segmentation masks. This precludes the use of the learned policy on a real robot. Therefore, we propose a novel approach that enables a direct deployment of the trained policy on real robots. We have designed visual auxiliary tasks, a tailored reward scheme, and a new powerful simulator to facilitate domain randomization. The policy is fine-tuned on images collected from real-world environments. We have evaluated the method on a mobile robot in a real office environment. The training took ~30 hours on a single GPU. In 30 navigation experiments, the robot reached a 0.3-meter neighborhood of the goal in more than 86.7% of cases. This result makes the proposed method directly applicable to tasks like mobile manipulation.

13 citations


Proceedings ArticleDOI
TL;DR: This work proposes a multi-objective symbolic regression approach that is driven by both the training data and the prior knowledge of the properties the desired model should manifest, and outperforms standard symbolic regression by several orders of magnitude in terms of the mean squared deviation from a reference model.
Abstract: In symbolic regression, the search for analytic models is typically driven purely by the prediction error observed on the training data samples. However, when the data samples do not sufficiently cover the input space, the prediction error does not provide sufficient guidance toward desired models. Standard symbolic regression techniques then yield models that are partially incorrect, for instance, in terms of their steady-state characteristics or local behavior. If these properties were considered already during the search process, more accurate and relevant models could be produced. We propose a multi-objective symbolic regression approach that is driven by both the training data and the prior knowledge of the properties the desired model should manifest. The properties given in the form of formal constraints are internally represented by a set of discrete data samples on which candidate models are exactly checked. The proposed approach was experimentally evaluated on three test problems with results clearly demonstrating its capability to evolve realistic models that fit the training data well while complying with the prior knowledge of the desired model characteristics at the same time. It outperforms standard symbolic regression by several orders of magnitude in terms of the mean squared deviation from a reference model.

10 citations


Proceedings ArticleDOI
01 May 2020
TL;DR: This work proposes a generic and computationally efficient optimization method which is based on constraint programming that takes into account the kinematics of the robots and guarantees that the motions of the Robots are collision-free while minimizing the overall makespan.
Abstract: The coordination of multiple robots operating simultaneously in the same workspace requires the integration of task allocation and motion scheduling. We focus on tasks in which the robot’s actions are not confined to small volumes, but can also occupy a large time-varying portion of the workspace, such as in welding along a line. The optimization of such tasks presents a considerable challenge mainly due to the fact that different variants of task execution exist, for instance, there can be multiple starting points of lines or closed curves, differentfilling patterns of areas, etc. We propose a generic and computationally efficient optimization method which is based on constraint programming. It takes into account the kinematics of the robots and guarantees that the motions of the robots are collision-free while minimizing the overall makespan. We evaluate our approach on several use-cases of varying complexity: cutting, additive manufacturing, spot welding, inserting and tightening bolts, performed by a dual-arm robot. In terms of the makespan, the result is superior to task execution by one robot arm as well as by two arms not working simultaneously.

9 citations


Journal ArticleDOI
07 Jul 2020
TL;DR: An object-based pose graph is proposed to solve the problem of mapping in indoor dynamic environments with mobile robots and a new method to capture the probability of different objects over time is proposed.
Abstract: Relying on static representations of the environment limits the use of mapping methods in most real-world tasks. Real-world environments are dynamic and undergo changes that need to be handled through map adaptation. In this work, an object-based pose graph is proposed to solve the problem of mapping in indoor dynamic environments with mobile robots. In contrast to state-of-the art methods where binary classifications between movable and static objects are used, we propose a new method to capture the probability of different objects over time. Object probability represents how likely it is to find a specific object in its previous location and it gives a quantification of how movable specific objects are. In addition, grouping object probabilities according to object class allows us to evaluate the movability of different object classes. We validate our object-based pose graph in real-world dynamic environments. Results in mapping and map adaptation with a real robot show efficient map maintenance through several mapping sessions and results in object classification according to movability show an improvement compared to binary classification.

6 citations


Posted Content
TL;DR: A novel model-based agent that learns a latent Koopman representation from images that allows DeepKoCo to plan efficiently using linear control methods, such as linear model predictive control, making the proposed agent more amenable for real-life applications.
Abstract: This paper presents DeepKoCo, a novel model-based agent that learns a latent Koopman representation from images. This representation allows DeepKoCo to plan efficiently using linear control methods, such as linear model predictive control. Compared to traditional agents, DeepKoCo is invariant to task-irrelevant dynamics, thanks to the use of a tailored lossy autoencoder network that allows DeepKoCo to learn latent dynamics that reconstruct and predict only observed costs, rather than all observed dynamics. As our results show, DeepKoCo achieves a similar final performance as traditional model-free methods on complex control tasks, while being considerably more robust to distractor dynamics, making the proposed agent more amenable for real-life applications.

5 citations


Proceedings ArticleDOI
24 Oct 2020
TL;DR: In this article, the authors address the problem of finding objects in partially known indoor environments using the knowledge of the floor plan and the mapped objects, and consider object-object and object-room co-occurrences as prior information for identifying promising locations where an unmapped object can be present.
Abstract: The ability to search for objects is a precondition for various robotic tasks. In this paper, we address the problem of finding objects in partially known indoor environments. Using the knowledge of the floor plan and the mapped objects, we consider object-object and object-room co-occurrences as prior information for identifying promising locations where an unmapped object can be present. We propose an efficient search strategy that determines the best pose of the robot based on the analysis of the candidate locations. We optimize the probability of finding the target object and the distance travelled through a cost function.To evaluate our method, several experiments in simulated and real-world environments were performed. The results show that the robot successfully finds the target object in the environment while covering only a small portion of the search space. The real-world experiments with the TurtleBot 2 mobile robot validate the proposed approach and demonstrate that the method performs well also in real environments.

3 citations


Proceedings ArticleDOI
01 Dec 2020
TL;DR: In this article, a bio-inspired adaptive feedback mechanism that uses a forward model and an inverse model that can be learned on-the-fly is presented for directed locomotion in modular robots with controllers based on central pattern generators.
Abstract: This study is motivated by evolutionary robot systems where robot bodies and brains evolve simultaneously. In such systems robot ‘birth’ must be followed by ‘infant learning’ by a learning method that works for various morphologies evolution may produce. Here we address the task of directed locomotion in modular robots with controllers based on Central Pattern Generators. We present a bio-inspired adaptive feedback mechanism that uses a forward model and an inverse model that can be learned on-the-fly. We compare two versions (a simple and a sophisticated one) of this concept to a traditional (open-loop) controller using Bayesian optimization as a learning algorithm. The experimental results show that the sophisticated version outperforms the simple one and the traditional controller. It leads to a better performance and more robust controllers that better cope with noise.

2 citations


Journal ArticleDOI
TL;DR: A gradient-based initial training phase is used to quickly learn both a state representation and an initial policy, followed by a gradient-free optimization of only the final action selection parameters that enables the policy to improve in a stable manner to a performance level not obtained by gradient- based optimization alone.

2 citations


Posted Content
TL;DR: DeepKoCo as mentioned in this paper uses a tailored lossy autoencoder network to learn a latent Koopman representation from images, which allows DeepKoCo to plan efficiently using linear control methods, such as linear model predictive control.
Abstract: This paper presents DeepKoCo, a novel model-based agent that learns a latent Koopman representation from images. This representation allows DeepKoCo to plan efficiently using linear control methods, such as linear model predictive control. Compared to traditional agents, DeepKoCo learns task-relevant dynamics, thanks to the use of a tailored lossy autoencoder network that allows DeepKoCo to learn latent dynamics that reconstruct and predict only observed costs, rather than all observed dynamics. As our results show, DeepKoCo achieves similar final performance as traditional model-free methods on complex control tasks while being considerably more robust to distractor dynamics, making the proposed agent more amenable for real-life applications.

Book ChapterDOI
09 Nov 2020
TL;DR: This work develops a data-driven approach using tactile data that computes the object pose in a self-supervised manner after the object-finger contact is established and shows that this approach can estimate object poses with at least 2 cm translational and 20 degrees rotational accuracy despite changed object properties and unsuccessful grasps.
Abstract: We consider the problem of estimating an object’s pose in the absence of visual feedback after contact with robotic fingers during grasping has been made Information about the object’s pose facilitates precise placement of the object after a successful grasp If the grasp fails, then knowing the pose of the object after the grasping attempt is made can also help re-grasp the object We develop a data-driven approach using tactile data that computes the object pose in a self-supervised manner after the object-finger contact is established Additionally, we evaluate the effects of various feature representations, machine learning algorithms, and object properties on the pose estimation accuracy Unlike other existing approaches, our method does not require any prior knowledge about the object and does not make any assumptions about grasp stability In experiments, we show that our approach can estimate object poses with at least 2 cm translational and \(20^{\circ }\) rotational accuracy despite changed object properties and unsuccessful grasps