Karl Henrik Johansson
Other affiliations: University of Groningen, Northeastern University, University of Sannio ...read more
Bio: Karl Henrik Johansson is an academic researcher from Royal Institute of Technology. The author has contributed to research in topics: Control theory & Wireless sensor network. The author has an hindex of 88, co-authored 1089 publications receiving 33751 citations. Previous affiliations of Karl Henrik Johansson include University of Groningen & Northeastern University.
Papers published on a yearly basis
TL;DR: The controller updates considered here are event-driven, depending on the ratio of a certain measurement error with respect to the norm of a function of the state, and are applied to a first order agreement problem.
Abstract: Event-driven strategies for multi-agent systems are motivated by the future use of embedded microprocessors with limited resources that will gather information and actuate the individual agent controller updates. The controller updates considered here are event-driven, depending on the ratio of a certain measurement error with respect to the norm of a function of the state, and are applied to a first order agreement problem. A centralized formulation is considered first and then its distributed counterpart, in which agents require knowledge only of their neighbors' states for the controller implementation. The results are then extended to a self-triggered setup, where each agent computes its next update time at the previous one, without having to keep track of the state error that triggers the actuation between two consecutive update instants. The results are illustrated through simulation examples.
01 Dec 2012
TL;DR: An introduction to event- and self-triggered control systems where sensing and actuation is performed when needed and how these control strategies can be implemented using existing wireless communication technology is shown.
Abstract: Recent developments in computer and communication technologies have led to a new type of large-scale resource-constrained wireless embedded control systems. It is desirable in these systems to limit the sensor and control computation and/or communication to instances when the system needs attention. However, classical sampled-data control is based on performing sensing and actuation periodically rather than when the system needs attention. This paper provides an introduction to event- and self-triggered control systems where sensing and actuation is performed when needed. Event-triggered control is reactive and generates sensor sampling and control actuation when, for instance, the plant state deviates more than a certain threshold from a desired value. Self-triggered control, on the other hand, is proactive and computes the next sampling or actuation instance ahead of time. The basics of these control strategies are introduced together with a discussion on the differences between state feedback and output feedback for event-triggered control. It is also shown how event- and self-triggered control can be implemented using existing wireless communication technology. Some applications to wireless control in process industry are discussed as well.
TL;DR: A novel control strategy for multi-agent coordination with event-based broadcasting is presented, in which each agent decides itself when to transmit its current state to its neighbors and the local control laws are based on these sampled state measurements.
Abstract: A novel control strategy for multi-agent coordination with event-based broadcasting is presented. In particular, each agent decides itself when to transmit its current state to its neighbors and the local control laws are based on these sampled state measurements. Three scenarios are analyzed: Networks of single-integrator agents with and without communication delays, and networks of double-integrator agents. The novel event-based scheduling strategy bounds each agent's measurement error by a time-dependent threshold. For each scenario it is shown that the proposed control strategy guarantees either asymptotic convergence to average consensus or convergence to a ball centered at the average consensus. Moreover, it is shown that the inter-event intervals are lower-bounded by a positive constant. Numerical simulations show the effectiveness of the novel event-based control strategy and how it compares to time-scheduled control.
TL;DR: The quadruple-tank process is ideal for illustrating many concepts in multivariable control, particularly performance limitations due toMultivariable right half-plane zeros, which have an appealing physical interpretation.
Abstract: A multivariable laboratory process that consists of four interconnected water tanks is presented. The linearized dynamics of the system have a multivariable zero that is possible to move along the real axis by changing a valve. The zero can be placed in both the left and the right half-plane. In this way the quadruple-tank process is ideal for illustrating many concepts in multivariable control, particularly performance limitations due to multivariable right half-plane zeros. The location and the direction of the zero have an appealing physical interpretation. Accurate models are derived from both physical and experimental data and decentralized control is demonstrated on the process.
TL;DR: Lyapunov's theorem on stability via linearization and LaSalle's invariance principle are generalized to hybrid automata and a class of hybrids whose solutions depend continuously on the initial state is characterized.
Abstract: Hybrid automata provide a language for modeling and analyzing digital and analogue computations in real-time systems. Hybrid automata are studied here from a dynamical systems perspective. Necessary and sufficient conditions for existence and uniqueness of solutions are derived and a class of hybrid automata whose solutions depend continuously on the initial state is characterized. The results on existence, uniqueness, and continuity serve as a starting point for stability analysis. Lyapunov's theorem on stability via linearization and LaSalle's invariance principle are generalized to hybrid automata.
TL;DR: There is, I think, something ethereal about i —the square root of minus one, which seems an odd beast at that time—an intruder hovering on the edge of reality.
Abstract: There is, I think, something ethereal about i —the square root of minus one. I remember first hearing about it at school. It seemed an odd beast at that time—an intruder hovering on the edge of reality. Usually familiarity dulls this sense of the bizarre, but in the case of i it was the reverse: over the years the sense of its surreal nature intensified. It seemed that it was impossible to write mathematics that described the real world in …
TL;DR: Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis.
Abstract: Machine Learning is the study of methods for programming computers to learn. Computers are applied to a wide range of tasks, and for most of these it is relatively easy for programmers to design and implement the necessary software. However, there are many tasks for which this is difficult or impossible. These can be divided into four general categories. First, there are problems for which there exist no human experts. For example, in modern automated manufacturing facilities, there is a need to predict machine failures before they occur by analyzing sensor readings. Because the machines are new, there are no human experts who can be interviewed by a programmer to provide the knowledge necessary to build a computer system. A machine learning system can study recorded data and subsequent machine failures and learn prediction rules. Second, there are problems where human experts exist, but where they are unable to explain their expertise. This is the case in many perceptual tasks, such as speech recognition, hand-writing recognition, and natural language understanding. Virtually all humans exhibit expert-level abilities on these tasks, but none of them can describe the detailed steps that they follow as they perform them. Fortunately, humans can provide machines with examples of the inputs and correct outputs for these tasks, so machine learning algorithms can learn to map the inputs to the outputs. Third, there are problems where phenomena are changing rapidly. In finance, for example, people would like to predict the future behavior of the stock market, of consumer purchases, or of exchange rates. These behaviors change frequently, so that even if a programmer could construct a good predictive computer program, it would need to be rewritten frequently. A learning program can relieve the programmer of this burden by constantly modifying and tuning a set of learned prediction rules. Fourth, there are applications that need to be customized for each computer user separately. Consider, for example, a program to filter unwanted electronic mail messages. Different users will need different filters. It is unreasonable to expect each user to program his or her own rules, and it is infeasible to provide every user with a software engineer to keep the rules up-to-date. A machine learning system can learn which mail messages the user rejects and maintain the filtering rules automatically. Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis. Statistics focuses on understanding the phenomena that have generated the data, often with the goal of testing different hypotheses about those phenomena. Data mining seeks to find patterns in the data that are understandable by people. Psychological studies of human learning aspire to understand the mechanisms underlying the various learning behaviors exhibited by people (concept learning, skill acquisition, strategy change, etc.).
01 Jan 2006
TL;DR: Probability distributions of linear models for regression and classification are given in this article, along with a discussion of combining models and combining models in the context of machine learning and classification.
Abstract: Probability Distributions.- Linear Models for Regression.- Linear Models for Classification.- Neural Networks.- Kernel Methods.- Sparse Kernel Machines.- Graphical Models.- Mixture Models and EM.- Approximate Inference.- Sampling Methods.- Continuous Latent Variables.- Sequential Data.- Combining Models.
01 Jan 2009
22 Jan 2006
TL;DR: Some of the major results in random graphs and some of the more challenging open problems are reviewed, including those related to the WWW.
Abstract: We will review some of the major results in random graphs and some of the more challenging open problems. We will cover algorithmic and structural questions. We will touch on newer models, including those related to the WWW.