55 research outputs found
Observer design for piecewise smooth and switched systems via contraction theory
The aim of this paper is to present the application of an approach to study
contraction theory recently developed for piecewise smooth and switched
systems. The approach that can be used to analyze incremental stability
properties of so-called Filippov systems (or variable structure systems) is
based on the use of regularization, a procedure to make the vector field of
interest differentiable before analyzing its properties. We show that by using
this extension of contraction theory to nondifferentiable vector fields, it is
possible to design observers for a large class of piecewise smooth systems
using not only Euclidean norms, as also done in previous literature, but also
non-Euclidean norms. This allows greater flexibility in the design and
encompasses the case of both piecewise-linear and piecewise-smooth (nonlinear)
systems. The theoretical methodology is illustrated via a set of representative
examples.Comment: Preprint accepted to IFAC World Congress 201
Data-driven design of complex network structures to promote synchronization
We consider the problem of optimizing the interconnection graphs of complex
networks to promote synchronization. When traditional optimization methods are
inapplicable, due to uncertain or unknown node dynamics, we propose a
data-driven approach leveraging datasets of relevant examples. We analyze two
case studies, with linear and nonlinear node dynamics. First, we show how
including node dynamics in the objective function makes the optimal graphs
heterogeneous. Then, we compare various design strategies, finding that the
best either utilize data samples close to a specific Pareto front or a
combination of a neural network and a genetic algorithm, with statistically
better performance than the best examples in the datasets
JAQ of All Trades: Job Mismatch, Firm Productivity and Managerial Quality
We develop a novel measure of job-worker allocation quality (JAQ) by ex-
ploiting employer-employee data with machine learning techniques. Based on
our measure, the quality of job-worker matching correlates positively with
individual labor earnings and firm productivity, as well as with market com-
petition, non-family firm status, and employees’ human capital. Management
plays a key role in job-worker matching: when managerial hirings and firings
persistently raise management quality, the matching of rank-and-file workers
to their jobs improves. JAQ can be constructed from any employer-employee
data set including workers’ occupations, and used to explore research ques-
tions in corporate finance and organization economics
Learning-based cognitive architecture for enhancing coordination in human groups
As interactions with autonomous agents-ranging from robots in physical
settings to avatars in virtual and augmented realities-become more prevalent,
developing advanced cognitive architectures is critical for enhancing the
dynamics of human-avatar groups. This paper presents a
reinforcement-learning-based cognitive architecture, trained via a sim-to-real
approach, designed to improve synchronization in periodic motor tasks, crucial
for applications in group rehabilitation and sports training. Extensive
numerical validation consistently demonstrates improvements in synchronization.
Theoretical derivations and numerical investigations are complemented by
preliminary experiments with real participants, showing that our avatars can
integrate seamlessly into human groups, often being indistinguishable from
humans
Control-Tutored Reinforcement Learning: Towards the Integration of Data-Driven and Model-Based Control
We present an architecture where a feedback controller derived on an approximate model of the environment assists the learning process to enhance its data efficiency. This architecture, which we term as Control-Tutored Q-learning (CTQL), is presented in two alternative flavours. The former is based on defining the reward function so that a Boolean condition can be used to determine when the control tutor policy is adopted, while the latter, termed as probabilistic CTQL (pCTQL), is instead based on executing calls to the tutor with a certain probability during learning. Both approaches are validated, and thoroughly benchmarked against Q-Learning, by considering the stabilization of an inverted pendulum as defined in OpenAI Gym as a representative problem
CT-DQN: Control-Tutored Deep Reinforcement Learning
One of the major challenges in Deep Reinforcement Learning for control is the need for extensive training to learn the policy. Motivated by this, we present the design of the Control-Tutored Deep Q-Networks (CT-DQN) algorithm, a Deep Reinforcement Learning algorithm that leverages a control tutor, i.e., an exogenous control law, to reduce learning time. The tutor can be designed using an approximate model of the system, without any assumption about the knowledge of the system’s dynamics. There is no expectation that it will be able to achieve the control objective if used stand-alone. During learning, the tutor occasionally suggests an action, thus partially guiding exploration. We validate our approach on three scenarios from OpenAI Gym: the inverted pendulum, lunar lander, and car racing. We demonstrate that CT-DQN is able to achieve better or equivalent data efficiency with respect to the classic function approximation solutions
- …
