55 research outputs found

    Observer design for piecewise smooth and switched systems via contraction theory

    Full text link
    The aim of this paper is to present the application of an approach to study contraction theory recently developed for piecewise smooth and switched systems. The approach that can be used to analyze incremental stability properties of so-called Filippov systems (or variable structure systems) is based on the use of regularization, a procedure to make the vector field of interest differentiable before analyzing its properties. We show that by using this extension of contraction theory to nondifferentiable vector fields, it is possible to design observers for a large class of piecewise smooth systems using not only Euclidean norms, as also done in previous literature, but also non-Euclidean norms. This allows greater flexibility in the design and encompasses the case of both piecewise-linear and piecewise-smooth (nonlinear) systems. The theoretical methodology is illustrated via a set of representative examples.Comment: Preprint accepted to IFAC World Congress 201

    Data-driven design of complex network structures to promote synchronization

    Full text link
    We consider the problem of optimizing the interconnection graphs of complex networks to promote synchronization. When traditional optimization methods are inapplicable, due to uncertain or unknown node dynamics, we propose a data-driven approach leveraging datasets of relevant examples. We analyze two case studies, with linear and nonlinear node dynamics. First, we show how including node dynamics in the objective function makes the optimal graphs heterogeneous. Then, we compare various design strategies, finding that the best either utilize data samples close to a specific Pareto front or a combination of a neural network and a genetic algorithm, with statistically better performance than the best examples in the datasets

    JAQ of All Trades: Job Mismatch, Firm Productivity and Managerial Quality

    Get PDF
    We develop a novel measure of job-worker allocation quality (JAQ) by ex- ploiting employer-employee data with machine learning techniques. Based on our measure, the quality of job-worker matching correlates positively with individual labor earnings and firm productivity, as well as with market com- petition, non-family firm status, and employees’ human capital. Management plays a key role in job-worker matching: when managerial hirings and firings persistently raise management quality, the matching of rank-and-file workers to their jobs improves. JAQ can be constructed from any employer-employee data set including workers’ occupations, and used to explore research ques- tions in corporate finance and organization economics

    Learning-based cognitive architecture for enhancing coordination in human groups

    Full text link
    As interactions with autonomous agents-ranging from robots in physical settings to avatars in virtual and augmented realities-become more prevalent, developing advanced cognitive architectures is critical for enhancing the dynamics of human-avatar groups. This paper presents a reinforcement-learning-based cognitive architecture, trained via a sim-to-real approach, designed to improve synchronization in periodic motor tasks, crucial for applications in group rehabilitation and sports training. Extensive numerical validation consistently demonstrates improvements in synchronization. Theoretical derivations and numerical investigations are complemented by preliminary experiments with real participants, showing that our avatars can integrate seamlessly into human groups, often being indistinguishable from humans

    Control-Tutored Reinforcement Learning: Towards the Integration of Data-Driven and Model-Based Control

    Get PDF
    We present an architecture where a feedback controller derived on an approximate model of the environment assists the learning process to enhance its data efficiency. This architecture, which we term as Control-Tutored Q-learning (CTQL), is presented in two alternative flavours. The former is based on defining the reward function so that a Boolean condition can be used to determine when the control tutor policy is adopted, while the latter, termed as probabilistic CTQL (pCTQL), is instead based on executing calls to the tutor with a certain probability during learning. Both approaches are validated, and thoroughly benchmarked against Q-Learning, by considering the stabilization of an inverted pendulum as defined in OpenAI Gym as a representative problem

    CT-DQN: Control-Tutored Deep Reinforcement Learning

    Get PDF
    One of the major challenges in Deep Reinforcement Learning for control is the need for extensive training to learn the policy. Motivated by this, we present the design of the Control-Tutored Deep Q-Networks (CT-DQN) algorithm, a Deep Reinforcement Learning algorithm that leverages a control tutor, i.e., an exogenous control law, to reduce learning time. The tutor can be designed using an approximate model of the system, without any assumption about the knowledge of the system’s dynamics. There is no expectation that it will be able to achieve the control objective if used stand-alone. During learning, the tutor occasionally suggests an action, thus partially guiding exploration. We validate our approach on three scenarios from OpenAI Gym: the inverted pendulum, lunar lander, and car racing. We demonstrate that CT-DQN is able to achieve better or equivalent data efficiency with respect to the classic function approximation solutions
    corecore