Matching Items (6)
Filtering by

Clear all filters

156573-Thumbnail Image.png
Description
In this thesis, different H∞ observers for time-delay systems are implemented and

their performances are compared. Equations that can be used to calculate observer gains are mentioned. Different methods that can be used to implement observers for time-delay systems are illustrated. Various stable and unstable systems are used and H∞ bounds

In this thesis, different H∞ observers for time-delay systems are implemented and

their performances are compared. Equations that can be used to calculate observer gains are mentioned. Different methods that can be used to implement observers for time-delay systems are illustrated. Various stable and unstable systems are used and H∞ bounds are calculated using these observer designing methods. Delays are assumed to be known constants for all systems. H∞ gains are calculated numerically using disturbance signals and performances of observers are compared.

The primary goal of this thesis is to implement the observer for Time Delay Systems designed using SOS and compare its performance with existing H∞ optimal observers. These observers are more general than other observers for time-delay systems as they make corrections to the delayed state as well along with the present state. The observer dynamics can be represented by an ODE coupled with a PDE. Results shown in this thesis show that this type of observers performs better than other H∞ observers. Sub-optimal observer-based state feedback system is also generated and simulated using the SOS observer. The simulation results show that the closed loop system converges very quickly, and the observer can be used to design full state-feedback closed loop system.
ContributorsTalati, Rushabh Vikram (Author) / Peet, Matthew (Thesis advisor) / Berman, Spring (Committee member) / Rivera, Daniel (Committee member) / Arizona State University (Publisher)
Created2018
154785-Thumbnail Image.png
Description
A computational framework based on convex optimization is presented for stability analysis of systems described by Partial Differential Equations (PDEs). Specifically, two forms of linear PDEs with spatially distributed polynomial coefficients are considered.

The first class includes linear coupled PDEs with one spatial variable. Parabolic, elliptic or hyperbolic PDEs with

A computational framework based on convex optimization is presented for stability analysis of systems described by Partial Differential Equations (PDEs). Specifically, two forms of linear PDEs with spatially distributed polynomial coefficients are considered.

The first class includes linear coupled PDEs with one spatial variable. Parabolic, elliptic or hyperbolic PDEs with Dirichlet, Neumann, Robin or mixed boundary conditions can be reformulated in order to be used by the framework. As an example, the reformulation is presented for systems governed by Schr¨odinger equation, parabolic type, relativistic heat conduction PDE and acoustic wave equation, hyperbolic types. The second form of PDEs of interest are scalar-valued with two spatial variables. An extra spatial variable allows consideration of problems such as local stability of fluid flows in channels and dynamics of population over two dimensional domains.

The approach does not involve discretization and is based on using Sum-of-Squares (SOS) polynomials and positive semi-definite matrices to parameterize operators which are positive on function spaces. Applying the parameterization to construct Lyapunov functionals with negative derivatives allows to express stability conditions as a set of LinearMatrix Inequalities (LMIs). The MATLAB package SOSTOOLS was used to construct the LMIs. The resultant LMIs then can be solved using existent Semi-Definite Programming (SDP) solvers such as SeDuMi or MOSEK. Moreover, the proposed approach allows to calculate bounds on the rate of decay of the solution norm.

The methodology is tested using several numerical examples and compared with the results obtained from simulation using standard methods of numerical discretization and analytic solutions.
ContributorsMeyer, Evgeny (Author) / Peet, Matthew (Thesis advisor) / Berman, Spring (Committee member) / Rivera, Daniel (Committee member) / Arizona State University (Publisher)
Created2016
154349-Thumbnail Image.png
Description
In this thesis, we focus on some of the NP-hard problems in control theory. Thanks to the converse Lyapunov theory, these problems can often be modeled as optimization over polynomials. To avoid the problem of intractability, we establish a trade off between accuracy and complexity. In particular, we develop a

In this thesis, we focus on some of the NP-hard problems in control theory. Thanks to the converse Lyapunov theory, these problems can often be modeled as optimization over polynomials. To avoid the problem of intractability, we establish a trade off between accuracy and complexity. In particular, we develop a sequence of tractable optimization problems - in the form of Linear Programs (LPs) and/or Semi-Definite Programs (SDPs) - whose solutions converge to the exact solution of the NP-hard problem. However, the computational and memory complexity of these LPs and SDPs grow exponentially with the progress of the sequence - meaning that improving the accuracy of the solutions requires solving SDPs with tens of thousands of decision variables and constraints. Setting up and solving such problems is a significant challenge. The existing optimization algorithms and software are only designed to use desktop computers or small cluster computers - machines which do not have sufficient memory for solving such large SDPs. Moreover, the speed-up of these algorithms does not scale beyond dozens of processors. This in fact is the reason we seek parallel algorithms for setting-up and solving large SDPs on large cluster- and/or super-computers.

We propose parallel algorithms for stability analysis of two classes of systems: 1) Linear systems with a large number of uncertain parameters; 2) Nonlinear systems defined by polynomial vector fields. First, we develop a distributed parallel algorithm which applies Polya's and/or Handelman's theorems to some variants of parameter-dependent Lyapunov inequalities with parameters defined over the standard simplex. The result is a sequence of SDPs which possess a block-diagonal structure. We then develop a parallel SDP solver which exploits this structure in order to map the computation, memory and communication to a distributed parallel environment. Numerical tests on a supercomputer demonstrate the ability of the algorithm to efficiently utilize hundreds and potentially thousands of processors, and analyze systems with 100+ dimensional state-space. Furthermore, we extend our algorithms to analyze robust stability over more complicated geometries such as hypercubes and arbitrary convex polytopes. Our algorithms can be readily extended to address a wide variety of problems in control such as Hinfinity synthesis for systems with parametric uncertainty and computing control Lyapunov functions.
ContributorsKamyar, Reza (Author) / Peet, Matthew (Thesis advisor) / Berman, Spring (Committee member) / Rivera, Daniel (Committee member) / Artemiadis, Panagiotis (Committee member) / Fainekos, Georgios (Committee member) / Arizona State University (Publisher)
Created2016
157578-Thumbnail Image.png
Description
Numerous works have addressed the control of multi-robot systems for coverage, mapping, navigation, and task allocation problems. In addition to classical microscopic approaches to multi-robot problems, which model the actions and decisions of individual robots, lately, there has been a focus on macroscopic or Eulerian approaches. In these approaches, the

Numerous works have addressed the control of multi-robot systems for coverage, mapping, navigation, and task allocation problems. In addition to classical microscopic approaches to multi-robot problems, which model the actions and decisions of individual robots, lately, there has been a focus on macroscopic or Eulerian approaches. In these approaches, the population of robots is represented as a continuum that evolves according to a mean-field model, which is directly designed such that the corresponding robot control policies produce target collective behaviours.



This dissertation presents a control-theoretic analysis of three types of mean-field models proposed in the literature for modelling and control of large-scale multi-agent systems, including robotic swarms. These mean-field models are Kolmogorov forward equations of stochastic processes, and their analysis is motivated by the fact that as the number of agents tends to infinity, the empirical measure associated with the agents converges to the solution of these models. Hence, the problem of transporting a swarm of agents from one distribution to another can be posed as a control problem for the forward equation of the process that determines the time evolution of the swarm density.



First, this thesis considers the case in which the agents' states evolve on a finite state space according to a continuous-time Markov chain (CTMC), and the forward equation is an ordinary differential equation (ODE). Defining the agents' task transition rates as the control parameters, the finite-time controllability, asymptotic controllability, and stabilization of the forward equation are investigated. Second, the controllability and stabilization problem for systems of advection-diffusion-reaction partial differential equations (PDEs) is studied in the case where the control parameters include the agents' velocity as well as transition rates. Third, this thesis considers a controllability and optimal control problem for the forward equation in the more general case where the agent dynamics are given by a nonlinear discrete-time control system. Beyond these theoretical results, this thesis also considers numerical optimal transport for control-affine systems. It is shown that finite-volume approximations of the associated PDEs lead to well-posed transport problems on graphs as long as the control system is controllable everywhere.
ContributorsElamvazhuthi, Karthik (Author) / Berman, Spring Melody (Thesis advisor) / Kawski, Matthias (Committee member) / Kuiper, Hendrik (Committee member) / Mignolet, Marc (Committee member) / Peet, Matthew (Committee member) / Arizona State University (Publisher)
Created2019
158221-Thumbnail Image.png
Description
The problem of modeling and controlling the distribution of a multi-agent system has recently evolved into an interdisciplinary effort. When the agent population is very large, i.e., at least on the order of hundreds of agents, it is important that techniques for analyzing and controlling the system scale well with

The problem of modeling and controlling the distribution of a multi-agent system has recently evolved into an interdisciplinary effort. When the agent population is very large, i.e., at least on the order of hundreds of agents, it is important that techniques for analyzing and controlling the system scale well with the number of agents. One scalable approach to characterizing the behavior of a multi-agent system is possible when the agents' states evolve over time according to a Markov process. In this case, the density of agents over space and time is governed by a set of difference or differential equations known as a {\it mean-field model}, whose parameters determine the stochastic control policies of the individual agents. These models often have the advantage of being easier to analyze than the individual agent dynamics. Mean-field models have been used to describe the behavior of chemical reaction networks, biological collectives such as social insect colonies, and more recently, swarms of robots that, like natural swarms, consist of hundreds or thousands of agents that are individually limited in capability but can coordinate to achieve a particular collective goal.

This dissertation presents a control-theoretic analysis of mean-field models for which the agent dynamics are governed by either a continuous-time Markov chain on an arbitrary state space, or a discrete-time Markov chain on a continuous state space. Three main problems are investigated. First, the problem of stabilization is addressed, that is, the design of transition probabilities/rates of the Markov process (the agent control parameters) that make a target distribution, satisfying certain conditions, invariant. Such a control approach could be used to achieve desired multi-agent distributions for spatial coverage and task allocation. However, the convergence of the multi-agent distribution to the designed equilibrium does not imply the convergence of the individual agents to fixed states. To prevent the agents from continuing to transition between states once the target distribution is reached, and thus potentially waste energy, the second problem addressed within this dissertation is the construction of feedback control laws that prevent agents from transitioning once the equilibrium distribution is reached. The third problem addressed is the computation of optimized transition probabilities/rates that maximize the speed at which the system converges to the target distribution.
ContributorsBiswal, Shiba (Author) / Berman, Spring (Thesis advisor) / Fainekos, Georgios (Committee member) / Lanchier, Nicolas (Committee member) / Mignolet, Marc (Committee member) / Peet, Matthew (Committee member) / Arizona State University (Publisher)
Created2020
193022-Thumbnail Image.png
Description
When solving analysis, estimation, and control problems for Partial Differential Equations (PDEs) via computational methods, one must resolve three main challenges: (a) the lack of a universal parametric representation of PDEs; (b) handling unbounded differential operators that appear as parameters; and (c), enforcing auxiliary constraints such as Boundary conditions and

When solving analysis, estimation, and control problems for Partial Differential Equations (PDEs) via computational methods, one must resolve three main challenges: (a) the lack of a universal parametric representation of PDEs; (b) handling unbounded differential operators that appear as parameters; and (c), enforcing auxiliary constraints such as Boundary conditions and continuity conditions. To address these challenges, an alternative representation of PDEs called the `Partial Integral Equation' (PIE) representation is proposed in this work. Primarily, the PIE representation alleviates the problem of the lack of a universal parametrization of PDEs since PIEs have, at most, $12$ Partial Integral (PI) operators as parameters. Naturally, this also resolves the challenges in handling unbounded operators because PI operators are bounded linear operators. Furthermore, for admissible PDEs, the PIE representation is unique and has no auxiliary constraints --- resolving the last of the $3$ main challenges. The PIE representation for a PDE is obtained by finding a unique unitary map from the states of the PIE to the states of the PDE. This map shows a PDE and its associated PIE have equivalent system properties, including well-posedness, internal stability, and I/O behavior. Furthermore, this unique map also allows us to construct a well-defined dual representation that can be used to solve optimal control problems for a PDE. Using the equivalent PIE representation of a PDE, mathematical and computational tools are developed to solve standard problems in Control theory for PDEs. In particular, problems such as a test for internal stability, Input-to-Output (I/O) $L_2$-gain, $\hinf$-optimal state observer design, and $\hinf$-optimal full state-feedback controller design are solved using convex-optimization and Lyapunov methods for linear PDEs in one spatial dimension. Once the PIE associated with a PDE is obtained, Lyapunov functions (or storage functions) are parametrized by positive PI operators to obtain a solvable convex formulation of the above-stated control problems. Lastly, the methods proposed here are applied to various PDE systems to demonstrate the application.
ContributorsShivakumar, Sachin (Author) / Peet, Matthew (Thesis advisor) / Nedich, Angelia (Committee member) / Marvi, Hamidreza (Committee member) / Platte, Rodrigo (Committee member) / Berman, Spring (Committee member) / Arizona State University (Publisher)
Created2024