Matching Items (4)
Filtering by

Clear all filters

187348-Thumbnail Image.png
Description
The introduction of assistive/autonomous features in cyber-physical systems, e.g., self-driving vehicles, have paved the way to a relatively new field of system analysis for safety-critical applications, along with the topic of controlling systems with performance and safety guarantees. The different works in this thesis explore and design methodologies that focus

The introduction of assistive/autonomous features in cyber-physical systems, e.g., self-driving vehicles, have paved the way to a relatively new field of system analysis for safety-critical applications, along with the topic of controlling systems with performance and safety guarantees. The different works in this thesis explore and design methodologies that focus on the analysis of nonlinear dynamical systems via set-membership approximations, as well as the development of controllers and estimators that can give worst-case performance guarantees, especially when the sensor data containing information on system outputs is prone to data drops and delays. For analyzing the distinguishability of nonlinear systems, building upon the idea of set membership over-approximation of the nonlinear systems, a novel optimization-based method for multi-model affine abstraction (i.e., simultaneous set-membership over-approximation of multiple models) is designed. This work solves for the existence of set-membership over-approximations of a pair of different nonlinear models such that the different systems can be distinguished/discriminated within a guaranteed detection time under worst-case uncertainties and approximation errors. Specifically, by combining mesh-based affine abstraction methods with T-distinguishability analysis in the literature yields a bilevel bilinear optimization problem, whereby leveraging robust optimization techniques and a suitable change of variables result in a sufficient linear program that can obtain a tractable solution with T-distinguishability guarantees. Moreover, the thesis studied the designs of controllers and estimators with performance guarantees, and specifically, path-dependent feedback controllers and bounded-error estimators for time-varying affine systems are proposed that are subject to delayed observations or missing data. To model the delayed/missing data, two approaches are explored; a fixed-length language and an automaton-based model. Furthermore, controllers/estimators that satisfy the equalized recovery property (a weaker form of invariance with time-varying finite bounds) are synthesized whose feedback gains can be adapted based on the observed path, i.e., the history of observed data patterns up to the latest available time step. Finally, a robust kinodynamic motion planning algorithm is also developed with collision avoidance and probabilistic completeness guarantees. In particular, methods based on fixed and flexible invariant tubes are designed such that the planned motion/trajectories can reject bounded disturbances using noisy observations.
ContributorsHassaan, Syed Muhammad (Author) / Yong, Sze Zheng (Thesis advisor) / Rivera, Daniel (Committee member) / Marvi, Hamidreza (Committee member) / Lee, Hyunglae (Committee member) / Berman, Spring (Committee member) / Arizona State University (Publisher)
Created2023
158221-Thumbnail Image.png
Description
The problem of modeling and controlling the distribution of a multi-agent system has recently evolved into an interdisciplinary effort. When the agent population is very large, i.e., at least on the order of hundreds of agents, it is important that techniques for analyzing and controlling the system scale well with

The problem of modeling and controlling the distribution of a multi-agent system has recently evolved into an interdisciplinary effort. When the agent population is very large, i.e., at least on the order of hundreds of agents, it is important that techniques for analyzing and controlling the system scale well with the number of agents. One scalable approach to characterizing the behavior of a multi-agent system is possible when the agents' states evolve over time according to a Markov process. In this case, the density of agents over space and time is governed by a set of difference or differential equations known as a {\it mean-field model}, whose parameters determine the stochastic control policies of the individual agents. These models often have the advantage of being easier to analyze than the individual agent dynamics. Mean-field models have been used to describe the behavior of chemical reaction networks, biological collectives such as social insect colonies, and more recently, swarms of robots that, like natural swarms, consist of hundreds or thousands of agents that are individually limited in capability but can coordinate to achieve a particular collective goal.

This dissertation presents a control-theoretic analysis of mean-field models for which the agent dynamics are governed by either a continuous-time Markov chain on an arbitrary state space, or a discrete-time Markov chain on a continuous state space. Three main problems are investigated. First, the problem of stabilization is addressed, that is, the design of transition probabilities/rates of the Markov process (the agent control parameters) that make a target distribution, satisfying certain conditions, invariant. Such a control approach could be used to achieve desired multi-agent distributions for spatial coverage and task allocation. However, the convergence of the multi-agent distribution to the designed equilibrium does not imply the convergence of the individual agents to fixed states. To prevent the agents from continuing to transition between states once the target distribution is reached, and thus potentially waste energy, the second problem addressed within this dissertation is the construction of feedback control laws that prevent agents from transitioning once the equilibrium distribution is reached. The third problem addressed is the computation of optimized transition probabilities/rates that maximize the speed at which the system converges to the target distribution.
ContributorsBiswal, Shiba (Author) / Berman, Spring (Thesis advisor) / Fainekos, Georgios (Committee member) / Lanchier, Nicolas (Committee member) / Mignolet, Marc (Committee member) / Peet, Matthew (Committee member) / Arizona State University (Publisher)
Created2020
168490-Thumbnail Image.png
Description
Modern life is full of challenging optimization problems that we unknowingly attempt to solve. For instance, a common dilemma often encountered is the decision of picking a parking spot while trying to minimize both the distance to the goal destination and time spent searching for parking; one strategy is to

Modern life is full of challenging optimization problems that we unknowingly attempt to solve. For instance, a common dilemma often encountered is the decision of picking a parking spot while trying to minimize both the distance to the goal destination and time spent searching for parking; one strategy is to drive as close as possible to the goal destination but risk a penalty cost if no parking spaces can be found. Optimization problems of this class all have underlying time-varying processes that can be altered by a decision/input to minimize some cost. Such optimization problems are commonly solved by a class of methods called Dynamic Programming (DP) that breaks down a complex optimization problem into a simpler family of sub-problems. In the 1950s Richard Bellman introduced a class of DP methods that broke down Multi-Stage Optimization Problems (MSOP) into a nested sequence of ``tail problems”. Bellman showed that for any MSOP with a cost function that satisfies a condition called additive separability, the solution to the tail problem of the MSOP initialized at time-stage k>0 can be used to solve the tail problem initialized at time-stage k-1. Therefore, by recursively solving each tail problem of the MSOP, a solution to the original MSOP can be found. This dissertation extends Bellman`s theory to a broader class of MSOPs involving non-additively separable costs by introducing a new state augmentation solution method and generalizing the Bellman Equation. This dissertation also considers the analogous continuous-time counterpart to discrete-time MSOPs, called Optimal Control Problems (OCPs). OCPs can be solved by solving a nonlinear Partial Differential Equation (PDE) called the Hamilton-Jacobi-Bellman (HJB) PDE. Unfortunately, it is rarely possible to obtain an analytical solution to the HJB PDE. This dissertation proposes a method for approximately solving the HJB PDE based on Sum-Of-Squares (SOS) programming. This SOS algorithm can be used to synthesize controllers, hence solving the OCP, and also compute outer bounds of reachable sets of dynamical systems. This methodology is then extended to infinite time horizons, by proposing SOS algorithms that yield Lyapunov functions that can approximate regions of attraction and attractor sets of nonlinear dynamical systems arbitrarily well.
ContributorsJones, Morgan (Author) / Peet, Matthew M (Thesis advisor) / Nedich, Angelia (Committee member) / Kawski, Matthias (Committee member) / Mignolet, Marc (Committee member) / Berman, Spring (Committee member) / Arizona State University (Publisher)
Created2021
193022-Thumbnail Image.png
Description
When solving analysis, estimation, and control problems for Partial Differential Equations (PDEs) via computational methods, one must resolve three main challenges: (a) the lack of a universal parametric representation of PDEs; (b) handling unbounded differential operators that appear as parameters; and (c), enforcing auxiliary constraints such as Boundary conditions and

When solving analysis, estimation, and control problems for Partial Differential Equations (PDEs) via computational methods, one must resolve three main challenges: (a) the lack of a universal parametric representation of PDEs; (b) handling unbounded differential operators that appear as parameters; and (c), enforcing auxiliary constraints such as Boundary conditions and continuity conditions. To address these challenges, an alternative representation of PDEs called the `Partial Integral Equation' (PIE) representation is proposed in this work. Primarily, the PIE representation alleviates the problem of the lack of a universal parametrization of PDEs since PIEs have, at most, $12$ Partial Integral (PI) operators as parameters. Naturally, this also resolves the challenges in handling unbounded operators because PI operators are bounded linear operators. Furthermore, for admissible PDEs, the PIE representation is unique and has no auxiliary constraints --- resolving the last of the $3$ main challenges. The PIE representation for a PDE is obtained by finding a unique unitary map from the states of the PIE to the states of the PDE. This map shows a PDE and its associated PIE have equivalent system properties, including well-posedness, internal stability, and I/O behavior. Furthermore, this unique map also allows us to construct a well-defined dual representation that can be used to solve optimal control problems for a PDE. Using the equivalent PIE representation of a PDE, mathematical and computational tools are developed to solve standard problems in Control theory for PDEs. In particular, problems such as a test for internal stability, Input-to-Output (I/O) $L_2$-gain, $\hinf$-optimal state observer design, and $\hinf$-optimal full state-feedback controller design are solved using convex-optimization and Lyapunov methods for linear PDEs in one spatial dimension. Once the PIE associated with a PDE is obtained, Lyapunov functions (or storage functions) are parametrized by positive PI operators to obtain a solvable convex formulation of the above-stated control problems. Lastly, the methods proposed here are applied to various PDE systems to demonstrate the application.
ContributorsShivakumar, Sachin (Author) / Peet, Matthew (Thesis advisor) / Nedich, Angelia (Committee member) / Marvi, Hamidreza (Committee member) / Platte, Rodrigo (Committee member) / Berman, Spring (Committee member) / Arizona State University (Publisher)
Created2024