Matching Items (29)
Filtering by

Clear all filters

156044-Thumbnail Image.png
Description
In a collaborative environment where multiple robots and human beings are expected

to collaborate to perform a task, it becomes essential for a robot to be aware of multiple

agents working in its work environment. A robot must also learn to adapt to

different agents in the workspace and conduct its interaction based

In a collaborative environment where multiple robots and human beings are expected

to collaborate to perform a task, it becomes essential for a robot to be aware of multiple

agents working in its work environment. A robot must also learn to adapt to

different agents in the workspace and conduct its interaction based on the presence

of these agents. A theoretical framework was introduced which performs interaction

learning from demonstrations in a two-agent work environment, and it is called

Interaction Primitives.

This document is an in-depth description of the new state of the art Python

Framework for Interaction Primitives between two agents in a single as well as multiple

task work environment and extension of the original framework in a work environment

with multiple agents doing a single task. The original theory of Interaction

Primitives has been extended to create a framework which will capture correlation

between more than two agents while performing a single task. The new state of the

art Python framework is an intuitive, generic, easy to install and easy to use python

library which can be applied to use the Interaction Primitives framework in a work

environment. This library was tested in simulated environments and controlled laboratory

environment. The results and benchmarks of this library are available in the

related sections of this document.
ContributorsKumar, Ashish, M.S (Author) / Amor, Hani Ben (Thesis advisor) / Zhang, Yu (Committee member) / Yang, Yezhou (Committee member) / Arizona State University (Publisher)
Created2017
157457-Thumbnail Image.png
Description
The construction industry is very mundane and tiring for workers without the assistance of machines. This challenge has changed the trend of construction industry tremendously by motivating the development of robots that can replace human workers. This thesis presents a computed torque controller that is designed to produce movements by

The construction industry is very mundane and tiring for workers without the assistance of machines. This challenge has changed the trend of construction industry tremendously by motivating the development of robots that can replace human workers. This thesis presents a computed torque controller that is designed to produce movements by a small-scale, 5 degree-of-freedom (DOF) robotic arm that are useful for construction operations, specifically bricklaying. A software framework for the robotic arm with motion and path planning features and different control capabilities has also been developed using the Robot Operating System (ROS).

First, a literature review of bricklaying construction activity and existing robots’ performance is discussed. After describing an overview of the required robot structure, a mathematical model is presented for the 5-DOF robotic arm. A model-based computed torque controller is designed for the nonlinear dynamic robotic arm, taking into consideration the dynamic and kinematic properties of the arm. For sustainable growth of this technology so that it is affordable to the masses, it is important that the energy consumption by the robot is optimized. In this thesis, the trajectory of the robotic arm is optimized using sequential quadratic programming. The results of the energy optimization procedure are also analyzed for different possible trajectories.

A construction testbed setup is simulated in the ROS platform to validate the designed controllers and optimized robot trajectories on different experimental scenarios. A commercially available 5-DOF robotic arm is modeled in the ROS simulators Gazebo and Rviz. The path and motion planning is performed using the Moveit-ROS interface and also implemented on a physical small-scale robotic arm. A Matlab-ROS framework for execution of different controllers on the physical robot is described. Finally, the results of the controller simulation and experiments are discussed in detail.
ContributorsGandhi, Sushrut (Author) / Berman, Spring (Thesis advisor) / Marvi, Hamidreza (Committee member) / Yong, Sze Zheng (Committee member) / Arizona State University (Publisher)
Created2019
156496-Thumbnail Image.png
Description
Soft Poly-Limb (SPL) is a pneumatically driven, wearable, soft continuum robotic arm designed to aid humans with medical conditions, such as cerebral palsy, paraplegia, cervical spondylotic myelopathy, perform activities of daily living. To support user's tasks, the SPL acts as an additional limb extending from the human body which can

Soft Poly-Limb (SPL) is a pneumatically driven, wearable, soft continuum robotic arm designed to aid humans with medical conditions, such as cerebral palsy, paraplegia, cervical spondylotic myelopathy, perform activities of daily living. To support user's tasks, the SPL acts as an additional limb extending from the human body which can be controlled to perform safe and compliant mobile manipulation in three-dimensional space. The SPL is inspired by invertebrate limbs, such as the elephant trunk and the arms of the octopus. In this work, various geometrical and physical parameters of the SPL are identified, and behavior of the actuators that comprise it are studied by varying their parameters through novel quasi-static computational models. As a result, this study provides a set of engineering design rules to create soft actuators for continuum soft robotic arms by understanding how varying parameters affect the actuator's motion as a function of the input pressure. A prototype of the SPL is fabricated to analyze the accuracy of these computational models by performing linear expansion, bending and arbitrary pose tests. Furthermore, combinations of the parameters based on the application of the SPL are determined to affect the weight, payload capacity, and stiffness of the arm. Experimental results demonstrate the accuracy of the proposed computational models and help in understanding the behavior of soft compliant actuators. Finally, based on the set functional requirements for the assistance of impaired users, results show the effectiveness of the SPL in performing tasks for activities of daily living.
ContributorsNuthi, Sai Gautham (Author) / Polygerinos, Panagiotis (Thesis advisor) / Lee, Hyunglae (Committee member) / Yong, Sze Zheng (Committee member) / Arizona State University (Publisher)
Created2018
156560-Thumbnail Image.png
Description
This work considers the design of separating input signals in order to discriminate among a finite number of uncertain nonlinear models. Each nonlinear model corresponds to a system operating mode, unobserved intents of other drivers or robots, or to fault types or attack strategies, etc., and the separating inputs are

This work considers the design of separating input signals in order to discriminate among a finite number of uncertain nonlinear models. Each nonlinear model corresponds to a system operating mode, unobserved intents of other drivers or robots, or to fault types or attack strategies, etc., and the separating inputs are designed such that the output trajectories of all the nonlinear models are guaranteed to be distinguishable from each other under any realization of uncertainties in the initial condition, model discrepancies or noise. I propose a two-step approach. First, using an optimization-based approach, we over-approximate nonlinear dynamics by uncertain affine models, as abstractions that preserve all its system behaviors such that any discrimination guarantees for the affine abstraction also hold for the original nonlinear system. Then, I propose a novel solution in the form of a mixed-integer linear program (MILP) to the active model discrimination problem for uncertain affine models, which includes the affine abstraction and thus, the nonlinear models. Finally, I demonstrate the effectiveness of our approach for identifying the intention of other vehicles in a highway lane changing scenario. For the abstraction, I explore two approaches. In the first approach, I construct the bounding planes using a Mixed-Integer Nonlinear Problem (MINLP) formulation of the given system with appropriately designed constraints. For the second approach, I solve a linear programming (LP) problem that over-approximates the nonlinear function at only the grid points of a mesh with a given resolution and then accounting for the entire domain via an appropriate correction term. To achieve a desired approximation accuracy, we also iteratively subdivide the domain into subregions. This method applies to nonlinear functions with different degrees of smoothness, including Lipschitz continuous functions, and improves on existing approaches by enabling the use of tighter bounds. Finally, we compare the effectiveness of this approach with the existing optimization-based methods in simulation and illustrate its applicability for estimator design.
ContributorsSingh, Kanishka Raj (Author) / Yong, Sze Zheng (Thesis advisor) / Artemiadis, Panagiotis (Committee member) / Berman, Spring (Committee member) / Arizona State University (Publisher)
Created2018
156771-Thumbnail Image.png
Description
Reinforcement learning (RL) is a powerful methodology for teaching autonomous agents complex behaviors and skills. A critical component in most RL algorithms is the reward function -- a mathematical function that provides numerical estimates for desirable and undesirable states. Typically, the reward function must be hand-designed by a human expert

Reinforcement learning (RL) is a powerful methodology for teaching autonomous agents complex behaviors and skills. A critical component in most RL algorithms is the reward function -- a mathematical function that provides numerical estimates for desirable and undesirable states. Typically, the reward function must be hand-designed by a human expert and, as a result, the scope of a robot's autonomy and ability to safely explore and learn in new and unforeseen environments is constrained by the specifics of the designed reward function. In this thesis, I design and implement a stateful collision anticipation model with powerful predictive capability based upon my research of sequential data modeling and modern recurrent neural networks. I also develop deep reinforcement learning methods whose rewards are generated by self-supervised training and intrinsic signals. The main objective is to work towards the development of resilient robots that can learn to anticipate and avoid damaging interactions by combining visual and proprioceptive cues from internal sensors. The introduced solutions are inspired by pain pathways in humans and animals, because such pathways are known to guide decision-making processes and promote self-preservation. A new "robot dodge ball' benchmark is introduced in order to test the validity of the developed algorithms in dynamic environments.
ContributorsRichardson, Trevor W (Author) / Ben Amor, Heni (Thesis advisor) / Yang, Yezhou (Committee member) / Srivastava, Siddharth (Committee member) / Arizona State University (Publisher)
Created2018
155378-Thumbnail Image.png
Description
To ensure system integrity, robots need to proactively avoid any unwanted physical perturbation that may cause damage to the underlying hardware. In this thesis work, we investigate a machine learning approach that allows robots to anticipate impending physical perturbations from perceptual cues. In contrast to other approaches that require knowledge

To ensure system integrity, robots need to proactively avoid any unwanted physical perturbation that may cause damage to the underlying hardware. In this thesis work, we investigate a machine learning approach that allows robots to anticipate impending physical perturbations from perceptual cues. In contrast to other approaches that require knowledge about sources of perturbation to be encoded before deployment, our method is based on experiential learning. Robots learn to associate visual cues with subsequent physical perturbations and contacts. In turn, these extracted visual cues are then used to predict potential future perturbations acting on the robot. To this end, we introduce a novel deep network architecture which combines multiple sub- networks for dealing with robot dynamics and perceptual input from the environment. We present a self-supervised approach for training the system that does not require any labeling of training data. Extensive experiments in a human-robot interaction task show that a robot can learn to predict physical contact by a human interaction partner without any prior information or labeling. Furthermore, the network is able to successfully predict physical contact from either depth stream input or traditional video input or using both modalities as input.
ContributorsSur, Indranil (Author) / Amor, Heni B (Thesis advisor) / Fainekos, Georgios (Committee member) / Yang, Yezhou (Committee member) / Arizona State University (Publisher)
Created2017
168739-Thumbnail Image.png
Description
Visual navigation is a useful and important task for a variety of applications. As the preva­lence of robots increase, there is an increasing need for energy-­efficient navigation methods as well. Many aspects of efficient visual navigation algorithms have been implemented in the lit­erature, but there is a lack of work

Visual navigation is a useful and important task for a variety of applications. As the preva­lence of robots increase, there is an increasing need for energy-­efficient navigation methods as well. Many aspects of efficient visual navigation algorithms have been implemented in the lit­erature, but there is a lack of work on evaluation of the efficiency of the image sensors. In this thesis, two methods are evaluated: adaptive image sensor quantization for traditional camera pipelines as well as new event­-based sensors for low­-power computer vision.The first contribution in this thesis is an evaluation of performing varying levels of sen­sor linear and logarithmic quantization with the task of visual simultaneous localization and mapping (SLAM). This unconventional method can provide efficiency benefits with a trade­ off between accuracy of the task and energy-­efficiency. A new sensor quantization method, gradient­-based quantization, is introduced to improve the accuracy of the task. This method only lowers the bit level of parts of the image that are less likely to be important in the SLAM algorithm since lower bit levels signify better energy­-efficiency, but worse task accuracy. The third contribution is an evaluation of the efficiency and accuracy of event­-based camera inten­sity representations for the task of optical flow. The results of performing a learning based optical flow are provided for each of five different reconstruction methods along with ablation studies. Lastly, the challenges of an event feature­-based SLAM system are presented with re­sults demonstrating the necessity for high quality and high­ resolution event data. The work in this thesis provides studies useful for examining trade­offs for an efficient visual navigation system with traditional and event vision sensors. The results of this thesis also provide multiple directions for future work.
ContributorsChristie, Olivia Catherine (Author) / Jayasuriya, Suren (Thesis advisor) / Chakrabarti, Chaitali (Committee member) / Yang, Yezhou (Committee member) / Arizona State University (Publisher)
Created2022
187693-Thumbnail Image.png
Description
Simultaneous localization and mapping (SLAM) has traditionally relied on low-level geometric or optical features. However, these features-based SLAM methods often struggle with feature-less or repetitive scenes. Additionally, low-level features may not provide sufficient information for robot navigation and manipulation, leaving robots without a complete understanding of the 3D spatial world.

Simultaneous localization and mapping (SLAM) has traditionally relied on low-level geometric or optical features. However, these features-based SLAM methods often struggle with feature-less or repetitive scenes. Additionally, low-level features may not provide sufficient information for robot navigation and manipulation, leaving robots without a complete understanding of the 3D spatial world. Advanced information is necessary to address these limitations. Fortunately, recent developments in learning-based 3D reconstruction allow robots to not only detect semantic meanings, but also recognize the 3D structure of objects from a few images. By combining this 3D structural information, SLAM can be improved from a low-level approach to a structure-aware approach. This work propose a novel approach for multi-view 3D reconstruction using recurrent transformer. This approach allows robots to accumulate information from multiple views and encode them into a compact latent space. The resulting latent representations are then decoded to produce 3D structural landmarks, which can be used to improve robot localization and mapping.
ContributorsHuang, Chi-Yao (Author) / Yang, Yezhou (Thesis advisor) / Turaga, Pavan (Committee member) / Jayasuriya, Suren (Committee member) / Arizona State University (Publisher)
Created2023
187764-Thumbnail Image.png
Description
This thesis considers the problem of multi-robot task allocation with inter-agent distance constraints, e.g., due to the presence of physical tethers or communication requirements, that must be satisfied at all times. Specifically, three optimization-based formulations are explored: (i) a “Naive Method” that leverages the classical multiple traveling salesman (mTSP) formulation

This thesis considers the problem of multi-robot task allocation with inter-agent distance constraints, e.g., due to the presence of physical tethers or communication requirements, that must be satisfied at all times. Specifically, three optimization-based formulations are explored: (i) a “Naive Method” that leverages the classical multiple traveling salesman (mTSP) formulation to find solutions that are then filtered out when the inter-agent distance constraints are violated, (ii) a “Timed Method” thatconstructs a new formulation that explicitly accounts for robot timings, including the inter-agent distance constraints, and (iii) an “Improved Naive Method” that reformulates the Naive Method with a novel graph-traversal algorithm to produce tours that, unlike the Naive Method, allow backtracking and also introduces a more systematic approach to filter out solutions that violate inter-agent distance constraints. The effectiveness of the approaches to return task allocations that satisfy the constraints are demonstrated and compared in simulation experiments.
ContributorsGoodwin, Walter Alexander (Author) / Yong, Sze Zheng (Thesis advisor) / Grewal, Anoop (Thesis advisor) / Xu, Zhe (Committee member) / Arizona State University (Publisher)
Created2023
158800-Thumbnail Image.png
Description
Bicycle stabilization has become a popular topic because of its complex dynamic behavior and the large body of bicycle modeling research. Riding a bicycle requires accurately performing several tasks, such as balancing and navigation which may be difficult for disabled people. Their problems could be partially reduced by providing steering

Bicycle stabilization has become a popular topic because of its complex dynamic behavior and the large body of bicycle modeling research. Riding a bicycle requires accurately performing several tasks, such as balancing and navigation which may be difficult for disabled people. Their problems could be partially reduced by providing steering assistance. For stabilization of these highly maneuverable and efficient machines, many control techniques have been applied – achieving interesting results, but with some limitations which includes strict environmental requirements. This thesis expands on the work of Randlov and Alstrom, using reinforcement learning for bicycle self-stabilization with robotic steering. This thesis applies the deep deterministic policy gradient algorithm, which can handle continuous action spaces which is not possible for Q-learning technique. The research involved algorithm training on virtual environments followed by simulations to assess its results. Furthermore, hardware testing was also conducted on Arizona State University’s RISE lab Smart bicycle platform for testing its self-balancing performance. Detailed analysis of the bicycle trial runs are presented. Validation of testing was done by plotting the real-time states and actions collected during the outdoor testing which included the roll angle of bicycle. Further improvements in regard to model training and hardware testing are also presented.
ContributorsTurakhia, Shubham (Author) / Zhang, Wenlong (Thesis advisor) / Yong, Sze Zheng (Committee member) / Ren, Yi (Committee member) / Arizona State University (Publisher)
Created2020