Matching Items (55)
Filtering by

Clear all filters

152349-Thumbnail Image.png
Description
As robots are increasingly migrating out of factories and research laboratories and into our everyday lives, they should move and act in environments designed for humans. For this reason, the need of anthropomorphic movements is of utmost importance. The objective of this thesis is to solve the inverse kinematics problem

As robots are increasingly migrating out of factories and research laboratories and into our everyday lives, they should move and act in environments designed for humans. For this reason, the need of anthropomorphic movements is of utmost importance. The objective of this thesis is to solve the inverse kinematics problem of redundant robot arms that results to anthropomorphic configurations. The swivel angle of the elbow was used as a human arm motion parameter for the robot arm to mimic. The swivel angle is defined as the rotation angle of the plane defined by the upper and lower arm around a virtual axis that connects the shoulder and wrist joints. Using kinematic data recorded from human subjects during every-day life tasks, the linear sensorimotor transformation model was validated and used to estimate the swivel angle, given the desired end-effector position. Defining the desired swivel angle simplifies the kinematic redundancy of the robot arm. The proposed method was tested with an anthropomorphic redundant robot arm and the computed motion profiles were compared to the ones of the human subjects. This thesis shows that the method computes anthropomorphic configurations for the robot arm, even if the robot arm has different link lengths than the human arm and starts its motion at random configurations.
ContributorsWang, Yuting (Author) / Artemiadis, Panagiotis (Thesis advisor) / Mignolet, Marc (Committee member) / Santos, Veronica J (Committee member) / Arizona State University (Publisher)
Created2013
152324-Thumbnail Image.png
Description
With robots being used extensively in various areas, a certain degree of robot autonomy has always been found desirable. In applications like planetary exploration, autonomous path planning and navigation are considered essential. But every now and then, a need to modify the robot's operation arises, a need for a human

With robots being used extensively in various areas, a certain degree of robot autonomy has always been found desirable. In applications like planetary exploration, autonomous path planning and navigation are considered essential. But every now and then, a need to modify the robot's operation arises, a need for a human to provide it some supervisory parameters that modify the degree of autonomy or allocate extra tasks to the robot. In this regard, this thesis presents an approach to include a provision to accept and incorporate such human inputs and modify the navigation functions of the robot accordingly. Concepts such as applying kinematical constraints while planning paths, traversing of unknown areas with an intent of maximizing field of view, performing complex tasks on command etc. have been examined and implemented. The approaches have been tested in Robot Operating System (ROS), using robots such as the iRobot Create, Personal Robotics (PR2) etc. Simulations and experimental demonstrations have proved that this approach is feasible for solving some of the existing problems and that it certainly can pave way to further research for enhancing functionality.
ContributorsVemprala, Sai Hemachandra (Author) / Saripalli, Srikanth (Thesis advisor) / Fainekos, Georgios (Committee member) / Turaga, Pavan (Committee member) / Arizona State University (Publisher)
Created2013
151787-Thumbnail Image.png
Description
Electromyogram (EMG)-based control interfaces are increasingly used in robot teleoperation, prosthetic devices control and also in controlling robotic exoskeletons. Over the last two decades researchers have come up with a plethora of decoding functions to map myoelectric signals to robot motions. However, this requires a lot of training and validation

Electromyogram (EMG)-based control interfaces are increasingly used in robot teleoperation, prosthetic devices control and also in controlling robotic exoskeletons. Over the last two decades researchers have come up with a plethora of decoding functions to map myoelectric signals to robot motions. However, this requires a lot of training and validation data sets, while the parameters of the decoding function are specific for each subject. In this thesis we propose a new methodology that doesn't require training and is not user-specific. The main idea is to supplement the decoding functional error with the human ability to learn inverse model of an arbitrary mapping function. We have shown that the subjects gradually learned the control strategy and their learning rates improved. We also worked on identifying an optimized control scheme that would be even more effective and easy to learn for the subjects. Optimization was done by taking into account that muscles act in synergies while performing a motion task. The low-dimensional representation of the neural activity was used to control a two-dimensional task. Results showed that in the case of reduced dimensionality mapping, the subjects were able to learn to control the device in a slower pace, however they were able to reach and retain the same level of controllability. To summarize, we were able to build an EMG-based controller for robot devices that would work for any subject, without any training or decoding function, suggesting human-embedded controllers for robotic devices.
ContributorsAntuvan, Chris Wilson (Author) / Artemiadis, Panagiotis (Thesis advisor) / Muthuswamy, Jitendran (Committee member) / Santos, Veronica J (Committee member) / Arizona State University (Publisher)
Created2013
151793-Thumbnail Image.png
Description
Linear Temporal Logic is gaining increasing popularity as a high level specification language for robot motion planning due to its expressive power and scalability of LTL control synthesis algorithms. This formalism, however, requires expert knowledge and makes it inaccessible to non-expert users. This thesis introduces a graphical specification environment to

Linear Temporal Logic is gaining increasing popularity as a high level specification language for robot motion planning due to its expressive power and scalability of LTL control synthesis algorithms. This formalism, however, requires expert knowledge and makes it inaccessible to non-expert users. This thesis introduces a graphical specification environment to create high level motion plans to control robots in the field by converting a visual representation of the motion/task plan into a Linear Temporal Logic (LTL) specification. The visual interface is built on the Android tablet platform and provides functionality to create task plans through a set of well defined gestures and on screen controls. It uses the notion of waypoints to quickly and efficiently describe the motion plan and enables a variety of complex Linear Temporal Logic specifications to be described succinctly and intuitively by the user without the need for the knowledge and understanding of LTL specification. Thus, it opens avenues for its use by personnel in military, warehouse management, and search and rescue missions. This thesis describes the construction of LTL for various scenarios used for robot navigation using the visual interface developed and leverages the use of existing LTL based motion planners to carry out the task plan by a robot.
ContributorsSrinivas, Shashank (Author) / Fainekos, Georgios (Thesis advisor) / Baral, Chitta (Committee member) / Burleson, Winslow (Committee member) / Arizona State University (Publisher)
Created2013
151803-Thumbnail Image.png
Description
Humans have an inherent capability of performing highly dexterous and skillful tasks with their arms, involving maintaining posture, movement and interacting with the environment. The latter requires for them to control the dynamic characteristics of the upper limb musculoskeletal system. Inertia, damping and stiffness, a measure of mechanical impedance, gives

Humans have an inherent capability of performing highly dexterous and skillful tasks with their arms, involving maintaining posture, movement and interacting with the environment. The latter requires for them to control the dynamic characteristics of the upper limb musculoskeletal system. Inertia, damping and stiffness, a measure of mechanical impedance, gives a strong representation of these characteristics. Many previous studies have shown that the arm posture is a dominant factor for determining the end point impedance in a horizontal plane (transverse plane). The objective of this thesis is to characterize end point impedance of the human arm in the three dimensional (3D) space. Moreover, it investigates and models the control of the arm impedance due to increasing levels of muscle co-contraction. The characterization is done through experimental trials where human subjects maintained arm posture, while perturbed by a robot arm. Moreover, the subjects were asked to control the level of their arm muscles' co-contraction, using visual feedback of their muscles' activation, in order to investigate the effect of the muscle co-contraction on the arm impedance. The results of this study showed a very interesting, anisotropic increase of the arm stiffness due to muscle co-contraction. This can lead to very useful conclusions about the arm biomechanics as well as many implications for human motor control and more specifically the control of arm impedance through muscle co-contraction. The study finds implications for the EMG-based control of robots that physically interact with humans.
ContributorsPatel, Harshil Naresh (Author) / Artemiadis, Panagiotis (Thesis advisor) / Berman, Spring (Committee member) / Helms Tillery, Stephen (Committee member) / Arizona State University (Publisher)
Created2013
152234-Thumbnail Image.png
Description
One of the main challenges in planetary robotics is to traverse the shortest path through a set of waypoints. The shortest distance between any two waypoints is a direct linear traversal. Often times, there are physical restrictions that prevent a rover form traversing straight to a waypoint. Thus, knowledge of

One of the main challenges in planetary robotics is to traverse the shortest path through a set of waypoints. The shortest distance between any two waypoints is a direct linear traversal. Often times, there are physical restrictions that prevent a rover form traversing straight to a waypoint. Thus, knowledge of the terrain is needed prior to traversal. The Digital Terrain Model (DTM) provides information about the terrain along with waypoints for the rover to traverse. However, traversing a set of waypoints linearly is burdensome, as the rovers would constantly need to modify their orientation as they successively approach waypoints. Although there are various solutions to this problem, this research paper proposes the smooth traversability of the rover using splines as a quick and easy implementation to traverse a set of waypoints. In addition, a rover was used to compare the smoothness of the linear traversal along with the spline interpolations. The data collected illustrated that spline traversals had a less rate of change in the velocity over time, indicating that the rover performed smoother than with linear paths.
ContributorsKamasamudram, Anurag (Author) / Saripalli, Srikanth (Thesis advisor) / Fainekos, Georgios (Thesis advisor) / Turaga, Pavan (Committee member) / Arizona State University (Publisher)
Created2013
152179-Thumbnail Image.png
Description
As the complexity of robotic systems and applications grows rapidly, development of high-performance, easy to use, and fully integrated development environments for those systems is inevitable. Model-Based Design (MBD) of dynamic systems using engineering software such as Simulink® from MathWorks®, SciCos from Metalau team and SystemModeler® from Wolfram® is quite

As the complexity of robotic systems and applications grows rapidly, development of high-performance, easy to use, and fully integrated development environments for those systems is inevitable. Model-Based Design (MBD) of dynamic systems using engineering software such as Simulink® from MathWorks®, SciCos from Metalau team and SystemModeler® from Wolfram® is quite popular nowadays. They provide tools for modeling, simulation, verification and in some cases automatic code generation for desktop applications, embedded systems and robots. For real-world implementation of models on the actual hardware, those models should be converted into compilable machine code either manually or automatically. Due to the complexity of robotic systems, manual code translation from model to code is not a feasible optimal solution so we need to move towards automated code generation for such systems. MathWorks® offers code generation facilities called Coder® products for this purpose. However in order to fully exploit the power of model-based design and code generation tools for robotic applications, we need to enhance those software systems by adding and modifying toolboxes, files and other artifacts as well as developing guidelines and procedures. In this thesis, an effort has been made to propose a guideline as well as a Simulink® library, StateFlow® interface API and a C/C++ interface API to complete this toolchain for NAO humanoid robots. Thus the model of the hierarchical control architecture can be easily and properly converted to code and built for implementation.
ContributorsRaji Kermani, Ramtin (Author) / Fainekos, Georgios (Thesis advisor) / Lee, Yann-Hang (Committee member) / Sarjoughian, Hessam S. (Committee member) / Arizona State University (Publisher)
Created2013
137772-Thumbnail Image.png
Description
As robots become more prevalent, the need is growing for efficient yet stable control systems for applications with humans in the loop. As such, it is a challenge for scientists and engineers to develop robust and agile systems that are capable of detecting instability in teleoperated systems. Despite how much

As robots become more prevalent, the need is growing for efficient yet stable control systems for applications with humans in the loop. As such, it is a challenge for scientists and engineers to develop robust and agile systems that are capable of detecting instability in teleoperated systems. Despite how much research has been done to characterize the spatiotemporal parameters of human arm motions for reaching and gasping, not much has been done to characterize the behavior of human arm motion in response to control errors in a system. The scope of this investigation is to investigate human corrective actions in response to error in an anthropomorphic teleoperated robot limb. Characterizing human corrective actions contributes to the development of control strategies that are capable of mitigating potential instabilities inherent in human-machine control interfaces. Characterization of human corrective actions requires the simulation of a teleoperated anthropomorphic armature and the comparison of a human subject's arm kinematics, in response to error, against the human arm kinematics without error. This was achieved using OpenGL software to simulate a teleoperated robot arm and an NDI motion tracking system to acquire the subject's arm position and orientation. Error was intermittently and programmatically introduced to the virtual robot's joints as the subject attempted to reach for several targets located around the arm. The comparison of error free human arm kinematics to error prone human arm kinematics revealed an addition of a bell shaped velocity peak into the human subject's tangential velocity profile. The size, extent, and location of the additional velocity peak depended on target location and join angle error. Some joint angle and target location combinations do not produce an additional peak but simply maintain the end effector velocity at a low value until the target is reached. Additional joint angle error parameters and degrees of freedom are needed to continue this investigation.
ContributorsBevilacqua, Vincent Frank (Author) / Artemiadis, Panagiotis (Thesis director) / Santello, Marco (Committee member) / Trimble, Steven (Committee member) / Barrett, The Honors College (Contributor) / Mechanical and Aerospace Engineering Program (Contributor)
Created2013-05
137748-Thumbnail Image.png
Description
I worked on the human-machine interface to improve human physical capability. This work was done in the Human Oriented Robotics and Control Lab (HORC) towards the creation of an advanced, EMG-controlled exoskeleton. The project was new, and any work on the human- machine interface needs the physical interface itself. So

I worked on the human-machine interface to improve human physical capability. This work was done in the Human Oriented Robotics and Control Lab (HORC) towards the creation of an advanced, EMG-controlled exoskeleton. The project was new, and any work on the human- machine interface needs the physical interface itself. So I designed and fabricated a human-robot coupling device with a novel safety feature. The validation testing of this coupling proved very successful, and the device was granted a provisional patent as well as published to facilitate its spread to other human-machine interface applications, where it could be of major benefit. I then employed this coupling in experimentation towards understanding impedance, with the end goal being the creation of an EMG-based impedance exoskeleton control system. I modified a previously established robot-to-human perturbation method for use in my novel, three- dimensional (3D) impedance measurement experiment. Upon execution of this experiment, I was able to successfully characterize passive, static human arm stiffness in 3D, and in doing so validated the aforementioned method. This establishes an important foundation for promising future work on understanding impedance and the creation of the proposed control scheme, thereby furthering the field of human-robot interaction.
ContributorsO'Neill, Gerald D. (Author) / Artemiadis, Panagiotis (Thesis director) / Santello, Marco (Committee member) / Santos, Veronica (Committee member) / Barrett, The Honors College (Contributor) / Mechanical and Aerospace Engineering Program (Contributor)
Created2013-05
137299-Thumbnail Image.png
Description
This thesis focused on grasping tasks with the goal of investigating, analyzing, and quantifying human catching trends by way of a mathematical model. The aim of this project was to study human trends in a dynamic grasping task (catching a rolling ball), relate those discovered trends to kinematic characteristics of

This thesis focused on grasping tasks with the goal of investigating, analyzing, and quantifying human catching trends by way of a mathematical model. The aim of this project was to study human trends in a dynamic grasping task (catching a rolling ball), relate those discovered trends to kinematic characteristics of the object, and use this relation to control a robot hand in real time. As an ultimate goal, it was hoped that this research will aide in furthering the bio-inspiration in robot control methods. To achieve the above goal, firstly a tactile sensing glove was developed. This instrument allowed for in depth study of human reactionary grasping movements when worn by subjects during experimentation. This sensing glove system recorded force data from the palm and motion data from four fingers. From these data sets, temporal trends were established relating to when subjects initiated grasping during each trial. Moreover, optical tracking was implemented to study the kinematics of the moving object during human experiments and also to close the loop during the control of the robot hand. Ultimately, a mathematical bio-inspired model was created. This was embodied in a two-term decreasing power function which related the temporal trend of wait time to the ball initial acceleration. The wait time is defined as the time between when the experimental conductor releases the ball and when the subject begins to initiate grasping by closing their fingers, over a distance of four feet. The initial acceleration is the first acceleration value of the object due to the force provided when the conductor throws the object. The distance over which the ball was thrown was incorporated into the model. This is discussed in depth within the thesis. Overall, the results presented here show promise for bio-inspired control schemes in the successful application of robotic devices. This control methodology will ideally be developed to move robotic prosthesis past discrete tasks and into more complicated activities.
ContributorsCard, Dillon (Co-author) / Mincieli, Jennifer (Co-author) / Artemiadis, Panagiotis (Thesis director) / Santos, Veronica (Committee member) / Middleton, James (Committee member) / Barrett, The Honors College (Contributor) / School of Sustainability (Contributor) / Mechanical and Aerospace Engineering Program (Contributor) / W. P. Carey School of Business (Contributor)
Created2014-05