Matching Items (99)
Filtering by

Clear all filters

152071-Thumbnail Image.png
Description
The development of advanced, anthropomorphic artificial hands aims to provide upper extremity amputees with improved functionality for activities of daily living. However, many state-of-the-art hands have a large number of degrees of freedom that can be challenging to control in an intuitive manner. Automated grip responses could be built into

The development of advanced, anthropomorphic artificial hands aims to provide upper extremity amputees with improved functionality for activities of daily living. However, many state-of-the-art hands have a large number of degrees of freedom that can be challenging to control in an intuitive manner. Automated grip responses could be built into artificial hands in order to enhance grasp stability and reduce the cognitive burden on the user. To this end, three studies were conducted to understand how human hands respond, passively and actively, to unexpected perturbations of a grasped object along and about different axes relative to the hand. The first study investigated the effect of magnitude, direction, and axis of rotation on precision grip responses to unexpected rotational perturbations of a grasped object. A robust "catch-up response" (a rapid, pulse-like increase in grip force rate previously reported only for translational perturbations) was observed whose strength scaled with the axis of rotation. Using two haptic robots, we then investigated the effects of grip surface friction, axis, and direction of perturbation on precision grip responses for unexpected translational and rotational perturbations for three different hand-centric axes. A robust catch-up response was observed for all axes and directions for both translational and rotational perturbations. Grip surface friction had no effect on the stereotypical catch-up response. Finally, we characterized the passive properties of the precision grip-object system via robot-imposed impulse perturbations. The hand-centric axis associated with the greatest translational stiffness was different than that for rotational stiffness. This work expands our understanding of the passive and active features of precision grip, a hallmark of human dexterous manipulation. Biological insights such as these could be used to enhance the functionality of artificial hands and the quality of life for upper extremity amputees.
ContributorsDe Gregorio, Michael (Author) / Santos, Veronica J. (Thesis advisor) / Artemiadis, Panagiotis K. (Committee member) / Santello, Marco (Committee member) / Sugar, Thomas (Committee member) / Helms Tillery, Stephen I. (Committee member) / Arizona State University (Publisher)
Created2013
151742-Thumbnail Image.png
Description
This research is focused on two separate but related topics. The first uses an electroencephalographic (EEG) brain-computer interface (BCI) to explore the phenomenon of motor learning transfer. The second takes a closer look at the EEG-BCI itself and tests an alternate way of mapping EEG signals into machine commands. We

This research is focused on two separate but related topics. The first uses an electroencephalographic (EEG) brain-computer interface (BCI) to explore the phenomenon of motor learning transfer. The second takes a closer look at the EEG-BCI itself and tests an alternate way of mapping EEG signals into machine commands. We test whether motor learning transfer is more related to use of shared neural structures between imagery and motor execution or to more generalized cognitive factors. Using an EEG-BCI, we train one group of participants to control the movements of a cursor using embodied motor imagery. A second group is trained to control the cursor using abstract motor imagery. A third control group practices moving the cursor using an arm and finger on a touch screen. We hypothesized that if motor learning transfer is related to the use of shared neural structures then the embodied motor imagery group would show more learning transfer than the abstract imaging group. If, on the other hand, motor learning transfer results from more general cognitive processes, then the abstract motor imagery group should also demonstrate motor learning transfer to the manual performance of the same task. Our findings support that motor learning transfer is due to the use of shared neural structures between imaging and motor execution of a task. The abstract group showed no motor learning transfer despite being better at EEG-BCI control than the embodied group. The fact that more participants were able to learn EEG-BCI control using abstract imagery suggests that abstract imagery may be more suitable for EEG-BCIs for some disabilities, while embodied imagery may be more suitable for others. In Part 2, EEG data collected in the above experiment was used to train an artificial neural network (ANN) to map EEG signals to machine commands. We found that our open-source ANN using spectrograms generated from SFFTs is fundamentally different and in some ways superior to Emotiv's proprietary method. Our use of novel combinations of existing technologies along with abstract and embodied imagery facilitates adaptive customization of EEG-BCI control to meet needs of individual users.
Contributorsda Silva, Flavio J. K (Author) / Mcbeath, Michael K (Thesis advisor) / Helms Tillery, Stephen (Committee member) / Presson, Clark (Committee member) / Sugar, Thomas (Committee member) / Arizona State University (Publisher)
Created2013
153498-Thumbnail Image.png
Description
Myoelectric control is lled with potential to signicantly change human-robot interaction.

Humans desire compliant robots to safely interact in dynamic environments

associated with daily activities. As surface electromyography non-invasively measures

limb motion intent and correlates with joint stiness during co-contractions,

it has been identied as a candidate for naturally controlling such robots. However,

state-of-the-art myoelectric

Myoelectric control is lled with potential to signicantly change human-robot interaction.

Humans desire compliant robots to safely interact in dynamic environments

associated with daily activities. As surface electromyography non-invasively measures

limb motion intent and correlates with joint stiness during co-contractions,

it has been identied as a candidate for naturally controlling such robots. However,

state-of-the-art myoelectric interfaces have struggled to achieve both enhanced

functionality and long-term reliability. As demands in myoelectric interfaces trend

toward simultaneous and proportional control of compliant robots, robust processing

of multi-muscle coordinations, or synergies, plays a larger role in the success of the

control scheme. This dissertation presents a framework enhancing the utility of myoelectric

interfaces by exploiting motor skill learning and

exible muscle synergies for

reliable long-term simultaneous and proportional control of multifunctional compliant

robots. The interface is learned as a new motor skill specic to the controller,

providing long-term performance enhancements without requiring any retraining or

recalibration of the system. Moreover, the framework oers control of both motion

and stiness simultaneously for intuitive and compliant human-robot interaction. The

framework is validated through a series of experiments characterizing motor learning

properties and demonstrating control capabilities not seen previously in the literature.

The results validate the approach as a viable option to remove the trade-o

between functionality and reliability that have hindered state-of-the-art myoelectric

interfaces. Thus, this research contributes to the expansion and enhancement of myoelectric

controlled applications beyond commonly perceived anthropomorphic and

\intuitive control" constraints and into more advanced robotic systems designed for

everyday tasks.
ContributorsIson, Mark (Author) / Artemiadis, Panagiotis (Thesis advisor) / Santello, Marco (Committee member) / Greger, Bradley (Committee member) / Berman, Spring (Committee member) / Sugar, Thomas (Committee member) / Fainekos, Georgios (Committee member) / Arizona State University (Publisher)
Created2015
153240-Thumbnail Image.png
Description
Human running requires extensive training and conditioning for an individual to maintain high speeds (greater than 10mph) for an extended duration of time. Studies have shown that running at peak speeds generates a high metabolic cost due to the use of large muscle groups in the legs associated with

Human running requires extensive training and conditioning for an individual to maintain high speeds (greater than 10mph) for an extended duration of time. Studies have shown that running at peak speeds generates a high metabolic cost due to the use of large muscle groups in the legs associated with the human gait cycle. Applying supplemental external and internal forces to the human body during the gait cycle has been shown to decrease the metabolic cost for walking, allowing individuals to carry additional weight and walk further distances. Significant research has been conducted to reduce the metabolic cost of walking, however, there are few if any documented studies that focus specifically on reducing the metabolic cost associated with high speed running. Three mechanical systems were designed to work in concert with the human user to decrease metabolic cost and increase the range and speeds at which a human can run.

The methods of design require a focus on mathematical modeling, simulations, and metabolic cost. Mathematical modeling and simulations are used to aid in the design process of robotic systems and metabolic testing is regarded as the final analysis process to determine the true effectiveness of robotic prototypes. Metabolic data, (VO2) is the volumetric consumption of oxygen, per minute, per unit mass (ml/min/kg). Metabolic testing consists of analyzing the oxygen consumption of a test subject while performing a task naturally and then comparing that data with analyzed oxygen consumption of the same task while using an assistive device.

Three devices were designed and tested to augment high speed running. The first device, AirLegs V1, is a mostly aluminum exoskeleton with two pneumatic linear actuators connecting from the lower back directly to the user's thighs, allowing the device to induce a torque on the leg by pushing and pulling on the user's thigh during running. The device also makes use of two smaller pneumatic linear actuators which drive cables connecting to small lever arms at the back of the heel, inducing a torque at the ankles. Device two, AirLegs V2, is also pneumatically powered but is considered to be a soft suit version of the first device. It uses cables to interface the forces created by actuators located vertically on the user's back. These cables then connect to the back of the user's knees resulting in greater flexibility and range of motion of the legs. Device three, a Jet Pack, produces an external force against the user's torso to propel a user forward and upward making it easier to run. Third party testing, pilot demonstrations and timed trials have demonstrated that all three of the devices effectively reduce the metabolic cost of running below that of natural running with no device.
ContributorsKerestes, Jason (Author) / Sugar, Thomas (Thesis advisor) / Redkar, Sangram (Committee member) / Rogers, Bradley (Committee member) / Arizona State University (Publisher)
Created2014
153807-Thumbnail Image.png
Description
Brain Computer Interfaces are becoming the next generation controllers not only in the medical devices for disabled individuals but also in the gaming and entertainment industries. In order to build an effective Brain Computer Interface, which accurately translates the user thoughts into machine commands, it is important to have robust

Brain Computer Interfaces are becoming the next generation controllers not only in the medical devices for disabled individuals but also in the gaming and entertainment industries. In order to build an effective Brain Computer Interface, which accurately translates the user thoughts into machine commands, it is important to have robust and fail proof signal processing and machine learning modules which operate on the raw EEG signals and estimate the current thought of the user.

In this thesis, several techniques used to perform EEG signal pre-processing, feature extraction and signal classification have been discussed, implemented, validated and verified; efficient supervised machine learning models, for the EEG motor imagery signal classification are identified. To further improve the performance of system unsupervised feature learning techniques have been investigated by pre-training the Deep Learning models. Use of pre-training stacked autoencoders have been proposed to solve the problems caused by random initialization of weights in neural networks.

Motor Imagery (imaginary hand and leg movements) signals are acquire using the Emotiv EEG headset. Different kinds of features like mean signal, band powers, RMS of the signal have been extracted and supplied to the machine learning (ML) stage, wherein, several ML techniques like LDA, KNN, SVM, Logistic regression and Neural Networks are applied and validated. During the validation phase the performances of various techniques are compared and some important observations are reported. Further, deep Learning techniques like autoencoding have been used to perform unsupervised feature learning. The reliability of the features is analyzed by performing classification by using the ML techniques mentioned earlier. The performance of the neural networks has been further improved by pre-training the network in an unsupervised fashion using stacked autoencoders and supplying the stacked autoencoders’ network parameters as initial parameters to the neural network. All the findings in this research, during each phase (pre-processing, feature extraction, classification) are directly relevant and can be used by the BCI research community for building motor imagery based BCI applications.

Additionally, this thesis attempts to develop, test, and compare the performance of an alternative method for classifying human driving behavior. This thesis proposes the use of driver affective states to know the driving behavior. The purpose of this part of the thesis was to classify the EEG data collected from several subjects while driving simulated vehicle and compare the classification results with those obtained by classifying the driving behavior using vehicle parameters collected simultaneously from all the subjects. The objective here is to see if the drivers’ mental state is reflected in his driving behavior.
ContributorsManchala, Vamsi Krishna (Author) / Redkar, Sangram (Thesis advisor) / Rogers, Bradley (Committee member) / Sugar, Thomas (Committee member) / Arizona State University (Publisher)
Created2015
Description
Optical Fibers coupled to laser light sources, and Light Emitting Diodes are the two classes of technologies used for optogenetic experiments. Arizona State University's Flexible Display Center fabricates novel flexible Organic Light Emitting Diodes(OLEDs). These OLEDs have the capability of being monolithically fabricated over flexible, transparent plastic substrates and having

Optical Fibers coupled to laser light sources, and Light Emitting Diodes are the two classes of technologies used for optogenetic experiments. Arizona State University's Flexible Display Center fabricates novel flexible Organic Light Emitting Diodes(OLEDs). These OLEDs have the capability of being monolithically fabricated over flexible, transparent plastic substrates and having power efficient ways of addressing high density arrays of LEDs. This thesis critically evaluates the technology by identifying the key advantages, current limitations and experimentally assessing the technology in in-vivo and in-vitro animal models. For in-vivo testing, the emitted light from a flat OLED panel was directly used to stimulate the neo-cortex in the M1 region of transgenic mice expressing ChR2 (B6.Cg-Tg (Thy1-ChR2/EYFP) 9Gfng/J). An alternative stimulation paradigm using a collimating optical system coupled with an optical fiber was used for stimulating neurons in layer 5 of the motor cortex in the same transgenic mice. EMG activity was recorded from the contralateral vastus lateralis muscles. In vitro testing of the OLEDs was done in primary cortical neurons in culture transfected with blue light sensitive ChR2. The neurons were cultured on a microelectrode array for taking neuronal recordings.
ContributorsShah, Ankur (Author) / Muthuswamy, Jitendran (Thesis advisor) / Greger, Bradley (Committee member) / Blain Christen, Jennifer (Committee member) / Arizona State University (Publisher)
Created2015
154164-Thumbnail Image.png
Description
Epilepsy is a group of disorders that cause seizures in approximately 2.2 million people in the United States. Over 30% of these patients have epilepsies that do not respond to treatment with anti-epileptic drugs. For this population, focal resection surgery could offer long-term seizure freedom. Surgery candidates undergo a myriad

Epilepsy is a group of disorders that cause seizures in approximately 2.2 million people in the United States. Over 30% of these patients have epilepsies that do not respond to treatment with anti-epileptic drugs. For this population, focal resection surgery could offer long-term seizure freedom. Surgery candidates undergo a myriad of tests and monitoring to determine where and when seizures occur. The “gold standard” method for focus identification involves the placement of electrocorticography (ECoG) grids in the sub-dural space, followed by continual monitoring and visual inspection of the patient’s cortical activity. This process, however, is highly subjective and uses dated technology. Multiple studies were performed to investigate how the evaluation process could benefit from an algorithmic adjust using current ECoG technology, and how the use of new microECoG technology could further improve the process.

Computational algorithms can quickly and objectively find signal characteristics that may not be detectable with visual inspection, but many assume the data are stationary and/or linear, which biological data are not. An empirical mode decomposition (EMD) based algorithm was developed to detect potential seizures and tested on data collected from eight patients undergoing monitoring for focal resection surgery. EMD does not require linearity or stationarity and is data driven. The results suggest that a biological data driven algorithm could serve as a useful tool to objectively identify changes in cortical activity associated with seizures.

Next, the use of microECoG technology was investigated. Though both ECoG and microECoG grids are composed of electrodes resting on the surface of the cortex, changing the diameter of the electrodes creates non-trivial changes in the physics of the electrode-tissue interface that need to be accounted for. Experimenting with different recording configurations showed that proper grounding, referencing, and amplification are critical to obtain high quality neural signals from microECoG grids.

Finally, the relationship between data collected from the cortical surface with micro and macro electrodes was studied. Simultaneous recordings of the two electrode types showed differences in power spectra that suggest the inclusion of activity, possibly from deep structures, by macroelectrodes that is not accessible by microelectrodes.
ContributorsAshmont, Kari Rich (Author) / Greger, Bradley (Thesis advisor) / Helms Tillery, Stephen (Committee member) / Buneo, Christopher (Committee member) / Adelson, P David (Committee member) / Dudek, F Edward (Committee member) / Arizona State University (Publisher)
Created2015
153889-Thumbnail Image.png
Description
Robust and stable decoding of neural signals is imperative for implementing a useful neuroprosthesis capable of carrying out dexterous tasks. A nonhuman primate (NHP) was trained to perform combined flexions of the thumb, index and middle fingers in addition to individual flexions and extensions of the same digits. An array

Robust and stable decoding of neural signals is imperative for implementing a useful neuroprosthesis capable of carrying out dexterous tasks. A nonhuman primate (NHP) was trained to perform combined flexions of the thumb, index and middle fingers in addition to individual flexions and extensions of the same digits. An array of microelectrodes was implanted in the hand area of the motor cortex of the NHP and used to record action potentials during finger movements. A Support Vector Machine (SVM) was used to classify which finger movement the NHP was making based upon action potential firing rates. The effect of four feature selection techniques, Wilcoxon signed-rank test, Relative Importance, Principal Component Analysis, and Mutual Information Maximization was compared based on SVM classification performance. SVM classification was used to examine the functional parameters of (i) efficacy (ii) endurance to simulated failure and (iii) longevity of classification. The effect of using isolated-neuron and multi-unit firing rates was compared as the feature vector supplied to the SVM. The best classification performance was on post-implantation day 36, when using multi-unit firing rates the worst classification accuracy resulted from features selected with Wilcoxon signed-rank test (51.12 ± 0.65%) and the best classification accuracy resulted from Mutual Information Maximization (93.74 ± 0.32%). On this day when using single-unit firing rates, the classification accuracy from the Wilcoxon signed-rank test was 88.85 ± 0.61 % and Mutual Information Maximization was 95.60 ± 0.52% (degrees of freedom =10, level of chance =10%)
ContributorsPadmanaban, Subash (Author) / Greger, Bradley (Thesis advisor) / Santello, Marco (Thesis advisor) / Helms Tillery, Stephen (Committee member) / Arizona State University (Publisher)
Created2015
155935-Thumbnail Image.png
Description
Object manipulation is a common sensorimotor task that humans perform to interact with the physical world. The first aim of this dissertation was to characterize and identify the role of feedback and feedforward mechanisms for force control in object manipulation by introducing a new feature based on force trajectories to

Object manipulation is a common sensorimotor task that humans perform to interact with the physical world. The first aim of this dissertation was to characterize and identify the role of feedback and feedforward mechanisms for force control in object manipulation by introducing a new feature based on force trajectories to quantify the interaction between feedback- and feedforward control. This feature was applied on two grasp contexts: grasping the object at either (1) predetermined or (2) self-selected grasp locations (“constrained” and “unconstrained”, respectively), where unconstrained grasping is thought to involve feedback-driven force corrections to a greater extent than constrained grasping. This proposition was confirmed by force feature analysis. The second aim of this dissertation was to quantify whether force control mechanisms differ between dominant and non-dominant hands. The force feature analysis demonstrated that manipulation by the dominant hand relies on feedforward control more than the non-dominant hand. The third aim was to quantify coordination mechanisms underlying physical interaction by dyads in object manipulation. The results revealed that only individuals with worse solo performance benefit from interpersonal coordination through physical couplings, whereas the better individuals do not. This work showed that naturally emerging leader-follower roles, whereby the leader in dyadic manipulation exhibits significant greater force changes than the follower. Furthermore, brain activity measured through electroencephalography (EEG) could discriminate leader and follower roles as indicated power modulation in the alpha frequency band over centro-parietal areas. Lastly, this dissertation suggested that the relation between force and motion (arm impedance) could be an important means for communicating intended movement direction between biological agents.
ContributorsMojtahedi, Keivan (Author) / Santello, Marco (Thesis advisor) / Greger, Bradley (Committee member) / Artemiadis, Panagiotis (Committee member) / Helms Tillery, Stephen (Committee member) / Buneo, Christopher (Committee member) / Arizona State University (Publisher)
Created2017
156321-Thumbnail Image.png
Description
The sensor industry is a growing industry that has been predicted by Allied Market Research to be a multi-billion industry by 2022. One of the many key drives behind this rapid growth in the sensor industry is the increase incorporation of sensors into portable electrical devices. The value

The sensor industry is a growing industry that has been predicted by Allied Market Research to be a multi-billion industry by 2022. One of the many key drives behind this rapid growth in the sensor industry is the increase incorporation of sensors into portable electrical devices. The value for sensor technologies are increased when the sensors are developed into innovative measuring system for application uses in the Aerospace, Defense, and Healthcare industries. While sensors are not new, their increased performance, size reduction, and decrease in cost has opened the door for innovative sensor combination for portable devices that could be worn or easily moved around. With this opportunity for further development of sensor use through concept engineering development, three concept projects for possible innovative portable devices was undertaken in this research. One project was the development of a pulse oximeter devise with fingerprint recognition. The second project was prototyping a portable Bluetooth strain gage monitoring system. The third project involved sensors being incorporated onto flexible printed circuit board (PCB) for improved comfort of wearable devices. All these systems were successfully tested in lab.
ContributorsNichols, Kevin William (Author) / Redkar, Sangram (Thesis advisor) / Rogers, Brad (Committee member) / Sugar, Thomas (Committee member) / Arizona State University (Publisher)
Created2018