Theses and Dissertations
Displaying 1 - 3 of 3
Filtering by
- Creators: Davulcu, Hasan
Description
Machine learning models convert raw data in the form of video, images, audio,
text, etc. into feature representations that are convenient for computational process-
ing. Deep neural networks have proven to be very efficient feature extractors for a
variety of machine learning tasks. Generative models based on deep neural networks
introduce constraints on the feature space to learn transferable and disentangled rep-
resentations. Transferable feature representations help in training machine learning
models that are robust across different distributions of data. For example, with the
application of transferable features in domain adaptation, models trained on a source
distribution can be applied to a data from a target distribution even though the dis-
tributions may be different. In style transfer and image-to-image translation, disen-
tangled representations allow for the separation of style and content when translating
images.
This thesis examines learning transferable data representations in novel deep gen-
erative models. The Semi-Supervised Adversarial Translator (SAT) utilizes adversar-
ial methods and cross-domain weight sharing in a neural network to extract trans-
ferable representations. These transferable interpretations can then be decoded into
the original image or a similar image in another domain. The Explicit Disentangling
Network (EDN) utilizes generative methods to disentangle images into their core at-
tributes and then segments sets of related attributes. The EDN can separate these
attributes by controlling the ow of information using a novel combination of losses
and network architecture. This separation of attributes allows precise modi_cations
to speci_c components of the data representation, boosting the performance of ma-
chine learning tasks. The effectiveness of these models is evaluated across domain
adaptation, style transfer, and image-to-image translation tasks.
text, etc. into feature representations that are convenient for computational process-
ing. Deep neural networks have proven to be very efficient feature extractors for a
variety of machine learning tasks. Generative models based on deep neural networks
introduce constraints on the feature space to learn transferable and disentangled rep-
resentations. Transferable feature representations help in training machine learning
models that are robust across different distributions of data. For example, with the
application of transferable features in domain adaptation, models trained on a source
distribution can be applied to a data from a target distribution even though the dis-
tributions may be different. In style transfer and image-to-image translation, disen-
tangled representations allow for the separation of style and content when translating
images.
This thesis examines learning transferable data representations in novel deep gen-
erative models. The Semi-Supervised Adversarial Translator (SAT) utilizes adversar-
ial methods and cross-domain weight sharing in a neural network to extract trans-
ferable representations. These transferable interpretations can then be decoded into
the original image or a similar image in another domain. The Explicit Disentangling
Network (EDN) utilizes generative methods to disentangle images into their core at-
tributes and then segments sets of related attributes. The EDN can separate these
attributes by controlling the ow of information using a novel combination of losses
and network architecture. This separation of attributes allows precise modi_cations
to speci_c components of the data representation, boosting the performance of ma-
chine learning tasks. The effectiveness of these models is evaluated across domain
adaptation, style transfer, and image-to-image translation tasks.
ContributorsEusebio, Jose Miguel Ang (Author) / Panchanathan, Sethuraman (Thesis advisor) / Davulcu, Hasan (Committee member) / Venkateswara, Hemanth (Committee member) / Arizona State University (Publisher)
Created2018
Description
Parents fulfill a pivotal role in early childhood development of social and communication
skills. In children with autism, the development of these skills can be delayed. Applied
behavioral analysis (ABA) techniques have been created to aid in skill acquisition.
Among these, pivotal response treatment (PRT) has been empirically shown to foster
improvements. Research into PRT implementation has also shown that parents can be
trained to be effective interventionists for their children. The current difficulty in PRT
training is how to disseminate training to parents who need it, and how to support and
motivate practitioners after training.
Evaluation of the parents’ fidelity to implementation is often undertaken using video
probes that depict the dyadic interaction occurring between the parent and the child during
PRT sessions. These videos are time consuming for clinicians to process, and often result
in only minimal feedback for the parents. Current trends in technology could be utilized to
alleviate the manual cost of extracting data from the videos, affording greater
opportunities for providing clinician created feedback as well as automated assessments.
The naturalistic context of the video probes along with the dependence on ubiquitous
recording devices creates a difficult scenario for classification tasks. The domain of the
PRT video probes can be expected to have high levels of both aleatory and epistemic
uncertainty. Addressing these challenges requires examination of the multimodal data
along with implementation and evaluation of classification algorithms. This is explored
through the use of a new dataset of PRT videos.
The relationship between the parent and the clinician is important. The clinician can
provide support and help build self-efficacy in addition to providing knowledge and
modeling of treatment procedures. Facilitating this relationship along with automated
feedback not only provides the opportunity to present expert feedback to the parent, but
also allows the clinician to aid in personalizing the classification models. By utilizing a
human-in-the-loop framework, clinicians can aid in addressing the uncertainty in the
classification models by providing additional labeled samples. This will allow the system
to improve classification and provides a person-centered approach to extracting
multimodal data from PRT video probes.
skills. In children with autism, the development of these skills can be delayed. Applied
behavioral analysis (ABA) techniques have been created to aid in skill acquisition.
Among these, pivotal response treatment (PRT) has been empirically shown to foster
improvements. Research into PRT implementation has also shown that parents can be
trained to be effective interventionists for their children. The current difficulty in PRT
training is how to disseminate training to parents who need it, and how to support and
motivate practitioners after training.
Evaluation of the parents’ fidelity to implementation is often undertaken using video
probes that depict the dyadic interaction occurring between the parent and the child during
PRT sessions. These videos are time consuming for clinicians to process, and often result
in only minimal feedback for the parents. Current trends in technology could be utilized to
alleviate the manual cost of extracting data from the videos, affording greater
opportunities for providing clinician created feedback as well as automated assessments.
The naturalistic context of the video probes along with the dependence on ubiquitous
recording devices creates a difficult scenario for classification tasks. The domain of the
PRT video probes can be expected to have high levels of both aleatory and epistemic
uncertainty. Addressing these challenges requires examination of the multimodal data
along with implementation and evaluation of classification algorithms. This is explored
through the use of a new dataset of PRT videos.
The relationship between the parent and the clinician is important. The clinician can
provide support and help build self-efficacy in addition to providing knowledge and
modeling of treatment procedures. Facilitating this relationship along with automated
feedback not only provides the opportunity to present expert feedback to the parent, but
also allows the clinician to aid in personalizing the classification models. By utilizing a
human-in-the-loop framework, clinicians can aid in addressing the uncertainty in the
classification models by providing additional labeled samples. This will allow the system
to improve classification and provides a person-centered approach to extracting
multimodal data from PRT video probes.
ContributorsCopenhaver Heath, Corey D (Author) / Panchanathan, Sethuraman (Thesis advisor) / McDaniel, Troy (Committee member) / Venkateswara, Hemanth (Committee member) / Davulcu, Hasan (Committee member) / Gaffar, Ashraf (Committee member) / Arizona State University (Publisher)
Created2019
Description
Access to real-time situational information including the relative position and motion of surrounding objects is critical for safe and independent travel. Object or obstacle (OO) detection at a distance is primarily a task of the visual system due to the high resolution information the eyes are able to receive from afar. As a sensory organ in particular, the eyes have an unparalleled ability to adjust to varying degrees of light, color, and distance. Therefore, in the case of a non-visual traveler, someone who is blind or low vision, access to visual information is unattainable if it is positioned beyond the reach of the preferred mobility device or outside the path of travel. Although, the area of assistive technology in terms of electronic travel aids (ETA’s) has received considerable attention over the last two decades; surprisingly, the field has seen little work in the area focused on augmenting rather than replacing current non-visual travel techniques, methods, and tools. Consequently, this work describes the design of an intuitive tactile language and series of wearable tactile interfaces (the Haptic Chair, HaptWrap, and HapBack) to deliver real-time spatiotemporal data. The overall intuitiveness of the haptic mappings conveyed through the tactile interfaces are evaluated using a combination of absolute identification accuracy of a series of patterns and subjective feedback through post-experiment surveys. Two types of spatiotemporal representations are considered: static patterns representing object location at a single time instance, and dynamic patterns, added in the HaptWrap, which represent object movement over a time interval. Results support the viability of multi-dimensional haptics applied to the body to yield an intuitive understanding of dynamic interactions occurring around the navigator during travel. Lastly, it is important to point out that the guiding principle of this work centered on providing the navigator with spatial knowledge otherwise unattainable through current mobility techniques, methods, and tools, thus, providing the \emph{navigator} with the information necessary to make informed navigation decisions independently, at a distance.
ContributorsDuarte, Bryan Joiner (Author) / McDaniel, Troy (Thesis advisor) / Davulcu, Hasan (Committee member) / Li, Baoxin (Committee member) / Venkateswara, Hemanth (Committee member) / Arizona State University (Publisher)
Created2020