Matching Items (28)
Filtering by

Clear all filters

147948-Thumbnail Image.png
Description

When examining the average college campus, it becomes obvious that students feel rushed from one place to another as they try to participate in class, clubs, and extracurricular activities. One way that students can feel more comfortable and relaxed around campus is to introduce the aspect of gaming. Studies show

When examining the average college campus, it becomes obvious that students feel rushed from one place to another as they try to participate in class, clubs, and extracurricular activities. One way that students can feel more comfortable and relaxed around campus is to introduce the aspect of gaming. Studies show that “Moderate videogame play has been found to contribute to emotional stability” (Jones, 2014). This demonstrates that the stress of college can be mitigated by introducing the ability to interact with video games. This same concept has been applied in the workplace, where studies have shown that “Gaming principles such as challenges, competition, rewards and personalization keep employees engaged and learning” (Clark, 2020). This means that if we manage to gamify the college experience, students will be more engaged which will increase and stabilize the retention rate of colleges which utilize this type of experience. Gaming allows students to connect with their peers in a casual environment while also allowing them to find resources around campus and find new places to eat and relax. We plan to gamify the college experience by introducing augmented reality in the form of an app. Augmented reality is “. . . a technology that combines virtual information with the real world” (Chen, 2019). College students will be able to utilize the resources and amenities available to them on campus while completing quests that help them within the application. This demonstrates the ability for video games to engage students using artificial tasks but real actions and experiences which help them feel more connected to campus. Our Founders Lab team has developed and tested an AR application that can be used to connect students with their campus and the resources available to them.

ContributorsRangarajan, Padmapriya (Co-author) / Klein, Jonathan (Co-author) / Li, Shimei (Co-author) / Byrne, Jared (Thesis director) / Pierce, John (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
135660-Thumbnail Image.png
Description
This paper presents work that was done to create a system capable of facial expression recognition (FER) using deep convolutional neural networks (CNNs) and test multiple configurations and methods. CNNs are able to extract powerful information about an image using multiple layers of generic feature detectors. The extracted information can

This paper presents work that was done to create a system capable of facial expression recognition (FER) using deep convolutional neural networks (CNNs) and test multiple configurations and methods. CNNs are able to extract powerful information about an image using multiple layers of generic feature detectors. The extracted information can be used to understand the image better through recognizing different features present within the image. Deep CNNs, however, require training sets that can be larger than a million pictures in order to fine tune their feature detectors. For the case of facial expression datasets, none of these large datasets are available. Due to this limited availability of data required to train a new CNN, the idea of using naïve domain adaptation is explored. Instead of creating and using a new CNN trained specifically to extract features related to FER, a previously trained CNN originally trained for another computer vision task is used. Work for this research involved creating a system that can run a CNN, can extract feature vectors from the CNN, and can classify these extracted features. Once this system was built, different aspects of the system were tested and tuned. These aspects include the pre-trained CNN that was used, the layer from which features were extracted, normalization used on input images, and training data for the classifier. Once properly tuned, the created system returned results more accurate than previous attempts on facial expression recognition. Based on these positive results, naïve domain adaptation is shown to successfully leverage advantages of deep CNNs for facial expression recognition.
ContributorsEusebio, Jose Miguel Ang (Author) / Panchanathan, Sethuraman (Thesis director) / McDaniel, Troy (Committee member) / Venkateswara, Hemanth (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2016-05
136785-Thumbnail Image.png
Description
This paper presents the design and evaluation of a haptic interface for augmenting human-human interpersonal interactions by delivering facial expressions of an interaction partner to an individual who is blind using a visual-to-tactile mapping of facial action units and emotions. Pancake shaftless vibration motors are mounted on the back of

This paper presents the design and evaluation of a haptic interface for augmenting human-human interpersonal interactions by delivering facial expressions of an interaction partner to an individual who is blind using a visual-to-tactile mapping of facial action units and emotions. Pancake shaftless vibration motors are mounted on the back of a chair to provide vibrotactile stimulation in the context of a dyadic (one-on-one) interaction across a table. This work explores the design of spatiotemporal vibration patterns that can be used to convey the basic building blocks of facial movements according to the Facial Action Unit Coding System. A behavioral study was conducted to explore the factors that influence the naturalness of conveying affect using vibrotactile cues.
ContributorsBala, Shantanu (Author) / Panchanathan, Sethuraman (Thesis director) / McDaniel, Troy (Committee member) / Barrett, The Honors College (Contributor) / Computer Science and Engineering Program (Contributor) / Department of Psychology (Contributor)
Created2014-05
147553-Thumbnail Image.png
Description

Obesity rates among adults have steadily grown in recent decades all the way up to<br/>42.4% in 2018. This is a 12% increase from the turn of the century (Center for Disease Control<br/>and Prevention, 2021). A major reason for this rise is increased consumption of processed,<br/>high-calorie foods. People eat these foods

Obesity rates among adults have steadily grown in recent decades all the way up to<br/>42.4% in 2018. This is a 12% increase from the turn of the century (Center for Disease Control<br/>and Prevention, 2021). A major reason for this rise is increased consumption of processed,<br/>high-calorie foods. People eat these foods at a young age and develop bad eating habits that can<br/>last for the rest of their lives. It is essential to intervene early and help adolescents form<br/>balanced, healthy eating habits before bad habits are already formed. Our solution to this<br/>problem is Green Gamers. Green Gamers combines adolescent’s passion for gaming with<br/>healthy eating via in-game rewards for healthy eating. People will be able to purchase healthy<br/>food items, such as a bag of carrots, and on the packaging there will be a QR code. They will<br/>then be able to scan the code on our website, and earn points which will unlock in-game items<br/>and other rewards. Video game rewards act as effective motivators for you people to eat more<br/>healthy foods. After the solution was formulated, a preliminary survey was conducted to<br/>confirm that video game related rewards would inspire children to eat more healthy foods.<br/>Based on those results, we are currently in the process of running a secondary market research<br/>campaign to learn if gift card rewards are a stronger motivator. Our end goal for Green Gamers<br/>would be to partner with large gaming studios and food producers. This would allow us access to<br/>many gaming franchises, so that rewards are available from a wide variety of games: making the<br/>platform appealing to a diverse audience of gamers. Similarly, a relationship with large food<br/>producers would give us the ability to place QR codes on a greater assortment of healthy food<br/>items. Although no relationships with large companies have been forged yet, we plan to utilize<br/>funding to test our concept on small focus groups in schools

ContributorsKim, Hwan (Co-author) / Wong, Brendan (Co-author) / Davis, Ben (Co-author) / Mckearney, Jack (Co-author) / Byrne, Jared (Thesis director) / Hall, Rick (Committee member) / Computer Science and Engineering Program (Contributor) / Dean, W.P. Carey School of Business (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
131274-Thumbnail Image.png
Description
Emotion recognition in conversation has applications within numerous domains such as affective computing and medicine. Recent methods for emotion recognition jointly utilize conversational data over several modalities including audio, video, and text. However, state-of-the-art frameworks for this task do not focus on the feature extraction and feature fusion steps of

Emotion recognition in conversation has applications within numerous domains such as affective computing and medicine. Recent methods for emotion recognition jointly utilize conversational data over several modalities including audio, video, and text. However, state-of-the-art frameworks for this task do not focus on the feature extraction and feature fusion steps of this process. This thesis aims to improve the state-of-the-art method by incorporating two components to better accomplish these steps. By doing so, we are able to produce improved representations for the text modality and better model the relationships between all modalities. This paper proposes two methods which focus on these concepts and provide improved accuracy over the state-of-the-art framework for multimodal emotion recognition in dialogue.
ContributorsRawal, Siddharth (Author) / Baral, Chitta (Thesis director) / Shah, Shrikant (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2020-05
131135-Thumbnail Image.png
Description
Accurate pose initialization and pose estimation are crucial requirements in on-orbit space assembly and various other autonomous on-orbit tasks. However, pose initialization and pose estimation are much more difficult to do accurately and consistently in space. This is primarily due to not only the variable lighting conditions present in space,

Accurate pose initialization and pose estimation are crucial requirements in on-orbit space assembly and various other autonomous on-orbit tasks. However, pose initialization and pose estimation are much more difficult to do accurately and consistently in space. This is primarily due to not only the variable lighting conditions present in space, but also the power requirements mandated by space-flyable hardware. This thesis investigates leveraging a deep learning approach for monocular one-shot pose initialization and pose estimation. A convolutional neural network was used to estimate the 6D pose of an assembly truss object. This network was trained by utilizing synthetic imagery generated from a simulation testbed. Furthermore, techniques to quantify model uncertainty of the deep learning model were investigated and applied in the task of in-space pose estimation and pose initialization. The feasibility of this approach on low-power computational platforms was also tested. The results demonstrate that accurate pose initialization and pose estimation can be conducted using a convolutional neural network. In addition, the results show that the model uncertainty can be obtained from the network. Lastly, the use of deep learning for pose initialization and pose estimation in addition with uncertainty quantification was demonstrated to be feasible on low-power compute platforms.
ContributorsKailas, Siva Maneparambil (Author) / Ben Amor, Heni (Thesis director) / Detry, Renaud (Committee member) / Economics Program in CLAS (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2020-05
131461-Thumbnail Image.png
Description
Spotify, one of the most popular music streaming services, has many
algorithms for recommending new music to users. However, at the
core of their recommendations is the collaborative filtering algorithm,
which recommends music based on what other people with similar
tastes have listened to [1]. While this can produce highly relevant
content recommendations, it tends

Spotify, one of the most popular music streaming services, has many
algorithms for recommending new music to users. However, at the
core of their recommendations is the collaborative filtering algorithm,
which recommends music based on what other people with similar
tastes have listened to [1]. While this can produce highly relevant
content recommendations, it tends to promote only popular content
[2]. The popularity bias inherent in collaborative-filtering based
systems can overlook music that fits a user’s taste, simply because
nobody else is listening to it. One possible solution to this problem is
to recommend music based on features of the music itself, and
recommend songs which have similar features. Here, a method for
extracting high-level features representing the mood of a song is
presented, with the aim of tailoring music recommendations to an
individual's mood, and providing music recommendations with
diversity in popularity.
ContributorsGomez, Luis Angel (Author) / Kevin, Burger (Thesis director) / Alberto, Hernández (Committee member) / Arts, Media and Engineering Sch T (Contributor) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2020-05
131212-Thumbnail Image.png
Description
In recent years, the development of new Machine Learning models has allowed for new technological advancements to be introduced for practical use across the world. Multiple studies and experiments have been conducted to create new variations of Machine Learning models with different algorithms to determine if potential systems would prove

In recent years, the development of new Machine Learning models has allowed for new technological advancements to be introduced for practical use across the world. Multiple studies and experiments have been conducted to create new variations of Machine Learning models with different algorithms to determine if potential systems would prove to be successful. Even today, there are still many research initiatives that are continuing to develop new models in the hopes to discover potential solutions for problems such as autonomous driving or determining the emotional value from a single sentence. One of the current popular research topics for Machine Learning is the development of Facial Expression Recognition systems. These Machine Learning models focus on classifying images of human faces that are expressing different emotions through facial expressions. In order to develop effective models to perform Facial Expression Recognition, researchers have gone on to utilize Deep Learning models, which are a more advanced implementation of Machine Learning models, known as Neural Networks. More specifically, the use of Convolutional Neural Networks has proven to be the most effective models for achieving highly accurate results at classifying images of various facial expressions. Convolutional Neural Networks are Deep Learning models that are capable of processing visual data, such as images and videos, and can be used to identify various facial expressions. The purpose of this project, I focused on learning about the important concepts of Machine Learning, Deep Learning, and Convolutional Neural Networks to implement a Convolutional Neural Network that was previously developed by a recommended research paper.
ContributorsFrace, Douglas R (Author) / Demakethepalli Venkateswara, Hemanth Kumar (Thesis director) / McDaniel, Troy (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2020-05
130975-Thumbnail Image.png
Description
Immunotherapy is an effective treatment for cancer which enables the patient's immune system to recognize tumor cells as pathogens. In order to design an individualized treatment, the t cell receptors (TCR) which bind to a tumor's unique antigens need to be determined. We created a convolutional neural network to predict

Immunotherapy is an effective treatment for cancer which enables the patient's immune system to recognize tumor cells as pathogens. In order to design an individualized treatment, the t cell receptors (TCR) which bind to a tumor's unique antigens need to be determined. We created a convolutional neural network to predict the binding affinity between a given TCR and antigen to enable this.
ContributorsCai, Michael Ray (Author) / Lee, Heewook (Thesis director) / Meuth, Ryan (Committee member) / Computer Science and Engineering Program (Contributor, Contributor) / Barrett, The Honors College (Contributor)
Created2020-12
132574-Thumbnail Image.png
Description
Convolutional neural networks boast a myriad of applications in artificial intelligence, but one of the most common uses for such networks is image extraction. The ability of convolutional layers to extract and combine data features for the purpose of image analysis can be leveraged for pose estimation on an object

Convolutional neural networks boast a myriad of applications in artificial intelligence, but one of the most common uses for such networks is image extraction. The ability of convolutional layers to extract and combine data features for the purpose of image analysis can be leveraged for pose estimation on an object - detecting the presence and attitude of corners and edges allows a convolutional neural network to identify how an object is positioned. This task can assist in working to grasp an object correctly in robotics applications, or to track an object more accurately in 3D space. However, the effectiveness of pose estimation may change based on properties of the object; the pose of a complex object, complexity being determined by internal occlusions, similar faces, etcetera, can be difficult to resolve.
This thesis is part of a collaboration between ASU’s Interactive Robotics Laboratory and NASA’s Jet Propulsion Laboratory. In this thesis, the training pipeline from Sharma’s paper “Pose Estimation for Non-Cooperative Spacecraft Rendezvous Using Convolutional Neural Networks” was modified to perform pose estimation on a complex object - specifically, a segment of a hollow truss. After initial attempts to replicate the architecture used in the paper and train solely on synthetic images, a combination of synthetic dataset generation and transfer learning on an ImageNet-pretrained AlexNet model was implemented to mitigate the difficulty of gathering large amounts of real-world data. Experimentation with pose estimation accuracy and hyperparameters of the model resulted in gradual test accuracy improvement, and future work is suggested to improve pose estimation for complex objects with some form of rotational symmetry.
ContributorsDsouza, Susanna Roshini (Author) / Ben Amor, Hani (Thesis director) / Maneparambil, Kailasnath (Committee member) / Computer Science and Engineering Program (Contributor, Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2019-05