Matching Items (25)
Filtering by

Clear all filters

148075-Thumbnail Image.png
Description

University Devils is a Founders Lab Thesis group looking to find a way for post-secondary institutions to increase the number of and diversity of incoming applications through the utilization of gaming and gaming approaches in the recruitment process while staying low-cost. This propelling question guided the group through their work.

University Devils is a Founders Lab Thesis group looking to find a way for post-secondary institutions to increase the number of and diversity of incoming applications through the utilization of gaming and gaming approaches in the recruitment process while staying low-cost. This propelling question guided the group through their work. The team’s work primarily focused on recruitment efforts at Arizona State University, but the concept can be modified and applied at other post-secondary institutions. The initial research showed that Arizona State University’s recruitment focused on visiting the high schools of prospective students and providing campus tours to interested students. A proposed alternative solution to aid in recruitment efforts through the utilization of gaming was to create an online multiplayer game that prospective students could play from their own homes. The basic premise of the game is that one player is selected to be “the Professor” while the other players are part of “the Students.” To complete the game, the Students must complete a set of tasks while the Professor applies various obstacles to prevent the Students from winning. When a Student completes their objectives, they win and the game ends. The game was created using Unity. The group has completed a proof-of-concept of the proposed game and worked to advertise and market the game to students via social media. The team’s efforts have gained traction, and the group continues to work to gain traction and bring the idea to more prospective students.

ContributorsDong, Edmund Engsun (Co-author) / Ouellette, Abigail (Co-author) / Cole, Tyler (Co-author) / Byrne, Jared (Thesis director) / Pierce, John (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
136312-Thumbnail Image.png
Description
While not officially recognized as an addictive activity by the Diagnostic and Statistical Manual of Mental Disorders, video game addiction has well-documented resources pointing to its effects on physiological and mental health for both addict and those close to the addict. With the rise of eSports, treating video game addiction

While not officially recognized as an addictive activity by the Diagnostic and Statistical Manual of Mental Disorders, video game addiction has well-documented resources pointing to its effects on physiological and mental health for both addict and those close to the addict. With the rise of eSports, treating video game addiction has become trickier as a passionate and growing fan base begins to act as a culture not unlike traditional sporting. These concerns call for a better understanding of what constitutes a harmful addiction to video games as its heavy practice becomes more financially viable and accepted into mainstream culture.
ContributorsGohil, Abhishek Bhagirathsinh (Author) / Kashiwagi, Dean (Thesis director) / Kashiwagi, Jacob (Committee member) / Barrett, The Honors College (Contributor) / Computer Science and Engineering Program (Contributor)
Created2015-05
135660-Thumbnail Image.png
Description
This paper presents work that was done to create a system capable of facial expression recognition (FER) using deep convolutional neural networks (CNNs) and test multiple configurations and methods. CNNs are able to extract powerful information about an image using multiple layers of generic feature detectors. The extracted information can

This paper presents work that was done to create a system capable of facial expression recognition (FER) using deep convolutional neural networks (CNNs) and test multiple configurations and methods. CNNs are able to extract powerful information about an image using multiple layers of generic feature detectors. The extracted information can be used to understand the image better through recognizing different features present within the image. Deep CNNs, however, require training sets that can be larger than a million pictures in order to fine tune their feature detectors. For the case of facial expression datasets, none of these large datasets are available. Due to this limited availability of data required to train a new CNN, the idea of using naïve domain adaptation is explored. Instead of creating and using a new CNN trained specifically to extract features related to FER, a previously trained CNN originally trained for another computer vision task is used. Work for this research involved creating a system that can run a CNN, can extract feature vectors from the CNN, and can classify these extracted features. Once this system was built, different aspects of the system were tested and tuned. These aspects include the pre-trained CNN that was used, the layer from which features were extracted, normalization used on input images, and training data for the classifier. Once properly tuned, the created system returned results more accurate than previous attempts on facial expression recognition. Based on these positive results, naïve domain adaptation is shown to successfully leverage advantages of deep CNNs for facial expression recognition.
ContributorsEusebio, Jose Miguel Ang (Author) / Panchanathan, Sethuraman (Thesis director) / McDaniel, Troy (Committee member) / Venkateswara, Hemanth (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2016-05
137137-Thumbnail Image.png
Description
Speech recognition in games is rarely seen. This work presents a project, a 2D computer game named "The Emblems" which utilizes speech recognition as input. The game itself is a two person strategy game whose goal is to defeat the opposing player's army. This report focuses on the speech-recognition aspect

Speech recognition in games is rarely seen. This work presents a project, a 2D computer game named "The Emblems" which utilizes speech recognition as input. The game itself is a two person strategy game whose goal is to defeat the opposing player's army. This report focuses on the speech-recognition aspect of the project. The players interact on a turn-by-turn basis by speaking commands into the computer's microphone. When the computer recognizes a command, it will respond accordingly by having the player's unit perform an action on screen.
ContributorsNguyen, Jordan Ngoc (Author) / Kobayashi, Yoshihiro (Thesis director) / Maciejewski, Ross (Committee member) / Barrett, The Honors College (Contributor) / Computing and Informatics Program (Contributor) / Computer Science and Engineering Program (Contributor)
Created2014-05
137149-Thumbnail Image.png
Description
The project, "The Emblems: OpenGL" is a 2D strategy game that incorporates Speech Recognition for control and OpenGL for computer graphics. Players control their own army by voice commands and try to eliminate the opponent's army. This report focuses on the 2D art and visual aspects of the project. There

The project, "The Emblems: OpenGL" is a 2D strategy game that incorporates Speech Recognition for control and OpenGL for computer graphics. Players control their own army by voice commands and try to eliminate the opponent's army. This report focuses on the 2D art and visual aspects of the project. There are different sprites for the player's army units and icons within the game. The game also has a grid for easy unit placement.
ContributorsHsia, Allen (Author) / Kobayashi, Yoshihiro (Thesis director) / Maciejewski, Ross (Committee member) / Barrett, The Honors College (Contributor) / Computer Science and Engineering Program (Contributor)
Created2014-05
148262-Thumbnail Image.png
Description

This thesis is based on bringing together three different components: non-Euclidean geometric worlds, virtual reality, and environmental puzzles in video games. While all three exist in their own right in the world of video games, as well as combined in pairs, there are virtually no examples of all three together.

This thesis is based on bringing together three different components: non-Euclidean geometric worlds, virtual reality, and environmental puzzles in video games. While all three exist in their own right in the world of video games, as well as combined in pairs, there are virtually no examples of all three together. Non-Euclidean environmental puzzle games have existed for around 10 years in various forms, short environmental puzzle games in virtual reality have come into existence in around the past five years, and non-Euclidean virtual reality exists mainly as non-video game short demos from the past few years. This project seeks to be able to bring these components together to create a proof of concept for how a game like this should function, particularly the integration of non-Euclidean virtual reality in the context of a video game. To do this, a Unity package which uses a custom system for creating worlds in a non-Euclidean way rather than Unity’s built-in components such as for transforms, collisions, and rendering was used. This was used in conjunction with the SteamVR implementation with Unity to create a cohesive and immersive player experience.

ContributorsVerhagen, Daniel William (Author) / Kobayashi, Yoshihiro (Thesis director) / Nelson, Brian (Committee member) / Computer Science and Engineering Program (Contributor, Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
131274-Thumbnail Image.png
Description
Emotion recognition in conversation has applications within numerous domains such as affective computing and medicine. Recent methods for emotion recognition jointly utilize conversational data over several modalities including audio, video, and text. However, state-of-the-art frameworks for this task do not focus on the feature extraction and feature fusion steps of

Emotion recognition in conversation has applications within numerous domains such as affective computing and medicine. Recent methods for emotion recognition jointly utilize conversational data over several modalities including audio, video, and text. However, state-of-the-art frameworks for this task do not focus on the feature extraction and feature fusion steps of this process. This thesis aims to improve the state-of-the-art method by incorporating two components to better accomplish these steps. By doing so, we are able to produce improved representations for the text modality and better model the relationships between all modalities. This paper proposes two methods which focus on these concepts and provide improved accuracy over the state-of-the-art framework for multimodal emotion recognition in dialogue.
ContributorsRawal, Siddharth (Author) / Baral, Chitta (Thesis director) / Shah, Shrikant (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2020-05
131135-Thumbnail Image.png
Description
Accurate pose initialization and pose estimation are crucial requirements in on-orbit space assembly and various other autonomous on-orbit tasks. However, pose initialization and pose estimation are much more difficult to do accurately and consistently in space. This is primarily due to not only the variable lighting conditions present in space,

Accurate pose initialization and pose estimation are crucial requirements in on-orbit space assembly and various other autonomous on-orbit tasks. However, pose initialization and pose estimation are much more difficult to do accurately and consistently in space. This is primarily due to not only the variable lighting conditions present in space, but also the power requirements mandated by space-flyable hardware. This thesis investigates leveraging a deep learning approach for monocular one-shot pose initialization and pose estimation. A convolutional neural network was used to estimate the 6D pose of an assembly truss object. This network was trained by utilizing synthetic imagery generated from a simulation testbed. Furthermore, techniques to quantify model uncertainty of the deep learning model were investigated and applied in the task of in-space pose estimation and pose initialization. The feasibility of this approach on low-power computational platforms was also tested. The results demonstrate that accurate pose initialization and pose estimation can be conducted using a convolutional neural network. In addition, the results show that the model uncertainty can be obtained from the network. Lastly, the use of deep learning for pose initialization and pose estimation in addition with uncertainty quantification was demonstrated to be feasible on low-power compute platforms.
ContributorsKailas, Siva Maneparambil (Author) / Ben Amor, Heni (Thesis director) / Detry, Renaud (Committee member) / Economics Program in CLAS (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2020-05
131140-Thumbnail Image.png
Description
Intelli-Trail is a game where the player plays as a small blue man with the simple goal of reaching the purple door. The player will primarily interact with the game through combat. The game itself will react to the patterns in the players behavior to progressively become harder

Intelli-Trail is a game where the player plays as a small blue man with the simple goal of reaching the purple door. The player will primarily interact with the game through combat. The game itself will react to the patterns in the players behavior to progressively become harder for the player to win.
ContributorsCurtin, Ethan (Author) / Gonzalez Sanchez, Javier (Thesis director) / Baron, Tyler (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2020-05
131461-Thumbnail Image.png
Description
Spotify, one of the most popular music streaming services, has many
algorithms for recommending new music to users. However, at the
core of their recommendations is the collaborative filtering algorithm,
which recommends music based on what other people with similar
tastes have listened to [1]. While this can produce highly relevant
content recommendations, it tends

Spotify, one of the most popular music streaming services, has many
algorithms for recommending new music to users. However, at the
core of their recommendations is the collaborative filtering algorithm,
which recommends music based on what other people with similar
tastes have listened to [1]. While this can produce highly relevant
content recommendations, it tends to promote only popular content
[2]. The popularity bias inherent in collaborative-filtering based
systems can overlook music that fits a user’s taste, simply because
nobody else is listening to it. One possible solution to this problem is
to recommend music based on features of the music itself, and
recommend songs which have similar features. Here, a method for
extracting high-level features representing the mood of a song is
presented, with the aim of tailoring music recommendations to an
individual's mood, and providing music recommendations with
diversity in popularity.
ContributorsGomez, Luis Angel (Author) / Kevin, Burger (Thesis director) / Alberto, Hernández (Committee member) / Arts, Media and Engineering Sch T (Contributor) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2020-05