Matching Items (23)
Filtering by

Clear all filters

155515-Thumbnail Image.png
Description
Currently, educational games are designed with the educational content as the primary factor driving the design of the game. While this may seem to be the optimal approach, this design paradigm causes multiple issues. For one, the games themselves are often not engaging as game design principles were put aside

Currently, educational games are designed with the educational content as the primary factor driving the design of the game. While this may seem to be the optimal approach, this design paradigm causes multiple issues. For one, the games themselves are often not engaging as game design principles were put aside in favor of increasing the educational value of the game. The other issue is that the code base of the game is mostly or completely unusable for any other games as the game mechanics are too strongly connected to the educational content being taught. This means that the mechanics are impossible to reuse in future projects without major revisions, and starting over is often more time and cost efficient.

This thesis presents the Content Agnostic Game Engineering (CAGE) model for designing educational games. CAGE is a way to separate the educational content from the game mechanics without compromising the educational value of the game. This is done by designing mechanics that can have multiple educational contents layered on top of them which can be switched out at any time. CAGE allows games to be designed with a game design first approach which allows them to maintain higher engagement levels. In addition, since the mechanics are not tied to the educational content several different educational topics can reuse the same set of mechanics without requiring major revisions to the existing code.

Results show that CAGE greatly reduces the amount of code needed to make additional versions of educational games, and speeds up the development process. The CAGE model is also shown to not induce high levels of cognitive load, allowing for more in depth topic work than was attempted in this thesis. However, engagement was low and switching the active content does interrupt the game flow considerably. Altering the difficulty of the game in real time in response to the affective state of the player was not shown to increase engagement. Potential causes of the issues with CAGE games and potential fixes are discussed.
ContributorsBaron, Tyler John (Author) / Amresh, Ashish (Thesis advisor) / Nelson, Brian C (Committee member) / Niemczyk, Mary (Committee member) / Arizona State University (Publisher)
Created2017
171730-Thumbnail Image.png
Description
Emerging body movement detection and gesture recognition software have opened a gateway of possibilities to make technology more intuitive, engaging, and accessible for people. A vast areaof natural user interfaces is leveraging body motion tracking and gesture recognition technologies and a human’s readily expressive body to extend interactions with software

Emerging body movement detection and gesture recognition software have opened a gateway of possibilities to make technology more intuitive, engaging, and accessible for people. A vast areaof natural user interfaces is leveraging body motion tracking and gesture recognition technologies and a human’s readily expressive body to extend interactions with software beyond mouse clicks and scrolls. However, these interfaces have been limited by hardware and software expenses, high development time and costs, and learning curves. This paper explores different approaches to providing both software developers and designers with easier ways to incorporate computer vision-based body and gesture detection solutions into the development of embodied experiences without suppressing creativity. Gesture.js is a JavaScript framework as a service (FaaS) that is both a thin library on top of the Document Object Model (DOM) consisting of a collection of tools for developing embodied-enabled applications on the web and a landmark computation and processing application programming interface. It wraps MediaPipe, an open-source collection of machine-learning solutions that perform inference over arbitrary sensory data, and additional landmark processing frameworks such as KalidoKit, a 3D model rigging solution, and ports the necessary information through either an object-oriented or an API-oriented implementation. It also comes with its web-based graphical interface for easy connection between Gesture.js and other application clients with little to no JavaScript code. This thesis also details a collection of example applications that demonstrate the usability, capacity, and potential of this framework.
ContributorsFowler, Azaria (Author) / Gowda, Tejaswi (Thesis advisor) / Kuznetsov, Anastasia (Committee member) / Kobayashi, Yoshihiro (Committee member) / Arizona State University (Publisher)
Created2022
158908-Thumbnail Image.png
Description
While significant qualitative, user study-focused research has been done on augmented reality, relatively few studies have been conducted on multiple, co-located synchronously collaborating users in augmented reality. Recognizing the need for more collaborative user studies in augmented reality and the value such studies present, a user study is conducted of

While significant qualitative, user study-focused research has been done on augmented reality, relatively few studies have been conducted on multiple, co-located synchronously collaborating users in augmented reality. Recognizing the need for more collaborative user studies in augmented reality and the value such studies present, a user study is conducted of collaborative decision-making in augmented reality to investigate the following research question: “Does presenting data visualizations in augmented reality influence the collaborative decision-making behaviors of a team?” This user study evaluates how viewing data visualizations with augmented reality headsets impacts collaboration in small teams compared to viewing together on a single 2D desktop monitor as a baseline. Teams of two participants performed closed and open-ended evaluation tasks to collaboratively analyze data visualized in both augmented reality and on a desktop monitor. Multiple means of collecting and analyzing data were employed to develop a well-rounded context for results and conclusions, including software logging of participant interactions, qualitative analysis of video recordings of participant sessions, and pre- and post-study participant questionnaires. The results indicate that augmented reality doesn’t significantly change the quantity of team member communication but does impact the means and strategies participants use to collaborate.
ContributorsKintscher, Michael (Author) / Bryan, Chris (Thesis advisor) / Amresh, Ashish (Thesis advisor) / Hansford, Dianne (Committee member) / Johnson, Erik (Committee member) / Arizona State University (Publisher)
Created2020