Matching Items (26)
Filtering by

Clear all filters

156468-Thumbnail Image.png
Description
With the emergence of edge computing paradigm, many applications such as image recognition and augmented reality require to perform machine learning (ML) and artificial intelligence (AI) tasks on edge devices. Most AI and ML models are large and computational heavy, whereas edge devices are usually equipped with limited computational and

With the emergence of edge computing paradigm, many applications such as image recognition and augmented reality require to perform machine learning (ML) and artificial intelligence (AI) tasks on edge devices. Most AI and ML models are large and computational heavy, whereas edge devices are usually equipped with limited computational and storage resources. Such models can be compressed and reduced in order to be placed on edge devices, but they may loose their capability and may not generalize and perform well compared to large models. Recent works used knowledge transfer techniques to transfer information from a large network (termed teacher) to a small one (termed student) in order to improve the performance of the latter. This approach seems to be promising for learning on edge devices, but a thorough investigation on its effectiveness is lacking.

The purpose of this work is to provide an extensive study on the performance (both in terms of accuracy and convergence speed) of knowledge transfer, considering different student-teacher architectures, datasets and different techniques for transferring knowledge from teacher to student.

A good performance improvement is obtained by transferring knowledge from both the intermediate layers and last layer of the teacher to a shallower student. But other architectures and transfer techniques do not fare so well and some of them even lead to negative performance impact. For example, a smaller and shorter network, trained with knowledge transfer on Caltech 101 achieved a significant improvement of 7.36\% in the accuracy and converges 16 times faster compared to the same network trained without knowledge transfer. On the other hand, smaller network which is thinner than the teacher network performed worse with an accuracy drop of 9.48\% on Caltech 101, even with utilization of knowledge transfer.
ContributorsSistla, Ragini (Author) / Zhao, Ming (Thesis advisor, Committee member) / Li, Baoxin (Committee member) / Tong, Hanghang (Committee member) / Arizona State University (Publisher)
Created2018
135660-Thumbnail Image.png
Description
This paper presents work that was done to create a system capable of facial expression recognition (FER) using deep convolutional neural networks (CNNs) and test multiple configurations and methods. CNNs are able to extract powerful information about an image using multiple layers of generic feature detectors. The extracted information can

This paper presents work that was done to create a system capable of facial expression recognition (FER) using deep convolutional neural networks (CNNs) and test multiple configurations and methods. CNNs are able to extract powerful information about an image using multiple layers of generic feature detectors. The extracted information can be used to understand the image better through recognizing different features present within the image. Deep CNNs, however, require training sets that can be larger than a million pictures in order to fine tune their feature detectors. For the case of facial expression datasets, none of these large datasets are available. Due to this limited availability of data required to train a new CNN, the idea of using naïve domain adaptation is explored. Instead of creating and using a new CNN trained specifically to extract features related to FER, a previously trained CNN originally trained for another computer vision task is used. Work for this research involved creating a system that can run a CNN, can extract feature vectors from the CNN, and can classify these extracted features. Once this system was built, different aspects of the system were tested and tuned. These aspects include the pre-trained CNN that was used, the layer from which features were extracted, normalization used on input images, and training data for the classifier. Once properly tuned, the created system returned results more accurate than previous attempts on facial expression recognition. Based on these positive results, naïve domain adaptation is shown to successfully leverage advantages of deep CNNs for facial expression recognition.
ContributorsEusebio, Jose Miguel Ang (Author) / Panchanathan, Sethuraman (Thesis director) / McDaniel, Troy (Committee member) / Venkateswara, Hemanth (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2016-05
Description
Technical innovation has always played a part in live theatre, whether in the form of mechanical pieces like lifts and trapdoors to the more recent integration of digital media. The advances of the art form encourage the development of technology, and at the same time, technological development enables the advancement

Technical innovation has always played a part in live theatre, whether in the form of mechanical pieces like lifts and trapdoors to the more recent integration of digital media. The advances of the art form encourage the development of technology, and at the same time, technological development enables the advancement of theatrical expression. As mechanics, lighting, sound, and visual media have made their way into the spotlight, advances in theatrical robotics continue to push for their inclusion in the director's toolbox. However, much of the technology available is gated by high prices and unintuitive interfaces, designed for large troupes and specialized engineers, making it difficult to access for small schools and students new to the medium. As a group of engineering students with a vested interest in the development of the arts, this thesis team designed a system that will enable troupes from any background to participate in the advent of affordable automation. The intended result of this thesis project was to create a robotic platform that interfaces with custom software, receiving commands and transmitting position data, and to design that software so that a user can define intuitive cues for their shows. In addition, a new pathfinding algorithm was developed to support free-roaming automation in a 2D space. The final product consisted of a relatively inexpensive (< $2000) free-roaming platform, made entirely with COTS and standard materials, and a corresponding control system with cue design, wireless path following, and position tracking. This platform was built to support 1000 lbs, and includes integrated emergency stopping. The software allows for custom cue design, speed variation, and dynamic path following. Both the blueprints and the source code for the platform and control system have been released to open-source repositories, to encourage further development in the area of affordable automation. The platform itself was donated to the ASU School of Theater.
ContributorsHollenbeck, Matthew D. (Co-author) / Wiebel, Griffin (Co-author) / Winnemann, Christopher (Thesis director) / Christensen, Stephen (Committee member) / Computer Science and Engineering Program (Contributor) / School of Film, Dance and Theatre (Contributor) / Barrett, The Honors College (Contributor)
Created2018-05
132967-Thumbnail Image.png
Description
Classical planning is a field of Artificial Intelligence concerned with allowing autonomous agents to make reasonable decisions in complex environments. This work investigates
the application of deep learning and planning techniques, with the aim of constructing generalized plans capable of solving multiple problem instances. We construct a Deep Neural Network that,

Classical planning is a field of Artificial Intelligence concerned with allowing autonomous agents to make reasonable decisions in complex environments. This work investigates
the application of deep learning and planning techniques, with the aim of constructing generalized plans capable of solving multiple problem instances. We construct a Deep Neural Network that, given an abstract problem state, predicts both (i) the best action to be taken from that state and (ii) the generalized “role” of the object being manipulated. The neural network was tested on two classical planning domains: the blocks world domain and the logistic domain. Results indicate that neural networks are capable of making such
predictions with high accuracy, indicating a promising new framework for approaching generalized planning problems.
ContributorsNakhleh, Julia Blair (Author) / Srivastava, Siddharth (Thesis director) / Fainekos, Georgios (Committee member) / Computer Science and Engineering Program (Contributor) / School of International Letters and Cultures (Contributor) / Barrett, The Honors College (Contributor)
Created2019-05
133339-Thumbnail Image.png
Description
Medical records are increasingly being recorded in the form of electronic health records (EHRs), with a significant amount of patient data recorded as unstructured natural language text. Consequently, being able to extract and utilize clinical data present within these records is an important step in furthering clinical care. One important

Medical records are increasingly being recorded in the form of electronic health records (EHRs), with a significant amount of patient data recorded as unstructured natural language text. Consequently, being able to extract and utilize clinical data present within these records is an important step in furthering clinical care. One important aspect within these records is the presence of prescription information. Existing techniques for extracting prescription information — which includes medication names, dosages, frequencies, reasons for taking, and mode of administration — from unstructured text have focused on the application of rule- and classifier-based methods. While state-of-the-art systems can be effective in extracting many types of information, they require significant effort to develop hand-crafted rules and conduct effective feature engineering. This paper presents the use of a bidirectional LSTM with CRF tagging model initialized with precomputed word embeddings for extracting prescription information from sentences without requiring significant feature engineering. The experimental results, run on the i2b2 2009 dataset, achieve an F1 macro measure of 0.8562, and scores above 0.9449 on four of the six categories, indicating significant potential for this model.
ContributorsRawal, Samarth Chetan (Author) / Baral, Chitta (Thesis director) / Anwar, Saadat (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2018-05
135018-Thumbnail Image.png
Description
The software element of home and small business networking solutions has failed to keep pace with annual development of newer and faster hardware. The software running on these devices is an afterthought, oftentimes equipped with minimal features, an obtuse user interface, or both. At the same time, this past year

The software element of home and small business networking solutions has failed to keep pace with annual development of newer and faster hardware. The software running on these devices is an afterthought, oftentimes equipped with minimal features, an obtuse user interface, or both. At the same time, this past year has seen the rise of smart home assistants that represent the next step in human-computer interaction with their advanced use of natural language processing. This project seeks to quell the issues with the former by exploring a possible fusion of a powerful, feature-rich software-defined networking stack and the incredible natural language processing tools of smart home assistants. To accomplish these ends, a piece of software was developed to leverage the powerful natural language processing capabilities of one such smart home assistant, the Amazon Echo. On one end, this software interacts with Amazon Web Services to retrieve information about a user's speech patterns and key information contained in their speech. On the other end, the software joins that information with its previous session state to intelligently translate speech into a series of commands for the separate components of a networking stack. The software developed for this project empowers a user to quickly make changes to several facets of their networking gear or acquire information about it with just their language \u2014 no terminals, java applets, or web configuration interfaces needed, thus circumventing clunky UI's or jumping from shell to shell. It is the author's hope that showing how networking equipment can be configured in this innovative way will draw more attention to the current failings of networking equipment and inspire a new series of intuitive user interfaces.
ContributorsHermens, Ryan Joseph (Author) / Meuth, Ryan (Thesis director) / Burger, Kevin (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2016-12
148325-Thumbnail Image.png
Description

The NBA yields billions of dollars each year and serves as a pastime and hobby for millions of Americans. However, many people do not have the time to watch several 2-hour games every week, especially when only a fraction of the game is actually exciting footage. The goal of Sports

The NBA yields billions of dollars each year and serves as a pastime and hobby for millions of Americans. However, many people do not have the time to watch several 2-hour games every week, especially when only a fraction of the game is actually exciting footage. The goal of Sports Summary is to take the ``fluff'' out of these games and create a distilled summary that includes only the most exciting and relevant events. The Sports Summary model records visual and auditory data, camera angles, and game clock readings and correlates it with the game's play-by-play data. On average, a game of more than 2 hours long is shortened to a summary of less than 20 minutes. This summary is then uploaded to the Sports Summary website, where users can filter by the type of event, giving more autonomy and a more comprehensive viewing experience than highlight reels. Additionally, the website allows for users to submit footage they would like to watch for processing and later viewing. Sports Summary creates an enjoyable and accessible way to watch games.

ContributorsZimmerman, Kenna Marleen (Author) / Espanol, Malena (Thesis director) / Dahlberg, Samantha (Committee member) / Pasha, Mirjeta (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148059-Thumbnail Image.png
Description

Generating an astounding $110.7 billion annually in domestic revenue alone [1], the world of accounting is one deceptively lacking automation of its most business-critical processes. While accounting tools do exist for the common person, especially when it is time to pay their taxes, such innovations scarcely exist for many larger

Generating an astounding $110.7 billion annually in domestic revenue alone [1], the world of accounting is one deceptively lacking automation of its most business-critical processes. While accounting tools do exist for the common person, especially when it is time to pay their taxes, such innovations scarcely exist for many larger industrial tasks. Exceedingly common business events, such as Business Combinations, are surprisingly manual tasks despite their $1.1 trillion valuation in 2020 [2]. This work presents the twin accounting solutions TurboGAAP and TurboIFRS: an unprecedented leap into these murky waters in an attempt to automate and streamline these gigantic accounting tasks once entrusted only to teams of experienced accountants.
A first-to-market approach to a trillion-dollar problem, TurboGAAP and TurboIFRS are the answers for years of demands from the accounting sector that established corporations have never solved.

ContributorsPreston, Michael Ernest (Co-author) / Capuano, Bailey (Co-author) / Kuhler, Madison (Co-author) / Chen, Yinong (Thesis director) / Hunt, Neil (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
147796-Thumbnail Image.png
Description

As much as SARS-CoV-2 has altered the way humans live since the beginning of 2020,<br/>this virus's deadly nature has required clinical testing to meet 2020's demands of higher<br/>throughput, higher accuracy and higher efficiency. Information technology has allowed<br/>institutions, like Arizona State University (ASU), to make strategic and operational changes to<br/>combat the

As much as SARS-CoV-2 has altered the way humans live since the beginning of 2020,<br/>this virus's deadly nature has required clinical testing to meet 2020's demands of higher<br/>throughput, higher accuracy and higher efficiency. Information technology has allowed<br/>institutions, like Arizona State University (ASU), to make strategic and operational changes to<br/>combat the SARS-CoV-2 pandemic. At ASU, information technology was one of the six facets<br/>identified in the ongoing review of the ASU Biodesign Clinical Testing Laboratory (ABCTL)<br/>among business, communications, management/training, law, and clinical analysis. The first<br/>chapter of this manuscript covers the background of clinical laboratory automation and details<br/>the automated laboratory workflow to perform ABCTL’s COVID-19 diagnostic testing. The<br/>second chapter discusses the usability and efficiency of key information technology systems of<br/>the ABCTL. The third chapter explains the role of quality control and data management within<br/>ABCTL’s use of information technology. The fourth chapter highlights the importance of data<br/>modeling and 10 best practices when responding to future public health emergencies.

ContributorsKandan, Mani (Co-author) / Leung, Michael (Co-author) / Woo, Sabrina (Co-author) / Knox, Garrett (Co-author) / Compton, Carolyn (Thesis director) / Dudley, Sean (Committee member) / Computer Science and Engineering Program (Contributor) / Department of Information Systems (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148017-Thumbnail Image.png
Description

Generating an astounding $110.7 billion annually in domestic revenue alone [1], the world of accounting is one deceptively lacking automation of its most business-critical processes. While accounting tools do exist for the common person, especially when it is time to pay their taxes, such innovations scarcely exist for many larger

Generating an astounding $110.7 billion annually in domestic revenue alone [1], the world of accounting is one deceptively lacking automation of its most business-critical processes. While accounting tools do exist for the common person, especially when it is time to pay their taxes, such innovations scarcely exist for many larger industrial tasks. Exceedingly common business events, such as Business Combinations, are surprisingly manual tasks despite their $1.1 trillion valuation in 2020 [2]. This work presents the twin accounting solutions TurboGAAP and TurboIFRS: an unprecedented leap into these murky waters in an attempt to automate and streamline these gigantic accounting tasks once entrusted only to teams of experienced accountants.
A first-to-market approach to a trillion-dollar problem, TurboGAAP and TurboIFRS are the answers for years of demands from the accounting sector that established corporations have never solved.

ContributorsKuhler, Madison Frances (Co-author) / Capuano, Bailey (Co-author) / Preston, Michael (Co-author) / Chen, Yinong (Thesis director) / Hunt, Neil (Committee member) / Computer Science and Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05