Matching Items (291)
Filtering by

Clear all filters

148169-Thumbnail Image.png
Description

This thesis was conducted to study and analyze the fund allocation process adopted by different states in the United States to reduce the impact of the Covid-19 virus. Seven different states and their funding methodologies were compared against the case count within the state. The study also focused on development

This thesis was conducted to study and analyze the fund allocation process adopted by different states in the United States to reduce the impact of the Covid-19 virus. Seven different states and their funding methodologies were compared against the case count within the state. The study also focused on development of a physical distancing index based on three significant attributes. This index was then compared to the expenditure and case counts to support decision making.
A regression model was developed to analyze and compare how different states case counts played out against the regression model and the risk index.

ContributorsJaisinghani, Shaurya (Author) / Mirchandani, Pitu (Thesis director) / Clough, Michael (Committee member) / McCarville, Daniel R. (Committee member) / Industrial, Systems & Operations Engineering Prgm (Contributor) / Department of Information Systems (Contributor) / Industrial, Systems & Operations Engineering Prgm (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
147971-Thumbnail Image.png
Description

This survey takes information on a participant’s beliefs on privacy security, the general digital knowledge, demographics, and willingness-to-pay points on if they would delete information on their social media, to see how an information treatment affects those payment points. This information treatment is meant to make half of the participants

This survey takes information on a participant’s beliefs on privacy security, the general digital knowledge, demographics, and willingness-to-pay points on if they would delete information on their social media, to see how an information treatment affects those payment points. This information treatment is meant to make half of the participants think about the deeper ramifications of the information they reveal. The initial hypothesis is that this information will make people want to pay more to remove their information from the web, but the results find a surprising negative correlation with the treatment.

ContributorsDeitrick, Noah Sumner (Author) / Silverman, Daniel (Thesis director) / Kuminoff, Nicolai (Committee member) / School of Mathematical and Statistical Sciences (Contributor) / Economics Program in CLAS (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
147992-Thumbnail Image.png
Description

The research presented in this Honors Thesis provides development in machine learning models which predict future states of a system with unknown dynamics, based on observations of the system. Two case studies are presented for (1) a non-conservative pendulum and (2) a differential game dictating a two-car uncontrolled intersection scenario.

The research presented in this Honors Thesis provides development in machine learning models which predict future states of a system with unknown dynamics, based on observations of the system. Two case studies are presented for (1) a non-conservative pendulum and (2) a differential game dictating a two-car uncontrolled intersection scenario. In the paper we investigate how learning architectures can be manipulated for problem specific geometry. The result of this research provides that these problem specific models are valuable for accurate learning and predicting the dynamics of physics systems.<br/><br/>In order to properly model the physics of a real pendulum, modifications were made to a prior architecture which was sufficient in modeling an ideal pendulum. The necessary modifications to the previous network [13] were problem specific and not transferrable to all other non-conservative physics scenarios. The modified architecture successfully models real pendulum dynamics. This case study provides a basis for future research in augmenting the symplectic gradient of a Hamiltonian energy function to provide a generalized, non-conservative physics model.<br/><br/>A problem specific architecture was also utilized to create an accurate model for the two-car intersection case. The Costate Network proved to be an improvement from the previously used Value Network [17]. Note that this comparison is applied lightly due to slight implementation differences. The development of the Costate Network provides a basis for using characteristics to decompose functions and create a simplified learning problem.<br/><br/>This paper is successful in creating new opportunities to develop physics models, in which the sample cases should be used as a guide for modeling other real and pseudo physics. Although the focused models in this paper are not generalizable, it is important to note that these cases provide direction for future research.

ContributorsMerry, Tanner (Author) / Ren, Yi (Thesis director) / Zhang, Wenlong (Committee member) / Mechanical and Aerospace Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148001-Thumbnail Image.png
Description

High-entropy alloys possessing mechanical, chemical, and electrical properties that far exceed those of conventional alloys have the potential to make a significant impact on many areas of engineering. Identifying element combinations and configurations to form these alloys, however, is a difficult, time-consuming, computationally intensive task. Machine learning has revolutionized many

High-entropy alloys possessing mechanical, chemical, and electrical properties that far exceed those of conventional alloys have the potential to make a significant impact on many areas of engineering. Identifying element combinations and configurations to form these alloys, however, is a difficult, time-consuming, computationally intensive task. Machine learning has revolutionized many different fields due to its ability to generalize well to different problems and produce computationally efficient, accurate predictions regarding the system of interest. In this thesis, we demonstrate the effectiveness of machine learning models applied to toy cases representative of simplified physics that are relevant to high-entropy alloy simulation. We show these models are effective at learning nonlinear dynamics for single and multi-particle cases and that more work is needed to accurately represent complex cases in which the system dynamics are chaotic. This thesis serves as a demonstration of the potential benefits of machine learning applied to high-entropy alloy simulations to generate fast, accurate predictions of nonlinear dynamics.

ContributorsDaly, John H (Author) / Ren, Yi (Thesis director) / Zhuang, Houlong (Committee member) / Mechanical and Aerospace Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148088-Thumbnail Image.png
Description

Colorimetric assays are an important tool in point-of-care testing that offers several advantages to traditional testing methods such as rapid response times and inexpensive costs. A factor that currently limits the portability and accessibility of these assays are methods that can objectively determine the results of these assays. Current solutions

Colorimetric assays are an important tool in point-of-care testing that offers several advantages to traditional testing methods such as rapid response times and inexpensive costs. A factor that currently limits the portability and accessibility of these assays are methods that can objectively determine the results of these assays. Current solutions consist of creating a test reader that standardizes the conditions the strip is under before being measured in some way. However, this increases the cost and decreases the portability of these assays. The focus of this study is to create a machine learning algorithm that can objectively determine results of colorimetric assays under varying conditions. To ensure the flexibility of a model to several types of colorimetric assays, three models were trained on the same convolutional neural network with different datasets. The images these models are trained on consist of positive and negative images of ETG, fentanyl, and HPV Antibodies test strips taken under different lighting and background conditions. A fourth model is trained on an image set composed of all three strip types. The results from these models show it is able to predict positive and negative results to a high level of accuracy.

ContributorsFisher, Rachel (Author) / Blain Christen, Jennifer (Thesis director) / Anderson, Karen (Committee member) / School of Life Sciences (Contributor) / Harrington Bioengineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
147514-Thumbnail Image.png
Description

Since the global financial crisis of 2007-8, interest in worker-cooperatives and alternative forms of organization has surged. Mondragon, located in the Basque region of Spain, represents the largest federation of worker-cooperatives around the world, consisting of 98 cooperatives and 143 subsidiaries, which earned a total revenue of $14.5 billion in

Since the global financial crisis of 2007-8, interest in worker-cooperatives and alternative forms of organization has surged. Mondragon, located in the Basque region of Spain, represents the largest federation of worker-cooperatives around the world, consisting of 98 cooperatives and 143 subsidiaries, which earned a total revenue of $14.5 billion in 2019. While previous attempts to establish a similar model have historically reached limited success, Mondragon has achieved a unique balance of remaining economically viable, on the one hand, and staying true to its founding principles of democratic governance, on the other. This paper sets out to analyze the democratic structure and the cooperative culture at the heart of the Mondragon model, as well as the new type of human relationship that it fosters. In particular, this relationship is one in which individual well-being is bound up with communal well-being that avoids the antagonistic clash between the capital and labor.

ContributorsSvejda, Sam (Author) / Simhony, Avital (Thesis director) / McNamara, Peter (Committee member) / Department of Economics (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
147516-Thumbnail Image.png
Description

Lithium ion batteries are quintessential components of modern life. They are used to power smart devices — phones, tablets, laptops, and are rapidly becoming major elements in the automotive industry. Demand projections for lithium are skyrocketing with production struggling to keep up pace. This drive is due mostly to the

Lithium ion batteries are quintessential components of modern life. They are used to power smart devices — phones, tablets, laptops, and are rapidly becoming major elements in the automotive industry. Demand projections for lithium are skyrocketing with production struggling to keep up pace. This drive is due mostly to the rapid adoption of electric vehicles; sales of electric vehicles in 2020 are more than double what they were only a year prior. With such staggering growth it is important to understand how lithium is sourced and what that means for the environment. Will production even be capable of meeting the demand as more industries make use of this valuable element? How will the environmental impact of lithium affect growth? This thesis attempts to answer these questions as the world looks to a decade of rapid growth for lithium ion batteries.

ContributorsMelton, John (Author) / Brian, Jennifer (Thesis director) / Karwat, Darshawn (Committee member) / Chemical Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
151802-Thumbnail Image.png
Description
The complexity of the systems that software engineers build has continuously grown since the inception of the field. What has not changed is the engineers' mental capacity to operate on about seven distinct pieces of information at a time. The widespread use of UML has led to more abstract software

The complexity of the systems that software engineers build has continuously grown since the inception of the field. What has not changed is the engineers' mental capacity to operate on about seven distinct pieces of information at a time. The widespread use of UML has led to more abstract software design activities, however the same cannot be said for reverse engineering activities. The introduction of abstraction to reverse engineering will allow the engineer to move farther away from the details of the system, increasing his ability to see the role that domain level concepts play in the system. In this thesis, we present a technique that facilitates filtering of classes from existing systems at the source level based on their relationship to concepts in the domain via a classification method using machine learning. We showed that concepts can be identified using a machine learning classifier based on source level metrics. We developed an Eclipse plugin to assist with the process of manually classifying Java source code, and collecting metrics and classifications into a standard file format. We developed an Eclipse plugin to act as a concept identifier that visually indicates a class as a domain concept or not. We minimized the size of training sets to ensure a useful approach in practice. This allowed us to determine that a training set of 7:5 to 10% is nearly as effective as a training set representing 50% of the system. We showed that random selection is the most consistent and effective means of selecting a training set. We found that KNN is the most consistent performer among the learning algorithms tested. We determined the optimal feature set for this classification problem. We discussed two possible structures besides a one to one mapping of domain knowledge to implementation. We showed that classes representing more than one concept are simply concepts at differing levels of abstraction. We also discussed composite concepts representing a domain concept implemented by more than one class. We showed that these composite concepts are difficult to detect because the problem is NP-complete.
ContributorsCarey, Maurice (Author) / Colbourn, Charles (Thesis advisor) / Collofello, James (Thesis advisor) / Davulcu, Hasan (Committee member) / Sarjoughian, Hessam S. (Committee member) / Ye, Jieping (Committee member) / Arizona State University (Publisher)
Created2013
151627-Thumbnail Image.png
Description
Text classification, in the artificial intelligence domain, is an activity in which text documents are automatically classified into predefined categories using machine learning techniques. An example of this is classifying uncategorized news articles into different predefined categories such as "Business", "Politics", "Education", "Technology" , etc. In this thesis, supervised machine

Text classification, in the artificial intelligence domain, is an activity in which text documents are automatically classified into predefined categories using machine learning techniques. An example of this is classifying uncategorized news articles into different predefined categories such as "Business", "Politics", "Education", "Technology" , etc. In this thesis, supervised machine learning approach is followed, in which a module is first trained with pre-classified training data and then class of test data is predicted. Good feature extraction is an important step in the machine learning approach and hence the main component of this text classifier is semantic triplet based features in addition to traditional features like standard keyword based features and statistical features based on shallow-parsing (such as density of POS tags and named entities). Triplet {Subject, Verb, Object} in a sentence is defined as a relation between subject and object, the relation being the predicate (verb). Triplet extraction process, is a 5 step process which takes input corpus as a web text document(s), each consisting of one or many paragraphs, from RSS feeds to lists of extremist website. Input corpus feeds into the "Pronoun Resolution" step, which uses an heuristic approach to identify the noun phrases referenced by the pronouns. The next step "SRL Parser" is a shallow semantic parser and converts the incoming pronoun resolved paragraphs into annotated predicate argument format. The output of SRL parser is processed by "Triplet Extractor" algorithm which forms the triplet in the form {Subject, Verb, Object}. Generalization and reduction of triplet features is the next step. Reduced feature representation reduces computing time, yields better discriminatory behavior and handles curse of dimensionality phenomena. For training and testing, a ten- fold cross validation approach is followed. In each round SVM classifier is trained with 90% of labeled (training) data and in the testing phase, classes of remaining 10% unlabeled (testing) data are predicted. Concluding, this paper proposes a model with semantic triplet based features for story classification. The effectiveness of the model is demonstrated against other traditional features used in the literature for text classification tasks.
ContributorsKarad, Ravi Chandravadan (Author) / Davulcu, Hasan (Thesis advisor) / Corman, Steven (Committee member) / Sen, Arunabha (Committee member) / Arizona State University (Publisher)
Created2013
148322-Thumbnail Image.png
Description

The field of biomedical research relies on the knowledge of binding interactions between various proteins of interest to create novel molecular targets for therapeutic purposes. While many of these interactions remain a mystery, knowledge of these properties and interactions could have significant medical applications in terms of understanding cell signaling

The field of biomedical research relies on the knowledge of binding interactions between various proteins of interest to create novel molecular targets for therapeutic purposes. While many of these interactions remain a mystery, knowledge of these properties and interactions could have significant medical applications in terms of understanding cell signaling and immunological defenses. Furthermore, there is evidence that machine learning and peptide microarrays can be used to make reliable predictions of where proteins could interact with each other without the definitive knowledge of the interactions. In this case, a neural network was used to predict the unknown binding interactions of TNFR2 onto LT-ɑ and TRAF2, and PD-L1 onto CD80, based off of the binding data from a sampling of protein-peptide interactions on a microarray. The accuracy and reliability of these predictions would rely on future research to confirm the interactions of these proteins, but the knowledge from these methods and predictions could have a future impact with regards to rational and structure-based drug design.

ContributorsPoweleit, Andrew Michael (Author) / Woodbury, Neal (Thesis director) / Diehnelt, Chris (Committee member) / Chiu, Po-Lin (Committee member) / School of Molecular Sciences (Contributor, Contributor) / Barrett, The Honors College (Contributor)
Created2021-05