Matching Items (330)
Filtering by

Clear all filters

147992-Thumbnail Image.png
Description

The research presented in this Honors Thesis provides development in machine learning models which predict future states of a system with unknown dynamics, based on observations of the system. Two case studies are presented for (1) a non-conservative pendulum and (2) a differential game dictating a two-car uncontrolled intersection scenario.

The research presented in this Honors Thesis provides development in machine learning models which predict future states of a system with unknown dynamics, based on observations of the system. Two case studies are presented for (1) a non-conservative pendulum and (2) a differential game dictating a two-car uncontrolled intersection scenario. In the paper we investigate how learning architectures can be manipulated for problem specific geometry. The result of this research provides that these problem specific models are valuable for accurate learning and predicting the dynamics of physics systems.<br/><br/>In order to properly model the physics of a real pendulum, modifications were made to a prior architecture which was sufficient in modeling an ideal pendulum. The necessary modifications to the previous network [13] were problem specific and not transferrable to all other non-conservative physics scenarios. The modified architecture successfully models real pendulum dynamics. This case study provides a basis for future research in augmenting the symplectic gradient of a Hamiltonian energy function to provide a generalized, non-conservative physics model.<br/><br/>A problem specific architecture was also utilized to create an accurate model for the two-car intersection case. The Costate Network proved to be an improvement from the previously used Value Network [17]. Note that this comparison is applied lightly due to slight implementation differences. The development of the Costate Network provides a basis for using characteristics to decompose functions and create a simplified learning problem.<br/><br/>This paper is successful in creating new opportunities to develop physics models, in which the sample cases should be used as a guide for modeling other real and pseudo physics. Although the focused models in this paper are not generalizable, it is important to note that these cases provide direction for future research.

ContributorsMerry, Tanner (Author) / Ren, Yi (Thesis director) / Zhang, Wenlong (Committee member) / Mechanical and Aerospace Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
Description

The PPP Loan Program was created by the CARES Act and carried out by the Small Business Administration (SBA) to provide support to small businesses in maintaining their payroll during the Coronavirus pandemic. This program was approved for $350 billion, but this amount was expanded by an additional $320 billion

The PPP Loan Program was created by the CARES Act and carried out by the Small Business Administration (SBA) to provide support to small businesses in maintaining their payroll during the Coronavirus pandemic. This program was approved for $350 billion, but this amount was expanded by an additional $320 billion to meet the demand by struggling businesses, since initial funding was exhausted under two weeks.<br/><br/>Significant controversy surrounds the program. In December 2020, the Department of Justice reported 90 individuals were charged for fraudulent use of funds, totaling $250 million. The loans, which were intended for small business, were actually approved for 450 public companies. Furthermore, the methods of approval are<br/>shrouded in mystery. In an effort to be transparent, the SBA has released information about loan recipients. Conveniently, the SBA has released information of all recipients. Detailed information was released for 661,218 recipients who have received a PPP loan in excess of $150,000. These recipients are the central point of this research.<br/><br/>This research sought to answer two primary questions: how did the SBA determine which loans, and therefore which industries are approved, and did the industries most affected by the pandemic receive the most in PPP loans, as intended by Congress? It was determined that, generally, PPP Loans were approved on the basis of employment percentages relative to the individual state. Furthermore, in general, the loans approved were approved fairly, with respect to the size of the industry. The loans, when adjusted for GDP and Employment factors, yielded a clear ranking that prioritized vulnerable industries first.<br/><br/>However, significant questions remain. The effectiveness of the PPP has been hindered by unclear incentives and negative outcomes, characterized by a government program that has essentially been rushed into service. Furthermore, limitations of available data to regress and compare the SBA's approved loans are not representative of small business.

ContributorsMaglanoc, Julian (Author) / Kenchington, David (Thesis director) / Cassidy, Nancy (Committee member) / Department of Finance (Contributor) / Dean, W.P. Carey School of Business (Contributor) / School of Accountancy (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148001-Thumbnail Image.png
Description

High-entropy alloys possessing mechanical, chemical, and electrical properties that far exceed those of conventional alloys have the potential to make a significant impact on many areas of engineering. Identifying element combinations and configurations to form these alloys, however, is a difficult, time-consuming, computationally intensive task. Machine learning has revolutionized many

High-entropy alloys possessing mechanical, chemical, and electrical properties that far exceed those of conventional alloys have the potential to make a significant impact on many areas of engineering. Identifying element combinations and configurations to form these alloys, however, is a difficult, time-consuming, computationally intensive task. Machine learning has revolutionized many different fields due to its ability to generalize well to different problems and produce computationally efficient, accurate predictions regarding the system of interest. In this thesis, we demonstrate the effectiveness of machine learning models applied to toy cases representative of simplified physics that are relevant to high-entropy alloy simulation. We show these models are effective at learning nonlinear dynamics for single and multi-particle cases and that more work is needed to accurately represent complex cases in which the system dynamics are chaotic. This thesis serves as a demonstration of the potential benefits of machine learning applied to high-entropy alloy simulations to generate fast, accurate predictions of nonlinear dynamics.

ContributorsDaly, John H (Author) / Ren, Yi (Thesis director) / Zhuang, Houlong (Committee member) / Mechanical and Aerospace Engineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
Description

The Covid-19 pandemic has made a significant impact on both the stock market and the<br/>global economy. The resulting volatility in stock prices has provided an opportunity to examine<br/>the Efficient Market Hypothesis. This study aims to gain insights into the efficiency of markets<br/>based on stock price performance in the Covid era.

The Covid-19 pandemic has made a significant impact on both the stock market and the<br/>global economy. The resulting volatility in stock prices has provided an opportunity to examine<br/>the Efficient Market Hypothesis. This study aims to gain insights into the efficiency of markets<br/>based on stock price performance in the Covid era. Specifically, it investigates the market’s<br/>ability to anticipate significant events during the Covid-19 timeline beginning November 1, 2019<br/><br/>and ending March 31, 2021. To examine the efficiency of markets, our team created a Stay-at-<br/>Home Portfolio, experiencing economic tailwinds from the Covid lockdowns, and a Pandemic<br/><br/>Loser Portfolio, experiencing economic headwinds from the Covid lockdowns. Cumulative<br/>returns of each portfolio are benchmarked to the cumulative returns of the S&P 500. The results<br/>showed that the Efficient Market Hypothesis is likely to be valid, although a definitive<br/>conclusion cannot be made based on the scope of the analysis. There are recommendations for<br/>further research surrounding key events that may be able to draw a more direct conclusion.

ContributorsBrock, Matt Ian (Co-author) / Beneduce, Trevor (Co-author) / Craig, Nicko (Co-author) / Hertzel, Michael (Thesis director) / Mindlin, Jeff (Committee member) / Department of Finance (Contributor) / Economics Program in CLAS (Contributor) / WPC Graduate Programs (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
Description

Robots are often used in long-duration scenarios, such as on the surface of Mars,where they may need to adapt to environmental changes. Typically, robots have been built specifically for single tasks, such as moving boxes in a warehouse

Robots are often used in long-duration scenarios, such as on the surface of Mars,where they may need to adapt to environmental changes. Typically, robots have been built specifically for single tasks, such as moving boxes in a warehouse or surveying construction sites. However, there is a modern trend away from human hand-engineering and toward robot learning. To this end, the ideal robot is not engineered,but automatically designed for a specific task. This thesis focuses on robots which learn path-planning algorithms for specific environments. Learning is accomplished via genetic programming. Path-planners are represented as Python code, which is optimized via Pareto evolution. These planners are encouraged to explore curiously and efficiently. This research asks the questions: “How can robots exhibit life-long learning where they adapt to changing environments in a robust way?”, and “How can robots learn to be curious?”.

ContributorsSaldyt, Lucas P (Author) / Ben Amor, Heni (Thesis director) / Pavlic, Theodore (Committee member) / Computer Science and Engineering Program (Contributor, Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148088-Thumbnail Image.png
Description

Colorimetric assays are an important tool in point-of-care testing that offers several advantages to traditional testing methods such as rapid response times and inexpensive costs. A factor that currently limits the portability and accessibility of these assays are methods that can objectively determine the results of these assays. Current solutions

Colorimetric assays are an important tool in point-of-care testing that offers several advantages to traditional testing methods such as rapid response times and inexpensive costs. A factor that currently limits the portability and accessibility of these assays are methods that can objectively determine the results of these assays. Current solutions consist of creating a test reader that standardizes the conditions the strip is under before being measured in some way. However, this increases the cost and decreases the portability of these assays. The focus of this study is to create a machine learning algorithm that can objectively determine results of colorimetric assays under varying conditions. To ensure the flexibility of a model to several types of colorimetric assays, three models were trained on the same convolutional neural network with different datasets. The images these models are trained on consist of positive and negative images of ETG, fentanyl, and HPV Antibodies test strips taken under different lighting and background conditions. A fourth model is trained on an image set composed of all three strip types. The results from these models show it is able to predict positive and negative results to a high level of accuracy.

ContributorsFisher, Rachel (Author) / Blain Christen, Jennifer (Thesis director) / Anderson, Karen (Committee member) / School of Life Sciences (Contributor) / Harrington Bioengineering Program (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
147956-Thumbnail Image.png
Description

Music streaming services have affected the music industry from both a financial and legal standpoint. Their current business model affects stakeholders such as artists, users, and investors. These services have been scrutinized recently for their imperfect royalty distribution model. Covid-19 has made these discussions even more relevant as touring income

Music streaming services have affected the music industry from both a financial and legal standpoint. Their current business model affects stakeholders such as artists, users, and investors. These services have been scrutinized recently for their imperfect royalty distribution model. Covid-19 has made these discussions even more relevant as touring income has come to a halt for musicians and the live entertainment industry. <br/>Under the current per-stream model, it is becoming exceedingly hard for artists to make a living off of streams. This forces artists to tour heavily as well as cut corners to create what is essentially “disposable art”. Rapidly releasing multiple projects a year has become the norm for many modern artists. This paper will examine the licensing framework, royalty payout issues, and propose a solution.

ContributorsKoudssi, Zakaria Corley (Author) / Sadusky, Brian (Thesis director) / Koretz, Lora (Committee member) / Dean, W.P. Carey School of Business (Contributor) / Department of Finance (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
Description

The COVID-19 pandemic has and will continue to radically shift the workplace. An increasing percentage of the workforce desires flexible working options and, as such, firms are likely to require less office space going forward. Additionally, the economic downturn caused by the pandemic provides an opportunity for companies to secure

The COVID-19 pandemic has and will continue to radically shift the workplace. An increasing percentage of the workforce desires flexible working options and, as such, firms are likely to require less office space going forward. Additionally, the economic downturn caused by the pandemic provides an opportunity for companies to secure favorable rent rates on new lease agreements. This project aims to evaluate and measure Company X’s potential cost savings from terminating current leases and downsizing office space in five selected cities. Along with city-specific real estate market research and forecasts, we employ a four-stage model of Company X’s real estate negotiation process to analyze whether existing lease agreements in these cities should be renewed or terminated.

ContributorsHegardt, Brandon Michael (Co-author) / Saker, Logan (Co-author) / Patterson, Jack (Co-author) / Ries, Sarah (Co-author) / Simonson, Mark (Thesis director) / Hertzel, Michael (Committee member) / Department of Finance (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
Description
In many classication problems data samples cannot be collected easily, example in drug trials, biological experiments and study on cancer patients. In many situations the data set size is small and there are many outliers. When classifying such data, example cancer vs normal patients the consequences of mis-classication are probably

In many classication problems data samples cannot be collected easily, example in drug trials, biological experiments and study on cancer patients. In many situations the data set size is small and there are many outliers. When classifying such data, example cancer vs normal patients the consequences of mis-classication are probably more important than any other data type, because the data point could be a cancer patient or the classication decision could help determine what gene might be over expressed and perhaps a cause of cancer. These mis-classications are typically higher in the presence of outlier data points. The aim of this thesis is to develop a maximum margin classier that is suited to address the lack of robustness of discriminant based classiers (like the Support Vector Machine (SVM)) to noise and outliers. The underlying notion is to adopt and develop a natural loss function that is more robust to outliers and more representative of the true loss function of the data. It is demonstrated experimentally that SVM's are indeed susceptible to outliers and that the new classier developed, here coined as Robust-SVM (RSVM), is superior to all studied classier on the synthetic datasets. It is superior to the SVM in both the synthetic and experimental data from biomedical studies and is competent to a classier derived on similar lines when real life data examples are considered.
ContributorsGupta, Sidharth (Author) / Kim, Seungchan (Thesis advisor) / Welfert, Bruno (Committee member) / Li, Baoxin (Committee member) / Arizona State University (Publisher)
Created2011
150158-Thumbnail Image.png
Description
Multi-label learning, which deals with data associated with multiple labels simultaneously, is ubiquitous in real-world applications. To overcome the curse of dimensionality in multi-label learning, in this thesis I study multi-label dimensionality reduction, which extracts a small number of features by removing the irrelevant, redundant, and noisy information while considering

Multi-label learning, which deals with data associated with multiple labels simultaneously, is ubiquitous in real-world applications. To overcome the curse of dimensionality in multi-label learning, in this thesis I study multi-label dimensionality reduction, which extracts a small number of features by removing the irrelevant, redundant, and noisy information while considering the correlation among different labels in multi-label learning. Specifically, I propose Hypergraph Spectral Learning (HSL) to perform dimensionality reduction for multi-label data by exploiting correlations among different labels using a hypergraph. The regularization effect on the classical dimensionality reduction algorithm known as Canonical Correlation Analysis (CCA) is elucidated in this thesis. The relationship between CCA and Orthonormalized Partial Least Squares (OPLS) is also investigated. To perform dimensionality reduction efficiently for large-scale problems, two efficient implementations are proposed for a class of dimensionality reduction algorithms, including canonical correlation analysis, orthonormalized partial least squares, linear discriminant analysis, and hypergraph spectral learning. The first approach is a direct least squares approach which allows the use of different regularization penalties, but is applicable under a certain assumption; the second one is a two-stage approach which can be applied in the regularization setting without any assumption. Furthermore, an online implementation for the same class of dimensionality reduction algorithms is proposed when the data comes sequentially. A Matlab toolbox for multi-label dimensionality reduction has been developed and released. The proposed algorithms have been applied successfully in the Drosophila gene expression pattern image annotation. The experimental results on some benchmark data sets in multi-label learning also demonstrate the effectiveness and efficiency of the proposed algorithms.
ContributorsSun, Liang (Author) / Ye, Jieping (Thesis advisor) / Li, Baoxin (Committee member) / Liu, Huan (Committee member) / Mittelmann, Hans D. (Committee member) / Arizona State University (Publisher)
Created2011