Matching Items (18)
Filtering by

Clear all filters

136255-Thumbnail Image.png
Description
Over the course of six months, we have worked in partnership with Arizona State University and a leading producer of semiconductor chips in the United States market (referred to as the "Company"), lending our skills in finance, statistics, model building, and external insight. We attempt to design models that hel

Over the course of six months, we have worked in partnership with Arizona State University and a leading producer of semiconductor chips in the United States market (referred to as the "Company"), lending our skills in finance, statistics, model building, and external insight. We attempt to design models that help predict how much time it takes to implement a cost-saving project. These projects had previously been considered only on the merit of cost savings, but with an added dimension of time, we hope to forecast time according to a number of variables. With such a forecast, we can then apply it to an expense project prioritization model which relates time and cost savings together, compares many different projects simultaneously, and returns a series of present value calculations over different ranges of time. The goal is twofold: assist with an accurate prediction of a project's time to implementation, and provide a basis to compare different projects based on their present values, ultimately helping to reduce the Company's manufacturing costs and improve gross margins. We believe this approach, and the research found toward this goal, is most valuable for the Company. Two coaches from the Company have provided assistance and clarified our questions when necessary throughout our research. In this paper, we begin by defining the problem, setting an objective, and establishing a checklist to monitor our progress. Next, our attention shifts to the data: making observations, trimming the dataset, framing and scoping the variables to be used for the analysis portion of the paper. Before creating a hypothesis, we perform a preliminary statistical analysis of certain individual variables to enrich our variable selection process. After the hypothesis, we run multiple linear regressions with project duration as the dependent variable. After regression analysis and a test for robustness, we shift our focus to an intuitive model based on rules of thumb. We relate these models to an expense project prioritization tool developed using Microsoft Excel software. Our deliverables to the Company come in the form of (1) a rules of thumb intuitive model and (2) an expense project prioritization tool.
ContributorsAl-Assi, Hashim (Co-author) / Chiang, Robert (Co-author) / Liu, Andrew (Co-author) / Ludwick, David (Co-author) / Simonson, Mark (Thesis director) / Hertzel, Michael (Committee member) / Barrett, The Honors College (Contributor) / Department of Information Systems (Contributor) / Department of Finance (Contributor) / Department of Economics (Contributor) / Department of Supply Chain Management (Contributor) / School of Accountancy (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Mechanical and Aerospace Engineering Program (Contributor) / WPC Graduate Programs (Contributor)
Created2015-05
133184-Thumbnail Image.png
Description
Cryptocurrencies are notorious for its volatility. But with its incredible rise in price, Bitcoin keep being on the top among the trending topics on social media. Although doubts continue to rise with price, Bloomberg even make critics on Bitcoin as ‘the biggest bubble in the history’, some investors still hold

Cryptocurrencies are notorious for its volatility. But with its incredible rise in price, Bitcoin keep being on the top among the trending topics on social media. Although doubts continue to rise with price, Bloomberg even make critics on Bitcoin as ‘the biggest bubble in the history’, some investors still hold strong enthusiasm and confidence towards Bitcoin. As contradicting opinions increase, it is worthy to dive into discussions on social media and use a scientific method to evaluate public’s non-negligible role in crypto price fluctuation.

Sentiment analysis, which is a notably method in text mining, can be used to extract the sentiment from people’s opinion. It then provides us with valuable perception on a topic from the public’s attitude, which create more opportunities for deeper analysis and prediction.

The thesis aims to investigate public’s sentiment towards Bitcoin through analyzing 10 million Bitcoin related tweets and assigning sentiment points on tweets, then using sentiment fluctuation as a factor to predict future crypto fluctuation. Price prediction is achieved by using a machine learning model called Recurrent Neural Network which automatically learns the pattern and generate following results with memory. The analysis revels slight connection between sentiment and crypto currency and the Neural Network model showed a strong connection between sentiment score and future price prediction.
ContributorsZhu, Xiaoyu (Author) / Benjamin, Victor (Thesis director) / Qinglai, He (Committee member) / Department of Information Systems (Contributor) / Department of Supply Chain Management (Contributor) / Barrett, The Honors College (Contributor)
Created2018-12
133957-Thumbnail Image.png
Description
Coherent vortices are ubiquitous structures in natural flows that affect mixing and transport of substances and momentum/energy. Being able to detect these coherent structures is important for pollutant mitigation, ecological conservation and many other aspects. In recent years, mathematical criteria and algorithms have been developed to extract these coherent structures

Coherent vortices are ubiquitous structures in natural flows that affect mixing and transport of substances and momentum/energy. Being able to detect these coherent structures is important for pollutant mitigation, ecological conservation and many other aspects. In recent years, mathematical criteria and algorithms have been developed to extract these coherent structures in turbulent flows. In this study, we will apply these tools to extract important coherent structures and analyze their statistical properties as well as their implications on kinematics and dynamics of the flow. Such information will aide representation of small-scale nonlinear processes that large-scale models of natural processes may not be able to resolve.
ContributorsCass, Brentlee Jerry (Author) / Tang, Wenbo (Thesis director) / Kostelich, Eric (Committee member) / Department of Information Systems (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2018-05
132832-Thumbnail Image.png
Description
Exchange traded funds (ETFs) in many ways are similar to more traditional closed-end mutual funds, although thee differ in a crucial way. ETFs rely on a creation and redemption feature to achieve their functionality and this mechanism is designed to minimize the deviations that occur between the ETF’s listed price

Exchange traded funds (ETFs) in many ways are similar to more traditional closed-end mutual funds, although thee differ in a crucial way. ETFs rely on a creation and redemption feature to achieve their functionality and this mechanism is designed to minimize the deviations that occur between the ETF’s listed price and the net asset value of the ETF’s underlying assets. However while this does cause ETF deviations to be generally lower than their mutual fund counterparts, as our paper explores this process does not eliminate these deviations completely. This article builds off an earlier paper by Engle and Sarkar (2006) that investigates these properties of premiums (discounts) of ETFs from their fair market value. And looks to see if these premia have changed in the last 10 years. Our paper then diverges from the original and takes a deeper look into the standard deviations of these premia specifically.

Our findings show that over 70% of an ETFs standard deviation of premia can be explained through a linear combination consisting of two variables: a categorical (Domestic[US], Developed, Emerging) and a discrete variable (time-difference from US). This paper also finds that more traditional metrics such as market cap, ETF price volatility, and even 3rd party market indicators such as the economic freedom index and investment freedom index are insignificant predictors of an ETFs standard deviation of premia when combined with the categorical variable. These findings differ somewhat from existing literature which indicate that these factors should have a significant impact on the predictive ability of an ETFs standard deviation of premia.
ContributorsZhang, Jingbo (Co-author, Co-author) / Henning, Thomas (Co-author) / Simonson, Mark (Thesis director) / Licon, L. Wendell (Committee member) / Department of Finance (Contributor) / Department of Information Systems (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2019-05
137296-Thumbnail Image.png
Description
This thesis studies the area of sentiment analysis and its general uses, benefits, and limitations. Social networking, blogging, and online forums have turned the Web into a vast repository of comments on many topics. Sentiment analysis is the process of using software to analyze social media to gauge the attitudes

This thesis studies the area of sentiment analysis and its general uses, benefits, and limitations. Social networking, blogging, and online forums have turned the Web into a vast repository of comments on many topics. Sentiment analysis is the process of using software to analyze social media to gauge the attitudes or sentiments of the users/authors concerning a particular subject. Sentiment analysis works by processing (data mining) unstructured textual evidence using natural language processing and machine learning to determine a positive, negative, or neutral measurement. When utilized correctly, sentiment analysis has the potential to glean valuable insights into consumers' minds, which in turn leads to increased revenue and improved customer satisfaction for businesses. This paper looks at four industries in which sentiment analysis is being used or being considered: retail/services, politics, healthcare, and finances. The goal of the thesis will be to explore whether sentiment analysis has been used successfully for economic or social benefit and whether it is a practical solution for analyzing consumer opinion.
ContributorsSoumya, Saswati (Author) / Uday, Kulkarni (Thesis director) / Brooks, Daniel (Committee member) / Barrett, The Honors College (Contributor) / Department of Economics (Contributor) / WPC Graduate Programs (Contributor) / Department of Information Systems (Contributor)
Created2014-05
Description
The object of the present study is to examine methods in which the company can optimize their costs on third-party suppliers whom oversee other third-party trade labor. The third parties in scope of this study are suspected to overstaff their workforce, thus overcharging the company. We will introduce a complex

The object of the present study is to examine methods in which the company can optimize their costs on third-party suppliers whom oversee other third-party trade labor. The third parties in scope of this study are suspected to overstaff their workforce, thus overcharging the company. We will introduce a complex spreadsheet model that will propose a proper project staffing level based on key qualitative variables and statistics. Using the model outputs, the Thesis team proposes a headcount solution for the company and problem areas to focus on, going forward. All sources of information come from company proprietary and confidential documents.
ContributorsLoo, Andrew (Co-author) / Brennan, Michael (Co-author) / Sheiner, Alexander (Co-author) / Hertzel, Michael (Thesis director) / Simonson, Mark (Committee member) / Barrett, The Honors College (Contributor) / Department of Information Systems (Contributor) / Department of Finance (Contributor) / Department of Supply Chain Management (Contributor) / WPC Graduate Programs (Contributor) / School of Accountancy (Contributor)
Created2014-05
134373-Thumbnail Image.png
Description
Our research encompassed the prospect draft in baseball and looked at what type of player teams drafted to maximize value. We wanted to know which position returned the best value to the team that drafted them, and which level is safer to draft players from, college or high school. We

Our research encompassed the prospect draft in baseball and looked at what type of player teams drafted to maximize value. We wanted to know which position returned the best value to the team that drafted them, and which level is safer to draft players from, college or high school. We decided to look at draft data from 2006-2010 for the first ten rounds of players selected. Because there is only a monetary cap on players drafted in the first ten rounds we restricted our data to these players. Once we set up the parameters we compiled a spreadsheet of these players with both their signing bonuses and their wins above replacement (WAR). This allowed us to see how much a team was spending per win at the major league level. After the data was compiled we made pivot tables and graphs to visually represent our data and better understand the numbers. We found that the worst position that MLB teams could draft would be high school second baseman. They returned the lowest WAR of any player that we looked at. In general though high school players were more costly to sign and had lower WARs than their college counterparts making them, on average, a worse pick value wise. The best position you could pick was college shortstops. They had the trifecta of the best signability of all players, along with one of the highest WARs and lowest signing bonuses. These were three of the main factors that you want with your draft pick and they ranked near the top in all three categories. This research can help give guidelines to Major League teams as they go to select players in the draft. While there are always going to be exceptions to trends, by following the enclosed research teams can minimize risk in the draft.
ContributorsValentine, Robert (Co-author) / Johnson, Ben (Co-author) / Eaton, John (Thesis director) / Goegan, Brian (Committee member) / Department of Finance (Contributor) / Department of Economics (Contributor) / Department of Information Systems (Contributor) / School of Accountancy (Contributor) / Barrett, The Honors College (Contributor)
Created2017-05
134082-Thumbnail Image.png
Description
Over the past several decades, analytics have become more and more prevalent in the game of baseball. Statistics are used in nearly every facet of the game. Each team develops its own processes, hoping to gain a competitive advantage over the rest of the league. One area of the game

Over the past several decades, analytics have become more and more prevalent in the game of baseball. Statistics are used in nearly every facet of the game. Each team develops its own processes, hoping to gain a competitive advantage over the rest of the league. One area of the game that has struggled to produce definitive analytics is amateur scouting. This project seeks to resolve this problem through the creation of a new statistic, Valued Plate Appearance Index (VPI). The problem is identified through analysis that was performed to determine whether any correlation exists between performances at the country's top amateur baseball league, the Cape Cod League, and performances in Major League Baseball. After several stats were analyzed, almost no correlation was determined between the two. This essentially means that teams have no way to statistically analyze Cape Cod League performance and project future statistics. An inherent contextual error in these amateur statistics prevents them from correlating. The project seeks to close that contextual gap and create concrete, encompassing values to illustrate a player's offensive performance in the Cape League. To solve for this problem, data was collected from the 2017 CCBL season. In addition to VPI, Valued Plate Appearance Approach (VPA) and Valued Plate Appearance Result (VPR) were created to better depict a player's all-around performance in each plate appearance. VPA values the quality of a player's approach in each plate appearance. VPR values the quality of the contact result, excluding factors out of the hitter's control. This statistic isolates player performance as well as eliminates luck that cannot normally be taken into account. This paper results in the segmentation of players from the 2017 CCBL into four different groups, which project how they will perform as they transition into professional baseball. These groups and the creation of these statistics could be essential tools in the evaluation and projection of amateur players by Major League clubs for years to come.
ContributorsLothrop, Joseph Kent (Author) / Eaton, John (Thesis director) / McIntosh, Daniel (Committee member) / Department of Information Systems (Contributor) / Department of Marketing (Contributor) / Barrett, The Honors College (Contributor)
Created2017-12
134845-Thumbnail Image.png
Description
The purpose of this thesis was to develop a tool to provide information and data for design teams to use throughout the mobile application design process. Ideally, this would enable teams to see patterns in iterative design, and ultimately use data-driven analysis to make their own decisions. The initial problem

The purpose of this thesis was to develop a tool to provide information and data for design teams to use throughout the mobile application design process. Ideally, this would enable teams to see patterns in iterative design, and ultimately use data-driven analysis to make their own decisions. The initial problem was a lack of available information offered by mobile application design teams—the initial goal being to work closely with design teams to learn their decision-making methodology. However, every team that was reached out to responded with rejection, presenting a new problem: a lack of access to quality information regarding the decision-making process for mobile applications. This problem was addressed by the development of an ethical hacking script that retrieves reviews in bulk from the Google Play Store using Python. The project was a success—by feeding an application’s unique Play Store ID, the script retrieves a user-specified amount of reviews (up to millions) for that mobile application and the 4 “recommended” applications from the Play Store. Ultimately, this thesis proved that protected reviews on the Play Store can be ethically retrieved and used for data-driven decision making and identifying patterns in an application’s iterative design. This script provides an automated tool for teams to “put a finger on the pulse” of their target applications.
ContributorsDyer, Mitchell Patrick (Author) / Lin, Elva (Thesis director) / Giles, Charles (Committee member) / Department of Information Systems (Contributor) / Barrett, The Honors College (Contributor)
Created2016-12
135606-Thumbnail Image.png
Description
League of Legends is a Multiplayer Online Battle Arena (MOBA) game. MOBA games are generally formatted where two teams of five, each player controlling a character (champion), will try to take each other's base as quickly as possible. Currently, with about 70 million, League of Legends is number one in

League of Legends is a Multiplayer Online Battle Arena (MOBA) game. MOBA games are generally formatted where two teams of five, each player controlling a character (champion), will try to take each other's base as quickly as possible. Currently, with about 70 million, League of Legends is number one in the digital entertainment industry with $1.63 billion dollars of revenue in year 2015. This research analysis scopes in on the niche of the "Jungler" role between different tiers of player in League of Legends. I uncovered differences in player strategy that may explain the achievement of high rank using data aggregation through Riot Games' API, data slicing with time-sensitive data, random sampling, clustering by tiers, graphical techniques to display the cluster, distribution analysis and finally, a comprehensive factor analysis on the data's implications.
ContributorsPoon, Alex (Author) / Clark, Joseph (Thesis director) / Simon, Alan (Committee member) / Department of Information Systems (Contributor) / Department of Management (Contributor) / Barrett, The Honors College (Contributor)
Created2016-05