Matching Items (22)
Filtering by

Clear all filters

135547-Thumbnail Image.png
Description
The Experimental Data Processing (EDP) software is a C++ GUI-based application to streamline the process of creating a model for structural systems based on experimental data. EDP is designed to process raw data, filter the data for noise and outliers, create a fitted model to describe that data, complete a

The Experimental Data Processing (EDP) software is a C++ GUI-based application to streamline the process of creating a model for structural systems based on experimental data. EDP is designed to process raw data, filter the data for noise and outliers, create a fitted model to describe that data, complete a probabilistic analysis to describe the variation between replicates of the experimental process, and analyze reliability of a structural system based on that model. In order to help design the EDP software to perform the full analysis, the probabilistic and regression modeling aspects of this analysis have been explored. The focus has been on creating and analyzing probabilistic models for the data, adding multivariate and nonparametric fits to raw data, and developing computational techniques that allow for these methods to be properly implemented within EDP. For creating a probabilistic model of replicate data, the normal, lognormal, gamma, Weibull, and generalized exponential distributions have been explored. Goodness-of-fit tests, including the chi-squared, Anderson-Darling, and Kolmogorov-Smirnoff tests, have been used in order to analyze the effectiveness of any of these probabilistic models in describing the variation of parameters between replicates of an experimental test. An example using Young's modulus data for a Kevlar-49 Swath stress-strain test was used in order to demonstrate how this analysis is performed within EDP. In order to implement the distributions, numerical solutions for the gamma, beta, and hypergeometric functions were implemented, along with an arbitrary precision library to store numbers that exceed the maximum size of double-precision floating point digits. To create a multivariate fit, the multilinear solution was created as the simplest solution to the multivariate regression problem. This solution was then extended to solve nonlinear problems that can be linearized into multiple separable terms. These problems were solved analytically with the closed-form solution for the multilinear regression, and then by using a QR decomposition to solve numerically while avoiding numerical instabilities associated with matrix inversion. For nonparametric regression, or smoothing, the loess method was developed as a robust technique for filtering noise while maintaining the general structure of the data points. The loess solution was created by addressing concerns associated with simpler smoothing methods, including the running mean, running line, and kernel smoothing techniques, and combining the ability of each of these methods to resolve those issues. The loess smoothing method involves weighting each point in a partition of the data set, and then adding either a line or a polynomial fit within that partition. Both linear and quadratic methods were applied to a carbon fiber compression test, showing that the quadratic model was more accurate but the linear model had a shape that was more effective for analyzing the experimental data. Finally, the EDP program itself was explored to consider its current functionalities for processing data, as described by shear tests on carbon fiber data, and the future functionalities to be developed. The probabilistic and raw data processing capabilities were demonstrated within EDP, and the multivariate and loess analysis was demonstrated using R. As the functionality and relevant considerations for these methods have been developed, the immediate goal is to finish implementing and integrating these additional features into a version of EDP that performs a full streamlined structural analysis on experimental data.
ContributorsMarkov, Elan Richard (Author) / Rajan, Subramaniam (Thesis director) / Khaled, Bilal (Committee member) / Chemical Engineering Program (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Ira A. Fulton School of Engineering (Contributor) / Barrett, The Honors College (Contributor)
Created2016-05
136550-Thumbnail Image.png
Description
The NFL is one of largest and most influential industries in the world. In America there are few companies that have a stronger hold on the American culture and create such a phenomena from year to year. In this project aimed to develop a strategy that helps an NFL team

The NFL is one of largest and most influential industries in the world. In America there are few companies that have a stronger hold on the American culture and create such a phenomena from year to year. In this project aimed to develop a strategy that helps an NFL team be as successful as possible by defining which positions are most important to a team's success. Data from fifteen years of NFL games was collected and information on every player in the league was analyzed. First there needed to be a benchmark which describes a team as being average and then every player in the NFL must be compared to that average. Based on properties of linear regression using ordinary least squares this project aims to define such a model that shows each position's importance. Finally, once such a model had been established then the focus turned to the NFL draft in which the goal was to find a strategy of where each position needs to be drafted so that it is most likely to give the best payoff based on the results of the regression in part one.
ContributorsBalzer, Kevin Ryan (Author) / Goegan, Brian (Thesis director) / Dassanayake, Maduranga (Committee member) / Barrett, The Honors College (Contributor) / Economics Program in CLAS (Contributor) / School of Mathematical and Statistical Sciences (Contributor)
Created2015-05
136406-Thumbnail Image.png
Description
In this paper, I analyze representations of nature in popular film, using the feminist / deconstructionist concept of a dualism to structure my critique. Using Val Plumwood’s analysis of the logical structure of dualism and the 5 ‘features of a dualism’ that she identifies, I critique 5 popular movies –

In this paper, I analyze representations of nature in popular film, using the feminist / deconstructionist concept of a dualism to structure my critique. Using Val Plumwood’s analysis of the logical structure of dualism and the 5 ‘features of a dualism’ that she identifies, I critique 5 popular movies – Star Wars, Lord of the Rings, Brave, Grizzly Man, and Planet Earth – by locating within each of them one of the 5 features and explaining how the movie functions to reinforce the Nature/Culture dualism . By showing how the Nature/Culture dualism shapes and is shaped by popular cinema, I show how “Nature” is a social construct, created as part of this very dualism, and reified through popular culture. I conclude with the introduction of a number of ‘subversive’ pieces of visual art that undermine and actively deconstruct the Nature/Culture dualism and show to the viewer a more honest presentation of the non-human world.
ContributorsBarton, Christopher Joseph (Author) / Broglio, Ron (Thesis director) / Minteer, Ben (Committee member) / Barrett, The Honors College (Contributor) / School of Sustainability (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / School of Geographical Sciences and Urban Planning (Contributor)
Created2015-05
133177-Thumbnail Image.png
Description
From 2007 to 2017, the state of California experienced two major droughts that required significant governmental action to decrease urban water demand. The purpose of this project is to isolate and explore the effects of these policy changes on water use during and after these droughts, and to see how

From 2007 to 2017, the state of California experienced two major droughts that required significant governmental action to decrease urban water demand. The purpose of this project is to isolate and explore the effects of these policy changes on water use during and after these droughts, and to see how these policies interact with hydroclimatic variability. As explanatory variables in multiple linear regression (MLR) models, water use policies were found to be significant at both the zip code and city levels. Policies that specifically target behavioral changes were significant mathematical drivers of water use in city-level models. Policy data was aggregated into a timeline and coded based on categories including user type, whether the policy was voluntary or mandatory, the targeted water use type, and whether the change in question concerns active or passive conservation. The analyzed policies include but are not limited to state drought declarations, regulatory municipal ordinances, and incentive programs for household appliances. Spatial averages of available hydroclimatic data have been computed and validated using inverse distance weighting methods. The data was aggregated at the zip code level to be comparable to the available water use data for use in MLR models. Factors already known to affect water use, such as temperature, precipitation, income, and water stress, were brought into the MLR models as explanatory variables. After controlling for these factors, the timeline policies were brought into the model as coded variables to test their effect on water demand during the years 2000-2017. Clearly identifying which policy traits are effective will inform future policymaking in cities aiming to conserve water. The findings suggest that drought-related policies impact per capita urban water use. The results of the city level MLR models indicate that implementation of mandatory policies that target water use behaviors effectively reduce water use. Temperature, income, unemployment, and the WaSSI were also observed to be mathematical drivers of water use. Interaction effects between policies and the WaSSI were statistically significant at both model scales.
ContributorsHjelmstad, Annika Margaret (Author) / Garcia, Margaret (Thesis director) / Larson, Kelli (Committee member) / Civil, Environmental and Sustainable Eng Program (Contributor, Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2018-12
132834-Thumbnail Image.png
Description
Exchange traded funds (ETFs) in many ways are similar to more traditional closed-end mutual
funds, although thee differ in a crucial way. ETFs rely on a creation and redemption feature to
achieve their functionality and this mechanism is designed to minimize the deviations that occur
between the ETF’s listed price and the net

Exchange traded funds (ETFs) in many ways are similar to more traditional closed-end mutual
funds, although thee differ in a crucial way. ETFs rely on a creation and redemption feature to
achieve their functionality and this mechanism is designed to minimize the deviations that occur
between the ETF’s listed price and the net asset value of the ETF’s underlying assets. However
while this does cause ETF deviations to be generally lower than their mutual fund counterparts,
as our paper explores this process does not eliminate these deviations completely. This article
builds off an earlier paper by Engle and Sarkar (2006) that investigates these properties of
premiums (discounts) of ETFs from their fair market value. And looks to see if these premia
have changed in the last 10 years. Our paper then diverges from the original and takes a deeper
look into the standard deviations of these premia specifically.
Our findings show that over 70% of an ETFs standard deviation of premia can be
explained through a linear combination consisting of two variables: a categorical (Domestic[US],
Developed, Emerging) and a discrete variable (time-difference from US). This paper also finds
that more traditional metrics such as market cap, ETF price volatility, and even 3rd party market
indicators such as the economic freedom index and investment freedom index are insignificant
predictors of an ETFs standard deviation of premia. These findings differ somewhat from
existing literature which indicate that these factors should have a significant impact on the
predictive ability of an ETFs standard deviation of premia.
ContributorsHenning, Thomas Louis (Co-author) / Zhang, Jingbo (Co-author) / Simonson, Mark (Thesis director) / Wendell, Licon (Committee member) / School of Mathematical and Statistical Sciences (Contributor) / Department of Finance (Contributor) / Barrett, The Honors College (Contributor)
Created2019-05
133570-Thumbnail Image.png
Description
In the last decade, the population of honey bees across the globe has declined sharply leaving scientists and bee keepers to wonder why? Amongst all nations, the United States has seen some of the greatest declines in the last 10 plus years. Without a definite explanation, Colony Collapse Disorder (CCD)

In the last decade, the population of honey bees across the globe has declined sharply leaving scientists and bee keepers to wonder why? Amongst all nations, the United States has seen some of the greatest declines in the last 10 plus years. Without a definite explanation, Colony Collapse Disorder (CCD) was coined to explain the sudden and sharp decline of the honey bee colonies that beekeepers were experiencing. Colony collapses have been rising higher compared to expected averages over the years, and during the winter season losses are even more severe than what is normally acceptable. There are some possible explanations pointing towards meteorological variables, diseases, and even pesticide usage. Despite the cause of CCD being unknown, thousands of beekeepers have reported their losses, and even numbers of infected colonies and colonies under certain stressors in the most recent years. Using the data that was reported to The United States Department of Agriculture (USDA), as well as weather data collected by The National Centers for Environmental Information (NOAA) and the National Centers for Environmental Information (NCEI), regression analysis was used to investigate honey bee colonies to find relationships between stressors in honey bee colonies and meteorological variables, and colony collapses during the winter months. The regression analysis focused on the winter season, or quarter 4 of the year, which includes the months of October, November, and December. In the model, the response variables was the percentage of colonies lost in quarter 4. Through the model, it was concluded that certain weather thresholds and the percentage increase of colonies under certain stressors were related to colony loss.
ContributorsVasquez, Henry Antony (Author) / Zheng, Yi (Thesis director) / Saffell, Erinanne (Committee member) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2018-05
137196-Thumbnail Image.png
Description
As society's energy crisis continues to become more imminent many industries and niches are seeking a new, sustainable and renewable source of electricity production. Similar to solar, wind and tidal energy, kinetic energy has the potential to generate electricity as an extremely renewable source of energy generation. While stationary bicycles

As society's energy crisis continues to become more imminent many industries and niches are seeking a new, sustainable and renewable source of electricity production. Similar to solar, wind and tidal energy, kinetic energy has the potential to generate electricity as an extremely renewable source of energy generation. While stationary bicycles can generate small amounts of electricity, the idea behind this project was to expand energy generation into the more common weight lifting side of exercising. The method for solving this problem was to find the average amount of power generated per user on a Smith machine and determine how much power was available from an accompanying energy generator. The generator consists of three phases: a copper coil and magnet generator, a full wave bridge rectifying circuit and a rheostat. These three phases working together formed a fully functioning controllable generator. The resulting issue with the kinetic energy generator was that the system was too inefficient to serve as a viable system for electricity generation. The electrical production of the generator only saved about 2 cents per year based on current Arizona electricity rates. In the end it was determined that the project was not a sustainable energy generation system and did not warrant further experimentation.
ContributorsO'Halloran, Ryan James (Author) / Middleton, James (Thesis director) / Hinrichs, Richard (Committee member) / Barrett, The Honors College (Contributor) / Mechanical and Aerospace Engineering Program (Contributor) / The Design School (Contributor) / School of Mathematical and Statistical Sciences (Contributor)
Created2014-05
134415-Thumbnail Image.png
Description
This paper will begin by initially discussing the potential uses and challenges of efficient and accurate traffic forecasting. The data we used includes traffic volume from seven locations on a busy Athens street in April and May of 2000. This data was used as part of a traffic forecasting competition.

This paper will begin by initially discussing the potential uses and challenges of efficient and accurate traffic forecasting. The data we used includes traffic volume from seven locations on a busy Athens street in April and May of 2000. This data was used as part of a traffic forecasting competition. Our initial observation, was that due to the volatility and oscillating nature of daily traffic volume, simple linear regression models will not perform well in predicting the time-series data. For this we present the Harmonic Time Series model. Such model (assuming all predictors are significant) will include a sinusoidal term for each time index within a period of data. Our assumption is that traffic volumes have a period of one week (which is evidenced by the graphs reproduced in our paper). This leads to a model that has 6,720 sine and cosine terms. This is clearly too many coefficients, so in an effort to avoid over-fitting and having an efficient model, we apply the sub-setting algorithm known as Adaptive Lass.
ContributorsMora, Juan (Author) / Kamarianakis, Ioannis (Thesis director) / Yu, Wanchunzi (Committee member) / W. P. Carey School of Business (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2017-05
Description
This paper attempts to introduce analytics and regression techniques into the National Hockey League. Hockey as a sport has been a slow adapter of analytics, and this can be attributed to poor data collection methods. Using data collected for hockeyreference.com, and R statistical software, the number of wins a team

This paper attempts to introduce analytics and regression techniques into the National Hockey League. Hockey as a sport has been a slow adapter of analytics, and this can be attributed to poor data collection methods. Using data collected for hockeyreference.com, and R statistical software, the number of wins a team experiences will be predicted using Goals For and Goals Against statistics from 2005-2017. The model showed statistical significance and strong normality throughout the data. The number of wins each team was expected to experience in 2016-2017 was predicted using the model and then compared to the actual number of games each team won. To further analyze the validity of the model, the expected playoff outcome for 2016-2017 was compared to the observed playoff outcome. The discussion focused on team's that did not fit the model or traditional analytics and expected forecasts. The possible discrepancies were analyzed using the Las Vegas Golden Knights as a case study. Possible next steps for data analysis are presented and the role of future technology and innovation in hockey analytics is discussed and predicted.
ContributorsVermeer, Brandon Elliot (Author) / Goegan, Brian (Thesis director) / Eaton, John (Committee member) / School of Mathematical and Statistical Sciences (Contributor) / Department of Finance (Contributor) / Barrett, The Honors College (Contributor)
Created2018-05
137727-Thumbnail Image.png
Description
Plastics continue to benefit society in innumerable ways, even though recent public focus on plastics has centered mostly on human health and environmental concerns, including their endocrine-disrupting properties and the long-term pollution they represent. The benefits of plastics are particularly apparent in medicine and public health. Plastics are versatile, cost-effective,

Plastics continue to benefit society in innumerable ways, even though recent public focus on plastics has centered mostly on human health and environmental concerns, including their endocrine-disrupting properties and the long-term pollution they represent. The benefits of plastics are particularly apparent in medicine and public health. Plastics are versatile, cost-effective, require less energy to produce than alternative materials like metal or glass, and can be manufactured to have many different properties. Due to these characteristics, polymers are used in diverse health applications like disposable syringes and intravenous bags, sterile packaging for medical instruments as well as in joint replacements, tissue engineering, etc. However, not all current uses of plastics are prudent and sustainable, as illustrated by the widespread, unwanted human exposure to endocrine-disrupting bisphenol A (BPA) and di-(2-ethylhexyl) phthalate (DEHP), problems arising from the large quantities of plastic being disposed of, and depletion of non-renewable petroleum resources as a result of the ever-increasing mass production of plastic consumer articles. Using the health-care sector as example, this review concentrates on the benefits and downsides of plastics and identifies opportunities to change the composition and disposal practices of these invaluable polymers for a more sustainable future consumption. It highlights ongoing efforts to phase out DEHP and BPA in the health-care and food industry and discusses biodegradable options for plastic packaging, opportunities for reducing plastic medical waste, and recycling in medical facilities in the quest to reap a maximum of benefits from polymers without compromising human health or the environment in the process.
ContributorsNorth, Emily Jean (Co-author) / Halden, Rolf (Co-author, Thesis director) / Mikhail, Chester (Committee member) / Hurlbut, Ben (Committee member) / Barrett, The Honors College (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Chemical Engineering Program (Contributor)
Created2013-05