Matching Items (48)
Filtering by

Clear all filters

147863-Thumbnail Image.png
Description

Over the years, advances in research have continued to decrease the size of computers from the size of<br/>a room to a small device that could fit in one’s palm. However, if an application does not require extensive<br/>computation power nor accessories such as a screen, the corresponding machine could be microscopic,<br/>only

Over the years, advances in research have continued to decrease the size of computers from the size of<br/>a room to a small device that could fit in one’s palm. However, if an application does not require extensive<br/>computation power nor accessories such as a screen, the corresponding machine could be microscopic,<br/>only a few nanometers big. Researchers at MIT have successfully created Syncells, which are micro-<br/>scale robots with limited computation power and memory that can communicate locally to achieve<br/>complex collective tasks. In order to control these Syncells for a desired outcome, they must each run a<br/>simple distributed algorithm. As they are only capable of local communication, Syncells cannot receive<br/>commands from a control center, so their algorithms cannot be centralized. In this work, we created a<br/>distributed algorithm that each Syncell can execute so that the system of Syncells is able to find and<br/>converge to a specific target within the environment. The most direct applications of this problem are in<br/>medicine. Such a system could be used as a safer alternative to invasive surgery or could be used to treat<br/>internal bleeding or tumors. We tested and analyzed our algorithm through simulation and visualization<br/>in Python. Overall, our algorithm successfully caused the system of particles to converge on a specific<br/>target present within the environment.

ContributorsMartin, Rebecca Clare (Author) / Richa, Andréa (Thesis director) / Lee, Heewook (Committee member) / Computer Science and Engineering Program (Contributor) / School of Mathematical and Statistical Sciences (Contributor, Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148156-Thumbnail Image.png
Description

This thesis project is part of a larger collaboration documenting the history of the ASU Biodesign Clinical Testing Laboratory (ABCTL). There are many different aspects that need to be considered when transforming to a clinical testing laboratory. This includes the different types of tests performed in the laboratory. In addition

This thesis project is part of a larger collaboration documenting the history of the ASU Biodesign Clinical Testing Laboratory (ABCTL). There are many different aspects that need to be considered when transforming to a clinical testing laboratory. This includes the different types of tests performed in the laboratory. In addition to the diagnostic polymerase chain reaction (PCR) test that is performed detecting the presence of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), antibody testing is also performed in clinical laboratories. Antibody testing is used to detect a previous infection. Antibodies are produced as part of the immune response against SARS-CoV-2. There are many different forms of antibody tests and their sensitives and specificities have been examined and reviewed in the literature. Antibody testing can be used to determine the seroprevalence of the disease which can inform policy decisions regarding public health strategies. The results from antibody testing can also be used for creating new therapeutics like vaccines. The ABCTL recognizes the shifting need of the community to begin testing for previous infections of SARS-CoV-2 and is developing new forms of antibody testing that can meet them.

ContributorsRuan, Ellen (Co-author) / Smetanick, Jennifer (Co-author) / Majhail, Kajol (Co-author) / Anderson, Laura (Co-author) / Breshears, Scott (Co-author) / Compton, Carolyn (Thesis director) / Magee, Mitch (Committee member) / School of Life Sciences (Contributor, Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148176-Thumbnail Image.png
Description

In this project, I examined the relationship between lockdowns implemented by COVID-19 and the activity of animals in urban areas. I hypothesized that animals became more active in urban areas during COVID-19 quarantine than they were before and I wanted to see if my hypothesis could be researched through Twitter

In this project, I examined the relationship between lockdowns implemented by COVID-19 and the activity of animals in urban areas. I hypothesized that animals became more active in urban areas during COVID-19 quarantine than they were before and I wanted to see if my hypothesis could be researched through Twitter crowdsourcing. I began by collecting tweets using python code, but upon examining all data output from code-based searches, I concluded that it is quicker and more efficient to use the advanced search on Twitter website. Based on my research, I can neither confirm nor deny if the appearance of wild animals is due to the COVID-19 lockdowns. However, I was able to discover a correlational relationship between these two factors in some research cases. Although my findings are mixed with regard to my original hypothesis, the impact that this phenomenon had on society cannot be denied.

ContributorsHeimlich, Kiana Raye (Author) / Dorn, Ronald (Thesis director) / Martin, Roberta (Committee member) / Donovan, Mary (Committee member) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
147953-Thumbnail Image.png
Description

Actuaries can analyze healthcare trends to determine if rates are reasonable and if reserves are adequate. In this talk, we will provide a framework of methods to analyze the healthcare trend during the pandemic. COVID-19 may influence future healthcare cost trends in many ways. First, direct COVID-19 costs may increase

Actuaries can analyze healthcare trends to determine if rates are reasonable and if reserves are adequate. In this talk, we will provide a framework of methods to analyze the healthcare trend during the pandemic. COVID-19 may influence future healthcare cost trends in many ways. First, direct COVID-19 costs may increase the amount of total experienced healthcare costs. However, with the implementation of social distancing, the amount of regularly scheduled care may be deferred to a future date. There are also many unknown factors regarding the transmission of the virus. Implementing epidemiology models allows us to predict infections by studying the dynamics of the disease. The correlation between infection amounts and hospitalization occupancies provide a methodology to estimate the amount of deferred and recouped amounts of regularly scheduled healthcare costs. Thus, the combination of the models allows to model the healthcare cost trend impact due to COVID-19.

ContributorsGabric, Lydia Joan (Author) / Zhou, Hongjuan (Thesis director) / Zicarelli, John (Committee member) / School of Mathematical and Statistical Sciences (Contributor, Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148207-Thumbnail Image.png
Description

Optimal foraging theory provides a suite of tools that model the best way that an animal will <br/>structure its searching and processing decisions in uncertain environments. It has been <br/>successful characterizing real patterns of animal decision making, thereby providing insights<br/>into why animals behave the way they do. However, it does

Optimal foraging theory provides a suite of tools that model the best way that an animal will <br/>structure its searching and processing decisions in uncertain environments. It has been <br/>successful characterizing real patterns of animal decision making, thereby providing insights<br/>into why animals behave the way they do. However, it does not speak to how animals make<br/>decisions that tend to be adaptive. Using simulation studies, prior work has shown empirically<br/>that a simple decision-making heuristic tends to produce prey-choice behaviors that, on <br/>average, match the predicted behaviors of optimal foraging theory. That heuristic chooses<br/>to spend time processing an encountered prey item if that prey item's marginal rate of<br/>caloric gain (in calories per unit of processing time) is greater than the forager's<br/>current long-term rate of accumulated caloric gain (in calories per unit of total searching<br/>and processing time). Although this heuristic may seem intuitive, a rigorous mathematical<br/>argument for why it tends to produce the theorized optimal foraging theory behavior has<br/>not been developed. In this thesis, an analytical argument is given for why this<br/>simple decision-making heuristic is expected to realize the optimal performance<br/>predicted by optimal foraging theory. This theoretical guarantee not only provides support<br/>for why such a heuristic might be favored by natural selection, but it also provides<br/>support for why such a heuristic might a reliable tool for decision-making in autonomous<br/>engineered agents moving through theatres of uncertain rewards. Ultimately, this simple<br/>decision-making heuristic may provide a recipe for reinforcement learning in small robots<br/>with little computational capabilities.

ContributorsCothren, Liliaokeawawa Kiyoko (Author) / Pavlic, Theodore (Thesis director) / Brewer, Naala (Committee member) / School of Mathematical and Statistical Sciences (Contributor, Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
148238-Thumbnail Image.png
Description

Dreadnought is a free-to-play multiplayer flight simulation in which two teams of 8 players each compete against one another to complete an objective. Each player controls a large-scale spaceship, various aspects of which can be customized to improve a player’s performance in a game. One such aspect is Officer Briefings,

Dreadnought is a free-to-play multiplayer flight simulation in which two teams of 8 players each compete against one another to complete an objective. Each player controls a large-scale spaceship, various aspects of which can be customized to improve a player’s performance in a game. One such aspect is Officer Briefings, which are passive abilities that grant ships additional capabilities. Two of these Briefings, known as Retaliator and Get My Good Side, have strong synergy when used together, which has led to the Dreadnought community’s claiming that the Briefings are too powerful and should be rebalanced to be more in line with the power levels of other Briefings. This study collected gameplay data with and without the use of these specific Officer Briefings to determine the precise impact on gameplay. Linear correlation matrices and inference on two means were used to determine performance impact. It was found that, although these Officer Briefings do improve an individual player’s performance in a game, they do not have a consistent impact on the player’s team performance, and that these Officer Briefings are therefore not in need of rebalancing.

ContributorsJacobs, Max I. (Author) / Schneider, Laurence (Thesis director) / Tran, Samantha (Committee member) / Mechanical and Aerospace Engineering Program (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
150135-Thumbnail Image.png
Description
It is common in the analysis of data to provide a goodness-of-fit test to assess the performance of a model. In the analysis of contingency tables, goodness-of-fit statistics are frequently employed when modeling social science, educational or psychological data where the interest is often directed at investigating the association among

It is common in the analysis of data to provide a goodness-of-fit test to assess the performance of a model. In the analysis of contingency tables, goodness-of-fit statistics are frequently employed when modeling social science, educational or psychological data where the interest is often directed at investigating the association among multi-categorical variables. Pearson's chi-squared statistic is well-known in goodness-of-fit testing, but it is sometimes considered to produce an omnibus test as it gives little guidance to the source of poor fit once the null hypothesis is rejected. However, its components can provide powerful directional tests. In this dissertation, orthogonal components are used to develop goodness-of-fit tests for models fit to the counts obtained from the cross-classification of multi-category dependent variables. Ordinal categories are assumed. Orthogonal components defined on marginals are obtained when analyzing multi-dimensional contingency tables through the use of the QR decomposition. A subset of these orthogonal components can be used to construct limited-information tests that allow one to identify the source of lack-of-fit and provide an increase in power compared to Pearson's test. These tests can address the adverse effects presented when data are sparse. The tests rely on the set of first- and second-order marginals jointly, the set of second-order marginals only, and the random forest method, a popular algorithm for modeling large complex data sets. The performance of these tests is compared to the likelihood ratio test as well as to tests based on orthogonal polynomial components. The derived goodness-of-fit tests are evaluated with studies for detecting two- and three-way associations that are not accounted for by a categorical variable factor model with a single latent variable. In addition the tests are used to investigate the case when the model misspecification involves parameter constraints for large and sparse contingency tables. The methodology proposed here is applied to data from the 38th round of the State Survey conducted by the Institute for Public Policy and Michigan State University Social Research (2005) . The results illustrate the use of the proposed techniques in the context of a sparse data set.
ContributorsMilovanovic, Jelena (Author) / Young, Dennis (Thesis advisor) / Reiser, Mark R. (Thesis advisor) / Wilson, Jeffrey (Committee member) / Eubank, Randall (Committee member) / Yang, Yan (Committee member) / Arizona State University (Publisher)
Created2011
151976-Thumbnail Image.png
Description
Parallel Monte Carlo applications require the pseudorandom numbers used on each processor to be independent in a probabilistic sense. The TestU01 software package is the standard testing suite for detecting stream dependence and other properties that make certain pseudorandom generators ineffective in parallel (as well as serial) settings. TestU01 employs

Parallel Monte Carlo applications require the pseudorandom numbers used on each processor to be independent in a probabilistic sense. The TestU01 software package is the standard testing suite for detecting stream dependence and other properties that make certain pseudorandom generators ineffective in parallel (as well as serial) settings. TestU01 employs two basic schemes for testing parallel generated streams. The first applies serial tests to the individual streams and then tests the resulting P-values for uniformity. The second turns all the parallel generated streams into one long vector and then applies serial tests to the resulting concatenated stream. Various forms of stream dependence can be missed by each approach because neither one fully addresses the multivariate nature of the accumulated data when generators are run in parallel. This dissertation identifies these potential faults in the parallel testing methodologies of TestU01 and investigates two different methods to better detect inter-stream dependencies: correlation motivated multivariate tests and vector time series based tests. These methods have been implemented in an extension to TestU01 built in C++ and the unique aspects of this extension are discussed. A variety of different generation scenarios are then examined using the TestU01 suite in concert with the extension. This enhanced software package is found to better detect certain forms of inter-stream dependencies than the original TestU01 suites of tests.
ContributorsIsmay, Chester (Author) / Eubank, Randall (Thesis advisor) / Young, Dennis (Committee member) / Kao, Ming-Hung (Committee member) / Lanchier, Nicolas (Committee member) / Reiser, Mark R. (Committee member) / Arizona State University (Publisher)
Created2013
150996-Thumbnail Image.png
Description
A least total area of triangle method was proposed by Teissier (1948) for fitting a straight line to data from a pair of variables without treating either variable as the dependent variable while allowing each of the variables to have measurement errors. This method is commonly called Reduced Major Axis

A least total area of triangle method was proposed by Teissier (1948) for fitting a straight line to data from a pair of variables without treating either variable as the dependent variable while allowing each of the variables to have measurement errors. This method is commonly called Reduced Major Axis (RMA) regression and is often used instead of Ordinary Least Squares (OLS) regression. Results for confidence intervals, hypothesis testing and asymptotic distributions of coefficient estimates in the bivariate case are reviewed. A generalization of RMA to more than two variables for fitting a plane to data is obtained by minimizing the sum of a function of the volumes obtained by drawing, from each data point, lines parallel to each coordinate axis to the fitted plane (Draper and Yang 1997; Goodman and Tofallis 2003). Generalized RMA results for the multivariate case obtained by Draper and Yang (1997) are reviewed and some investigations of multivariate RMA are given. A linear model is proposed that does not specify a dependent variable and allows for errors in the measurement of each variable. Coefficients in the model are estimated by minimization of the function of the volumes previously mentioned. Methods for obtaining coefficient estimates are discussed and simulations are used to investigate the distribution of coefficient estimates. The effects of sample size, sampling error and correlation among variables on the estimates are studied. Bootstrap methods are used to obtain confidence intervals for model coefficients. Residual analysis is considered for assessing model assumptions. Outlier and influential case diagnostics are developed and a forward selection method is proposed for subset selection of model variables. A real data example is provided that uses the methods developed. Topics for further research are discussed.
ContributorsLi, Jingjin (Author) / Young, Dennis (Thesis advisor) / Eubank, Randall (Thesis advisor) / Reiser, Mark R. (Committee member) / Kao, Ming-Hung (Committee member) / Yang, Yan (Committee member) / Arizona State University (Publisher)
Created2012
151128-Thumbnail Image.png
Description
This dissertation involves three problems that are all related by the use of the singular value decomposition (SVD) or generalized singular value decomposition (GSVD). The specific problems are (i) derivation of a generalized singular value expansion (GSVE), (ii) analysis of the properties of the chi-squared method for regularization parameter selection

This dissertation involves three problems that are all related by the use of the singular value decomposition (SVD) or generalized singular value decomposition (GSVD). The specific problems are (i) derivation of a generalized singular value expansion (GSVE), (ii) analysis of the properties of the chi-squared method for regularization parameter selection in the case of nonnormal data and (iii) formulation of a partial canonical correlation concept for continuous time stochastic processes. The finite dimensional SVD has an infinite dimensional generalization to compact operators. However, the form of the finite dimensional GSVD developed in, e.g., Van Loan does not extend directly to infinite dimensions as a result of a key step in the proof that is specific to the matrix case. Thus, the first problem of interest is to find an infinite dimensional version of the GSVD. One such GSVE for compact operators on separable Hilbert spaces is developed. The second problem concerns regularization parameter estimation. The chi-squared method for nonnormal data is considered. A form of the optimized regularization criterion that pertains to measured data or signals with nonnormal noise is derived. Large sample theory for phi-mixing processes is used to derive a central limit theorem for the chi-squared criterion that holds under certain conditions. Departures from normality are seen to manifest in the need for a possibly different scale factor in normalization rather than what would be used under the assumption of normality. The consequences of our large sample work are illustrated by empirical experiments. For the third problem, a new approach is examined for studying the relationships between a collection of functional random variables. The idea is based on the work of Sunder that provides mappings to connect the elements of algebraic and orthogonal direct sums of subspaces in a Hilbert space. When combined with a key isometry associated with a particular Hilbert space indexed stochastic process, this leads to a useful formulation for situations that involve the study of several second order processes. In particular, using our approach with two processes provides an independent derivation of the functional canonical correlation analysis (CCA) results of Eubank and Hsing. For more than two processes, a rigorous derivation of the functional partial canonical correlation analysis (PCCA) concept that applies to both finite and infinite dimensional settings is obtained.
ContributorsHuang, Qing (Author) / Eubank, Randall (Thesis advisor) / Renaut, Rosemary (Thesis advisor) / Cochran, Douglas (Committee member) / Gelb, Anne (Committee member) / Young, Dennis (Committee member) / Arizona State University (Publisher)
Created2012