Matching Items (2)
Filtering by

Clear all filters

156200-Thumbnail Image.png
Description
Modern, advanced statistical tools from data mining and machine learning have become commonplace in molecular biology in large part because of the “big data” demands of various kinds of “-omics” (e.g., genomics, transcriptomics, metabolomics, etc.). However, in other fields of biology where empirical data sets are conventionally smaller, more

Modern, advanced statistical tools from data mining and machine learning have become commonplace in molecular biology in large part because of the “big data” demands of various kinds of “-omics” (e.g., genomics, transcriptomics, metabolomics, etc.). However, in other fields of biology where empirical data sets are conventionally smaller, more traditional statistical methods of inference are still very effective and widely used. Nevertheless, with the decrease in cost of high-performance computing, these fields are starting to employ simulation models to generate insights into questions that have been elusive in the laboratory and field. Although these computational models allow for exquisite control over large numbers of parameters, they also generate data at a qualitatively different scale than most experts in these fields are accustomed to. Thus, more sophisticated methods from big-data statistics have an opportunity to better facilitate the often-forgotten area of bioinformatics that might be called “in-silicomics”.

As a case study, this thesis develops methods for the analysis of large amounts of data generated from a simulated ecosystem designed to understand how mammalian biomechanics interact with environmental complexity to modulate the outcomes of predator–prey interactions. These simulations investigate how other biomechanical parameters relating to the agility of animals in predator–prey pairs are better predictors of pursuit outcomes. Traditional modelling techniques such as forward, backward, and stepwise variable selection are initially used to study these data, but the number of parameters and potentially relevant interaction effects render these methods impractical. Consequently, new modelling techniques such as LASSO regularization are used and compared to the traditional techniques in terms of accuracy and computational complexity. Finally, the splitting rules and instances in the leaves of classification trees provide the basis for future simulation with an economical number of additional runs. In general, this thesis shows the increased utility of these sophisticated statistical techniques with simulated ecological data compared to the approaches traditionally used in these fields. These techniques combined with methods from industrial Design of Experiments will help ecologists extract novel insights from simulations that combine habitat complexity, population structure, and biomechanics.
ContributorsSeto, Christian (Author) / Pavlic, Theodore (Thesis advisor) / Li, Jing (Committee member) / Yan, Hao (Committee member) / Arizona State University (Publisher)
Created2018
Description

The first step in process improvement is to scope the problem, next is measure the current process, but if data is not readily available and cannot be manually collected, then a measurement system must be implemented. General Dynamics Mission Systems (GDMS) is a lean company that is always seeking to

The first step in process improvement is to scope the problem, next is measure the current process, but if data is not readily available and cannot be manually collected, then a measurement system must be implemented. General Dynamics Mission Systems (GDMS) is a lean company that is always seeking to improve. One of their current bottlenecks is the incoming inspection department. This department is responsible for finding defects on parts purchased and is critical to the high reliability product produced by GDMS. To stay competitive and hold their market share, a decision was made to optimize incoming inspection. This proved difficult because no data is being collected. Early steps in many process improvement methodologies, such as Define, Measure, Analyze, Improve and Control (DMAIC), include data collection; however, no measurement system was in place, resulting in no available data for improvement. The solution to this problem was to design and implement a Management Information System (MIS) that will track a variety of data. This will provide the company with data that will be used for analysis and improvement. The first stage of the MIS was developed in Microsoft Excel with Visual Basic for Applications because of the low cost and overall effectiveness of the software. Excel allows update to be made quickly, and allows GDMS to collect data immediately. Stage two would be moving the MIS to a more practicable software, such as Access or MySQL. This thesis is only focuses on stage one of the MIS, and GDMS will proceed with stage two.

ContributorsDiaz, Angel (Author) / McCarville, Daniel R. (Thesis director) / Pavlic, Theodore (Committee member) / Industrial, Systems (Contributor) / Barrett, The Honors College (Contributor)
Created2017-05