Matching Items (419)
Filtering by

Clear all filters

151681-Thumbnail Image.png
Description
Using experience, observations, data, current research, and writings in the field of volunteer management, it was determined there was a need to study the effects of leadership/management practices on the productivity outcomes of a volunteer construction workforce. A simple wood bench that would be tiled and painted was designed to

Using experience, observations, data, current research, and writings in the field of volunteer management, it was determined there was a need to study the effects of leadership/management practices on the productivity outcomes of a volunteer construction workforce. A simple wood bench that would be tiled and painted was designed to test the areas of Time, Waste, Quality, Safety, and Satisfaction of different volunteer groups. The challenge was bolstered by giving the teams no power tools and limited available resources. A simple design of experiment model was used to test highs and lows in the three management techniques of Instruction, Help, and Encouragement. Each scenario was tested multiple times. Data was collected, normalized and analyzed using statistical analysis software. A few significant findings were discovered. The first; the research showed that there was no significant correlation between the management practices of the leader and the satisfaction of the volunteers. The second; the research also showed when further analyzed into specific realistic scenarios that the organizations would be better to focus on high amounts of Help and Encouragement in order to maximize the productivity of their volunteer construction workforce. This is significant as it allows NPO's and governments to focus their attention where best suited to produce results. The results were shared and the study was further validated as "significant" by conducting interviews with experts in the construction nonprofit sector.
ContributorsPrigge, Diedrich (Author) / Sullivan, Kenneth (Thesis advisor) / Wiezel, Avi (Committee member) / Badger, William (Committee member) / Arizona State University (Publisher)
Created2013
152220-Thumbnail Image.png
Description
Many longitudinal studies, especially in clinical trials, suffer from missing data issues. Most estimation procedures assume that the missing values are ignorable or missing at random (MAR). However, this assumption leads to unrealistic simplification and is implausible for many cases. For example, an investigator is examining the effect of treatment

Many longitudinal studies, especially in clinical trials, suffer from missing data issues. Most estimation procedures assume that the missing values are ignorable or missing at random (MAR). However, this assumption leads to unrealistic simplification and is implausible for many cases. For example, an investigator is examining the effect of treatment on depression. Subjects are scheduled with doctors on a regular basis and asked questions about recent emotional situations. Patients who are experiencing severe depression are more likely to miss an appointment and leave the data missing for that particular visit. Data that are not missing at random may produce bias in results if the missing mechanism is not taken into account. In other words, the missing mechanism is related to the unobserved responses. Data are said to be non-ignorable missing if the probabilities of missingness depend on quantities that might not be included in the model. Classical pattern-mixture models for non-ignorable missing values are widely used for longitudinal data analysis because they do not require explicit specification of the missing mechanism, with the data stratified according to a variety of missing patterns and a model specified for each stratum. However, this usually results in under-identifiability, because of the need to estimate many stratum-specific parameters even though the eventual interest is usually on the marginal parameters. Pattern mixture models have the drawback that a large sample is usually required. In this thesis, two studies are presented. The first study is motivated by an open problem from pattern mixture models. Simulation studies from this part show that information in the missing data indicators can be well summarized by a simple continuous latent structure, indicating that a large number of missing data patterns may be accounted by a simple latent factor. Simulation findings that are obtained in the first study lead to a novel model, a continuous latent factor model (CLFM). The second study develops CLFM which is utilized for modeling the joint distribution of missing values and longitudinal outcomes. The proposed CLFM model is feasible even for small sample size applications. The detailed estimation theory, including estimating techniques from both frequentist and Bayesian perspectives is presented. Model performance and evaluation are studied through designed simulations and three applications. Simulation and application settings change from correctly-specified missing data mechanism to mis-specified mechanism and include different sample sizes from longitudinal studies. Among three applications, an AIDS study includes non-ignorable missing values; the Peabody Picture Vocabulary Test data have no indication on missing data mechanism and it will be applied to a sensitivity analysis; the Growth of Language and Early Literacy Skills in Preschoolers with Developmental Speech and Language Impairment study, however, has full complete data and will be used to conduct a robust analysis. The CLFM model is shown to provide more precise estimators, specifically on intercept and slope related parameters, compared with Roy's latent class model and the classic linear mixed model. This advantage will be more obvious when a small sample size is the case, where Roy's model experiences challenges on estimation convergence. The proposed CLFM model is also robust when missing data are ignorable as demonstrated through a study on Growth of Language and Early Literacy Skills in Preschoolers.
ContributorsZhang, Jun (Author) / Reiser, Mark R. (Thesis advisor) / Barber, Jarrett (Thesis advisor) / Kao, Ming-Hung (Committee member) / Wilson, Jeffrey (Committee member) / St Louis, Robert D. (Committee member) / Arizona State University (Publisher)
Created2013
152223-Thumbnail Image.png
Description
Nowadays product reliability becomes the top concern of the manufacturers and customers always prefer the products with good performances under long period. In order to estimate the lifetime of the product, accelerated life testing (ALT) is introduced because most of the products can last years even decades. Much research has

Nowadays product reliability becomes the top concern of the manufacturers and customers always prefer the products with good performances under long period. In order to estimate the lifetime of the product, accelerated life testing (ALT) is introduced because most of the products can last years even decades. Much research has been done in the ALT area and optimal design for ALT is a major topic. This dissertation consists of three main studies. First, a methodology of finding optimal design for ALT with right censoring and interval censoring have been developed and it employs the proportional hazard (PH) model and generalized linear model (GLM) to simplify the computational process. A sensitivity study is also given to show the effects brought by parameters to the designs. Second, an extended version of I-optimal design for ALT is discussed and then a dual-objective design criterion is defined and showed with several examples. Also in order to evaluate different candidate designs, several graphical tools are developed. Finally, when there are more than one models available, different model checking designs are discussed.
ContributorsYang, Tao (Author) / Pan, Rong (Thesis advisor) / Montgomery, Douglas C. (Committee member) / Borror, Connie (Committee member) / Rigdon, Steve (Committee member) / Arizona State University (Publisher)
Created2013
152189-Thumbnail Image.png
Description
This work presents two complementary studies that propose heuristic methods to capture characteristics of data using the ensemble learning method of random forest. The first study is motivated by the problem in education of determining teacher effectiveness in student achievement. Value-added models (VAMs), constructed as linear mixed models, use students’

This work presents two complementary studies that propose heuristic methods to capture characteristics of data using the ensemble learning method of random forest. The first study is motivated by the problem in education of determining teacher effectiveness in student achievement. Value-added models (VAMs), constructed as linear mixed models, use students’ test scores as outcome variables and teachers’ contributions as random effects to ascribe changes in student performance to the teachers who have taught them. The VAMs teacher score is the empirical best linear unbiased predictor (EBLUP). This approach is limited by the adequacy of the assumed model specification with respect to the unknown underlying model. In that regard, this study proposes alternative ways to rank teacher effects that are not dependent on a given model by introducing two variable importance measures (VIMs), the node-proportion and the covariate-proportion. These VIMs are novel because they take into account the final configuration of the terminal nodes in the constitutive trees in a random forest. In a simulation study, under a variety of conditions, true rankings of teacher effects are compared with estimated rankings obtained using three sources: the newly proposed VIMs, existing VIMs, and EBLUPs from the assumed linear model specification. The newly proposed VIMs outperform all others in various scenarios where the model was misspecified. The second study develops two novel interaction measures. These measures could be used within but are not restricted to the VAM framework. The distribution-based measure is constructed to identify interactions in a general setting where a model specification is not assumed in advance. In turn, the mean-based measure is built to estimate interactions when the model specification is assumed to be linear. Both measures are unique in their construction; they take into account not only the outcome values, but also the internal structure of the trees in a random forest. In a separate simulation study, under a variety of conditions, the proposed measures are found to identify and estimate second-order interactions.
ContributorsValdivia, Arturo (Author) / Eubank, Randall (Thesis advisor) / Young, Dennis (Committee member) / Reiser, Mark R. (Committee member) / Kao, Ming-Hung (Committee member) / Broatch, Jennifer (Committee member) / Arizona State University (Publisher)
Created2013
152244-Thumbnail Image.png
Description
Statistics is taught at every level of education, yet teachers often have to assume their students have no knowledge of statistics and start from scratch each time they set out to teach statistics. The motivation for this experimental study comes from interest in exploring educational applications of augmented reality (AR)

Statistics is taught at every level of education, yet teachers often have to assume their students have no knowledge of statistics and start from scratch each time they set out to teach statistics. The motivation for this experimental study comes from interest in exploring educational applications of augmented reality (AR) delivered via mobile technology that could potentially provide rich, contextualized learning for understanding concepts related to statistics education. This study examined the effects of AR experiences for learning basic statistical concepts. Using a 3 x 2 research design, this study compared learning gains of 252 undergraduate and graduate students from a pre- and posttest given before and after interacting with one of three types of augmented reality experiences, a high AR experience (interacting with three dimensional images coupled with movement through a physical space), a low AR experience (interacting with three dimensional images without movement), or no AR experience (two dimensional images without movement). Two levels of collaboration (pairs and no pairs) were also included. Additionally, student perceptions toward collaboration opportunities and engagement were compared across the six treatment conditions. Other demographic information collected included the students' previous statistics experience, as well as their comfort level in using mobile devices. The moderating variables included prior knowledge (high, average, and low) as measured by the student's pretest score. Taking into account prior knowledge, students with low prior knowledge assigned to either high or low AR experience had statistically significant higher learning gains than those assigned to a no AR experience. On the other hand, the results showed no statistical significance between students assigned to work individually versus in pairs. Students assigned to both high and low AR experience perceived a statistically significant higher level of engagement than their no AR counterparts. Students with low prior knowledge benefited the most from the high AR condition in learning gains. Overall, the AR application did well for providing a hands-on experience working with statistical data. Further research on AR and its relationship to spatial cognition, situated learning, high order skill development, performance support, and other classroom applications for learning is still needed.
ContributorsConley, Quincy (Author) / Atkinson, Robert K (Thesis advisor) / Nguyen, Frank (Committee member) / Nelson, Brian C (Committee member) / Arizona State University (Publisher)
Created2013
151913-Thumbnail Image.png
Description
In this mixed-methods study, I examined the relationship between professional development based on the Common Core State Standards for Mathematics and teacher knowledge, classroom practice, and student learning. Participants were randomly assigned to experimental and control groups. The 50-hour professional development treatment was administered to the treatment group during one

In this mixed-methods study, I examined the relationship between professional development based on the Common Core State Standards for Mathematics and teacher knowledge, classroom practice, and student learning. Participants were randomly assigned to experimental and control groups. The 50-hour professional development treatment was administered to the treatment group during one semester, and then a follow-up replication treatment was administered to the control group during the subsequent semester. Results revealed significant differences in teacher knowledge as a result of the treatment using two instruments. The Learning Mathematics for Teaching scales were used to detect changes in mathematical knowledge for teaching, and an online sorting task was used to detect changes in teachers' knowledge of their standards. Results also indicated differences in classroom practice between pairs of matched teachers selected to participate in classroom observations and interviews. No statistical difference was detected between the groups' student assessment scores using the district's benchmark assessment system. This efficacy study contributes to the literature in two ways. First, it provides an evidence base for a professional development model designed to promote effective implementation of the Common Core State Standards for Mathematics. Second, it addresses ways to impact and measure teachers' knowledge of curriculum in addition to their mathematical content knowledge. The treatment was designed to focus on knowledge of curriculum, but it also successfully impacted teachers' specialized content knowledge, knowledge of content and students, and knowledge of content and teaching.
ContributorsRimbey, Kimberly A (Author) / Middleton, James A. (Thesis advisor) / Sloane, Finbarr (Committee member) / Atkinson, Robert K (Committee member) / Arizona State University (Publisher)
Created2013
152039-Thumbnail Image.png
Description
An integral part of teacher development are teacher observations. Many teachers are observed once or twice a year to evaluate their performance and hold them accountable for meeting standards. Instructional coaches, however, observe and work with teachers to help them reflect on their performance, with the goal of improving their

An integral part of teacher development are teacher observations. Many teachers are observed once or twice a year to evaluate their performance and hold them accountable for meeting standards. Instructional coaches, however, observe and work with teachers to help them reflect on their performance, with the goal of improving their practice. Video-based evidence has long been used in connection with teacher reflection and as the technology necessary to record video has become more readily available, video recordings have found an increasing presence in teacher observations. In addition, more and more schools are turning to mobile technology to help record evidence during teacher observations. Several mobile applications have been developed, which are designed to help instructional coaches, administrators, and teachers make the most of teacher observations. This study looked at the use of the DataCapture mobile application to record video-based evidence in teacher observations as part of an instructional coaching program in a large public school district in the Southwestern United States. Six instructional coaches and two teachers participated in interviews at the end of the study period. Additional data was collected from the DataCapture mobile application and from a survey of instructional coaches conducted by the school district in connection with its Title I programs. Results show that instructional coaches feel that using video-based evidence for teacher reflection is effective in a number of ways. Teachers who have experienced seeing themselves on video also felt that video-based evidence is effective at improving teacher reflection, while teachers who have not yet experienced seeing themselves on video displayed extreme apprehensiveness about being video recorded in the classroom. Instructional coaches felt the DataCapture mobile application was beneficial in teacher evaluation, but there were several issues that impacted the use of the mobile application and video-based evidence, including logistics, time requirements, and administrative support. The discussion focuses on recommendations for successfully using video-based evidence in an instructional coaching context, as well as some suggestions for other researchers attempting to study how video-based evidence impacts teachers' ability to reflect on their own teaching.
ContributorsShewell, Justin Reed (Author) / Bitter, Gary (Thesis advisor) / Dawson, Edwin (Committee member) / Blair, Heidi (Committee member) / Arizona State University (Publisher)
Created2013
151988-Thumbnail Image.png
Description
Growing popularity of alternatively certifying teachers has created challenges for teacher preparation programs. Many non-traditional routes into classroom include no full-time mentor teacher. Absence of a mentor teacher in the classroom leaves teachers with a deficit. This study follows ten teachers on the intern certificate enrolled in both an alternative

Growing popularity of alternatively certifying teachers has created challenges for teacher preparation programs. Many non-traditional routes into classroom include no full-time mentor teacher. Absence of a mentor teacher in the classroom leaves teachers with a deficit. This study follows ten teachers on the intern certificate enrolled in both an alternative certification teacher preparation program and the Teach for America organization as they pursue a master's degree in education and state teaching certification from a large southwestern university. The five randomly chosen for the treatment group and the control group contained 1 male and 4 female teachers, some of whom teach at public schools and others at charter schools. All were secondary education language arts teachers ranging in age from 22- 29. The treatment used in this study is a job-embedded, professional development, software tool designed to help teachers track their classroom practices called MyiLOGS. The purpose of this action research project was to study the effect using MyiLOGS had on six of the nine areas evaluated by a modified version of the Teacher Advancement Program evaluation rubric, alignment with Opportunity To Learn constructs, and the tool's influence on the efficacy of these first year teachers. The data generated from this study indicate that the MyiLOGS tool did have a positive effect on the teachers' TAP evaluation performances. Also, the MyiLOGS tool had a large impact on the teachers' instruction as measured by the constructs of Opportunity to Learn and their teaching self-efficacy. Implications suggested the tool was an asset to these teachers because they tracked their data, became more reflective, and self-sufficient.
ContributorsRoggeman, Pamela (Author) / Puckett, Kathleen (Thesis advisor) / Kurz, Alexander (Committee member) / Mathur, Sarup (Committee member) / Arizona State University (Publisher)
Created2013
151992-Thumbnail Image.png
Description
Dimensionality assessment is an important component of evaluating item response data. Existing approaches to evaluating common assumptions of unidimensionality, such as DIMTEST (Nandakumar & Stout, 1993; Stout, 1987; Stout, Froelich, & Gao, 2001), have been shown to work well under large-scale assessment conditions (e.g., large sample sizes and item pools;

Dimensionality assessment is an important component of evaluating item response data. Existing approaches to evaluating common assumptions of unidimensionality, such as DIMTEST (Nandakumar & Stout, 1993; Stout, 1987; Stout, Froelich, & Gao, 2001), have been shown to work well under large-scale assessment conditions (e.g., large sample sizes and item pools; see e.g., Froelich & Habing, 2007). It remains to be seen how such procedures perform in the context of small-scale assessments characterized by relatively small sample sizes and/or short tests. The fact that some procedures come with minimum allowable values for characteristics of the data, such as the number of items, may even render them unusable for some small-scale assessments. Other measures designed to assess dimensionality do not come with such limitations and, as such, may perform better under conditions that do not lend themselves to evaluation via statistics that rely on asymptotic theory. The current work aimed to evaluate the performance of one such metric, the standardized generalized dimensionality discrepancy measure (SGDDM; Levy & Svetina, 2011; Levy, Xu, Yel, & Svetina, 2012), under both large- and small-scale testing conditions. A Monte Carlo study was conducted to compare the performance of DIMTEST and the SGDDM statistic in terms of evaluating assumptions of unidimensionality in item response data under a variety of conditions, with an emphasis on the examination of these procedures in small-scale assessments. Similar to previous research, increases in either test length or sample size resulted in increased power. The DIMTEST procedure appeared to be a conservative test of the null hypothesis of unidimensionality. The SGDDM statistic exhibited rejection rates near the nominal rate of .05 under unidimensional conditions, though the reliability of these results may have been less than optimal due to high sampling variability resulting from a relatively limited number of replications. Power values were at or near 1.0 for many of the multidimensional conditions. It was only when the sample size was reduced to N = 100 that the two approaches diverged in performance. Results suggested that both procedures may be appropriate for sample sizes as low as N = 250 and tests as short as J = 12 (SGDDM) or J = 19 (DIMTEST). When used as a diagnostic tool, SGDDM may be appropriate with as few as N = 100 cases combined with J = 12 items. The study was somewhat limited in that it did not include any complex factorial designs, nor were the strength of item discrimination parameters or correlation between factors manipulated. It is recommended that further research be conducted with the inclusion of these factors, as well as an increase in the number of replications when using the SGDDM procedure.
ContributorsReichenberg, Ray E (Author) / Levy, Roy (Thesis advisor) / Thompson, Marilyn S. (Thesis advisor) / Green, Samuel B. (Committee member) / Arizona State University (Publisher)
Created2013
151976-Thumbnail Image.png
Description
Parallel Monte Carlo applications require the pseudorandom numbers used on each processor to be independent in a probabilistic sense. The TestU01 software package is the standard testing suite for detecting stream dependence and other properties that make certain pseudorandom generators ineffective in parallel (as well as serial) settings. TestU01 employs

Parallel Monte Carlo applications require the pseudorandom numbers used on each processor to be independent in a probabilistic sense. The TestU01 software package is the standard testing suite for detecting stream dependence and other properties that make certain pseudorandom generators ineffective in parallel (as well as serial) settings. TestU01 employs two basic schemes for testing parallel generated streams. The first applies serial tests to the individual streams and then tests the resulting P-values for uniformity. The second turns all the parallel generated streams into one long vector and then applies serial tests to the resulting concatenated stream. Various forms of stream dependence can be missed by each approach because neither one fully addresses the multivariate nature of the accumulated data when generators are run in parallel. This dissertation identifies these potential faults in the parallel testing methodologies of TestU01 and investigates two different methods to better detect inter-stream dependencies: correlation motivated multivariate tests and vector time series based tests. These methods have been implemented in an extension to TestU01 built in C++ and the unique aspects of this extension are discussed. A variety of different generation scenarios are then examined using the TestU01 suite in concert with the extension. This enhanced software package is found to better detect certain forms of inter-stream dependencies than the original TestU01 suites of tests.
ContributorsIsmay, Chester (Author) / Eubank, Randall (Thesis advisor) / Young, Dennis (Committee member) / Kao, Ming-Hung (Committee member) / Lanchier, Nicolas (Committee member) / Reiser, Mark R. (Committee member) / Arizona State University (Publisher)
Created2013