Matching Items (7)
Filtering by

Clear all filters

152032-Thumbnail Image.png
Description
In order to analyze data from an instrument administered at multiple time points it is a common practice to form composites of the items at each wave and to fit a longitudinal model to the composites. The advantage of using composites of items is that smaller sample sizes are required

In order to analyze data from an instrument administered at multiple time points it is a common practice to form composites of the items at each wave and to fit a longitudinal model to the composites. The advantage of using composites of items is that smaller sample sizes are required in contrast to second order models that include the measurement and the structural relationships among the variables. However, the use of composites assumes that longitudinal measurement invariance holds; that is, it is assumed that that the relationships among the items and the latent variables remain constant over time. Previous studies conducted on latent growth models (LGM) have shown that when longitudinal metric invariance is violated, the parameter estimates are biased and that mistaken conclusions about growth can be made. The purpose of the current study was to examine the impact of non-invariant loadings and non-invariant intercepts on two longitudinal models: the LGM and the autoregressive quasi-simplex model (AR quasi-simplex). A second purpose was to determine if there are conditions in which researchers can reach adequate conclusions about stability and growth even in the presence of violations of invariance. A Monte Carlo simulation study was conducted to achieve the purposes. The method consisted of generating items under a linear curve of factors model (COFM) or under the AR quasi-simplex. Composites of the items were formed at each time point and analyzed with a linear LGM or an AR quasi-simplex model. The results showed that AR quasi-simplex model yielded biased path coefficients only in the conditions with large violations of invariance. The fit of the AR quasi-simplex was not affected by violations of invariance. In general, the growth parameter estimates of the LGM were biased under violations of invariance. Further, in the presence of non-invariant loadings the rejection rates of the hypothesis of linear growth increased as the proportion of non-invariant items and as the magnitude of violations of invariance increased. A discussion of the results and limitations of the study are provided as well as general recommendations.
ContributorsOlivera-Aguilar, Margarita (Author) / Millsap, Roger E. (Thesis advisor) / Levy, Roy (Committee member) / MacKinnon, David (Committee member) / West, Stephen G. (Committee member) / Arizona State University (Publisher)
Created2013
151957-Thumbnail Image.png
Description
Random Forests is a statistical learning method which has been proposed for propensity score estimation models that involve complex interactions, nonlinear relationships, or both of the covariates. In this dissertation I conducted a simulation study to examine the effects of three Random Forests model specifications in propensity score analysis. The

Random Forests is a statistical learning method which has been proposed for propensity score estimation models that involve complex interactions, nonlinear relationships, or both of the covariates. In this dissertation I conducted a simulation study to examine the effects of three Random Forests model specifications in propensity score analysis. The results suggested that, depending on the nature of data, optimal specification of (1) decision rules to select the covariate and its split value in a Classification Tree, (2) the number of covariates randomly sampled for selection, and (3) methods of estimating Random Forests propensity scores could potentially produce an unbiased average treatment effect estimate after propensity scores weighting by the odds adjustment. Compared to the logistic regression estimation model using the true propensity score model, Random Forests had an additional advantage in producing unbiased estimated standard error and correct statistical inference of the average treatment effect. The relationship between the balance on the covariates' means and the bias of average treatment effect estimate was examined both within and between conditions of the simulation. Within conditions, across repeated samples there was no noticeable correlation between the covariates' mean differences and the magnitude of bias of average treatment effect estimate for the covariates that were imbalanced before adjustment. Between conditions, small mean differences of covariates after propensity score adjustment were not sensitive enough to identify the optimal Random Forests model specification for propensity score analysis.
ContributorsCham, Hei Ning (Author) / Tein, Jenn-Yun (Thesis advisor) / Enders, Stephen G (Thesis advisor) / Enders, Craig K. (Committee member) / Mackinnon, David P (Committee member) / Arizona State University (Publisher)
Created2013
153904-Thumbnail Image.png
Description
Recent advances in hierarchical or multilevel statistical models and causal inference using the potential outcomes framework hold tremendous promise for mock and real jury research. These advances enable researchers to explore how individual jurors can exert a bottom-up effect on the jury’s verdict and how case-level features can exert a

Recent advances in hierarchical or multilevel statistical models and causal inference using the potential outcomes framework hold tremendous promise for mock and real jury research. These advances enable researchers to explore how individual jurors can exert a bottom-up effect on the jury’s verdict and how case-level features can exert a top-down effect on a juror’s perception of the parties at trial. This dissertation explains and then applies these technical advances to a pre-existing mock jury dataset to provide worked examples in an effort to spur the adoption of these techniques. In particular, the paper introduces two new cross-level mediated effects and then describes how to conduct ecological validity tests with these mediated effects. The first cross-level mediated effect, the a1b1 mediated effect, is the juror level mediated effect for a jury level manipulation. The second cross-level mediated effect, the a2bc mediated effect, is the unique contextual effect that being in a jury has on the individual the juror. When a mock jury study includes a deliberation versus non-deliberation manipulation, the a1b1 can be compared for the two conditions, enabling a general test of ecological validity. If deliberating in a group generally influences the individual, then the two indirect effects should be significantly different. The a2bc can also be interpreted as a specific test of how much changes in jury level means of this specific mediator effect juror level decision-making.
ContributorsLovis-McMahon, David (Author) / Schweitzer, Nicholas (Thesis advisor) / Saks, Michael (Thesis advisor) / Salerno, Jessica (Committee member) / MacKinnon, David (Committee member) / Arizona State University (Publisher)
Created2015
155625-Thumbnail Image.png
Description
The process of combining data is one in which information from disjoint datasets sharing at least a number of common variables is merged. This process is commonly referred to as data fusion, with the main objective of creating a new dataset permitting more flexible analyses than the separate analysis of

The process of combining data is one in which information from disjoint datasets sharing at least a number of common variables is merged. This process is commonly referred to as data fusion, with the main objective of creating a new dataset permitting more flexible analyses than the separate analysis of each individual dataset. Many data fusion methods have been proposed in the literature, although most utilize the frequentist framework. This dissertation investigates a new approach called Bayesian Synthesis in which information obtained from one dataset acts as priors for the next analysis. This process continues sequentially until a single posterior distribution is created using all available data. These informative augmented data-dependent priors provide an extra source of information that may aid in the accuracy of estimation. To examine the performance of the proposed Bayesian Synthesis approach, first, results of simulated data with known population values under a variety of conditions were examined. Next, these results were compared to those from the traditional maximum likelihood approach to data fusion, as well as the data fusion approach analyzed via Bayes. The assessment of parameter recovery based on the proposed Bayesian Synthesis approach was evaluated using four criteria to reflect measures of raw bias, relative bias, accuracy, and efficiency. Subsequently, empirical analyses with real data were conducted. For this purpose, the fusion of real data from five longitudinal studies of mathematics ability varying in their assessment of ability and in the timing of measurement occasions was used. Results from the Bayesian Synthesis and data fusion approaches with combined data using Bayesian and maximum likelihood estimation methods were reported. The results illustrate that Bayesian Synthesis with data driven priors is a highly effective approach, provided that the sample sizes for the fused data are large enough to provide unbiased estimates. Bayesian Synthesis provides another beneficial approach to data fusion that can effectively be used to enhance the validity of conclusions obtained from the merging of data from different studies.
ContributorsMarcoulides, Katerina M (Author) / Grimm, Kevin (Thesis advisor) / Levy, Roy (Thesis advisor) / MacKinnon, David (Committee member) / Suk, Hye Won (Committee member) / Arizona State University (Publisher)
Created2017
157544-Thumbnail Image.png
Description
Longitudinal recursive partitioning (LRP) is a tree-based method for longitudinal data. It takes a sample of individuals that were each measured repeatedly across time, and it splits them based on a set of covariates such that individuals with similar trajectories become grouped together into nodes. LRP does this by fitting

Longitudinal recursive partitioning (LRP) is a tree-based method for longitudinal data. It takes a sample of individuals that were each measured repeatedly across time, and it splits them based on a set of covariates such that individuals with similar trajectories become grouped together into nodes. LRP does this by fitting a mixed-effects model to each node every time that it becomes partitioned and extracting the deviance, which is the measure of node purity. LRP is implemented using the classification and regression tree algorithm, which suffers from a variable selection bias and does not guarantee reaching a global optimum. Additionally, fitting mixed-effects models to each potential split only to extract the deviance and discard the rest of the information is a computationally intensive procedure. Therefore, in this dissertation, I address the high computational demand, variable selection bias, and local optimum solution. I propose three approximation methods that reduce the computational demand of LRP, and at the same time, allow for a straightforward extension to recursive partitioning algorithms that do not have a variable selection bias and can reach the global optimum solution. In the three proposed approximations, a mixed-effects model is fit to the full data, and the growth curve coefficients for each individual are extracted. Then, (1) a principal component analysis is fit to the set of coefficients and the principal component score is extracted for each individual, (2) a one-factor model is fit to the coefficients and the factor score is extracted, or (3) the coefficients are summed. The three methods result in each individual having a single score that represents the growth curve trajectory. Therefore, now that the outcome is a single score for each individual, any tree-based method may be used for partitioning the data and group the individuals together. Once the individuals are assigned to their final nodes, a mixed-effects model is fit to each terminal node with the individuals belonging to it.

I conduct a simulation study, where I show that the approximation methods achieve the goals proposed while maintaining a similar level of out-of-sample prediction accuracy as LRP. I then illustrate and compare the methods using an applied data.
ContributorsStegmann, Gabriela (Author) / Grimm, Kevin (Thesis advisor) / Edwards, Michael (Committee member) / MacKinnon, David (Committee member) / McNeish, Daniel (Committee member) / Arizona State University (Publisher)
Created2019
171394-Thumbnail Image.png
Description
Mediation analysis is integral to psychology, investigating human behavior’s causal mechanisms. The diversity of explanations for human behavior has implications for the estimation and interpretation of statistical mediation models. Individuals can have similar observed outcomes while undergoing different causal processes or different observed outcomes while receiving the same treatment. Researchers

Mediation analysis is integral to psychology, investigating human behavior’s causal mechanisms. The diversity of explanations for human behavior has implications for the estimation and interpretation of statistical mediation models. Individuals can have similar observed outcomes while undergoing different causal processes or different observed outcomes while receiving the same treatment. Researchers can employ diverse strategies when studying individual differences in multiple mediation pathways, including individual fit measures and analysis of residuals. This dissertation investigates the use of individual residuals and fit measures to identify individual differences in multiple mediation pathways. More specifically, this study focuses on mediation model residuals in a heterogeneous population in which some people experience indirect effects through one mediator and others experience indirect effects through a different mediator. A simulation study investigates 162 conditions defined by effect size and sample size for three proposed methods: residual differences, delta z, and generalized Cook’s distance. Results indicate that analogs of Type 1 error rates are generally acceptable for the method of residual differences, but statistical power is limited. Likewise, neither delta z nor gCd could reliably distinguish between contrasts that had true effects and those that did not. The outcomes of this study reveal the potential for statistical measures of individual mediation. However, limitations related to unequal subpopulation variances, multiple dependent variables, the inherent relationship between direct effects and unestimated indirect effects, and minimal contrast effects require more research to develop a simple method that researchers can use on single data sets.
ContributorsSmyth, Heather Lynn (Author) / MacKinnon, David (Thesis advisor) / Tein, Jenn-Yun (Committee member) / McNeish, Daniel (Committee member) / Grimm, Kevin (Committee member) / Arizona State University (Publisher)
Created2022
190785-Thumbnail Image.png
Description
Psychologists report effect sizes in randomized controlled trials to facilitate interpretation and inform clinical or policy guidance. Since commonly used effect size measures (e.g., standardized mean difference) are not sensitive to heterogeneous treatment effects, methodologists have suggested the use of an alternative effect size δ, a between-subjects causal parameter describing

Psychologists report effect sizes in randomized controlled trials to facilitate interpretation and inform clinical or policy guidance. Since commonly used effect size measures (e.g., standardized mean difference) are not sensitive to heterogeneous treatment effects, methodologists have suggested the use of an alternative effect size δ, a between-subjects causal parameter describing the probability that the outcome of a random participant in the treatment group is better than the outcome of another random participant in the control group. Although this effect size is useful, researchers could mistakenly use δ to describe its within-subject analogue, ψ, the probability that an individual will do better under the treatment than the control. Hand’s paradox describes the situation where ψ and δ are on opposing sides of 0.5: δ may imply most are helped whereas the (unknown) underlying ψ indicates that most are harmed by the treatment. The current study used Monte Carlo simulations to investigate plausible situations under which Hand’s paradox does and does not occur, tracked the magnitude of the discrepancy between ψ and δ, and explored whether the size of the discrepancy could be reduced with a relevant covariate. The findings suggested that although the paradox should not occur under bivariate normal data conditions in the population, there could be sample cases with the paradox. The magnitude of the discrepancy between ψ and δ depended on both the size of the average treatment effect and the underlying correlation between the potential outcomes, ρ. Smaller effects led to larger discrepancies when ρ < 0 and ρ = 1, whereas larger effects led to larger discrepancies when 0 < ρ < 1. It was useful to consider a relevant covariate when calculating ψ and δ. Although ψ and δ were still discrepant within covariate levels, results indicated that conditioning upon relevant covariates is still useful in describing heterogeneous treatment effects.
ContributorsLiu, Xinran (Author) / Anderson, Samantha F (Thesis advisor) / McNeish, Daniel (Committee member) / MacKinnon, David (Committee member) / Arizona State University (Publisher)
Created2023