Matching Items (4)
Filtering by

Clear all filters

152051-Thumbnail Image.png
Description
Schennach (2007) has shown that the Empirical Likelihood (EL) estimator may not be asymptotically normal when a misspecified model is estimated. This problem occurs because the empirical probabilities of individual observations are restricted to be positive. I find that even the EL estimator computed without the restriction can fail to

Schennach (2007) has shown that the Empirical Likelihood (EL) estimator may not be asymptotically normal when a misspecified model is estimated. This problem occurs because the empirical probabilities of individual observations are restricted to be positive. I find that even the EL estimator computed without the restriction can fail to be asymptotically normal for misspecified models if the sample moments weighted by unrestricted empirical probabilities do not have finite population moments. As a remedy for this problem, I propose a group of alternative estimators which I refer to as modified EL (MEL) estimators. For correctly specified models, these estimators have the same higher order asymptotic properties as the EL estimator. The MEL estimators are obtained by the Generalized Method of Moments (GMM) applied to an exactly identified model. The simulation results provide promising evidence for these estimators. In the second chapter, I introduce an alternative group of estimators to the Generalized Empirical Likelihood (GEL) family. The new group is constructed by employing demeaned moment functions in the objective function while using the original moment functions in the constraints. This designation modifies the higher-order properties of estimators. I refer to these new estimators as Demeaned Generalized Empirical Likelihood (DGEL) estimators. Although Newey and Smith (2004) show that the EL estimator in the GEL family has fewer sources of bias and is higher-order efficient after bias-correction, the demeaned exponential tilting (DET) estimator in the DGEL group has those superior properties. In addition, if data are symmetrically distributed, every estimator in the DGEL family shares the same higher-order properties as the best member.  
ContributorsXiang, Jin (Author) / Ahn, Seung (Thesis advisor) / Wahal, Sunil (Thesis advisor) / Bharath, Sreedhar (Committee member) / Mehra, Rajnish (Committee member) / Tserlukevich, Yuri (Committee member) / Arizona State University (Publisher)
Created2013
152857-Thumbnail Image.png
Description
This dissertation applies the Bayesian approach as a method to improve the estimation efficiency of existing econometric tools. The first chapter suggests the Continuous Choice Bayesian (CCB) estimator which combines the Bayesian approach with the Continuous Choice (CC) estimator suggested by Imai and Keane (2004). Using simulation study, I provide

This dissertation applies the Bayesian approach as a method to improve the estimation efficiency of existing econometric tools. The first chapter suggests the Continuous Choice Bayesian (CCB) estimator which combines the Bayesian approach with the Continuous Choice (CC) estimator suggested by Imai and Keane (2004). Using simulation study, I provide two important findings. First, the CC estimator clearly has better finite sample properties compared to a frequently used Discrete Choice (DC) estimator. Second, the CCB estimator has better estimation efficiency when data size is relatively small and it still retains the advantage of the CC estimator over the DC estimator. The second chapter estimates baseball's managerial efficiency using a stochastic frontier function with the Bayesian approach. When I apply a stochastic frontier model to baseball panel data, the difficult part is that dataset often has a small number of periods, which result in large estimation variance. To overcome this problem, I apply the Bayesian approach to a stochastic frontier analysis. I compare the confidence interval of efficiencies from the Bayesian estimator with the classical frequentist confidence interval. Simulation results show that when I use the Bayesian approach, I achieve smaller estimation variance while I do not lose any reliability in a point estimation. Then, I apply the Bayesian stochastic frontier analysis to answer some interesting questions in baseball.
ContributorsChoi, Kwang-shin (Author) / Ahn, Seung (Thesis advisor) / Mehra, Rajnish (Committee member) / Park, Sungho (Committee member) / Arizona State University (Publisher)
Created2014
149591-Thumbnail Image.png
Description
I study the importance of financial factors and real exchange rate shocks in explaining business cycle fluctuations, which have been considered important in the literature as non-technological factors in explaining business cycle fluctuations. In the first chapter, I study the implications of fluctuations in corporate credit spreads for business cycle

I study the importance of financial factors and real exchange rate shocks in explaining business cycle fluctuations, which have been considered important in the literature as non-technological factors in explaining business cycle fluctuations. In the first chapter, I study the implications of fluctuations in corporate credit spreads for business cycle fluctuations. Motivated by the fact that corporate credit spreads are countercyclical, I build a simple model in which difference in default probabilities on corporate debts leads to the spread in interest rates paid by firms. In the model, firms differ in the variance of the firm-level productivity, which is in turn linked to the difference in the default probability. The key mechanism is that an increase in the variance of productivity for risky firms relative to safe firms leads to reallocation of capital away from risky firms toward safe firms and decrease in aggregate output and productivity. I embed the above mechanism into an otherwise standard growth model, calibrate it and numerically solve for the equilibrium. In my benchmark case, I find that shocks to variance of productivity for risky and safe firms account for about 66% of fluctuations in output and TFP in the U.S. economy. In the second chapter, I study the importance of shocks to the price of imports relative to the price of final goods, led by the real exchange rate shocks, in accounting for fluctuations in output and TFP in the Korean economy during the Asian crisis of 1997-98. Using the Korean data, I calibrate a standard small open economy model with taxes and tariffs on imported goods, and simulate it. I find that shocks to the price of imports are an important source of fluctuations in Korea's output and TFP in the Korean crisis episode. In particular, in my benchmark case, shocks to the price of imports account for about 55% of the output deviation (from trend), one third of the TFP deviation and three quarters of the labor deviation in 1998.
ContributorsKim, Seon Tae (Author) / Prescott, Edward C. (Thesis advisor) / Rogerson, Richard (Committee member) / Ahn, Seung (Committee member) / Low, Stuart (Committee member) / Arizona State University (Publisher)
Created2011
Description
This dissertation studies how forecasting performance can be improved in big data. The first chapter with Seung C. Ahn considers Partial Least Squares (PLS) estimation of a time-series forecasting model with data containing a large number of time series observations of many predictors. In the model, a subset or a

This dissertation studies how forecasting performance can be improved in big data. The first chapter with Seung C. Ahn considers Partial Least Squares (PLS) estimation of a time-series forecasting model with data containing a large number of time series observations of many predictors. In the model, a subset or a whole set of the latent common factors in predictors determine a target variable. First, the optimal number of the PLS factors for forecasting could be smaller than the number of the common factors relevant for the target variable. Second, as more than the optimal number of PLS factors is used, the out-of-sample explanatory power of the factors could decrease while their in-sample power may increase. Monte Carlo simulation results also confirm these asymptotic results. In addition, simulation results indicate that the out-of-sample forecasting power of the PLS factors is often higher when a smaller than the asymptotically optimal number of factors are used. Finally, the out-of-sample forecasting power of the PLS factors often decreases as the second, third, and more factors are added, even if the asymptotically optimal number of the factors is greater than one. The second chapter studies the predictive performance of various factor estimations comprehensively. Big data that consist of major U.S. macroeconomic and finance variables, are constructed. 148 target variables are forecasted, using 7 factor estimation methods with 11 information criteria. First, the number of factors used in forecasting is important and Incorporating more factors does not always provide better forecasting performance. Second, using consistently estimated number of factors does not necessarily improve predictive performance. The first PLS factor, which is not theoretically consistent, very often shows strong forecasting performance. Third, there is a large difference in the forecasting performance across different information criteria, even when the same factor estimation method is used. Therefore, the choice of factor estimation method, as well as the information criterion, is crucial in forecasting practice. Finally, the first PLS factor yields forecasting performance very close to the best result from the total combinations of the 7 factor estimation methods and 11 information criteria.
ContributorsBae, Juhui (Author) / Ahn, Seung (Thesis advisor) / Pruitt, Seth (Committee member) / Kuminoff, Nicolai (Committee member) / Ferraro, Domenico (Committee member) / Arizona State University (Publisher)
Created2021