nep-ecm New Economics Papers
on Econometrics
Issue of 2015‒04‒02
seventeen papers chosen by
Sune Karlsson
Örebro universitet

  1. Sparse Graphical Vector Autoregression: A Bayesian Approach By Roberto Casarin; Daniel Felix Ahelegbey; Monica Billio
  2. Testing for Flexible Nonlinear Trends with an Integrated or Stationary Noise Component By Pierre Perron; Mototsugu Shintani; Tomoyoshi Yabu
  3. Uniform Asymptotic Risk of Averaging GMM Estimator Robust to Misspecification, Second Version By Xu Cheng; Zhipeng Liao; Ruoyao Shi
  4. Two-sample Hotelling's T² statistics based on the functional Mahalanobis semi-distance By Joseph Esdras; Pedro Galeano; Rosa E. Lillo
  5. Information Criteria for Nonlinear Time Series Models By Rinke, Saskia; Sibbertsen, Philipp
  6. Testing the Constancy of Conditional Correlations in Multivariate GARCH-type Models (Extended Version with Appendix) By Anne Péguin-Feissolle; Bilel Sanhaji
  7. Testing Linearity Using Power Transforms of Regressors By YAE IN BAEK; Jin Seo Cho
  8. We provide mathematical proofs for the results in ¡°Testing Linearity Using Power Transforms of Regressors¡± by Baek, Cho, and Phillips (2014). By YAE IN BAEK; Jin Seo Cho; PETER C.B. PHILLIPS
  9. GMM and Indirect Inference: An appraisal of their connections and new results on their properties under second order identification By Prosper Donovon; Alastair R. Hall
  10. Maximum Likelihood Estimation of Dynamic Panel Threshold Models By Nelson Ramírez-Rondán
  11. Exploiting the Errors: A Simple Approach for Improved Volatility Forecasting By Tim Bollerslev; Andrew J. Patton; Rogier Quaedvlieg
  12. A Robust Statistics Approach to Minimum Variance Portfolio Optimization By Liusha Yang; Romain Couillet; Matthew R. McKay
  13. Assessing Macro Uncertainty In Real-Time When Data Are Subject To Revision By Michael P. Clements;
  14. A Simple Method to Estimate Large Fixed Effects Models Applied to Wage Determinants and Matching By Nikolas Mittag
  15. A Comprehensive Approach to Posterior Jointness Analysis in Bayesian Model Averaging Applications By Jesus Crespo Cuaresma; Bettina Grün; Paul Hofmarcher; Stefan Humer; Mathias Moser
  16. The Asymptotic Behaviour of the Residual Sum of Squares in Models with Multiple Break Points By Alastair R. Hall; Denise R. Osborn; Nikolaos Sakkas
  17. Constrained Optimization Approaches to Estimation of Structural Models: Comment By Fedor Iskhakov; John Rust; Bertel Schjerning

  1. By: Roberto Casarin (Department of Economics, University of Venice Cà Foscari); Daniel Felix Ahelegbey (Department of Economics, University of Venice Cà Foscari); Monica Billio (Department of Economics, University of Venice Cà Foscari)
    Abstract: In high-dimensional vector autoregressive (VAR) models, it is natural to have large number of predictors relative to the number of observations, and a lack of efficiency in estimation and forecasting. In this context, model selection is a difficult issue and standard procedures may often be inefficient. In this paper we aim to provide a solution to these problems. We introduce sparsity on the structure of temporal dependence of a graphical VAR and develop an efficient model selection approach. We follow a Bayesian approach and introduce prior restrictions to control the maximal number of explanatory variables for VAR models. We discuss the joint inference of the temporal dependence, the maximum lag order and the parameters of the model, and provide an efficient Markov chain Monte Carlo procedure. The efficiency of the proposed approach is showed on simulated experiments and real data to model and forecast selected US macroeconomic variables with many predictors.
    Keywords: High-dimensional Models, Large Vector Autoregression, Model Selection, Prior Distribution, Sparse Graphical Models.
    JEL: C11 C15 C52 E17 G17
    Date: 2014
    URL: http://d.repec.org/n?u=RePEc:ven:wpaper:2014:29&r=ecm
  2. By: Pierre Perron (Boston University); Mototsugu Shintani (University of Tokyo and Vanderbilt University); Tomoyoshi Yabu (Keio University)
    Abstract: This paper proposes a new test for the presence of a nonlinear deterministic trend approximated by a Fourier expansion in a univariate time series for which there is no prior knowledge as to whether the noise component is stationary or contains an autoregressive unit root. Our approach builds on the work of Perron and Yabu (2009a) and is based on a Feasible Generalized Least Squares procedure that uses a super-efficient estimator of the sum of the autoregressive coefficients α when α=1. The resulting Wald test statistic asymptotically follows a chi-square limit distribution in both the I(0) and I(1) cases. To improve the finite sample properties of the test, we use a bias corrected version of the OLS estimator of α proposed by Roy and Fuller (2001). We show that our procedure is substantially more powerful than currently available alternatives. We illustrate the usefulness of our method via an application to modeling the trend of global and hemispheric temperatures.
    Keywords: nonlinear trends, unit root, median-unbiased estimator, GLS procedure, super-efficient estimator
    JEL: C2
    Date: 2015–02–27
    URL: http://d.repec.org/n?u=RePEc:van:wpaper:vuecon-sub-15-00001&r=ecm
  3. By: Xu Cheng (Department of Economics, University of Pennsylvania); Zhipeng Liao (Department of Economics, UCLA); Ruoyao Shi (Department of Economics, UCLA)
    Abstract: This paper studies the averaging generalized method of moments (GMM) estimator that combines a conservative GMM estimator based on valid moment conditions and an aggressive GMM estimator based on both valid and possibly misspecified moment conditions, where the weight is the sample analog of an infeasible optimal weight. It is an alternative to pre-test estimators that switch between the conservative and aggressive estimators based on model specification tests. This averaging estimator is robust in the sense that it uniformly dominates the conservative estimator by reducing the risk under any degree of misspecification, whereas the pre-test estimators reduce the risk in parts of the parameter space and increase it in other parts. To establish uniform dominance of one estimator over another, we establish asymptotic theories on uniform approximations of the finite-sample risk differences between two estimators. These asymptotic results are developed along drifting sequences of data generating processes (DGPs) that model various degrees of local misspecification as well as global misspecification. Extending seminal results on the James-Stein estimator, the uniform dominance is established in non-Gaussian semiparametric nonlinear models. The proposed averaging estimator is applied to estimate the human capital production function in a life-cycle labor supply model.
    Keywords: Finite-Sample Risk, Generalized Shrinkage Estimator, GMM Misspecification, Model Averaging, Uniform Approximation
    JEL: C26 C36 C51 C52
    Date: 2013–08–01
    URL: http://d.repec.org/n?u=RePEc:pen:papers:15-017&r=ecm
  4. By: Joseph Esdras; Pedro Galeano; Rosa E. Lillo
    Abstract: The comparison of the means of two independent samples is one of the most popular problems in real-world data analysis. In the multivariate context, two-sample Hotelling's T² frequently used to test the equality of means of two independent Gaussian random samples assuming either the same or a different covariance matrix. In this paper, we derive two-sample Hotelling's T² from two functional distributions. The statistics that we propose are based on the functional Mahalanobis semi-distance and, under certain conditions, their asymptotic distributions are chisquared, regardless the distribution of the functional random samples. Additionally, we provide the link between the two-sample Hotelling's T² semi-distance and statistics based on the functional principal components semi-distance. A Monte Carlo study indicates that the twosample Hotelling's T² of power those based on the functional principal components semidistance. We analyze a data set of daily temperature records of 35 Canadian weather stations over a year with the goal of testing whether or not the mean temperature functions of the stations in the Eastern and Western Canada regions are equal. The results appear to indicate differences between both regions that are not found with statistics based on the functional principal components semi-distance.
    Keywords: Functional Behrens-Fisher problem, Functional data analysis, Functional Mahalanobis semi-distance, Functional principal components semi-distance, Hotelling's T² statistics
    Date: 2015–03
    URL: http://d.repec.org/n?u=RePEc:cte:wsrepe:ws1503&r=ecm
  5. By: Rinke, Saskia; Sibbertsen, Philipp
    Abstract: In this paper the performance of different information criteria for simultaneous model class and lag order selection is evaluated using simulation studies. We focus on the ability of the criteria to distinguish linear and nonlinear models. In the simulation studies, we consider three different versions of the commonly known criteria AIC, SIC and AICc. In addition, we also assess the performance of WIC and evaluate the impact of the error term variance estimator. Our results confirm the findings of different authors that AIC and AICc favor nonlinear over linear models, whereas weighted versions of WIC and all versions of SIC are able to successfully distinguish linear and nonlinear models. However, the discrimination between different nonlinear model classes is more difficult. Nevertheless, the lag order selection is reliable. In general, information criteria involving the unbiased error term variance estimator overfit less and should be preferred to using the usual ML estimator of the error term variance.
    Keywords: Information Criteria, Nonlinear Time Series, Threshold Models,Monte Carlo
    JEL: C15 C22
    Date: 2015–03
    URL: http://d.repec.org/n?u=RePEc:han:dpaper:dp-548&r=ecm
  6. By: Anne Péguin-Feissolle (Aix Marseille University (Aix-Marseille School of Economics), CNRS & EHESS, Aix-Marseille); Bilel Sanhaji (Aix Marseille University (Aix-Marseille School of Economics), CNRS & EHESS, Aix-Marseille)
    Abstract: We introduce two multivariate constant conditional correlation tests that require little knowledge of the functional relationship determining the conditional correlations. The first test is based on artificial neural networks and the second one is based on a Taylor expansion of each unknown conditional correlation. These new tests can be seen as general misspecification tests of a large set of multivariate GARCH-type models. We investigate the size and the power of these tests through Monte Carlo experiments. Moreover, we study their robustness to non-normality by simulating some models such as the GARCH?t and Beta?t?EGARCH models. We give some illustrative empirical examples based on financial data.
    Keywords: multivariate GARCH, neural network, Taylor expansion
    JEL: C22 C45 C58
    Date: 2015–03–10
    URL: http://d.repec.org/n?u=RePEc:aim:wpaimx:1516&r=ecm
  7. By: YAE IN BAEK (University of California, San Diego); Jin Seo Cho (Yonsei University)
    Abstract: We develop a method of testing linearity using power transforms of regressors, allowing for stationary processes and time trends. The linear model is a simplifying hypothesis that derives from the power transform model in three different ways, each producing its own identification problem. We call this modeling difficulty the trifold identification problem and show that it may be overcome using a test based on the quasi-likelihood ratio (QLR) statistic. More specifically, the QLR statistic may be approximated under each identification problem and the separate null approximations may be combined to produce a composite approximation that embodies the linear model hypothesis. The limit theory for the QLR test statistic depends on a Gaussian stochastic process. In the important special case of a linear time trend regressor and martingale difference errors asymptotic critical values of the test are provided. Test power is analyzed and an empirical application to crop-yield distributions is provided. The paper also considers generalizations of the Box-Cox transformation, which are associated with the QLR test statistic.
    Keywords: Box Cox transform; Gaussian stochastic process; Neglected nonlinearity; Power transformation; Quasi-likelihood ratio test; Trend exponent; Trifold identification problem.
    JEL: C12 C18 C46 C52
    Date: 2015–03
    URL: http://d.repec.org/n?u=RePEc:yon:wpaper:2015rwp-79&r=ecm
  8. By: YAE IN BAEK (University of California, San Diego); Jin Seo Cho (Yonsei University); PETER C.B. PHILLIPS (Yale University, University of Auckland, Singapore Management University & University of Southampton)
    JEL: C12 C18 C46 C52
    Date: 2015–03
    URL: http://d.repec.org/n?u=RePEc:yon:wpaper:2015rwp-79a&r=ecm
  9. By: Prosper Donovon; Alastair R. Hall
    Date: 2015
    URL: http://d.repec.org/n?u=RePEc:man:sespap:1505&r=ecm
  10. By: Nelson Ramírez-Rondán (Central Bank of Peru)
    Abstract: Threshold estimation methods are developed for dynamic panels with individual fixed specific effects covering short time periods. Maximum likelihood estimation of the threshold and the slope parameters is proposed using first difference transformations. Threshold estimate is shown to be consistent and it converges to a double-sided standard Brownian motion distribution, when the number of individuals grows to infinity for a fixed time period; and the slope estimates are consistent and asymptotically normally distributed. The method is applied to a sample of 72 countries and 8 periods of 5-year averages to determine the effect of inflation rate on long-run economic growth.
    Keywords: Threshold Models, Dynamic Panel Data, Maximum Likelihood Estimation, Inflation, Economic Growth
    JEL: C13 C23
    Date: 2015–03
    URL: http://d.repec.org/n?u=RePEc:apc:wpaper:2015-032&r=ecm
  11. By: Tim Bollerslev (Duke University, NBER and CREATES); Andrew J. Patton (Duke University); Rogier Quaedvlieg (Maastricht University)
    Abstract: We propose a new family of easy-to-implement realized volatility based forecasting models. The models exploit the asymptotic theory for high-frequency realized volatility estimation to improve the accuracy of the forecasts. By allowing the parameters of the models to vary explicitly with the (estimated) degree of measurement error, the models exhibit stronger persistence, and in turn generate more responsive forecasts, when the measurement error is relatively low. Implementing the new class of models for the S&P500 equity index and the individual constituents of the Dow Jones Industrial Average, we document significant improvements in the accuracy of the resulting forecasts compared to the forecasts from some of the most popular existing models that implicitly ignore the temporal variation in the magnitude of the realized volatility measurement errors.
    Keywords: Realized volatility, Forecasting, Measurement Errors, HAR, HARQ
    JEL: C22 C51 C53 C58
    Date: 2015–03–10
    URL: http://d.repec.org/n?u=RePEc:aah:create:2015-14&r=ecm
  12. By: Liusha Yang; Romain Couillet; Matthew R. McKay
    Abstract: We study the design of portfolios under a minimum risk criterion. The performance of the optimized portfolio relies on the accuracy of the estimated covariance matrix of the portfolio asset returns. For large portfolios, the number of available market returns is often of similar order to the number of assets, so that the sample covariance matrix performs poorly as a covariance estimator. Additionally, financial market data often contain outliers which, if not correctly handled, may further corrupt the covariance estimation. We address these shortcomings by studying the performance of a hybrid covariance matrix estimator based on Tyler's robust M-estimator and on Ledoit-Wolf's shrinkage estimator while assuming samples with heavy-tailed distribution. Employing recent results from random matrix theory, we develop a consistent estimator of (a scaled version of) the realized portfolio risk, which is minimized by optimizing online the shrinkage intensity. Our portfolio optimization method is shown via simulations to outperform existing methods both for synthetic and real market data.
    Date: 2015–03
    URL: http://d.repec.org/n?u=RePEc:arx:papers:1503.08013&r=ecm
  13. By: Michael P. Clements (ICMA Centre, Henley Business School, University of Reading);
    Abstract: Model-based estimates of future uncertainty are generally based on the in-sample fit of the model, as when Box-Jenkins prediction intervals are calculated. However, this approach will generate biased uncertainty estimates in real time when there are data revisions. A simple remedy is suggested, and used to generate more accurate prediction intervals for 25 macroeconomic variables, in line with the theory. A simulation study based on an empirically-estimated model of data revisions for US output growth is used to investigate small-sample properties.
    Keywords: in-sample uncertainty, out-of-sample uncertainty, real-time-vintage estimation
    JEL: C53
    Date: 2015–01
    URL: http://d.repec.org/n?u=RePEc:rdg:icmadp:icma-dp2015-02&r=ecm
  14. By: Nikolas Mittag
    Abstract: Models with high dimensional sets of fixed effects are frequently used to examine, among others, linked employer-employee data, student outcomes and migration. Estimating these models is computationally difficult, so simplifying assumptions that cause bias are often invoked to make computation feasible and specification tests are rarely conducted. I present a simple method to estimate large two-way fixed effects (TWFE) and worker-firm match effect models without additional assumptions. It computes the exact OLS solution including estimates of the fixed effects and makes testing feasible even with multi-way clustered errors. An application using German linked employer-employee data illustrates the advantages: The data reject the assumptions of simpler estimators and omitting match effects biases estimates including the returns to experience and the gender wage gap. Specification test detect both problems. Firm fixed effects, not match effects, are the main channel through which job transitions drive wage dynamics, which underlines the importance of firm heterogeneity for labor market dynamics.
    Keywords: multi-way fixed effects; linked employer-employee data; matching, wage dynamics;
    JEL: J31 J63 C23 C63
    Date: 2015–03
    URL: http://d.repec.org/n?u=RePEc:cer:papers:wp532&r=ecm
  15. By: Jesus Crespo Cuaresma (Department of Economics, Vienna University of Economics and Business); Bettina Grün (Department of Applied Statistics, Johannes Kepler University Linz); Paul Hofmarcher (Department of Economics, Vienna University of Economics and Business); Stefan Humer (Department of Economics, Vienna University of Economics and Business); Mathias Moser (Department of Economics, Vienna University of Economics and Business)
    Abstract: Posterior analysis in Bayesian model averaging (BMA) applications often includes the assessment of measures of jointness (joint inclusion) across covariates. We link the discussion of jointness measures in the econometric literature to the literature on association rules in data mining exercises. We analyze a group of alternative jointness measures that include those proposed in the BMA literature and several others put forward in the field of data mining. The way these measures address the joint exclusion of covariates appears particularly important in terms of the conclusions that can be drawn from them. Using a dataset of economic growth determinants, we assess how the measurement of jointness in BMA can affect inference about the structure of bivariate inclusion patterns across covariates.
    Keywords: Bayesian Model Averaging, Jointness, Robust Growth Determinants, Machine Learning, Association Rules
    JEL: C11 O40
    URL: http://d.repec.org/n?u=RePEc:wiw:wiwwuw:wuwp193&r=ecm
  16. By: Alastair R. Hall; Denise R. Osborn; Nikolaos Sakkas
    Date: 2015
    URL: http://d.repec.org/n?u=RePEc:man:sespap:1504&r=ecm
  17. By: Fedor Iskhakov (University of New South Wales); John Rust (Georgetown University); Bertel Schjerning (University of Copenhagen)
    Abstract: We revisit the comparison of mathematical programming with equilibrium constraints (MPEC) and nested fixed point (NFXP) algorithms for estimating structural dynamic models by Su and Judd (SJ, 2012). They used an inefficient version of the nested fixed point algorithm that relies on successive approximations. We re-do their comparison using the more efficient version of NFXP proposed by Rust (1987), which combines successive approximations and Newton-Kantorovich iterations to solve the fixed point problem (NFXP-NK). We show that MPEC and NFXP-NK are similar in performance when the sample size is relatively small. However, in problems with larger sample sizes, NFXP-NK outperforms MPEC by a significant margin.
    Keywords: Structural estimation, dynamic discrete choice, NFXP, MPEC, successive approximations, Newton-Kantorovich algorithm.
    Date: 2015–03–23
    URL: http://d.repec.org/n?u=RePEc:kud:kuiedp:1505&r=ecm

This nep-ecm issue is ©2015 by Sune Karlsson. It is provided as is without any express or implied warranty. It may be freely redistributed in whole or in part for any purpose. If distributed in part, please include this notice.
General information on the NEP project can be found at http://nep.repec.org. For comments please write to the director of NEP, Marco Novarese at <director@nep.repec.org>. Put “NEP” in the subject, otherwise your mail may be rejected.
NEP’s infrastructure is sponsored by the School of Economics and Finance of Massey University in New Zealand.