nep-ecm New Economics Papers
on Econometrics
Issue of 2020‒03‒23
twenty-one papers chosen by
Sune Karlsson
Örebro universitet

  1. Semi-parametric estimation of multi-valued treatment effects for the treated:estimating equations and sandwich estimators By Zetterqvist, Johan; Waernbaum, Ingeborg
  2. Identification and Estimation of Weakly Separable Models Without Monotonicity By Songnian Chen; Shakeeb Khan; Xun Tang
  3. Bayesian Inference in High-Dimensional Time-varying Parameter Models using Integrated Rotated Gaussian Approximations By Florian Huber; Gary Koop; Michael Pfarrhofer
  4. Kernel Conditional Moment Test via Maximum Moment Restriction By Krikamol Muandet; Wittawat Jitkrittum; Jonas K\"ubler
  5. Random-Coefficients Logit Demand Estimation with Zero-Valued Market Shares By Jean-Pierre H. Dubé; Ali Hortaçsu; Joonhwi Joo
  6. Multivariate Goodness-of-Fit Tests Based on Wasserstein Distance By Marc Hallin; Gilles Mordant; Johan Segers
  7. Unit-root test within a threshold ARMA framework By Kung-Sik Chan; Simone Giannerini; Greta Goracci; Howell Tong
  8. On bootstrapping tests of equal forecast accuracy for nested models By Firmin Doko Tchatoka; Qazi Haque
  9. Estimation and Inference about Tail Features with Tail Censored Data By Yulong Wang; Zhijie Xiao
  10. Backward CUSUM for Testing and Monitoring Structural Change By Sven Otto; J\"org Breitung
  11. Equal Predictive Ability Tests for Panel Data with an Application to OECD and IMF Forecasts By Oguzhan Akgun; Alain Pirotte; Giovanni Urga; Zhenlin Yang
  12. Scoring Functions for Multivariate Distributions and Level Sets By Xiaochun Meng; James W. Taylor; Souhaib Ben Taieb; Siran Li
  13. Adaptive exponential power distribution with moving estimator for nonstationary time series By Jarek Duda
  14. Estimating Economic Models with Testable Assumptions: Theory and Application to LATE By Moyu Liao
  15. Pollution and Expenditures in a Penalized Vector Spatial Autoregressive Time Series Model with Data-Driven Networks By Andree,Bo Pieter Johannes; Spencer,Phoebe Girouard; Azari,Sardar; Chamorro,Andres; Wang,Dieter; Dogo,Harun
  16. Non-stationary neural network for stock return prediction By Steven Y. K. Wong; Jennifer Chan; Lamiae Azizi; Richard Y. D. Xu
  17. Identification of Random Coefficient Latent Utility Models By Roy Allen; John Rehbeck
  18. Deep Learning, Jumps, and Volatility Bursts By Oksana Bashchenko; Alexis Marchal
  19. Ordered Beta Regression: A Parsimonious, Well-Fitting Model for Survey Sliders and Visual Analog Scales By Kubinec, Robert
  20. The More the Merrier? A Machine Learning Algorithm for Optimal Pooling of Panel Data By Marijn A. Bolhuis; Brett Rayner
  21. A mixture autoregressive model based on Gaussian and Student's $t$-distributions By Savi Virolainen

  1. By: Zetterqvist, Johan (Karolinska institutet); Waernbaum, Ingeborg (IFAU - Institute for Evaluation of Labour Market and Education Policy)
    Abstract: An estimand of interest in empirical studies with observational data is the average treatment effect of a multi-valued treatment in the treated subpopulation. We demonstrate three estimation approaches: outcome regression, inverse probability weighting and inverse probability weighted regression, where the latter estimator holds a so called doubly robust property. Here, we define the estimators in the framework of partial M-estimation and derive corresponding sandwich estimators of their variances. The finite sample properties of the estimators and the proposed variance estimators are evaluated in simulations that reproduce designs from a previous simulation study in the literature of multi-valued treatment effects. The proposed variance estimators are investigated and compared to a bootstrap estimator.
    Keywords: ATT; causal inference; inverse probability weighting; doubly robust; weighted ordinary least squares
    JEL: C14
    Date: 2020–03–05
    URL: http://d.repec.org/n?u=RePEc:hhs:ifauwp:2020_004&r=all
  2. By: Songnian Chen; Shakeeb Khan; Xun Tang
    Abstract: We study the identification and estimation of treatment effect parameters in non-separable models. In their seminal work, Vytlacil and Yildiz (2007) showed how to identify and estimate the average treatment effect of a dummy endogenous variable in a weakly separable model under a monotonicity condition. We also consider similar weakly separable models, but relax the monotonicity condition for identification. Our approach exploits information from the distribution of the outcome. To illustrate the advantage of our approach, we provide several examples of models where our approach can identify parameters of interest whereas existing methods cannot be applied because monotonicity fails. These include models with multiple unobserved disturbance terms such as the Roy model and multinomial choice models with dummy endogenous variables, as well as potential outcome models with endogenous random coefficients. Our proposed method applies to a wide class of models and is easy to implement. We also establish standard asymptotic properties such as consistency and asymptotic normality.
    Date: 2020–03
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2003.04337&r=all
  3. By: Florian Huber; Gary Koop; Michael Pfarrhofer
    Abstract: Researchers increasingly wish to estimate time-varying parameter (TVP) regressions which involve a large number of explanatory variables. Including prior information to mitigate over-parameterization concerns has led to many using Bayesian methods. However, Bayesian Markov Chain Monte Carlo (MCMC) methods can be very computationally demanding. In this paper, we develop computationally efficient Bayesian methods for estimating TVP models using an integrated rotated Gaussian approximation (IRGA). This exploits the fact that whereas constant coefficients on regressors are often important, most of the TVPs are often unimportant. Since Gaussian distributions are invariant to rotations we can split the the posterior into two parts: one involving the constant coefficients, the other involving the TVPs. Approximate methods are used on the latter and, conditional on these, the former are estimated with precision using MCMC methods. In empirical exercises involving artificial data and a large macroeconomic data set, we show the accuracy and computational benefits of IRGA methods.
    Date: 2020–02
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2002.10274&r=all
  4. By: Krikamol Muandet; Wittawat Jitkrittum; Jonas K\"ubler
    Abstract: We propose a new family of specification tests called kernel conditional moment (KCM) tests. Our tests are built on conditional moment embeddings (CMME)---a novel representation of conditional moment restrictions in a reproducing kernel Hilbert space (RKHS). After transforming the conditional moment restrictions into a continuum of unconditional counterparts, the test statistic is defined as the maximum moment restriction within the unit ball of the RKHS. We show that the CMME fully characterizes the original conditional moment restrictions, leading to consistency in both hypothesis testing and parameter estimation. The proposed test also has an analytic expression that is easy to compute as well as closed-form asymptotic distributions. Our empirical studies show that the KCM test has a promising finite-sample performance compared to existing tests.
    Date: 2020–02
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2002.09225&r=all
  5. By: Jean-Pierre H. Dubé; Ali Hortaçsu; Joonhwi Joo
    Abstract: Although typically overlooked, many purchase datasets exhibit a high incidence of products with zero sales. We propose a new estimator for the Random-Coefficients Logit demand system for purchase datasets with zero-valued market shares. The identification of the demand parameters is based on a pairwise-differencing approach that constructs moment conditions based on differences in demand between pairs of products. The corresponding estimator corrects non-parametrically for the potential selection of the incidence of zeros on unobserved aspects of demand. The estimator also corrects for the potential endogeneity of marketing variables both in demand and in the selection propensities. Monte Carlo simulations show that our proposed estimator provides reliable small-sample inference both with and without selection-on- unobservables. In an empirical case study, the proposed estimator not only generates different demand estimates than approaches that ignore selection in the incidence of zero shares, it also generates better out-of-sample fit of observed retail contribution margins.
    JEL: D12 L00 L66 L81 M3
    Date: 2020–02
    URL: http://d.repec.org/n?u=RePEc:nbr:nberwo:26795&r=all
  6. By: Marc Hallin; Gilles Mordant; Johan Segers
    Abstract: Goodness-of-fit tests based on the empirical Wasserstein distance are proposed for simple and composite null hypotheses involving general multivariate distributions. This includes the important problem of testing for multivariate normality with unspecified location and covariance and, more generally, testing for elliptical symmetry with given standard radial density, unspecified location and scatter parameters. The calculation of test statistics boils down to solving the well-studied semi-discrete optimal transport problem. Exact critical values can be computed for some important particular cases, such as null hypotheses of ellipticity with given standard radial density and unspecified location and scatter; else, approximate critical values are obtained via parametric bootstrap. Consistency is established, based on a result on the convergence to zero, uniformly over certain families of distributions, of the empirical Wasserstein distance---a novel result of independent interest. A simulation study establishes the practical feasibility and excellent performance of the proposed tests.
    Keywords: Copula; Elliptical distribution; Goodness-of- t; Multivariate normality; Optimal transport; Semi-discrete problem; Skew-t distribution; Wasserstein distance
    Date: 2020–03
    URL: http://d.repec.org/n?u=RePEc:eca:wpaper:2013/303372&r=all
  7. By: Kung-Sik Chan; Simone Giannerini; Greta Goracci; Howell Tong
    Abstract: We propose a new unit-root test based on Lagrange Multipliers, where we extend the null hypothesis to an integrated moving-average process (IMA(1,1)) and the alternative to a first-order threshold autoregressive moving-average process (TARMA(1,1)). This new theoretical framework provides tests with good size without pre-modelling steps. Moreover, leveraging on the versatile capability of the TARMA(1,1), our test has power against a wide range of linear and nonlinear alternatives. We prove the consistency and asymptotic similarity of the test. The proof of tightness of the test is of independent and general theoretical interest. Moreover, we propose a wild bootstrap version of the statistic. Our proposals outperform most existing tests in many contexts. We support the view that rejection does not necessarily imply nonlinearity so that unit-root tests should not be used uncritically to select a model. Finally, we present an application to real exchange rates.
    Date: 2020–02
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2002.09968&r=all
  8. By: Firmin Doko Tchatoka (School of Economics, University of Adelaide); Qazi Haque (Business School, The University of Western Australia)
    Abstract: The asymptotic distributions of the recursive out-of-sample forecast accuracy test statistics depend on stochastic integrals of Brownian motion when the models under comparison are nested. This often complicates their implementation in practice because the computation of their asymptotic critical values is costly. Hansen and Timmermann (2015, Econometrica) propose a Wald approximation of the commonly used recursive F-statistic and provide a simple characterization of the exact density of its asymptotic distribution. However, this characterization holds only when the larger model has one extra predictor or the forecast errors are homoscedastic. No such closed-form characterization is readily available when the nesting involves more than one predictor and heteroskedasticity is present. We first show both the recursive F-test and its Wald approximation have poor finite-sample properties, especially when the forecast horizon is greater than one. We then propose an hybrid bootstrap method consisting of a block moving bootstrap (which is nonparametric) and a residual based bootstrap for both statistics, and establish its validity. Simulations show that our hybrid bootstrap has good finite-sample performance, even in multi-step ahead forecasts with heteroscedastic or autocorrelated errors, and more than one predictor. The bootstrap method is illustrated on forecasting core inflation and GDP growth.
    Keywords: Out-of-sample forecasts; HAC estimator; Moving block bootstrap; Bootstrap consistency
    JEL: C12 C15 C32
    Date: 2020–02
    URL: http://d.repec.org/n?u=RePEc:adl:wpaper:2020-03&r=all
  9. By: Yulong Wang; Zhijie Xiao
    Abstract: This paper considers estimation and inference about tail features when the observations beyond some threshold are censored. We first show that ignoring such tail censoring could lead to substantial bias and size distortion, even if the censored probability is tiny. Second, we propose a new maximum likelihood estimator (MLE) based on the Pareto tail approximation and derive its asymptotic properties. Third, we provide a small sample modification to the MLE by resorting to Extreme Value theory. The MLE with this modification delivers excellent small sample performance, as shown by Monte Carlo simulations. We illustrate its empirical relevance by estimating (i) the tail index and the extreme quantiles of the US individual earnings with the Current Population Survey dataset and (ii) the tail index of the distribution of macroeconomic disasters and the coefficient of risk aversion using the dataset collected by Barro and Urs{\'u}a (2008). Our new empirical findings are substantially different from the existing literature.
    Date: 2020–02
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2002.09982&r=all
  10. By: Sven Otto; J\"org Breitung
    Abstract: It is well known that the conventional CUSUM test suffers from low power and large detection delay. We therefore propose two alternative detector statistics. The backward CUSUM detector sequentially cumulates the recursive residuals in reverse chronological order, whereas the stacked backward CUSUM detector considers a triangular array of backward cumulated residuals. While both the backward CUSUM detector and the stacked backward CUSUM detector are suitable for retrospective testing, only the stacked backward CUSUM detector can be monitored on-line. The limiting distributions of the maximum statistics under suitable sequences of alternatives are derived for retrospective testing and fixed endpoint monitoring. In the retrospective testing context, the local power of the tests is shown to be substantially higher than that for the conventional CUSUM test if a single break occurs after one third of the sample size. When applied to monitoring schemes, the detection delay of the stacked backward CUSUM is shown to be much shorter than that of the conventional monitoring CUSUM procedure. Moreover, an infinite horizon monitoring procedure and critical values are presented.
    Date: 2020–03
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2003.02682&r=all
  11. By: Oguzhan Akgun; Alain Pirotte; Giovanni Urga; Zhenlin Yang
    Abstract: This paper develops novel tests to compare the predictive ability of two forecasters using panels. We consider two different equal predictive ability (EPA) hypotheses. First hypothesis states that the predictive ability of two forecasters is equal on average over all periods and units. Under the second one, the EPA hypothesis holds jointly for all units. We study the asymptotic properties of proposed tests using sequential limits under strong and weak cross-sectional dependence. Their finite sample properties are investigated via Monte Carlo simulations. They are applied to compare the economic growth forecasts of OECD and IMF using data from OECD countries.
    Date: 2020–03
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2003.02803&r=all
  12. By: Xiaochun Meng; James W. Taylor; Souhaib Ben Taieb; Siran Li
    Abstract: Interest in predicting multivariate probability distributions is growing due to the increasing availability of rich datasets and computational developments. Scoring functions enable the comparison of forecast accuracy, and can potentially be used for estimation. A scoring function for multivariate distributions that has gained some popularity is the energy score. This is a generalization of the continuous ranked probability score (CRPS), which is widely used for univariate distributions. A little-known, alternative generalization is the multivariate CRPS (MCRPS). We propose a theoretical framework for scoring functions for multivariate distributions, which encompasses the energy score and MCRPS, as well as the quadratic score, which has also received little attention. We demonstrate how this framework can be used to generate new scores. For univariate distributions, it is well-established that the CRPS can be expressed as the integral over a quantile score. We show that, in a similar way, scoring functions for multivariate distributions can be "disintegrated" to obtain scoring functions for level sets. Using this, we present scoring functions for different types of level set, including those for densities and cumulative distributions. To compute the scoring functions, we propose a simple numerical algorithm. We illustrate our proposals using simulated and stock returns data.
    Date: 2020–02
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2002.09578&r=all
  13. By: Jarek Duda
    Abstract: While standard estimation assumes that all datapoints are from probability distribution of the same fixed parameters $\theta$, we will focus on maximum likelihood (ML) adaptive estimation for nonstationary time series: separately estimating parameters $\theta_T$ for each time $T$ based on the earlier values $(x_t)_{t
    Date: 2020–03
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2003.02149&r=all
  14. By: Moyu Liao
    Abstract: This paper studies the identification and estimation problem in incomplete economic models with testable assumptions. Testable assumptions give strong and interpretable empirical content to the models but they also carry the possibility that our distribution of observed outcome may reject these assumptions. A natural way is to find a set of relaxed assumptions $\tilde{A}$ that cannot be rejected by any reasonable distribution of observed outcome and preserves identified set of parameter of interest. The main contribution of this paper is to characterize the property of such relaxed assumption $\tilde{A}$ using a generalized definition of refutability and confirmability. A general estimation and inference procedure is proposed, and can be applied to most incomplete economic models. As the key application, I study the Imbens and Angrist Monotonicity assumption in potential outcome framework. I give a set of relaxed assumptions $\tilde{A}$ that can never be rejected and always preserve the identified set of local average treatment effect (LATE). The LATE is point identified and easy to estimate under $\tilde{A}$
    Date: 2020–02
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2002.10415&r=all
  15. By: Andree,Bo Pieter Johannes; Spencer,Phoebe Girouard; Azari,Sardar; Chamorro,Andres; Wang,Dieter; Dogo,Harun
    Abstract: This paper introduces a Spatial Vector Autoregressive Moving Average (SVARMA) model in which multiple cross-sectional time series are modeled as multivariate, possibly fat-tailed, spatial autoregressive ARMA processes. The estimation requires specifying the cross-sectional spillover channels through spatial weights matrices. the paper explores a kernel method to estimate the network topology based on similarities in the data. It discusses the model and estimation, focusing on a penalized Maximum Likelihood criterion. The empirical performance of the estimator is explored in a simulation study. The model is used to study a spatial time series of pollution and household expenditure data in Indonesia. The analysis finds that the new model improves in terms of implied density, and better neutralizes residual correlations than the VARMA, using fewer parameters. The results suggest that growth in household expenditures precedes pollution reduction, particularly after the expenditures of poorer households increase; that increasing pollution is followed by reduced growth in expenditures, particularly reducing the growth of poorer households; and that there are significant spillovers from bottom-up growth in expenditures. The paper does not find evidence for top-down growth spillovers. Feedback between the identified mechanisms may contribute to pollution-poverty traps and the results imply that pollution damages are economically significant.
    Keywords: Global Environment,Inequality,Brown Issues and Health,Air Quality&Clean Air,Pollution Management&Control,Health Service Management and Delivery
    Date: 2019–02–25
    URL: http://d.repec.org/n?u=RePEc:wbk:wbrwps:8757&r=all
  16. By: Steven Y. K. Wong (University of Technology Sydney); Jennifer Chan (University of Sydney); Lamiae Azizi (University of Sydney); Richard Y. D. Xu (University of Technology Sydney)
    Abstract: We consider the problem of neural network training in a time-varying context. Machine learning algorithms have excelled in problems that do not change over time. However, problems encountered in financial markets are often non-stationary. We propose the online early stopping algorithm and show that a neural network trained using this algorithm can track a function changing with unknown dynamics. We applied the proposed algorithm to the stock return prediction problem studied in Gu et al. (2019) and achieved mean rank correlation of 4.69%, almost twice as high as the expanding window approach. We also show that prominent factors, such as the size effect and momentum, exhibit time varying stock return predictiveness.
    Date: 2020–03
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2003.02515&r=all
  17. By: Roy Allen; John Rehbeck
    Abstract: This paper provides nonparametric identification results for random coefficient distributions in perturbed utility models. We cover discrete and continuous choice models. We establish identification using variation in mean quantities, and the results apply when an analyst observes aggregate demands but not whether goods are chosen together. We require exclusion restrictions and independence between random slope coefficients and random intercepts. We do not require regressors to have large supports or parametric assumptions.
    Date: 2020–02
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2003.00276&r=all
  18. By: Oksana Bashchenko (HEC Lausanne; Swiss Finance Institute); Alexis Marchal (EPFL; SFI)
    Abstract: We develop a new method that detects jumps nonparametrically in financial time series and significantly outperforms the current benchmark on simulated data. We use a long short- term memory (LSTM) neural network that is trained on labelled data generated by a process that experiences both jumps and volatility bursts. As a result, the network learns how to disentangle the two. Then it is applied to out-of-sample simulated data and delivers results that considerably differ from the benchmark: we obtain fewer spurious detection and identify a larger number of true jumps. When applied to real data, our approach for jump screening allows to extract a more precise signal about future volatility.
    Keywords: Jumps, Volatility Burst, High-Frequency Data, Deep Learning, LSTM
    JEL: C14 C32 C45 C58 G17
    Date: 2020–03
    URL: http://d.repec.org/n?u=RePEc:chf:rpseri:rp2010&r=all
  19. By: Kubinec, Robert (Princeton University)
    Abstract: I propose a new model, ordered beta regression, for data collected from human subjects using slider scales/visual analog scales with lower and upper bounds. This model employs the cutpoint technique popularized by ordered logit to simultaneously estimate the probability that the outcome is at the upper bound, lower bound, or any continuous number in between. This model is contrasted with existing approaches, including ordinary least squares (OLS) regression and the zero-one-inflated beta regression (ZOIB) model. Simulation evidence shows that the proposed model, relative to existing approaches, estimates effects with more accuracy while capturing the full uncertainty in the distribution. Furthermore, an analysis of data on U.S. public opinion towards college professors reveals that the proposed model is better able to combine variation across continuous and degenerate responses. The model can be fit with the R package brms.
    Date: 2020–03–02
    URL: http://d.repec.org/n?u=RePEc:osf:socarx:2sx6y&r=all
  20. By: Marijn A. Bolhuis; Brett Rayner
    Abstract: We leverage insights from machine learning to optimize the tradeoff between bias and variance when estimating economic models using pooled datasets. Specifically, we develop a simple algorithm that estimates the similarity of economic structures across countries and selects the optimal pool of countries to maximize out-of-sample prediction accuracy of a model. We apply the new alogrithm by nowcasting output growth with a panel of 102 countries and are able to significantly improve forecast accuracy relative to alternative pools. The algortihm improves nowcast performance for advanced economies, as well as emerging market and developing economies, suggesting that machine learning techniques using pooled data could be an important macro tool for many countries.
    Date: 2020–02–28
    URL: http://d.repec.org/n?u=RePEc:imf:imfwpa:20/44&r=all
  21. By: Savi Virolainen
    Abstract: We introduce a new mixture autoregressive model which combines Gaussian and Student's $t$ mixture components. The model has very attractive properties analogous to the Gaussian and Student's $t$ mixture autoregressive models, but it is more flexible as it enables to model series which consist of both conditionally homoscedastic Gaussian regimes and conditionally heteroscedastic Student's $t$ regimes. The usefulness of our model is demonstrated in an empirical application to the monthly U.S. interest rate spread between the 3-month Treasury bill rate and the effective federal funds rate.
    Date: 2020–03
    URL: http://d.repec.org/n?u=RePEc:arx:papers:2003.05221&r=all

This nep-ecm issue is ©2020 by Sune Karlsson. It is provided as is without any express or implied warranty. It may be freely redistributed in whole or in part for any purpose. If distributed in part, please include this notice.
General information on the NEP project can be found at http://nep.repec.org. For comments please write to the director of NEP, Marco Novarese at <director@nep.repec.org>. Put “NEP” in the subject, otherwise your mail may be rejected.
NEP’s infrastructure is sponsored by the School of Economics and Finance of Massey University in New Zealand.