
on Econometrics 
By:  Ping Yu (The University of Hong Kong); Qin Liao (The University of Hong Kong); Peter C.B. Phillips (Cowles Foundation, Yale University) 
Abstract:  We propose three new methods of inference for the threshold point in endogenous threshold regression and two speciï¬ cation tests designed to assess the presence of endogeneity and threshold eï¬€ects without necessarily relying on instrumentation of the covariates. The ï¬ rst inferential method is a parametric twostage least squares method and is suitable when instruments are available. The second and third methods are based on smoothing the objective function of the integrated diï¬€erence kernel estimator in diï¬€erent ways and these methods do not require instrumentation. All three methods are applicable irrespective of endogeneity of the threshold variable. The two speciï¬ cation tests are constructed using a scoretype principle. The threshold eï¬€ect test extends conventional parametric structural change tests to the nonparametric case. A wild bootstrap procedure is suggested to deliver ï¬ nite sample critical values for both tests. Simulations show good ï¬ nite sample performance of these procedures and the methods provide flexibility in testing and inference for practitioners working with threshold models. 
Keywords:  Threshold regression, Endogeneity, Identification, Confidence interval, 2SLS, IDKE, Secification testing, Bootstrap, Ustatistic 
JEL:  C21 C24 C26 
Date:  2019–11 
URL:  http://d.repec.org/n?u=RePEc:cwl:cwldpp:2209&r=all 
By:  Chib, Siddhartha (Olin Business School, Washington University in St. Louis); Shin, Minchul (Federal Reserve Bank of Philadelphia); Simoni, Anna (CREST, CNRS, Ecole Polytechnique) 
Abstract:  We provide a Bayesian analysis of models in which the unknown distribution of the outcomes is speciﬁed up to a set of conditional moment restrictions. This analysis is based on the nonparametric exponentially tilted empirical likelihood (ETEL) function, which is constructed to satisfy a sequence of unconditional moments, obtained from the conditional moments by an increasing (in sample size) vector of approximating functions (such as tensor splines based on the splines of each conditioning variable). The posterior distribution is shown to satisfy the Bernsteinvon Mises theorem, subject to a growth rate condition on the number of approximating functions, even under misspeciﬁcation of the conditional moments. A largesample theory for comparing different conditional moment models is also developed. The central result is that the marginal likelihood criterion selects the model that is less misspeciﬁed, that is, the model that is closer to the unknown true distribution in terms of the KullbackLeibler divergence. Several examples are provided to illustrate the framework and results. 
Keywords:  Bayesian inference; Bernsteinvon Mises theorem; Conditional moment restrictions; Exponentially tilted empirical likelihood; Marginal likelihood; Misspeciﬁcation; Posterior consistency 
JEL:  C11 C13 C14 C52 
Date:  2019–12–09 
URL:  http://d.repec.org/n?u=RePEc:fip:fedpwp:1951&r=all 
By:  Jin Seo Cho (Yonsei Univ); Matthew GreenwoodNimmo (Univ of Melbourne); Yong Cheol Shin (Univ of York) 
Abstract:  We consider estimation of and inference on the nonlinear autoregressive distributed lag (NARDL) model, which is a singleequation error correction model that allows for asymmetry with respect to positive and negative changes in the explanatory variable(s). We show that the NARDL model exhibits an asymptotic singularity issue that frustrates efforts to derive the asymptotic properties of the singlestep estimator. Consequently, we propose a twostep estimation framework, in which the parameters of the longrun relationship are estimated first using the fullymodified least squares estimator before the dynamic parameters are estimated by OLS in the second step. We show that our twostep estimators are consistent for the parameters of the NARDL model and we derive their limit distributions. We also develop Wald test statistics for the hypotheses of shortrun and longrun parameter asymmetry. We demonstrate the utility of our framework with an application to postwar dividendsmoothing in the U.S. 
Keywords:  Nonlinear Autoregressive Distributed Lag (NARDL) Model; FullyModified Least Squares Estimator; TwoStep Estimation; Wald Test Statistic; DividendSmoothing. 
JEL:  C22 G35 
Date:  2019–12 
URL:  http://d.repec.org/n?u=RePEc:yon:wpaper:2019rwp154&r=all 
By:  Hajivassiliou, Vassilis 
Abstract:  This paper proposes efficient estimation methods for panel data limited dependent variables (LDV) models possessing a variety of complications: nonignorable persistent heterogeneity; contemporaneous and intertemporal endogeneity; and observable and unobservable dynamics. An important problem handled by the novel framework of this paper involves contemporaneous and intertemporal simultaneity caused by social strategic interactive effects or contagion across economic agents over time. The paper first shows how a simple modification of estimators based on the Random Effects principle can preserve the consistency and asymptotic efficiency of the method in panel data despite nonignorable persistent heterogeneity driven by correlations between the individualspecific component of the error term and the regressors. The approach is extremely easy to implement and allows straightforward classical and omnibus tests of the significance of such correlations that lie behind the nonignorable persistent heterogeneity. The method applies to linear as well as nonlinear panel data models, static or dynamic. Two major extensions of the existing literature are that the method works for timeinvariant as well as timevarying regressors, and that these dependencies may be nonlinear functions of the regressors. The paper then combines this modified random effects approach with two simulationbased estimation strategies to overcome analytical as well as computational intractabilities in a widely applicable class of nonlinear models for panel data, namely the class of LDV models with contemporaneous and intertemporal endogeneity. The effectiveness of the estimation methods in providing asymptotically efficient estimates in such cases is illustrated with three discreteresponse econometric models for panel data. 
Keywords:  limited dependent variable models; simulationbased estimation; endogeneity; correlated random effects; initial conditions in nonlinear dynamic panel data models; strategic and social interaction; contagion 
JEL:  C51 C52 C15 
Date:  2019–09 
URL:  http://d.repec.org/n?u=RePEc:ehl:lserod:102843&r=all 
By:  Maurizio Daniele; Julie Schnaitmann 
Abstract:  We propose a regularized factoraugmented vector autoregressive (FAVAR) model that allows for sparsity in the factor loadings. In this framework, factors may only load on a subset of variables which simplifies the factor identification and their economic interpretation. We identify the factors in a datadriven manner without imposing specific relations between the unobserved factors and the underlying time series. Using our approach, the effects of structural shocks can be investigated on economically meaningful factors and on all observed time series included in the FAVAR model. We prove consistency for the estimators of the factor loadings, the covariance matrix of the idiosyncratic component, the factors, as well as the autoregressive parameters in the dynamic model. In an empirical application, we investigate the effects of a monetary policy shock on a broad range of economically relevant variables. We identify this shock using a joint identification of the factor model and the structural innovations in the VAR model. We find impulse response functions which are in line with economic rationale, both on the factor aggregates and observed time series level. 
Date:  2019–12 
URL:  http://d.repec.org/n?u=RePEc:arx:papers:1912.06049&r=all 
By:  Igor Kheifets (Instituto Tecnologico Autonomo de Mexico); Peter C.B. Phillips (Cowles Foundation, Yale University) 
Abstract:  Multicointegration is traditionally deï¬ ned as a particular long run relationship among variables in a parametric vector autoregressive model that introduces links between these variables and partial sums of the equilibrium errors. This paper departs from the parametric model, using a semiparametric formulation that reveals the explicit role that singularity of the long run conditional covariance matrix plays in determining multicointegration. The semiparametric framework has the advantage that short run dynamics do not need to be modeled and estimation by standard techniques such as fully modiï¬ ed least squares (FMOLS) on the original I(1) system is straightforward. The paper derives FMOLS limit theory in the multicointegrated setting, showing how faster rates of convergence are achieved in the direction of singularity and that the limit distribution depends on the distribution of the conditional onesided long run covariance estimator used in FMOLS estimation. Wald tests of restrictions on the regression coeï¬€icients have nonstandard limit theory which depends on nuisance parameters in general. The usual tests are shown to be conservative when the restrictions are isolated to the directions of singularity and, under certain conditions, are invariant to singularity otherwise. Simulations show that approximations derived in the paper work well in ï¬ nite samples. We illustrate our ï¬ ndings by analyzing ï¬ scal sustainability of the US government over the postwar period. 
Keywords:  Cointegration, Multicointegration, Fully modified regression, Singular long run variance matrix, Degenerate Wald test, Fiscal sustainability 
JEL:  C12 C13 C22 
Date:  2019–11 
URL:  http://d.repec.org/n?u=RePEc:cwl:cwldpp:2210&r=all 
By:  Halvarsson, Daniel (The Ratio Institute) 
Abstract:  This paper considers a flexible class of asymmetric double Pareto distributions (ADP) that allows for skewness and asymmetric heavy tails. The inference problem is examined for maximum likelihood. Consistency is proven for the general case when all parameters are unknown. After deriving the Fisher information matrix, asymptotic normality and efficiency are established for a restricted model with the location parameter known. The asymptotic properties of the estimators are then examined using Monte Carlo simulations. To assess its goodness of fit, the ADP is applied to companies’ growth rates, for which it is unequivocally favored over competing models 
Keywords:  Distribution Theory; Double Pareto Distribution; Maximum Likelihood; Firm Growth 
JEL:  C16 C46 
Date:  2019–12–18 
URL:  http://d.repec.org/n?u=RePEc:hhs:ratioi:0327&r=all 
By:  Simon Hetland (Department of Economics, University of Copenhagen, Denmark); Rasmus Søndergaard Pedersen (Department of Economics, University of Copenhagen, Denmark); Anders Rahbek (Department of Economics, University of Copenhagen, Denmark) 
Abstract:  In this paper we consider a multivariate generalized autoregressive conditional heteroskedastic (GARCH) class of models where the eigenvalues of the conditional covariance matrix are timevarying. The proposed dynamics of the eigenvalues is based on applying the general theory of dynamic conditional score models as proposed by Creal, Koopman and Lucas (2013) and Harvey (2013). We denote the obtained GARCH model with dynamic conditional eigenvalues (and constant conditional eigenvectors) as the ?GARCH model. We provide new results on asymptotic theory for the Gaussian QMLE, and for testing of reduced rank of the (G)ARCH loading matrices of the timevarying eigenvalues. The theory is applied to US data, where we ?find that the eigenvalue structure can be reduced similar to testing for the number in factors in volatility models. 
Keywords:  Multivariate GARCH; GOGARCH; Reduced Rank; Asymptotic Theory 
JEL:  C32 C51 C58 
Date:  2019–12–17 
URL:  http://d.repec.org/n?u=RePEc:kud:kuiedp:1913&r=all 
By:  Laura Liu (Indiana University, Bloomington, Indiana); Hyungsik Roger Moon (University of Southern California and Yonsei); Frank Schorfheide (University of Pennsylvania CEPR, NBER, and PIER) 
Abstract:  We use a dynamic panel Tobit model with heteroskedasticity to generate point, set, and density forecasts for a large crosssection of short time series of censored observations. Our fully Bayesian approach allows us to ﬂexibly estimate the crosssectional distribution of heterogeneous coeffients and then implicitly use this distribution as prior to construct Bayes forecasts for the individual time series. We construct set forecasts that explicitly target the average coverage probability for the crosssection. We present a novel application in which we forecast banklevel chargeoff rates for credit card and residential real estate loans, comparing various versions of the panel Tobit model. 
Keywords:  Bayesian inference, density forecasts, interval forecasts, loan chargeoffs, panel data, point forecasts, set forecasts, Tobit model 
JEL:  C11 C14 C23 C53 G21 
Date:  2019–05 
URL:  http://d.repec.org/n?u=RePEc:inu:caeprp:2019005&r=all 
By:  Mohamed CHIKHI; Claude DIEBOLT; Tapas MISHRA 
Abstract:  Despite an inherent share of unpredictability, asset prices such as in stock and Bitcoin markets are naturally driven by significant magnitudes of memory; depending on the strength of path dependence, prices in such markets can be (at least partially) predicted. Being able to predict asset prices is always a boon for investors, more so, if the forecasts are largely unconditional and can only be explained by the series’ own historical trajectories. Although memory dynamics have been exploited in forecasting stock prices, Bitcoin market pose additional challenge, because the lack of proper financial theoretic model limits the development of adequate theorydriven empirical construct. In this paper, we propose a class of autoregressive fractionally integrated moving average (ARFIMA) model with asymmetric exponential generalized autoregressive score (AEGAS) errors to accommodate a complex interplay of ‘memory’ to drive predictive performance (an outofsample forecasting). Our conditional variance includes leverage effects, jumps and fat tailskewness distribution, each of which affects magnitude of memory both the stock and Bitcoin price system would possess enabling us to build a true forecast function. We estimate several models using the Skewed Studentt maximum likelihood and find that the informational shocks in asset prices, in general, have permanent effects on returns. The ARFIMAAEGAS is appropriate for capturing volatility clustering for both negative (long ValueatRisk) and positive returns (short ValueatRisk). We show that this model has better predictive performance over competing models for both long and/or some short time horizons. The predictions from this model beats comfortably the random walk model. Accordingly, we find that the weak efficiency assumption of financial markets stands violated for all price returns studied over longer time horizon. 
Keywords:  Asset price; Forecasting; Memory; ARFIMAAEGAS; Leverage effects and jumps; Market Efficiency. 
JEL:  C14 C58 C22 G17 
Date:  2019 
URL:  http://d.repec.org/n?u=RePEc:ulp:sbbeta:201943&r=all 
By:  Cannon, Alex J. (Environment and Climate Change Canada) 
Abstract:  The goal of quantile regression is to estimate conditional quantiles for specified values of quantile probability using linear or nonlinear regression equations. These estimates are prone to "quantile crossing", where regression predictions for different quantile probabilities do not increase as probability increases. In the context of the environmental sciences, this could, for example, lead to estimates of the magnitude of a 10yr return period rainstorm that exceed the 20yr storm, or similar nonphysical results. This problem, as well as the potential for overfitting, is exacerbated for small to moderate sample sizes and for nonlinear quantile regression models. As a remedy, this study introduces a novel nonlinear quantile regression model, the monotone composite quantile regression neural network (MCQRNN), that (1) simultaneously estimates multiple noncrossing, nonlinear conditional quantile functions; (2) allows for optional monotonicity, positivity/nonnegativity, and generalized additive model constraints; and (3) can be adapted to estimate standard leastsquares regression and noncrossing expectile regression functions. First, the MCQRNN model is evaluated on synthetic data from multiple functions and error distributions using Monte Carlo simulations. MCQRNN outperforms the benchmark models, especially for nonnormal error distributions. Next, the MCQRNN model is applied to realworld climate data by estimating rainfall IntensityDurationFrequency (IDF) curves at locations in Canada. IDF curves summarize the relationship between the intensity and occurrence frequency of extreme rainfall over storm durations ranging from minutes to a day. Because annual maximum rainfall intensity is a nonnegative quantity that should increase monotonically as the occurrence frequency and storm duration decrease, monotonicity and nonnegativity constraints are key constraints in IDF curve estimation. In comparison to standard QRNN models, the ability of the MCQRNN model to incorporate these constraints, in addition to noncrossing, leads to more robust and realistic estimates of extreme rainfall. 
Date:  2017–12–05 
URL:  http://d.repec.org/n?u=RePEc:osf:eartha:wg7sn&r=all 
By:  Jan Bierlant; Sven Buitendag; Eustasio Del Barrio; Marc Hallin 
Abstract:  All multivariate extensions of the univariate theory of risk measurement run into the same fundamental problem of the absence, in dimension d > 1, of a canonical ordering of Rd. Based on measure transportation ideas, several attempts have been made recently in the statistical literature to overcome that conceptual diﬃculty. In Hallin (2017), the concepts of centeroutward distribution and quantile functions are developed as generalisations of the classical univariate concepts of distribution and quantile functions, along with their empirical versions. The centeroutward distribution function F± is a homeomorphic cyclically monotone mapping from Rd \ F−1 ± (0) to the open punctured unit ball Bd \ {0}, while its empirical counterpart F(n) ± is a cyclically monotone mapping from the sample to a regular grid over Bd. In dimension d = 1, F± reduces to 2F − 1, while F(n) ± generates the same sigmaﬁeld as traditional univariate ranks. The empirical F(n) ± ,however, involves a large number of ties, which is impractical in the context of risk measurement. We therefore propose a class of smooth approximations Fn,ξ (ξ a smoothness index) of F(n) ± as an alternative to the interpolation developed in del Barrio et al. (2018). This approximation allows for the computation of some new empirical risk measures, based either on the convex potential associated with the proposed transports, or on the volumes of the resulting empirical quantile regions. We also discuss the role of such transports in the evaluation of the risk associated with multivariate regularly varying distributions. Some simulations and applications to case studies illustrate the value of the approach. 
Date:  2019–12 
URL:  http://d.repec.org/n?u=RePEc:eca:wpaper:2013/297778&r=all 
By:  Dinghai Xu (Department of Economics, University of Waterloo) 
Abstract:  This paper investigates the “spurious almost integration” effect of volatility under a threshold GARCH structure with realized volatility measures. To closely examine the effect, the realized persistence of volatility is proposed to be used as a threshold trigger for volatility regimes. Under the threshold framework, general closedform solutions of moment conditions are derived, which provide a convenient way to theoretically examine the “spurious almost integration” effect and its associated impacts. We find that introducing the volatility persistencedriven threshold can capture regimespecific characteristics well. It performs better than the traditional GARCHtype models in terms of both insample fitting and outofsample forecasting. Based on our Monte Carlo and empirical results, in general we find that overlooking the relatively low persistence regime(s) could lead to some misleading conclusions. 
JEL:  C01 C58 
Date:  2019–12 
URL:  http://d.repec.org/n?u=RePEc:wat:wpaper:1903&r=all 
By:  Hui Chen; Winston Wei Dou; Leonid Kogan 
Abstract:  We introduce an informationbased fragility measure for GMM models that are potentially misspecified and unstable. A large fragility measure signifies a GMM model's lack of internal refutability (weak power of specification tests) and external validity (poor outofsample fit). The fragility of a set of modelimplied moment restrictions is tightly linked to the quantity of additional information the econometrician can obtain about the model parameters by imposing these restrictions. Our fragility measure can be computed at little cost even for complex dynamic structural models. We illustrate its applications via two models: a raredisaster risk model and a longrun risk model. 
JEL:  C52 D81 E32 G12 
Date:  2019–11 
URL:  http://d.repec.org/n?u=RePEc:nbr:nberwo:26418&r=all 
By:  Mark Bognanni; John Zito 
Abstract:  We develop a sequential Monte Carlo (SMC) algorithm for Bayesian inference in vector autoregressions with stochastic volatility (VARSV). The algorithm builds particle approximations to the sequence of the model’s posteriors, adapting the particles from one approximation to the next as the window of available data expands. The parallelizability of the algorithm’s computations allows the adaptations to occur rapidly. Our particular algorithm exploits the ability to marginalize many parameters from the posterior analytically and embeds a known Markov chain Monte Carlo (MCMC) algorithm for the model as an effective mutation kernel for fighting particle degeneracy. We show that, relative to using MCMC alone, our algorithm increases the precision of inference while reducing computing time by an order of magnitude when estimating a mediumscale VARSV model. 
Keywords:  Vector autoregressions; sequential Monte Carlo; RaoBlackwellization; particle filter; stochastic volatility 
JEL:  E17 C11 C51 C32 
Date:  2019–12–16 
URL:  http://d.repec.org/n?u=RePEc:fip:fedcwq:86647&r=all 
By:  Thomas Crossley (Institute for Fiscal Studies and Institute for Fiscal Studies, University of Essex); Peter Levell (Institute for Fiscal Studies and Institute for Fiscal Studies); Stavros Poupakis (Institute for Fiscal Studies and University of Essex) 
Abstract:  Researchers are often interested in the relationship between two variables, with no single data set containing both. A common strategy is to use proxies for the dependent variable that are common to two surveys to impute the dependent variable into the data set containing the independent variable. We show that commonly employed regression or matchingbased imputation procedures lead to inconsistent estimates. We o?er an easilyimplemented correction and correct asymptotic standard errors. We illustrate these with Monte Carlo experiments and empirical examples using data from the US Consumer Expenditure Survey (CE) and the Panel Study of Income Dynamics (PSID). 
Date:  2019–06–24 
URL:  http://d.repec.org/n?u=RePEc:ifs:ifsewp:19/16&r=all 
By:  Bryan T. Kelly; Asaf Manela; Alan Moreira 
Abstract:  Text data is ultrahigh dimensional, which makes machine learning techniques indispensable for textual analysis. Text is often selected—journalists, speechwriters, and others craft messages to target their audiences’ limited attention. We develop an economically motivated high dimensional selection model that improves learning from text (and from sparse counts data more generally). Our model is especially useful when the choice to include a phrase is more interesting than the choice of how frequently to repeat it. It allows for parallel estimation, making it computationally scalable. A first application revisits the partisanship of US congressional speech. We find that earlier spikes in partisanship manifested in increased repetition of different phrases, whereas the upward trend starting in the 1990s is due to entirely distinct phrase selection. Additional applications show how our model can backcast, nowcast, and forecast macroeconomic indicators using newspaper text, and that it substantially improves outofsample fit relative to alternative approaches. 
JEL:  C1 C4 C55 C58 E17 G12 G17 
Date:  2019–11 
URL:  http://d.repec.org/n?u=RePEc:nbr:nberwo:26517&r=all 
By:  Valérie Lechene (Institute for Fiscal Studies and University College London); Krishna Pendakur (Institute for Fiscal Studies and Simon Fraser University); Alexander Wolf (Institute for Fiscal Studies and ECARES) 
Abstract:  Individuals may be poor even if their household is not poor, because the intrahousehold distribution of resources may be unequal. We develop a model wherein the resource share of each person in a collective household  defined as their share of household consumption  may be estimated by simple linear regressions using offtheshelf consumer expenditure microdata. The model is a linear approximation of Dunbar, Lewbel and Pendakur (2013), whose nonlinear structural model can be computationally difficult. Our model allows for complex household types, including those with multiple adult men and/or women and single parent households. We also provide a simple linear pretest to check for model identification. Resource shares are obtained as nonlinear functions of estimated coefficients from OLS regressions. We apply the model to data from 12 countries, and investigate resource shares, gender gaps and individual poverty. We find that equal sharing  the implicit assumption underlying householdlevel poverty calculations  is always rejected. We also find evidence of large gender gaps in resource shares, and consequently in poverty rates, in a few countries. 
Date:  2019–07–11 
URL:  http://d.repec.org/n?u=RePEc:ifs:ifsewp:19/19&r=all 
By:  KAINOU Kazunari 
Abstract:  DifferenceInDifference(DID) is frequently used methodology in Policy Impact Assessment, but necessary assumptions to be confirmed for DID or ways to fulfill and ensure them are not clearly identified nor well developed yet. Especially in case of ignoring No Autocorrelations Assumptions(NACA) or Stable Unit Treatment Value Assumption(SUTVA) related problems may cause certain bias in the estimated assessment results. This paper shows that four majpr assumptions need to be confirmed in DID, namely Overalap, Conditional Independence, NACA and SUTVA, based on the inductive survay of academic papers in Economics, Sociology and so on, and identifies existing measures applicable for three major approaches of DID, experimental approach with rendomisation, statistical approach with matching or synthetic control group. Based on the inductive survay above, this paper proposes new methodology applocable for statistical approaches where both NACA and SUTVA related problems may happen. Assuming that secondary effect from treated group to control group are identical and other assumptions holds, checking the significance of constant terms in the result of regression analysis of rate of beforeafter indicator(BAI) and differenceindifference indicator(DIDI) with inverse DIDI for each control group samples provides solutions for both NACA and SUTVA related problems. In order to demonstrate the practicality of the new methodology, this paper tries treatment effect evaluation of Fukushima rice price before and after the East Japan Great Earthquake and Fukusima No.1 Plant Nuclear Accident where NACA and SUTVA related problems exist. And this paper quantified possible bias caused by four major assumptions related problems of DID and concluded that SUTVA related problems potentially cause largest bias in the results in this case. 
Date:  2019–11 
URL:  http://d.repec.org/n?u=RePEc:eti:rdpsjp:19065&r=all 
By:  Hirschauer, Norbert; Grüner, Sven; Mußhoff, Oliver; Becker, Claudia 
Abstract:  Replication crisis and debates about pvalues have raised doubts about what we can statistically infer from research findings, both in experimental and observational studies. With a view to the present debate on inferential errors, this paper systematizes and discusses experimental designs with regard to the inferences that can and  perhaps more important  that cannot be made from particular designs. 
Keywords:  economic experiments,ceteris paribus,confounders,control,inference,Internal/external validity,randomization,random sampling,superpopulation 
JEL:  B41 C18 C90 
Date:  2019 
URL:  http://d.repec.org/n?u=RePEc:zbw:ifwedp:201965&r=all 
By:  Bartosz Uniejewski; Rafal Weron 
Abstract:  Quantile Regression Averaging (QRA) has sparked interest in the electricity price forecasting community after its unprecedented success in the Global Energy Forecasting Competition 2014, where the top two winning teams in the price track used variants of QRA. However, recent studies have reported the method's vulnerability to low quality predictors when the set of regressors is larger than just a few. To address this issue, we consider a regularized variant of QRA, which utilizes the Least Absolute Shrinkage and Selection Operator (LASSO) to automatically select the relevant regressors. We evaluate the introduced technique – dubbed LASSO QRA or LQRA for short – using datasets from the Polish and Nordic power markets, a set of 25 point forecasts obtained for calibration windows of different lengths and 20 different values of the regularization parameter. By comparing against nearly 30 benchmarks, we provide evidence for its superior predictive performance in terms of the Kupiec test, the pinball score and the test for conditional predictive accuracy. 
Keywords:  Electricity price forecasting; Probabilistic forecast; Quantile Regression Averaging; LASSO; Kupiec test; Pinball score; Conditional predictive accuracy 
JEL:  C22 C32 C51 C52 C53 Q41 Q47 
Date:  2019–11–16 
URL:  http://d.repec.org/n?u=RePEc:wuu:wpaper:hsc1904&r=all 