nep-ecm New Economics Papers
on Econometrics
Issue of 2008‒03‒15
fifteen papers chosen by
Sune Karlsson
Orebro University

  1. An Alternative Sense of Asymptotic Efficiency By Mueller, Ulrich
  2. Bootstrap Inference in a Linear Equation Estimated by Instrumental Variables By Russell Davidson; James G. MacKinnon
  3. Tests for Unbalanced Error Component Models Under Local Misspecication By Walter Sosa Escudero; Anil K. Bera
  4. Thresholds, News Impact Surfaces and Dynamic Asymmetric Multivariate GARCH By Massimiliano Caporin; Michael McAleer
  5. Econometric Analysis of Structural Systems with Permanent and Transitory Shocks By Adrian R. Pagan; M. Hashem Pesaran
  6. Using Firm Optimization to Evaluate and Estimate Returns to Scale By Gorodnichenko, Yuriy
  7. Changing regime volatility: A fractionally integrated SETAR model By Gilles Dufrenot; Dominique Guegan; Anne Peguin-Feissolle
  8. Simulating interventions in graphical chain models for longitudinal data By Riccardo Borgoni; Peter W. F. Smith; Ann M. Berrington
  9. Cointegration implications of linear rational expectation models By Juselius, Mikael
  10. Engel elasticities, pseudo-maximum likelihood estimation and bootstrapped standard errors. A case study By Terje Skjerpen
  11. MIMIC Models, Cointegration and Error Correction: An Application to the French Shadow Economy By Buehn, Andreas; Schneider, Friedrich
  12. Stochastic Volatility: Origins and Overview By Neil Shephard; Torben G. Andersen
  13. Subset Selection from Large Datasets for Kriging Modeling By Rennen, G.
  14. Forecasting VaR and Expected shortfall using dynamical Systems : a risk Management Strategy, By Dominique Guegan; Cyril Caillault
  15. Hétérogénéité non observée dans les modèles de durée By Guillaume, HORNY

  1. By: Mueller, Ulrich
    Abstract: The paper studies the asymptotic efficiency and robustness of hypothesis tests when models of interest are defined in terms of a weak convergence property. The null and local alternatives induce different limiting distributions for a random element, and a test is considered robust if it controls asymptotic size for all data generating processes for which the random element has the null limiting distribution. Under weak regularity conditions, asymptotically robust and efficient tests are then simply given by efficient tests of the limiting problem--that is, with the limiting random element assumed observed--evaluated at sample analogues. These tests typically coincide with suitably robustified versions of optimal tests in canonical parametric versions of the model. This paper thus establishes an alternative and broader sense of asymptotic efficiency for many previously derived tests in econometrics, such as tests for unit roots, parameter stability tests and tests about regression coefficients under weak instruments, and it provides a concrete limit on the potential for more powerful tests in less parametric set-ups.
    Keywords: hypothesis tests; optimality; robustness; weak convergence
    JEL: C14 C12
    Date: 2008–03
    URL: http://d.repec.org/n?u=RePEc:pra:mprapa:7741&r=ecm
  2. By: Russell Davidson (McGill University); James G. MacKinnon (Queen's University)
    Abstract: We study several tests for the coefficient of the single right-hand-side endogenous variable in a linear equation estimated by instrumental variables. We show that writing all the test statistics -- Student's t, Anderson-Rubin, the LM statistic of Kleibergen and Moreira (K), and likelihood ratio (LR) -- as functions of six random quantities leads to a number of interesting results about the properties of the tests under weak-instrument asymptotics. We then propose several new procedures for bootstrapping the three non-exact test statistics and also a new conditional bootstrap version of the LR test. These use more efficient estimates of the parameters of the reduced-form equation than existing procedures. When the best of these new procedures is used, both the K and conditional bootstrap LR tests have excellent performance under the null. However, power considerations suggest that the latter is probably the method of choice.
    Keywords: bootstrap test, weak instruments, Anderson-Rubin test, conditional LR test, Wald test, K test
    JEL: C10 C12 C15 C30
    Date: 2008–03
    URL: http://d.repec.org/n?u=RePEc:qed:wpaper:1157&r=ecm
  3. By: Walter Sosa Escudero (Department of Economics Universidad de San Andrés); Anil K. Bera (Department of Economics University of Illinois)
    Abstract: This paper derives unbalanced versions of the tests statistics for ¯rst order serial correlation and random individual e®ects summarized in Sosa Escudero and Bera (2001), and updates their xttest1 routine. The derived tests statistics should be useful for applied researchers faced with the increasing availability of panel information where not every individual or country is observed for the full time span. Also, as it was the case of the previously available tests, the test statistics proposed here are based on the OLS residuals, and hence are computationally very simple.
    Keywords: error components model, unbalanced panel data, testing, misspecification.
    JEL: C12 C23 C52
    Date: 2008–03
    URL: http://d.repec.org/n?u=RePEc:dls:wpaper:0065&r=ecm
  4. By: Massimiliano Caporin (Università di Padova); Michael McAleer (University of Western Australia)
    Abstract: DAMGARCH extends the VARMA-GARCH model of Ling and McAleer (2003) by introducing multiple thresholds and time-dependent structure in the asymmetry of the conditional variances. DAMGARCH models the shocks affecting the conditional variances on the basis of an underlying multivariate distribution. It is possible to model explicitly asset-specific shocks and common innovations by partitioning the multivariate density support. This paper presents the model structure, describes the implementation issues, and provides the conditions for the existence of a unique stationary solution, and for consistency and asymptotic normality of the quasi-maximum likelihood estimators. The paper also provides analytical expressions for the news impact surface implied by DAMGARCH and an empirical example.
    Keywords: multivariate asymmetry, conditional variance, stationarity conditions, asymptotic theory, multivariate news impact curve
    Date: 2008
    URL: http://d.repec.org/n?u=RePEc:pad:wpaper:0064&r=ecm
  5. By: Adrian R. Pagan (School of Economics, The University of New South Wales); M. Hashem Pesaran (Faculty of Economics, University of Cambridge)
    Abstract: This paper considers the implications of the permanent/transitory decomposition of shocks for identification of structural models in the general case where the model might contain more than one permanent structural shock. It provides a simple and intuitive generalization of the influential work of Blanchard and Quah (1989), and shows that structural equations with known permanent shocks can not contain error correction terms, thereby freeing up the latter to be used as instruments in estimating their parameters. The approach is illustrated by a re-examination of the identification schemes used by Wickens and Motto (2001), Shapiro and Watson (1988), King, Plosser, Stock, Watson (1991), Gali (1992, 1999) and Fisher (2006).
    Keywords: Permanent shocks; structural identification; error correction models; IS-LM models
    JEL: C30 C32 E10
    Date: 2008–03
    URL: http://d.repec.org/n?u=RePEc:swe:wpaper:2008-04&r=ecm
  6. By: Gorodnichenko, Yuriy (University of California, Berkeley)
    Abstract: At the firm level, revenue and costs are well measured but prices and quantities are not. This paper shows that because of these data limitations estimates of returns to scale at the firm level are for the revenue function, not production function. Given this observation, the paper argues that, under weak assumptions, micro-level estimates of returns to scale are often inconsistent with profit maximization or imply implausibly large profits. The puzzle arises because popular estimators ignore heterogeneity and endogeneity in factor/product prices, assume perfect elasticity of factor supply curves or neglect the restrictions imposed by profit maximization (cost minimization) so that estimators are inconsistent or poorly identified. The paper argues that simple structural estimators can address these problems. Specifically, the paper proposes a full-information estimator that models the cost and the revenue functions simultaneously and accounts for unobserved heterogeneity in productivity and factor prices symmetrically. The strength of the proposed estimator is illustrated by Monte Carlo simulations and an empirical application. Finally, the paper discusses a number of implications of estimating revenue functions rather than production functions and demonstrates that the profit share in revenue is a robust non-parametric economic diagnostic for estimates of returns to scale.
    Keywords: production function, identification, returns to scale, covariance structures
    JEL: C23 C33 D24
    Date: 2008–02
    URL: http://d.repec.org/n?u=RePEc:iza:izadps:dp3368&r=ecm
  7. By: Gilles Dufrenot (GREQAM - Groupement de Recherche en Économie Quantitative d'Aix-Marseille - Université de la Méditerranée - Aix-Marseille II - Université Paul Cézanne - Aix-Marseille III - Ecole des Hautes Etudes en Sciences Sociales - CNRS : UMR6579); Dominique Guegan (CES - Centre d'économie de la Sorbonne - CNRS : UMR8174 - Université Panthéon-Sorbonne - Paris I); Anne Peguin-Feissolle (GREQAM - Groupement de Recherche en Économie Quantitative d'Aix-Marseille - Université de la Méditerranée - Aix-Marseille II - Université Paul Cézanne - Aix-Marseille III - Ecole des Hautes Etudes en Sciences Sociales - CNRS : UMR6579)
    Abstract: This paper presents a 2-regime SETAR model with different long-memory processes in both regimes. We briefly present the memory properties of this model and propose an estimation method. Such a process is applied to the absolute and squared returns of five stock indices. A comparison with simple FARIMA models is made using some forecastibility criteria. Our empirical results suggest that our model offers an interesting alternative competing framework to describe the persistent dynamics in modeling the returns.
    Keywords: SETAR - Long-memory - Stock indices - Forecasting
    Date: 2008–03–06
    URL: http://d.repec.org/n?u=RePEc:hal:papers:halshs-00185369_v1&r=ecm
  8. By: Riccardo Borgoni; Peter W. F. Smith; Ann M. Berrington
    Abstract: Simulating the outcome of an intervention is a central problem in many fields as this allows decision-makers to quantify the effect of any given strategy and, hence, to evaluate different schemes of actions. Simulation is particularly relevant in very large systems where the statistical model involves many variables that, possibly, interact with each other. In this case one usually has a large number of parameters whose interpretation becomes extremely difficult. Furthermore, in a real system, although one may have a unique target variable, there may be a number of variables which might, and often should, be logically considered predictors of the target outcome and, at the same time, responses of other variables of the system. An intervention taking place on a given variable, therefore, may affect the outcome either directly and indirectly though the way in which it affects other variables within the system. Graphical chain models are particularly helpful in depicting all of the paths through which an intervention may affect the final outcome. Furthermore, they identify all of the relevant conditional distributions and therefore they are particularly useful in driving the simulation process. Focussing on binary variables, we propose a method to simulate the effect of an intervention. Our approach, however, can be easily extended to continuous and mixed responses variables. We apply the proposed methodology to assess the effect that a policy intervention may have on poorer health in early adulthood using prospective data provided by the 1970 British Birth Cohort Study (BCS70).
    Keywords: chain graph, conditional approach, Gibbs Sampling, Simulation of interventions, age at motherhood, mental health
    JEL: C15 I18
    Date: 2008–02
    URL: http://d.repec.org/n?u=RePEc:mis:wpaper:20080301&r=ecm
  9. By: Juselius, Mikael (University of Helsinki, Department of Economics)
    Abstract: This paper derives the cointegration spaces that are implied by linear rational expectations models when data are I(1). The cointegration implications are easy to calculate and can be readily applied to test if the models are consistent with the long-run properties of the data. However, the restrictions on cointegration only form a subset of all the cross-equation restrictions that the models place on data. The approach is particularly useful in separating potentially data-consistent models from the remaining models within a large model family. Moreover, the approach provides useful information on the empirical shock structure of the data.
    Keywords: rational expectations; cointegration
    JEL: C52
    Date: 2008–03–11
    URL: http://d.repec.org/n?u=RePEc:hhs:bofrdp:2008_006&r=ecm
  10. By: Terje Skjerpen (Statistics Norway)
    Abstract: Estimation of standard errors of Engel elasticities within the framework of a linear structural model formulated on two-wave panel data is considered. The complete demand system is characterized by measurement errors in total expenditure and by latent preference variation. The estimation of the parameters as well as the standard errors of the estimates is based on the assumption that the variables are normally distributed. Considering a concrete case it is demonstrated that normality does not hold as a maintained assumption. In the light of this standard errors are estimated by means of bootstrapping. However, one obtains rather similar estimates of the standard errors of the Engel elasticities no matter whether one sticks to classical normal inference or perform non-parametric bootstrapping.
    Keywords: Engel elasticities; standard errors; classical normal theory; bootstrapping
    JEL: C13 C14 C15 C33 D12
    Date: 2008–03
    URL: http://d.repec.org/n?u=RePEc:ssb:dispap:532&r=ecm
  11. By: Buehn, Andreas (Dresden University of Technology); Schneider, Friedrich (University of Linz)
    Abstract: The analysis of economic loss attributed to the shadow economy has attracted much attention in recent years by both academics and policy makers. Often, multiple indicators multiple causes (MIMIC) models are applied to time series data estimating the size and development of the shadow economy for a particular country. This type of model derives information about the relationship between cause and indicator variables and a latent variable, here the shadow economy, from covariance structures. As most macroeconomic variables do not satisfy stationarity, long run information is lost when employing first differences. Arguably, this shortcoming is rooted in the lack of an appropriate MIMIC model which considers cointegration among variables. This paper develops a MIMIC model which estimates the cointegration equilibrium relationship and the error correction short run dynamics, thereby retaining information for the long run. Using France as our example, we demonstrate that this approach allows researchers to obtain more accurate estimates about the size and development of the shadow economy.
    Keywords: shadow economy, tax burden, regulation, unemployment, cointegration, error correction models, MIMIC models
    JEL: O17 O5 D78 H2 H11 H26
    Date: 2008–01
    URL: http://d.repec.org/n?u=RePEc:iza:izadps:dp3306&r=ecm
  12. By: Neil Shephard; Torben G. Andersen
    Date: 2008
    URL: http://d.repec.org/n?u=RePEc:sbs:wpsefe:2008fe23&r=ecm
  13. By: Rennen, G. (Tilburg University, Center for Economic Research)
    Abstract: When building a Kriging model, the general intuition is that using more data will always result in a better model. However, we show that when we have a large non-uniform dataset, using a uniform subset can have several advantages. Reducing the time necessary to fit the model, avoiding numerical inaccuracies and improving the robustness with respect to errors in the output data are some aspects which can be improved by using a uniform subset. We furthermore describe several new and current methods for selecting a uniform subset. These methods are tested and compared on several artificial datasets and one real life dataset. The comparison shows how the selected subsets affect different aspects of the resulting Kriging model. As none of the subset selection methods performs best on all criteria, the best method to choose depends on how the different aspects are valued. The comparison made in this paper can be used to facilitate the user in making a good choice.
    Keywords: Design of computer experiments;dispersion problem;Kriging model;large non-uniform datasets;radial basis functions;robustness;space filling;subset selection;uniformity.
    JEL: C0 C90
    Date: 2008
    URL: http://d.repec.org/n?u=RePEc:dgr:kubcen:200826&r=ecm
  14. By: Dominique Guegan (CES - Centre d'économie de la Sorbonne - CNRS : UMR8174 - Université Panthéon-Sorbonne - Paris I); Cyril Caillault (FORTIS Investments - Fortis investments)
    Abstract: Using non-parametric (copulas) and parametric models, we show that the bivariate distribution of an Asian portfolio is not stable along all the period under study. We suggest several dynamic models to compute two market risk measures, the Value at Risk and the Expected Shortfall: the RiskMetric methodology, the Multivariate GARCH models, the Multivariate Markov-Switching models, the empirical histogram and the dynamic copulas. We discuss the choice of the best method with respect to the policy management of bank supervisors. The copula approach seems to be a good compromise between all these models. It permits taking financial crises into account and obtaining a low capital requirement during the most important crises.
    Keywords: Value at Risk - Expected Shortfall - Copula - RiskMetrics - Risk management -<br />GARCH models - Switching models.
    Date: 2008–03–06
    URL: http://d.repec.org/n?u=RePEc:hal:papers:halshs-00185374_v1&r=ecm
  15. By: Guillaume, HORNY
    Abstract: Cet article est une revue de la littérature où le temps passé dans un état est une variable aléatoire issue d’un mélange continu de distributions. Elle s’est constituée à partir de l’estimation de fonctions de hasards et de méthodes d’approximations d’intégrales. Nous présentons d’abord le modèle de mélange de hasards proprotionnels et ses propriétés. Les conséquences des principaux résultats d’identification sont ensuite discutées. Nous présentons ensuite des méthodes d’estimations paramétriques, semi-paramétriques et bayésiennes, ainsi que les méthodes d’optimisation correspondantes.
    Keywords: modèles de durée, hasards proportionnels, vraisemblance pénalisée, vraisemblance partielle, méthodes bayésiennes
    JEL: C41 C51 C61 C11 C14
    Date: 2008–02–01
    URL: http://d.repec.org/n?u=RePEc:ctl:louvec:2007046&r=ecm

This nep-ecm issue is ©2008 by Sune Karlsson. It is provided as is without any express or implied warranty. It may be freely redistributed in whole or in part for any purpose. If distributed in part, please include this notice.
General information on the NEP project can be found at http://nep.repec.org. For comments please write to the director of NEP, Marco Novarese at <director@nep.repec.org>. Put “NEP” in the subject, otherwise your mail may be rejected.
NEP’s infrastructure is sponsored by the School of Economics and Finance of Massey University in New Zealand.