Search results

1 – 10 of over 2000
Book part
Publication date: 30 August 2019

Md. Nazmul Ahsan and Jean-Marie Dufour

Statistical inference (estimation and testing) for the stochastic volatility (SV) model Taylor (1982, 1986) is challenging, especially likelihood-based methods which are difficult…

Abstract

Statistical inference (estimation and testing) for the stochastic volatility (SV) model Taylor (1982, 1986) is challenging, especially likelihood-based methods which are difficult to apply due to the presence of latent variables. The existing methods are either computationally costly and/or inefficient. In this paper, we propose computationally simple estimators for the SV model, which are at the same time highly efficient. The proposed class of estimators uses a small number of moment equations derived from an ARMA representation associated with the SV model, along with the possibility of using “winsorization” to improve stability and efficiency. We call these ARMA-SV estimators. Closed-form expressions for ARMA-SV estimators are obtained, and no numerical optimization procedure or choice of initial parameter values is required. The asymptotic distributional theory of the proposed estimators is studied. Due to their computational simplicity, the ARMA-SV estimators allow one to make reliable – even exact – simulation-based inference, through the application of Monte Carlo (MC) test or bootstrap methods. We compare them in a simulation experiment with a wide array of alternative estimation methods, in terms of bias, root mean square error and computation time. In addition to confirming the enormous computational advantage of the proposed estimators, the results show that ARMA-SV estimators match (or exceed) alternative estimators in terms of precision, including the widely used Bayesian estimator. The proposed methods are applied to daily observations on the returns for three major stock prices (Coca-Cola, Walmart, Ford) and the S&P Composite Price Index (2000–2017). The results confirm the presence of stochastic volatility with strong persistence.

Details

Topics in Identification, Limited Dependent Variables, Partial Observability, Experimentation, and Flexible Modeling: Part A
Type: Book
ISBN: 978-1-78973-241-2

Keywords

Book part
Publication date: 29 March 2006

Jean-Marie Dufour and Pascale Valéry

In this paper, we consider the estimation of volatility parameters in the context of a linear regression where the disturbances follow a stochastic volatility (SV) model of order…

Abstract

In this paper, we consider the estimation of volatility parameters in the context of a linear regression where the disturbances follow a stochastic volatility (SV) model of order one with Gaussian log-volatility. The linear regression represents the conditional mean of the process and may have a fairly general form, including for example finite-order autoregressions. We provide a computationally simple two-step estimator available in closed form. Under general regularity conditions, we show that this two-step estimator is asymptotically normal. We study its statistical properties by simulation, compare it with alternative generalized method-of-moments (GMM) estimators, and present an application to the S&P composite index.

Details

Econometric Analysis of Financial and Economic Time Series
Type: Book
ISBN: 978-0-76231-274-0

Book part
Publication date: 23 June 2016

Bao Yong, Fan Yanqin, Su Liangjun and Zinde-Walsh Victoria

This paper examines Aman Ullah’s contributions to robust inference, finite sample econometrics, nonparametrics and semiparametrics, and panel and spatial models. His early works…

Abstract

This paper examines Aman Ullah’s contributions to robust inference, finite sample econometrics, nonparametrics and semiparametrics, and panel and spatial models. His early works on robust inference and finite sample theory were mostly motivated by his thesis advisor, Professor Anirudh Lal Nagar. They eventually led to his most original rethinking of many statistics and econometrics models that developed into the monograph Finite Sample Econometrics published in 2004. His desire to relax distributional and functional-form assumptions lead him in the direction of nonparametric estimation and he summarized his views in his most influential textbook Nonparametric Econometrics (with Adrian Pagan) published in 1999 that has influenced a whole generation of econometricians. His innovative contributions in the areas of seemingly unrelated regressions, parametric, semiparametric and nonparametric panel data models, and spatial models have also inspired a larger literature on nonparametric and semiparametric estimation and inference and spurred on research in robust estimation and inference in these and related areas.

Book part
Publication date: 16 December 2009

Zongwu Cai and Yongmiao Hong

This paper gives a selective review on some recent developments of nonparametric methods in both continuous and discrete time finance, particularly in the areas of nonparametric…

Abstract

This paper gives a selective review on some recent developments of nonparametric methods in both continuous and discrete time finance, particularly in the areas of nonparametric estimation and testing of diffusion processes, nonparametric testing of parametric diffusion models, nonparametric pricing of derivatives, nonparametric estimation and hypothesis testing for nonlinear pricing kernel, and nonparametric predictability of asset returns. For each financial context, the paper discusses the suitable statistical concepts, models, and modeling procedures, as well as some of their applications to financial data. Their relative strengths and weaknesses are discussed. Much theoretical and empirical research is needed in this area, and more importantly, the paper points to several aspects that deserve further investigation.

Details

Nonparametric Econometric Methods
Type: Book
ISBN: 978-1-84950-624-3

Article
Publication date: 1 March 1991

David Blake

The different types of estimators of rational expectations modelsare surveyed. A key feature is that the model′s solution has to be takeninto account when it is estimated. The two…

Abstract

The different types of estimators of rational expectations models are surveyed. A key feature is that the model′s solution has to be taken into account when it is estimated. The two ways of doing this, the substitution and errors‐in‐variables methods, give rise to different estimators. In the former case, a generalised least‐squares or maximum‐likelihood type estimator generally gives consistent and efficient estimates. In the latter case, a generalised instrumental variable (GIV) type estimator is needed. Because the substitution method involves more complicated restrictions and because it resolves the solution indeterminacy in a more arbitary fashion, when there are forward‐looking expectations, the errors‐in‐variables solution with the GIV estimator is the recommended combination.

Details

Journal of Economic Studies, vol. 18 no. 3
Type: Research Article
ISSN: 0144-3585

Keywords

Book part
Publication date: 30 December 2004

Thomas L. Marsh and Ron C. Mittelhammer

We formulate generalized maximum entropy estimators for the general linear model and the censored regression model when there is first order spatial autoregression in the…

Abstract

We formulate generalized maximum entropy estimators for the general linear model and the censored regression model when there is first order spatial autoregression in the dependent variable. Monte Carlo experiments are provided to compare the performance of spatial entropy estimators relative to classical estimators. Finally, the estimators are applied to an illustrative model allocating agricultural disaster payments.

Details

Spatial and Spatiotemporal Econometrics
Type: Book
ISBN: 978-0-76231-148-4

Book part
Publication date: 10 April 2019

Antonio Cosma, Andreï V. Kostyrka and Gautam Tripathi

We show how to use a smoothed empirical likelihood approach to conduct efficient semiparametric inference in models characterized as conditional moment equalities when data are…

Abstract

We show how to use a smoothed empirical likelihood approach to conduct efficient semiparametric inference in models characterized as conditional moment equalities when data are collected by variable probability sampling. Results from a simulation experiment suggest that the smoothed empirical likelihood based estimator can estimate the model parameters very well in small to moderately sized stratified samples.

Book part
Publication date: 19 December 2012

Francesco Bravo, Juan Carlos Escanciano and Taisuke Otsu

This chapter proposes a simple, fairly general, test for global identification of unconditional moment restrictions implied from point-identified conditional moment restrictions…

Abstract

This chapter proposes a simple, fairly general, test for global identification of unconditional moment restrictions implied from point-identified conditional moment restrictions. The test is a Hausman-type test based on the Hausdorff distance between an estimator that is consistent even under global identification failure of the unconditional moment restrictions, and an estimator of the identified set of the unconditional moment restrictions. The proposed test has a χ2 limiting distribution and is also able to detect weak identification. Some Monte Carlo experiments show that the proposed test has competitive finite sample properties already for moderate sample sizes.

Book part
Publication date: 19 December 2012

Nicky Grant

Principal component (PC) techniques are commonly used to improve the small sample properties of the linear instrumental variables (IV) estimator. Carrasco (2012) argue that PC…

Abstract

Principal component (PC) techniques are commonly used to improve the small sample properties of the linear instrumental variables (IV) estimator. Carrasco (2012) argue that PC type methods provide a natural ranking of instruments with which to reduce the size of the instrument set. This chapter shows how reducing the size of the instrument based on PC methods can lead to poor small sample properties of IV estimators. A new approach to ordering instruments termed ‘normalized principal components’ (NPCs) is introduced to overcome this problem. A simulation study shows the favourable small samples properties of IV estimators using NPC, methods to reduce the size of the instrument relative to PC. Using NPC we provide evidence that the IV setup in Angrist and Krueger (1992) may not suffer the weak instrument problem.

Details

Essays in Honor of Jerry Hausman
Type: Book
ISBN: 978-1-78190-308-7

Keywords

Book part
Publication date: 19 October 2020

Julian TszKin Chan

This chapter studies a snowball sampling method for social networks with endogenous peer selection. Snowball sampling is a sampling design which preserves the dependence structure…

Abstract

This chapter studies a snowball sampling method for social networks with endogenous peer selection. Snowball sampling is a sampling design which preserves the dependence structure of the network. It sequentially collects the information of vertices linked to the vertices collected in the previous iteration. The snowball samples suffer from a sample selection problem because of the endogenous peer selection. The author proposes a new estimation method that uses the relationship between samples in different iterations to correct selection. The author uses the snowball samples collected from Facebook to estimate the proportion of users who support the Umbrella Movement in Hong Kong.

1 – 10 of over 2000