Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Series TitleSeries Title
-
Reading LevelReading Level
-
YearFrom:-To:
-
More FiltersMore FiltersContent TypeItem TypeIs Full-Text AvailableSubjectPublisherSourceDonorLanguagePlace of PublicationContributorsLocation
Done
Filters
Reset
2,403
result(s) for
"Hansen, Christian"
Sort by:
High-Dimensional Methods and Inference on Structural and Treatment Effects
by
Belloni, Alexandre
,
Chernozhukov, Victor
,
Hansen, Christian
in
Abortion
,
Censuses
,
Confounding (Statistics)
2014
Data with a large number of variables relative to the sample size—“high-dimensional data”—are readily available and increasingly common in empirical economics. Highdimensional data arise through a combination of two phenomena. First, the data may be inherently high dimensional in that many different characteristics per observation are available. For example, the US Census collects information on hundreds of individual characteristics and scanner datasets record transaction-level data for households across a wide range of products. Second, even when the number of available variables is relatively small, researchers rarely know the exact functional form with which the small number of variables enter the model of interest. Researchers are thus faced with a large set of potential variables formed by different ways of interacting and transforming the underlying variables. This paper provides an overview of how innovations in “data mining” can be adapted and modified to provide high-quality inference about model parameters. Note that we use the term “data mining” in a modern sense which denotes a principled search for “true” predictive power that guards against false discovery and overfitting, does not erroneously equate in-sample fit to out-of-sample predictive ability, and accurately accounts for using the same data to examine many different hypotheses or models.
Journal Article
Pre-Event Trends in the Panel Event-Study Design
2019
We consider a linear panel event-study design in which unobserved confounds may be related both to the outcome and to the policy variable of interest. We provide sufficient conditions to identify the causal effect of the policy by exploiting covariates related to the policy only through the confounds. Our model implies a set of moment equations that are linear in parameters. The effect of the policy can be estimated by 2SLS, and causal inference is valid even when endogeneity leads to pre-event trends (“pre-trends”) in the outcome. Alternative approaches perform poorly in our simulations.
Journal Article
Post-Selection and Post-Regularization Inference in Linear Models with Many Controls and Instruments
by
Chernozhukov, Victor
,
Hansen, Christian
,
Spindler, Martin
in
Coefficients
,
Data mining
,
Economic analysis
2015
We consider estimation of and inference about coefficients on endogenous variables in a linear instrumental variables model where the number of instruments and exogenous control variables are each allowed to be larger than the sample size. We work within an approximately sparse framework that maintains that the signal available in the instruments and control variables may be effectively captured by a small number of the available variables. We provide a LASSO-based method for this setting which provides uniformly valid inference about the coefficients on endogenous variables. We illustrate the method through an application to demand estimation.
Journal Article
Double/Debiased/Neyman Machine Learning of Treatment Effects
by
Newey, Whitney
,
Chetverikov, Denis
,
Duflo, Esther
in
Artificial intelligence
,
Bias
,
Cognitive style
2017
Chernozhukov et al. (2016) provide a generic double/de-biased machine learning (ML) approach for obtaining valid inferential statements about focal parameters, using Neyman-orthogonal scores and cross-fitting, in settings where nuisance parameters are estimated using ML methods. In this note, we illustrate the application of this method in the context of estimating average treatment effects and average treatment effects on the treated using observational data.
Journal Article
Double/debiased machine learning for treatment and structural parameters
by
Robins, James
,
Duflo, Esther
,
Chernozhukov, Victor
in
Artificial intelligence
,
Averages
,
Bias
2018
We revisit the classic semi-parametric problem of inference on a low-dimensional parameter θ₀ in the presence of high-dimensional nuisance parameters η₀. We depart from the classical setting by allowing for η₀ to be so high-dimensional that the traditional assumptions (e.g. Donsker properties) that limit complexity of the parameter space for this object break down. To estimate η₀, we consider the use of statistical or machine learning (ML) methods, which are particularly well suited to estimation in modern, very high-dimensional cases. ML methods perform well by employing regularization to reduce variance and trading off regularization bias with overfitting in practice. However, both regularization bias and overfitting in estimating η₀ cause a heavy bias in estimators of θ₀ that are obtained by naively plugging ML estimators of η₀ into estimating equations for θ₀. This bias results in the naive estimator failing to be N-½ consistent, where N is the sample size. We show that the impact of regularization bias and overfitting on estimation of the parameter of interest θ₀ can be removed by using two simple, yet critical, ingredients: (1) using Neyman-orthogonal moments/scores that have reduced sensitivity with respect to nuisance parameters to estimate θ₀; (2) making use of cross-fitting, which provides an efficient form of data-splitting. We call the resulting set of methods double or debiased ML (DML). We verify that DML delivers point estimators that concentrate in an N-½-neighbourhood of the true parameter values and are approximately unbiased and normally distributed, which allows construction of valid confidence statements. The generic statistical theory of DML is elementary and simultaneously relies on only weak theoretical requirements, which will admit the use of a broad array of modern ML methods for estimating the nuisance parameters, such as random forests, lasso, ridge, deep neural nets, boosted trees, and various hybrids and ensembles of these methods. We illustrate the general theory by applying it to provide theoretical properties of the following: DML applied to learn the main regression parameter in a partially linear regression model; DML applied to learn the coefficient on an endogenous variable in a partially linear instrumental variables model; DML applied to learn the average treatment effect and the average treatment effect on the treated under unconfoundedness; DML applied to learn the local average treatment effect in an instrumental variables setting. In addition to these theoretical applications, we also illustrate the use of DML in three empirical examples.
Journal Article
Analytic helicity amplitudes for two-loop five-gluon scattering: the single-minus case
by
Peraro, Tiziano
,
Brønnum-Hansen, Christian
,
Badger, Simon
in
Amplitudes
,
Classical and Quantum Gravitation
,
Elementary Particles
2019
A
bstract
We present a compact analytic expression for the leading colour two-loop five-gluon amplitude in Yang-Mills theory with a single negative helicity and four positive helicities. The analytic result is reconstructed from numerical evaluations over finite fields. The numerical method combines integrand reduction, integration-by-parts identities and Laurent expansion into a basis of pentagon functions to compute the coefficients directly from six-dimensional generalised unitarity cuts.
Journal Article
Plausibly exogenous
by
Conley, Timothy G.
,
Rossi, Peter E.
,
Hansen, Christian B.
in
Accumulation
,
Approximation
,
Assets
2012
Instrumental variable (IV) methods are widely used to identify causal effects in models with endogenous explanatory variables. Often the instrument exclusion restriction that underlies the validity of the usual IV inference is suspect; that is, instruments are only plausibly exogenous. We present practical methods for performing inference while relaxing the exclusion restriction. We illustrate the approaches with empirical examples that examine the effect of 401(k) participation on asset accumulation, price elasticity of demand for margarine, and returns to schooling. We find that inference is informative even with a substantial relaxation of the exclusion restriction in two of the three cases.
Journal Article
Inference on Treatment Effects after Selection among High-Dimensional Controls
by
HANSEN, CHRISTIAN
,
BELLONI, ALEXANDRE
,
CHERNOZHUKOV, VICTOR
in
Abortion
,
Approximation
,
Approximations
2014
We propose robust methods for inference about the effect of a treatment variable on a scalar outcome in the presence of very many regressors in a model with possibly non-Gaussian and heteroscedastic disturbances. We allow for the number of regressors to be larger than the sample size. To make informative inference feasible, we require the model to be approximately sparse; that is, we require that the effect of confounding factors can be controlled for up to a small approximation error by including a relatively small number of variables whose identities are unknown. The latter condition makes it possible to estimate the treatment effect by selecting approximately the right set of regressors. We develop a novel estimation and uniformly valid inference method for the treatment effect in this setting, called the \"post-double-selection\" method. The main attractive feature of our method is that it allows for imperfect selection of the controls and provides confidence intervals that are valid uniformly across a large class of models. In contrast, standard post-model selection estimators fail to provide uniform inference even in simple cases with a small, fixed number of controls. Thus, our method resolves the problem of uniform inference after model selection for a large, interesting class of models. We also present a generalization of our method to a fully heterogeneous model with a binary treatment variable. We illustrate the use of the developed methods with numerical simulations and an application that considers the effect of abortion on crime rates.
Journal Article