(Machine) learning parameter regions

被引:2
作者
Olea, Jose Luis Montiel [1 ]
Nesbit, James [2 ]
机构
[1] Columbia Univ, Dept Econ, 420 West 118th St, New York, NY 10027 USA
[2] NYU, Dept Econ, 19 W 4th St,6th Floor, New York, NY 10012 USA
关键词
Machine learning; Supervised learning; Set-identified models; Structural vector autoregressions; UNIFORM CONFIDENCE BANDS; INFERENCE; IDENTIFICATION; SETS;
D O I
10.1016/j.jeconom.2020.06.008
中图分类号
F [经济];
学科分类号
02 ;
摘要
How many random points from an identified set, a confidence set, or a highest posterior density set suffice to describe them? This paper argues that taking random draws from a parameter region in order to approximate its shape is a supervised learning problem (analogous to sampling pixels of an image to recognize it). Misclassification error - a common criterion in machine learning - provides an off-the-shelf tool to assess the quality of a given approximation. We say a parameter region can be learned if there is an algorithm that yields a misclassification error of at most epsilon with probability at least 1 - delta, regardless of the sampling distribution. We show that learning a parameter region is possible if and only if its potential shapes are not too complex. Moreover, the tightest band that contains a d-dimensional parameter region is always learnable from the inside (in a sense we make precise), with at least max {(1 - epsilon) ln (1/delta), (3/16)d}/epsilon draws, but at most min{2d ln(2d/delta), exp(1)(2d+ln(1/delta))}/epsilon. These bounds grow linearly in the dimension of the parameter region, and are uniform with respect to its true shape. We illustrate the usefulness of our results using structural vector autoregressions. We show how many orthogonal matrices are necessary/sufficient to evaluate the impulse responses' identified set and how many 'shotgun plots' to report when conducting joint inference on impulse responses. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页码:716 / 744
页数:29
相关论文
共 41 条
[31]  
Mohri M., 2012, FDN MACHINE LEARNING
[32]   Bayesian and Frequentist Inference in Partially Identified Models [J].
Moon, Hyungsik Roger ;
Schorfheide, Frank .
ECONOMETRICA, 2012, 80 (02) :755-782
[33]  
Murphy KP, 2012, MACHINE LEARNING: A PROBABILISTIC PERSPECTIVE, P1
[34]   Simultaneous confidence bands: Theory, implementation, and an application to SVARs [J].
Olea, Jose Luis Montiel ;
Plagborg-Moller, Mikkel .
JOURNAL OF APPLIED ECONOMETRICS, 2019, 34 (01) :1-17
[35]   Squaring the Circle and Cubing the Sphere: Circular and Spherical Copulas [J].
Perlman, Michael D. ;
Wellner, Jon A. .
SYMMETRY-BASEL, 2011, 3 (03) :574-599
[36]  
Rockafellar R. T., 1998, VARIATIONAL ANAL, V317
[37]   Structural Vector Autoregressions: Theory of Identification and Algorithms for Inference [J].
Rubio-Ramirez, Juan F. ;
Waggoner, Daniel F. ;
Zha, Tao .
REVIEW OF ECONOMIC STUDIES, 2010, 77 (02) :665-696
[38]   What are the effects of monetary policy on output? Results from an agnostic identification procedure [J].
Uhlig, H .
JOURNAL OF MONETARY ECONOMICS, 2005, 52 (02) :381-419
[39]   A THEORY OF THE LEARNABLE [J].
VALIANT, LG .
COMMUNICATIONS OF THE ACM, 1984, 27 (11) :1134-1142
[40]  
Vapnik V., 1995, STAT LEARNING THEORY