(Machine) learning parameter regions

被引:2
作者
Olea, Jose Luis Montiel [1 ]
Nesbit, James [2 ]
机构
[1] Columbia Univ, Dept Econ, 420 West 118th St, New York, NY 10027 USA
[2] NYU, Dept Econ, 19 W 4th St,6th Floor, New York, NY 10012 USA
关键词
Machine learning; Supervised learning; Set-identified models; Structural vector autoregressions; UNIFORM CONFIDENCE BANDS; INFERENCE; IDENTIFICATION; SETS;
D O I
10.1016/j.jeconom.2020.06.008
中图分类号
F [经济];
学科分类号
02 ;
摘要
How many random points from an identified set, a confidence set, or a highest posterior density set suffice to describe them? This paper argues that taking random draws from a parameter region in order to approximate its shape is a supervised learning problem (analogous to sampling pixels of an image to recognize it). Misclassification error - a common criterion in machine learning - provides an off-the-shelf tool to assess the quality of a given approximation. We say a parameter region can be learned if there is an algorithm that yields a misclassification error of at most epsilon with probability at least 1 - delta, regardless of the sampling distribution. We show that learning a parameter region is possible if and only if its potential shapes are not too complex. Moreover, the tightest band that contains a d-dimensional parameter region is always learnable from the inside (in a sense we make precise), with at least max {(1 - epsilon) ln (1/delta), (3/16)d}/epsilon draws, but at most min{2d ln(2d/delta), exp(1)(2d+ln(1/delta))}/epsilon. These bounds grow linearly in the dimension of the parameter region, and are uniform with respect to its true shape. We illustrate the usefulness of our results using structural vector autoregressions. We show how many orthogonal matrices are necessary/sufficient to evaluate the impulse responses' identified set and how many 'shotgun plots' to report when conducting joint inference on impulse responses. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页码:716 / 744
页数:29
相关论文
共 41 条
  • [1] Uniform approximation of Vapnik-Chervonenkis classes
    Adams, Terrence M.
    Nobel, Andrew B.
    [J]. BERNOULLI, 2012, 18 (04) : 1310 - 1319
  • [2] Coarse decision making and overfitting
    Al-Najjar, Nabil I.
    Pai, Mallesh M.
    [J]. JOURNAL OF ECONOMIC THEORY, 2014, 150 : 467 - 486
  • [3] DECISION MAKERS AS STATISTICIANS: DIVERSITY, AMBIGUITY, AND LEARNING
    Al-Najjar, Nabil I.
    [J]. ECONOMETRICA, 2009, 77 (05) : 1371 - 1401
  • [4] Amir-Ahmadi P., 2020, QUANTATIVE EC
  • [5] [Anonymous], 2017, DISCRETIZING U UNPUB
  • [6] [Anonymous], 2014, UNDERSTANDING MACHIN
  • [7] Approximating hyper-rectangles: Learning and pseudorandom sets
    Auer, P
    Long, PM
    Srinivasan, A
    [J]. JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 1998, 57 (03) : 376 - 388
  • [8] Bar H., 2013, COMPUTATIONAL METHOD
  • [9] PROGRAM EVALUATION AND CAUSAL INFERENCE WITH HIGH-DIMENSIONAL DATA
    Belloni, A.
    Chernozhukov, V.
    Fernandez-Val, I.
    Hansen, C.
    [J]. ECONOMETRICA, 2017, 85 (01) : 233 - 298
  • [10] Inference on Treatment Effects after Selection among High-Dimensional ControlsaEuro
    Belloni, Alexandre
    Chernozhukov, Victor
    Hansen, Christian
    [J]. REVIEW OF ECONOMIC STUDIES, 2014, 81 (02) : 608 - 650