On Correctness of Automatic Differentiation for Non-Differentiable Functions

被引:0
|
作者
Lee, Wonyeol [1 ]
Yu, Hangyeol [1 ]
Rival, Xavier [2 ,3 ]
Yang, Hongseok [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Sch Comp Sci, Daejeon, South Korea
[2] INRIA Paris, Dept dInformat ENS, Paris, France
[3] PSL Univ, CNRS, Paris, France
基金
新加坡国家研究基金会; 欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Differentiation lies at the core of many machine-learning algorithms, and is well-supported by popular autodiff systems, such as TensorFlow and PyTorch. Originally, these systems have been developed to compute derivatives of differentiable functions, but in practice, they are commonly applied to functions with non-differentiabilities. For instance, neural networks using ReLU define non-differentiable functions in general, but the gradients of losses involving those functions are computed using autodiff systems in practice. This status quo raises a natural question: are autodiff systems correct in any formal sense when they are applied to such non-differentiable functions? In this paper, we provide a positive answer to this question. Using counterexamples, we first point out flaws in often-used informal arguments, such as: non-differentiabilities arising in deep learning do not cause any issues because they form a measure-zero set. We then investigate a class of functions, called PAP functions, that includes nearly all (possibly non-differentiable) functions in deep learning nowadays. For these PAP functions, we propose a new type of derivatives, called intensional derivatives, and prove that these derivatives always exist and coincide with standard derivatives for almost all inputs. We also show that these intensional derivatives are what most autodiff systems compute or try to compute essentially. In this way, we formally establish the correctness of autodiff systems applied to non-differentiable functions.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] ON WEIERSTRASS NON-DIFFERENTIABLE FUNCTION
    HATA, M
    COMPTES RENDUS DE L ACADEMIE DES SCIENCES SERIE I-MATHEMATIQUE, 1988, 307 (03): : 119 - 123
  • [22] Renormalisation of non-differentiable potentials
    J. Alexandre
    N. Defenu
    G. Grigolia
    I. G. Márián
    D. Mdinaradze
    A. Trombettoni
    Y. Turovtsi-Shiutev
    I. Nándori
    Journal of High Energy Physics, 2022
  • [23] On the local fractional derivative of everywhere non-differentiable continuous functions on intervals
    Liu, Cheng-shi
    COMMUNICATIONS IN NONLINEAR SCIENCE AND NUMERICAL SIMULATION, 2017, 42 : 229 - 235
  • [24] GLOBAL ASYMPTOTIC STABILITY OF NONLINEAR DIFFERENCE - EQUATIONS WITH NON-DIFFERENTIABLE FUNCTIONS
    SZIDAROVSZKY, F
    ECONOMICS LETTERS, 1990, 33 (04) : 333 - 336
  • [25] On the structure of the critical set of non-differentiable functions with a weak compactness condition
    Bonanno, Gabriele
    Marano, Salvatore A.
    APPLICABLE ANALYSIS, 2010, 89 (01) : 1 - 10
  • [26] Non-Differentiable Functions Defined in Terms of Classical Representations of Real Numbers
    Serbenyuk, S. O.
    JOURNAL OF MATHEMATICAL PHYSICS ANALYSIS GEOMETRY, 2018, 14 (02) : 197 - 213
  • [27] Isotonic Modeling with Non-Differentiable Loss Functions with Application to Lasso Regularization
    Painsky, Amichai
    Rosset, Saharon
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (02) : 308 - 321
  • [28] Non-differentiable symmetric duality
    Mond, B
    Schechter, M
    BULLETIN OF THE AUSTRALIAN MATHEMATICAL SOCIETY, 1996, 53 (02) : 177 - 188
  • [29] Renormalisation of non-differentiable potentials
    Alexandre, J.
    Defenu, N.
    Grigolia, G.
    Marian, I. G.
    Mdinaradze, D.
    Trombettoni, A.
    Turovtsi-Shiutev, Y.
    Nandori, I
    JOURNAL OF HIGH ENERGY PHYSICS, 2022, 2022 (07)
  • [30] THERMODYNAMICS OF NON-DIFFERENTIABLE SYSTEMS
    BOYLING, JB
    INTERNATIONAL JOURNAL OF THEORETICAL PHYSICS, 1974, 9 (06) : 379 - 392