Tensor Dropout for Robust Learning

被引:15
|
作者
Kolbeinsson, Arinbjorn [1 ]
Kossaifi, Jean [2 ]
Panagakis, Yannis [3 ]
Bulat, Adrian [4 ]
Anandkumar, Animashree [2 ]
Tzoulaki, Ioanna [5 ,6 ,7 ]
Matthews, Paul M. [5 ,6 ]
机构
[1] Imperial Coll London, Dept Epidemiol & Biostat, 4615 London, London, England
[2] NVIDIA, Santa Clara, CA USA
[3] Natl & Kapodistrian Univ Athens, Dept Informat & Telecommun, Athens 1584, Greece
[4] Samsung AI, Cambridge CB1 2RE, England
[5] Imperial Coll London, 4615 London, London, England
[6] Imperial Coll London, UK Dementia Res Inst, 4615 London, London, England
[7] Univ Ioannina, Med Sch, Ioannina, Greece
基金
英国工程与自然科学研究理事会; 英国医学研究理事会;
关键词
Tensors; Training; Robustness; Magnetic resonance imaging; Perturbation methods; Diseases; Deep learning; randomized tensor regression; robustness; stochastic regularization; tensor dropout; tensor methods; tensor regression; tensor regression layers; NEURAL-NETWORKS; BRAIN;
D O I
10.1109/JSTSP.2021.3064182
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
CNNs achieve high levels of performance by leveraging deep, over-parametrized neural architectures, trained on large datasets. However, they exhibit limited generalization abilities outside their training domain and lack robustness to corruptions such as noise and adversarial attacks. To improve robustness and obtain more computationally and memory efficient models, better inductive biases are needed. To provide such inductive biases, tensor layers have been successfully proposed to leverage multi-linear structure through higher-order computations. In this paper, we propose tensor dropout, a randomization technique that can be applied to tensor factorizations, such as those parametrizing tensor layers. In particular, we study tensor regression layers, parametrized by low-rank weight tensors and augmented with our proposed tensor dropout. We empirically show that our approach improves generalization for image classification on ImageNet and CIFAR-100. We also establish state-of-the-art accuracy for phenotypic trait prediction on the largest available dataset of brain MRI (U.K. Biobank), where multi-linear structure is paramount. In all cases, we demonstrate superior performance and significantly improved robustness, both to noisy inputs and to adversarial attacks. We establish the theoretical validity of our approach and the regularizing effect of tensor dropout by demonstrating the link between randomized tensor regression with tensor dropout and deterministic regularized tensor regression.
引用
收藏
页码:630 / 640
页数:11
相关论文
共 50 条
  • [41] Robust Skin Disease Classification by Distilling Deep Neural Network Ensemble for the Mobile Diagnosis of Herpes Zoster
    Back, Seunghyeok
    Lee, Seongju
    Shin, Sungho
    Yu, Yeonguk
    Yuk, Taekyeong
    Jong, Saepomi
    Ryu, Seungjun
    Lee, Kyoobin
    IEEE ACCESS, 2021, 9 : 20156 - 20169
  • [42] Robust Federated Learning With Noisy Communication
    Ang, Fan
    Chen, Li
    Zhao, Nan
    Chen, Yunfei
    Wang, Weidong
    Yu, F. Richard
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2020, 68 (06) : 3452 - 3464
  • [43] Learning to Reconstruct CT Images From the VVBP-Tensor
    Tao, Xi
    Wang, Yongbo
    Lin, Liyan
    Hong, Zixuan
    Ma, Jianhua
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2021, 40 (11) : 3030 - 3041
  • [44] A Comprehensive Analysis of Dropout Assisted Regularized Deep Learning Architectures for Dynamic Electricity Price Forecasting
    Joshi, Pooja
    Stordal, Stale
    Lien, Gudbrand
    Mishra, Deepti
    Haugom, Erik
    IEEE ACCESS, 2024, 12 : 177327 - 177341
  • [45] Scaling Learning-based Policy Optimization for Temporal Logic Tasks by Controller Network Dropout
    Hashemi, Navid
    Hoxha, Bardh
    Prokhorov, Danil
    Fainekos, Georgios
    Deshmukh, Jyotirmoy V.
    ACM TRANSACTIONS ON CYBER-PHYSICAL SYSTEMS, 2024, 8 (04)
  • [46] Checkerboard Dropout: A Structured Dropout With Checkerboard Pattern for Convolutional Neural Networks
    Nguyen, Khanh-Binh
    Choi, Jaehyuk
    Yang, Joon-Sung
    IEEE ACCESS, 2022, 10 : 76044 - 76054
  • [47] ROFL: RObust privacy preserving Federated Learning
    Chattopadhyay, Nandish
    Singh, Arpit
    Chattopadhyay, Anupam
    2022 IEEE 42ND INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS WORKSHOPS (ICDCSW), 2022, : 125 - 132
  • [48] Learning to Diversify for Robust Video Moment Retrieval
    Ge, Huilin
    Liu, Xiaolei
    Guo, Zihang
    Qiu, Zhiwen
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) : 2894 - 2904
  • [49] Laplacian Welsch Regularization for Robust Semisupervised Learning
    Ke, Jingchen
    Gong, Chen
    Liu, Tongliang
    Zhao, Lin
    Yang, Jian
    Tao, Dacheng
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (01) : 164 - 177
  • [50] Robust Deep Learning for IC Test Problems
    Chowdhury, Animesh Basak
    Tan, Benjamin
    Garg, Siddharth
    Karri, Ramesh
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (01) : 183 - 195