Sparseout: Controlling Sparsity in Deep Networks

被引:4
|
作者
Khan, Najeeb [1 ]
Stavness, Ian [1 ]
机构
[1] Univ Saskatchewan, Dept Comp Sci, Saskatoon, SK, Canada
来源
ADVANCES IN ARTIFICIAL INTELLIGENCE | 2019年 / 11489卷
基金
加拿大自然科学与工程研究理事会;
关键词
NEURAL-NETWORKS; CONNECTIVITY;
D O I
10.1007/978-3-030-18305-9_24
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dropout is commonly used to help reduce overfitting in deep neural networks. Sparsity is a potentially important property of neural networks, but is not explicitly controlled by Dropout-based regularization. In this work, we propose Sparseout a simple and efficient variant of Dropout that can be used to control the sparsity of the activations in a neural network. We theoretically prove that Sparseout is equivalent to an L-q penalty on the features of a generalized linear model and that Dropout is a special case of Sparseout for neural networks. We empirically demonstrate that Sparseout is computationally inexpensive and is able to control the desired level of sparsity in the activations. We evaluated Sparseout on image classification and language modelling tasks to see the effect of sparsity on these tasks. We found that sparsity of the activations is favorable for language modelling performance while image classification benefits from denser activations. Sparseout provides a way to investigate sparsity in state-of-the-art deep learning models. Source code for Sparseout could be found at https://github.com/najeebkhan/sparseout.
引用
收藏
页码:296 / 307
页数:12
相关论文
共 50 条
  • [1] Maximal Sparsity with Deep Networks?
    Xin, Bo
    Wang, Yizhou
    Gao, Wen
    Wang, Baoyuan
    Wipf, David
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [2] Addressing Sparsity in Deep Neural Networks
    Zhou, Xuda
    Du, Zidong
    Zhang, Shijin
    Zhang, Lei
    Lan, Huiying
    Liu, Shaoli
    Li, Ling
    Guo, Qi
    Chen, Tianshi
    Chen, Yunji
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2019, 38 (10) : 1858 - 1871
  • [3] Local and Global Sparsity for Deep Learning Networks
    Zhang, Long
    Zhao, Jieyu
    Shi, Xiangfu
    Ye, Xulun
    IMAGE AND GRAPHICS (ICIG 2017), PT II, 2017, 10667 : 74 - 85
  • [4] Learning Structured Sparsity in Deep Neural Networks
    Wen, Wei
    Wu, Chunpeng
    Wang, Yandan
    Chen, Yiran
    Li, Hai
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [5] Dynamic sparsity control in Deep Belief Networks
    Keyvanrad, Mohammad Ali
    Homayounpour, Mohammad Mehdi
    INTELLIGENT DATA ANALYSIS, 2017, 21 (04) : 963 - 979
  • [6] Deep belief networks with self-adaptive sparsity
    Qiao, Chen
    Yang, Lan
    Shi, Yan
    Fang, Hanfeng
    Kang, Yanmei
    APPLIED INTELLIGENCE, 2022, 52 (01) : 237 - 253
  • [7] TRANSFER KNOWLEDGE FOR HIGH SPARSITY IN DEEP NEURAL NETWORKS
    Liu, Wenran
    Chen, Xiaogang
    Ji, Xiangyang
    2017 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2017), 2017, : 1354 - 1358
  • [8] Combined Group and Exclusive Sparsity for Deep Neural Networks
    Yoon, Jaehong
    Hwang, Sung Ju
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [9] Deep belief networks with self-adaptive sparsity
    Chen Qiao
    Lan Yang
    Yan Shi
    Hanfeng Fang
    Yanmei Kang
    Applied Intelligence, 2022, 52 : 237 - 253
  • [10] Sparsity-Aware Caches to Accelerate Deep Neural Networks
    Ganesan, Vinod
    Sen, Sanchari
    Kumar, Pratyush
    Gala, Neel
    Veezhinathan, Kamakoti
    Raghunathan, Anand
    PROCEEDINGS OF THE 2020 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2020), 2020, : 85 - 90