Efficient feature selection for pre-trained vision transformers

被引:2
作者
Huang, Lan [1 ,2 ]
Zeng, Jia [1 ]
Yu, Mengqiang [1 ]
Ding, Weiping [3 ]
Bai, Xingyu [1 ]
Wang, Kangping [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Peoples R China
[2] Jilin Univ, Key Lab Symbol Computat & Knowledge Engn, Minist Educ, Changchun 130012, Peoples R China
[3] Nantong Univ, Sch Informat Sci & Technol, Nantong 226019, Peoples R China
关键词
Feature selection; Vision transformer; Model pruning;
D O I
10.1016/j.cviu.2025.104326
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Handcrafted layer-wise vision transformers have demonstrated remarkable performance in image classification. However, their high computational cost limits their practical applications. In this paper, we first identify and highlight the data-independent feature redundancy in pre-trained Vision Transformer (ViT) models. Based on this observation, we explore the feasibility of searching for the best substructure within the original pre-trained model. To this end, we propose EffiSelecViT, a novel pruning method aimed at reducing the computational cost of ViTs while preserving their accuracy. EffiSelecViT introduces importance scores for both self-attention heads and Multi-Layer Perceptron (MLP) neurons in pre-trained ViT models. L1 regularization is applied to constrain and learn these scores. In this simple way, components that are crucial for model performance are assigned higher scores, while those with lower scores are identified as less important and subsequently pruned. Experimental results demonstrate that EffiSelecViT can prune DeiT-B to retain only 64% of FLOPs while maintaining accuracy. This efficiency-accuracy trade-off is consistent across various ViT architectures. Furthermore, qualitative analysis reveals enhanced information expression in the pruned models, affirming the effectiveness and practicality of EffiSelecViT. The code is available at https://github.com/ZJ6789/EffiSelecViT.
引用
收藏
页数:10
相关论文
共 50 条
[31]   Hybrid-Recursive Feature Elimination for Efficient Feature Selection [J].
Jeon, Hyelynn ;
Oh, Sejong .
APPLIED SCIENCES-BASEL, 2020, 10 (09)
[32]   HeatViT: Hardware-Efficient Adaptive Token Pruning for Vision Transformers [J].
Dong, Peiyan ;
Sun, Mengshu ;
Lu, Alec ;
Xie, Yanyue ;
Liu, Kenneth ;
Kong, Zhenglun ;
Meng, Xin ;
Li, Zhengang ;
Lin, Xue ;
Fang, Zhenman ;
Wang, Yanzhi .
2023 IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE, HPCA, 2023, :442-455
[33]   Voting from Nearest Tasks: Meta-Vote Pruning of Pre-trained Models for Downstream Tasks [J].
Zhao, Haiyan ;
Zhou, Tianyi ;
Long, Guodong ;
Jiang, Jing ;
Zhang, Chengqi .
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II, 2023, 14170 :52-68
[34]   Feature fusion Vision Transformers using MLP-Mixer for enhanced deepfake detection [J].
Essa, Ehab .
NEUROCOMPUTING, 2024, 598
[35]   Vision Transformers (ViTs) for Feature Extraction and Classification of AI-Generated Visual Designs [J].
Yun, Qing .
IEEE ACCESS, 2025, 13 :69459-69477
[36]   An efficient feature selection algorithm for hybrid data [J].
Wang, Feng ;
Liang, Jiye .
NEUROCOMPUTING, 2016, 193 :33-41
[37]   Efficient feature selection techniques for sentiment analysis [J].
Avinash Madasu ;
Sivasankar Elango .
Multimedia Tools and Applications, 2020, 79 :6313-6335
[38]   Efficient Method for Feature Selection in Text Classification [J].
Sun, Jian ;
Zhang, Xiang ;
Liao, Dan ;
Chang, Victor .
2017 INTERNATIONAL CONFERENCE ON ENGINEERING AND TECHNOLOGY (ICET), 2017,
[39]   FEATURE SUBSET SELECTION FOR EFFICIENT ADABOOST TRAINING [J].
Sun, Chensheng ;
Hu, Jiwei ;
Lam, Kin-Man .
2011 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2011,
[40]   Efficient Feature Selection for Intrusion Detection Systems [J].
Ahmadi, S. Sareh ;
Rashad, Sherif ;
Elgazzar, Heba .
2019 IEEE 10TH ANNUAL UBIQUITOUS COMPUTING, ELECTRONICS & MOBILE COMMUNICATION CONFERENCE (UEMCON), 2019, :1029-1034