Say No to Redundant Information: Unsupervised Redundant Feature Elimination for Active Learning

被引:4
作者
Yang, Jiachen [1 ]
Ma, Shukun [1 ]
Zhang, Zhuo [1 ]
Li, Yang [1 ]
Xiao, Shuai [1 ]
Wen, Jiabao [1 ]
Lu, Wen [2 ]
Gao, Xinbo [3 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
[2] Xidian Univ, Sch Elect Engn, Xian 710126, Peoples R China
[3] Xidian Univ, Sch Elect Engn, State Key Lab Integrated Serv Networks, Xian 710126, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Data models; Labeling; Costs; Training; Redundancy; Computational modeling; Active learning; data issues; deep learning; information redundancy; label noise;
D O I
10.1109/TMM.2024.3371192
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The usual active learning is to sample unlabeled set by designing efficient sample information evaluation algorithms. However, information redundancy between candidate sets is often overlooked. This can cause similar data to be labeled repeatedly, producing ineffective gains for the model. In this paper, we proposed an Unsupervised Redundant Feature Elimination Active Learning module (URFEAL), which utilizes the information feature coincidence of the unlabeled set to eliminate information redundant data, thus guaranteeing the validity of each candidate data. URFEAL consists of feature clusterer and eliminator. The feature clusterer computes class boundaries based on feature densities to discretize each class of the candidate set, and the eliminator judges data similarity by overlapping degree to eliminate redundant data features. Furthermore, we propose an anti-noise sampling strategy Outlier Feature Elimination (OFE) in URFEAL to filter mislabeled sets for relabeling in the data sampling stage. We extensively evaluate our method by image classification and perform experimental validation on CIFAR-10, CIFAR-100 and CALTECH-101. The experimental results show that the improvements we make are especially significant for most existing active learning algorithms in the low data stage, which demonstrates the effectiveness and generality of URFEAL.
引用
收藏
页码:7721 / 7733
页数:13
相关论文
共 66 条
[1]   Contextual Diversity for Active Learning [J].
Agarwal, Sharat ;
Arora, Himanshu ;
Anand, Saket ;
Arora, Chetan .
COMPUTER VISION - ECCV 2020, PT XVI, 2020, 12361 :137-153
[2]   Active Learning for Deep Detection Neural Networks [J].
Aghdam, Hamed H. ;
Gonzalez-Garcia, Abel ;
van de Weijer, Joost ;
Lopez, Antonio M. .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3671-3679
[3]   Spatial Prior Fuzziness Pool-Based Interactive Classification of Hyperspectral Images [J].
Ahmad, Muhammad ;
Khan, Asad ;
Khan, Adil Mehmood ;
Mazzara, Manuel ;
Distefano, Salvatore ;
Sohaib, Ahmed ;
Nibouche, Omar .
REMOTE SENSING, 2019, 11 (09)
[4]  
Angluin D., 1988, Machine Learning, V2, P319, DOI 10.1023/A:1022821128753
[5]   The power of ensembles for active learning in image classification [J].
Beluch, William H. ;
Genewein, Tim ;
Nuernberger, Andreas ;
Koehler, Jan M. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9368-9377
[6]  
Bilgic M., 2009, P NIPS WORKSH AN NET, P1
[7]   Your "Flamingo" is My "Bird": Fine-Grained, or Not [J].
Chang, Dongliang ;
Pang, Kaiyue ;
Zheng, Yixiao ;
Ma, Zhanyu ;
Song, Yi-Zhe ;
Guo, Jun .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :11471-11480
[8]   Active learning with statistical models [J].
Cohn, DA ;
Ghahramani, Z ;
Jordan, MI .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 1996, 4 :129-145
[9]  
Collins B, 2008, LECT NOTES COMPUT SC, V5302, P86, DOI 10.1007/978-3-540-88682-2_8
[10]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848