Deep Convolution Neural Network sharing for the multi-label images classification

被引:26
作者
Coulibaly, Solemane [1 ,2 ]
Kamsu-Foguem, Bernard [1 ]
Kamissoko, Dantouma [2 ]
Traore, Daouda [2 ]
机构
[1] Toulouse Univ, Ecole Natl Ingn Tarbes ENIT, Lab Genie Prod Ecole, 47 Ave Azereix,BP 1629,65016, F-65000 Tarbes, France
[2] Univ Segou, BP 24 Segou, Segou, Mali
来源
MACHINE LEARNING WITH APPLICATIONS | 2022年 / 10卷
关键词
Images multi-label classification; Multi-branch neural network; Deep convolutional neural network; Multitask learning; Transfer learning; CLASSIFIERS;
D O I
10.1016/j.mlwa.2022.100422
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Addressing issues related to multi-label classification is relevant in many fields of applications. In this work. We present a multi-label classification architecture based on Multi-Branch Neural Network Model (MBNN) that permits the network to encode data from multiple semi-parallel subnetworks or layers outputs separately. Different types of neural networks can be used in the MBNN, but the proposal is made with Convolutional Neural Networks subnetworks, trained, and joined in classifying the outputs (i.e., labels). The proposed work makes it possible to perform incremental changes on existing Multitask Learning architectures for an adaptation to the multi-label classification. These transformations lead us to define two new architectures (neural network multi-outputs and neural network multi-features) using the feature extractors from the pre-trained neural networks. The empirical and statistical results verify that the proposed multibranch neural network architecture performs better than other simple multi-label classification architectures. Later, the "network with multifeatures"obtained the highest classification score than other deep neural networks with 83.31% of the f1-score for the Amazon rainforest dataset. The f1-score values are 88.81% for Pascal VOC 2007 dataset, 87.71% for Nuswide, and 88.64% for Pascal VOC 2012.
引用
收藏
页数:17
相关论文
共 112 条
[1]  
Abdullahi HS, 2017, 2017 SEVENTH INTERNATIONAL CONFERENCE ON INNOVATIVE COMPUTING TECHNOLOGY (INTECH 2017), P155, DOI 10.1109/INTECH.2017.8102436
[2]   On automated source selection for transfer learning in convolutional neural networks [J].
Afridi, Muhammad Jamal ;
Ross, Arun ;
Shapiro, Erik M. .
PATTERN RECOGNITION, 2018, 73 :65-75
[3]  
Ahmadibeni Ali, 2020, Applications of machine learning 2020, V30
[4]   A Deep Multi-task Contextual Attention Framework for Multi-modal Affect Analysis [J].
Akhtar, Md Shad ;
Chauhan, Dushyant Singh ;
Ekbal, Asif .
ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2020, 14 (03)
[5]   Recognition of COVID-19 disease from X-ray images by hybrid model consisting of 2D curvelet transform, chaotic salp swarm algorithm and deep learning technique [J].
Altan, Aytac ;
Karasu, Seckin .
CHAOS SOLITONS & FRACTALS, 2020, 140
[6]  
Alvarez-Melis D, 2020, Arxiv, DOI arXiv:2002.02923
[7]  
[Anonymous], 2017, Planet: Understanding the amazon from space
[8]  
[Anonymous], 2019, Improving image classification robustness through selective CNN-filters fine-tuning
[9]   Multimodal Machine Learning: A Survey and Taxonomy [J].
Baltrusaitis, Tadas ;
Ahuja, Chaitanya ;
Morency, Louis-Philippe .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (02) :423-443
[10]  
Bascol K, 2019, IEEE IMAGE PROC, P3043, DOI [10.1109/ICIP.2019.8803325, 10.1109/icip.2019.8803325]