A transfer learning-based network model integrating kernel convolution with graph attention mechanism for point cloud segmentation of livestock

被引:3
作者
Shi, Yuanyuan [1 ]
Wang, Yuxiao [1 ]
Ling, Yin [1 ,2 ,3 ]
Wu, Zhenfang [2 ,3 ]
Lin, Junyong [1 ]
Tian, Xuhong [1 ]
Qiong, Huang [1 ]
Zhang, Sumin [1 ,3 ]
Li, Zhiying [4 ]
机构
[1] South China Agr Univ, Coll Math & Informat, Guangzhou 510642, Peoples R China
[2] Natl Engn Res Ctr Swine Breeding Ind, Guangzhou 510642, Peoples R China
[3] State Key Lab Swine & Poultry Breeding Ind, Guangzhou 510642, Peoples R China
[4] South China Agr Univ, Coll Foreign Studies, Guangzhou 510642, Peoples R China
基金
中国国家自然科学基金;
关键词
Point cloud segmentation; Kernel convolution; Graph attention mechanism; Transfer learning;
D O I
10.1016/j.compag.2024.109325
中图分类号
S [农业科学];
学科分类号
09 ;
摘要
Non-contact body size measurement has become a hot research topic in intelligent livestock farming. In regard to body size measurement of large livestock, such as cattle, collecting a substantial number of point clouds is frequently involved. The direct calculation of all point clouds for body size measurement can be impacted as point clouds of different body parts may interfere with each other, which poses huge challenges for the positioning of key points and induces inaccurate positioning, resulting in measurement errors. The accuracy of body size measurement can be improved by segmenting point clouds of different body parts from each other, key measurement points can be precisely located, thus enhancing the accuracy of body size measurement. In this paper, we propose a network model initially trained for pig point cloud segmentation based on the Kernel Convolution integrated with Graph Attention Mechanism (KCGATNet for short), which, through transfer learning techniques, can also be used to achieve successful segmentation of various cattle point clouds using only 7 training samples. The model utilizes two core modules, Kernel Convolution (KC) and Point-based Graph Attention Mechanism (P-GAT), to extract local neighborhood features of point clouds. When using pig body point clouds as training data, it achieved precise segmentation of the head, ears, limbs, torso, and tail of pigs through a downsampling-upsampling architecture. On the test set of pig point clouds, Overall Accuracy (OA) reached 98.1% and mean Intersection over Union (mIoU) was up to 90.5%. Furthermore, when this model served as a pre-trained model and underwent transfer learning using 7 sets of annotated data of Simmental cattle, it achieved a mIoU of 90.1% on a test set of 93 Simmental cattle, 89.6% on a test set of 439 dairy buffalo, 90.2% on a test set of 103 Hereford cattle, and 90.0% on a test set of 119 Black Angus cattle. The experimental outcomes fully demonstrate the robustness of the proposed livestock point cloud segmentation model, KCGATNet. With transfer learning of a small sample size, it can reliably perform point cloud segmentation on a wide range of different breeds of quadrupedal livestock, saving a significant amount of time spent on manual annotation and improving the efficiency of livestock point cloud segmentation models.
引用
收藏
页数:14
相关论文
共 43 条
[1]   ATTENTION-BASED MULTI-SCALE GRAPH CONVOLUTION FOR POINT CLOUD SEMANTIC SEGMENTATION [J].
Akwensi, Perpetual Hope ;
Wang, Ruisheng .
2022 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2022), 2022, :7515-7518
[2]  
Anvekar T., 2023, P IEEE CVF C COMP VI, P4178
[3]   A Local-Global Feature Fusing Method for Point Clouds Semantic Segmentation [J].
Bi, Yuanwei ;
Zhang, Lujian ;
Liu, Yaowen ;
Huang, Yansen ;
Liu, Hao .
IEEE ACCESS, 2023, 11 :68776-68790
[4]   Pointwise Convolutional Neural Networks [J].
Binh-Son Hua ;
Minh-Khoi Tran ;
Yeung, Sai-Kit .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :984-993
[5]  
Brock A, 2016, Arxiv, DOI arXiv:1608.04236
[6]  
Chen LZ, 2019, Arxiv, DOI arXiv:1905.05442
[7]  
Chen XZ, 2017, Arxiv, DOI [arXiv:1611.07759, 10.48550/ARXIV.1611.07759]
[8]  
Chen YL, 2019, Arxiv, DOI arXiv:1908.02990
[9]  
Choy C, 2019, Arxiv, DOI [arXiv:1904.08755, 10.48550/arXiv.1904.08755]
[10]   ScanNet: Richly-annotated 3D Reconstructions of Indoor Scenes [J].
Dai, Angela ;
Chang, Angel X. ;
Savva, Manolis ;
Halber, Maciej ;
Funkhouser, Thomas ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2432-2443