Automated Individual Cattle Identification Using Video Data: A Unified Deep Learning Architecture Approach

被引:9
作者
Qiao, Yongliang [1 ]
Clark, Cameron [2 ]
Lomax, Sabrina [2 ]
Kong, He [1 ]
Su, Daobilige [3 ]
Sukkarieh, Salah [1 ]
机构
[1] Univ Sydney, Fac Engn, Australian Ctr Field Robot, Sydney, NSW, Australia
[2] Univ Sydney, Fac Sci, Sch Life & Environm Sci, Livestock Prod & Welf Grp, Sydney, NSW, Australia
[3] China Agr Univ, Coll Engn, Beijing, Peoples R China
来源
FRONTIERS IN ANIMAL SCIENCE | 2021年 / 2卷
关键词
cattle identification; deep learning; BiLSTM; self-attention; precision livestock farming; FULLY CONVOLUTIONAL NETWORKS; ACTION RECOGNITION; ATTENTION; LSTM; MODEL; COWS;
D O I
10.3389/fanim.2021.759147
中图分类号
S8 [畜牧、 动物医学、狩猎、蚕、蜂];
学科分类号
0905 ;
摘要
Individual cattle identification is a prerequisite and foundation for precision livestock farming. Existing methods for cattle identification require radio frequency or visual ear tags, all of which are prone to loss or damage. Here, we propose and implement a new unified deep learning approach to cattle identification using video analysis. The proposed deep learning framework is composed of a Convolutional Neural Network (CNN) and Bidirectional Long Short-Term Memory (BiLSTM) with a self-attention mechanism. More specifically, the Inception-V3 CNN was used to extract features from a cattle video dataset taken in a feedlot with rear-view. Extracted features were then fed to a BiLSTM layer to capture spatio-temporal information. Then, self-attention was employed to provide a different focus on the features captured by BiLSTM for the final step of cattle identification. We used a total of 363 rear-view videos from 50 cattle at three different times with an interval of 1 month between data collection periods. The proposed method achieved 93.3% identification accuracy using a 30-frame video length, which outperformed current state-of-the-art methods (Inception-V3, MLP, SimpleRNN, LSTM, and BiLSTM). Furthermore, two different attention schemes, namely, additive and multiplicative attention mechanisms were compared. Our results show that the additive attention mechanism achieved 93.3% accuracy and 91.0% recall, greater than multiplicative attention mechanism with 90.7% accuracy and 87.0% recall. Video length also impacted accuracy, with video sequence length up to 30-frames enhancing identification performance. Overall, our approach can capture key spatio-temporal features to improve cattle identification accuracy, enabling automated cattle identification for precision livestock farming.
引用
收藏
页数:14
相关论文
共 59 条
[1]   Visual Localisation and Individual Identification of Holstein Friesian Cattle via Deep Learning [J].
Andrew, William ;
Greatwood, Colin ;
Burghardt, Tilo .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2017), 2017, :2850-2859
[2]  
Andrew W, 2016, IEEE IMAGE PROC, P484, DOI 10.1109/ICIP.2016.7532404
[3]   The role of interoperable data standards in precision livestock farming in extensive livestock systems: A review [J].
Bahlo, Christiane ;
Dahlhaus, Peter ;
Thompson, Helen ;
Trotter, Mark .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2019, 156 :459-466
[4]   Precision livestock farming technologies for welfare management in intensive livestock systems [J].
Berckmans, D. .
REVUE SCIENTIFIQUE ET TECHNIQUE-OFFICE INTERNATIONAL DES EPIZOOTIES, 2014, 33 (01) :189-196
[5]   Computer vision system for measuring individual cow feed intake using RGB-D camera and deep learning algorithms [J].
Bezen, Ran ;
Edan, Yael ;
Halachmi, Ilan .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2020, 172
[6]  
Cai C, 2013, ASIAPAC SIGN INFO PR
[7]  
Chollet F, 2015, Keras: The Python Deep Learning library
[8]   Does Smart Farming Improve or Damage Animal Welfare? Technology and What Animals Want [J].
Dawkins, Marian Stamp .
FRONTIERS IN ANIMAL SCIENCE, 2021, 2
[9]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[10]  
Deng JF, 2020, Arxiv, DOI [arXiv:2002.00735, DOI 10.48550/ARXIV.2002.00735]