Underwater target perception algorithm based on pressure sequence generative adversarial network

被引:2
作者
Zhao, Jiang [1 ,2 ]
Wang, Shushan [1 ,2 ]
Jia, Xiyu [1 ,2 ]
Gao, Yu [3 ]
Zhu, Wei [1 ,2 ]
Ma, Feng [1 ,2 ]
Liu, Qiang [1 ,2 ]
机构
[1] Beijing Inst Technol, Inst Unmanned Underwater Syst, Sch Mechatron Engn, Beijing 100081, Peoples R China
[2] Beijing Inst Technol, State Key Lab Explos Sci & Technol, Beijing 100081, Peoples R China
[3] Harbin Inst Technol Shenzhen, Sch Comp Sci & Technol, Shenzhen 518055, Peoples R China
关键词
Underwater target perception; Deep learning; Transformer; GAN; GRU;
D O I
10.1016/j.oceaneng.2023.115547
中图分类号
U6 [水路运输]; P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
In recent years, underwater target perception algorithms based on deep learning have attracted extensive attention. Deep learning can independently learn to extract features from a large number of labelled data, which improves the robustness of underwater target perception accuracy. However, owing to the high collection cost in the underwater natural environment and the high time cost of simulated calculations, it is often unrealistic to provide a large number of labelled data. To solve this problem, this paper proposes an underwater target perception algorithm based on a generative adversarial network (GAN). This GAN uses the transformer model to augment the samples of a small number of simulated underwater pressure sequences and then establishes a multilayer gated recurrent unit (GRU) network to recognise the azimuth, distance, and velocity of underwater targets. The experimental results show that the method proposed in this paper can effectively realise underwater target perception, and with an accuracy of 97.86%, and the root mean square errors of the target distance, azimuth, and velocity estimations are 0.1244, 0.9828, and 0.8271, respectively.
引用
收藏
页数:16
相关论文
共 46 条
  • [21] Gulrajani I, 2017, ADV NEUR IN, V30
  • [22] Iwana B, 2021, Arxiv, DOI arXiv:2007.15951
  • [23] Kang CY, 2004, LECT NOTES COMPUT SC, V3173, P930
  • [24] Karras Tero, 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Proceedings, P8107, DOI 10.1109/CVPR42600.2020.00813
  • [25] Karras T, 2018, Arxiv, DOI arXiv:1710.10196
  • [26] Lee S.G., 2017, ARXIV
  • [27] Surrogate Rehabilitative Time Series Data for Image-based Deep Learning
    Lee, Tracey K. M. Eileen
    Kuah, Y. L.
    Leo, Kee-Hao
    Sanei, Saeid
    Chew, Effie
    Zhao, Ling
    [J]. 2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2019,
  • [28] [李悦 Li Yue], 2021, [声学学报, Acta Acustica], V46, P1013
  • [29] Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
    Liu, Ze
    Lin, Yutong
    Cao, Yue
    Hu, Han
    Wei, Yixuan
    Zhang, Zheng
    Lin, Stephen
    Guo, Baining
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 9992 - 10002
  • [30] Mirza M, 2014, Arxiv, DOI [arXiv:1411.1784, 10.48550/arXiv.1411.1784]