Memory-efficient Transformer-based network model for Traveling Salesman Problem

被引:18
|
作者
Yang, Hua [1 ]
Zhao, Minghao [1 ]
Yuan, Lei [2 ]
Yu, Yang [2 ]
Li, Zhenhua [1 ]
Gu, Ming [1 ]
机构
[1] Tsinghua Univ, Sch Software, Beijing, Peoples R China
[2] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing, Peoples R China
关键词
Combinatorial optimization; Deep reinforcement learning; TSP; Transformer; COMBINATORIAL OPTIMIZATION; ALGORITHMS;
D O I
10.1016/j.neunet.2023.02.014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Combinatorial optimization problems such as Traveling Salesman Problem (TSP) have a wide range of real-world applications in transportation, logistics, manufacturing. It has always been a difficult problem to solve large-scale TSP problems quickly because of memory usage limitations. Recent research shows that the Transformer model is a promising approach. However, the Transformer has several severe problems that prevent it from quickly solving TSP combinatorial optimization problems, such as quadratic time complexity, especially quadratic space complexity, and the inherent limitations of the encoder and decoder itself. To address these issues, we developed a memory-efficient Transformer-based network model for TSP combinatorial optimization problems, termed Tspformer, with two distinctive characteristics: (1) a sampled scaled dot-product attention mechanism with O(L log(L)) (L is the length of input sequences) time and space complexity, which is the most different between our work and other works. (2) due to the reduced space complexity, GPU/CPU memory usage is significantly reduced. Extensive experiments demonstrate that Tspformer significantly outperforms existing methods and provides a new solution to the TSP combinatorial optimization problems. Our Pytorch code will be publicly available on GitHub https://github.com/yhnju/tspFormer.(c) 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:589 / 597
页数:9
相关论文
共 50 条
  • [41] Transformer-based Point Cloud Generation Network
    Xu, Rui
    Hui, Le
    Han, Yuehui
    Qian, Jianjun
    Xie, Jin
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4169 - 4177
  • [42] TNPC: Transformer-based network for cloud classification☆
    Zhou, Wei
    Zhao, Yiheng
    Xiao, Yi
    Min, Xuanlin
    Yi, Jun
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 239
  • [43] Privacy Protection in Transformer-based Neural Network
    Lang, Jiaqi
    Li, Linjing
    Chen, Weiyun
    Zeng, Daniel
    2019 IEEE INTERNATIONAL CONFERENCE ON INTELLIGENCE AND SECURITY INFORMATICS (ISI), 2019, : 182 - 184
  • [44] Improved Biogeography-Based Optimization for the Traveling Salesman Problem
    Wu, Jinping
    Feng, Siling
    2017 2ND IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND APPLICATIONS (ICCIA), 2017, : 166 - 171
  • [45] Traveling salesman problem based on improved ant colony algorithm
    Zhang Hui
    Wang Xi-huai
    Xiao Jian-mei
    Proceedings of the 2007 Chinese Control and Decision Conference, 2007, : 492 - +
  • [46] Generalizing Graph Network Models for the Traveling Salesman Problem with Lin-Kernighan-Helsgaun Heuristics
    Li, Mingfei
    Tu, Shikui
    Xu, Lei
    NEURAL INFORMATION PROCESSING, ICONIP 2023, PT I, 2024, 14447 : 528 - 539
  • [47] A Novel Algorithm for Solving the Prize Collecting Traveling Salesman Problem Based on DNA Computing
    Wang, Zhao-Cai
    Liang, Kun
    Bao, Xiao-Guang
    Wu, Tun-Hua
    IEEE TRANSACTIONS ON NANOBIOSCIENCE, 2024, 23 (02) : 220 - 232
  • [48] Population-Based Incremental Learning Algorithm for a Serial Colored Traveling Salesman Problem
    Meng, Xianghu
    Li, Jun
    Zhou, MengChu
    Dai, Xianzhong
    Dou, Jianping
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2018, 48 (02): : 277 - 288
  • [49] A model induced max-min ant colony optimization for asymmetric traveling salesman problem
    Bai, Jie
    Yang, Gen-Ke
    Chen, Yu-Wang
    Hu, Li-Sheng
    Pan, Chang-Chun
    APPLIED SOFT COMPUTING, 2013, 13 (03) : 1365 - 1375
  • [50] Bangla-BERT: Transformer-Based Efficient Model for Transfer Learning and Language Understanding
    Kowsher, M.
    Sami, Abdullah A. S.
    Prottasha, Nusrat Jahan
    Arefin, Mohammad Shamsul
    Dhar, Pranab Kumar
    Koshiba, Takeshi
    IEEE ACCESS, 2022, 10 : 91855 - 91870