Prompt Guided Transformer for Multi-Task Dense Prediction

被引:6
|
作者
Lu, Yuxiang [1 ]
Sirejiding, Shalayiding [1 ]
Ding, Yue [1 ]
Wang, Chunlin [2 ]
Lu, Hongtao [1 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai 200240, Peoples R China
[2] Chuxiong Normal Univ, Sch Informat Sci & Technol, Chuxiong 675099, Peoples R China
关键词
Multi-task learning; dense prediction; prompting; vision transformer;
D O I
10.1109/TMM.2024.3349865
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Task-conditional architecture offers advantage in parameter efficiency but falls short in performance compared to state-of-the-art multi-decoder methods. How to trade off performance and model parameters is an important and difficult problem. In this paper, we introduce a simple and lightweight task-conditional model called Prompt Guided Transformer (PGT) to optimize this challenge. Our approach designs a Prompt-conditioned Transformer block, which incorporates task-specific prompts in the self-attention mechanism to achieve global dependency modeling and parameter-efficient feature adaptation across multiple tasks. This block is integrated into both the shared encoder and decoder, enhancing the capture of intra- and inter-task features. Moreover, we design a lightweight decoder to further reduce parameter usage, which accounts for only 2.7% of the total model parameters. Extensive experiments on two multi-task dense prediction benchmarks, PASCAL-Context and NYUD-v2, demonstrate that our approach achieves state-of-the-art results among task-conditional methods while using fewer parameters, and maintains a significant balance between performance and parameter size.
引用
收藏
页码:6375 / 6385
页数:11
相关论文
共 50 条
  • [21] Multi-task learning for pKa prediction
    Grigorios Skolidis
    Katja Hansen
    Guido Sanguinetti
    Matthias Rupp
    Journal of Computer-Aided Molecular Design, 2012, 26 : 883 - 895
  • [22] Adaptive Weight Generator for Multi-Task Image Recognition by Task Grouping Prompt
    Wu, Gaojie
    Zeng, Ling-an
    Meng, Jing-Ke
    Zheng, Wei-Shi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 9906 - 9919
  • [23] Metric-Guided Multi-task Learning
    Ren, Jinfu
    Liu, Yang
    Liu, Jiming
    FOUNDATIONS OF INTELLIGENT SYSTEMS (ISMIS 2020), 2020, 12117 : 21 - 31
  • [24] Bayesian Multi-Task Transfer Learning for Soft Prompt Tuning
    Lee, Haeju
    Jeong, Minchan
    Yun, Se-Young
    Kim, Kee-Eung
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 4942 - 4958
  • [25] Multi-Task Learning with Personalized Transformer for Review Recommendation
    Wang, Haiming
    Liu, Wei
    Yin, Jian
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2021, PT II, 2021, 13081 : 162 - 176
  • [26] TransNuSeg: A Lightweight Multi-task Transformer for Nuclei Segmentation
    He, Zhenqi
    Unberath, Mathias
    Ke, Jing
    Shen, Yiqing
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT IV, 2023, 14223 : 206 - 215
  • [27] DENSETRACKER: A MULTI-TASK DENSE NETWORK FOR VISUAL TRACKING
    Zhao, Fei
    Tang, Ming
    Wu, Yi
    Wang, Jinqiao
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 607 - 612
  • [28] Cross-task Attention Mechanism for Dense Multi-task Learning
    Lopes, Ivan
    Tuan-Hung Vu
    de Charette, Raoul
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 2328 - 2337
  • [29] Dual feature correlation guided multi-task learning for Alzheimer's disease prediction
    Tang, Shanshan
    Cao, Peng
    Huang, Min
    Liu, Xiaoli
    Zaiane, Osmar
    COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 140
  • [30] JOINEDTrans: Prior guided multi-task transformer for joint optic disc/cup segmentation and fovea detection
    He H.
    Qiu J.
    Lin L.
    Cai Z.
    Cheng P.
    Tang X.
    Computers in Biology and Medicine, 2024, 177