In-context learning enables multimodal large language models to classify cancer pathology images

被引:4
|
作者
Ferber, Dyke [1 ,2 ,3 ]
Woelflein, Georg [4 ]
Wiest, Isabella C. [3 ,5 ]
Ligero, Marta [3 ]
Sainath, Srividhya [3 ]
Ghaffari Laleh, Narmin [3 ]
El Nahhas, Omar S. M. [3 ]
Mueller-Franzes, Gustav [6 ]
Jaeger, Dirk [1 ,2 ]
Truhn, Daniel [6 ]
Kather, Jakob Nikolas [1 ,2 ,3 ,7 ]
机构
[1] Heidelberg Univ Hosp, Natl Ctr Tumor Dis NCT, Heidelberg, Germany
[2] Heidelberg Univ Hosp, Dept Med Oncol, Heidelberg, Germany
[3] Tech Univ Dresden, Else Kroener Fresenius Ctr Digital Hlth, Dresden, Germany
[4] Univ St Andrews, Sch Comp Sci, St Andrews, Scotland
[5] Heidelberg Univ, Med Fac Mannheim, Dept Med 2, Mannheim, Germany
[6] Univ Hosp Aachen, Dept Diagnost & Intervent Radiol, Aachen, Germany
[7] Univ Hosp Dresden, Dept Med 1, Dresden, Germany
关键词
D O I
10.1038/s41467-024-51465-9
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Medical image classification requires labeled, task-specific datasets which are used to train deep learning networks de novo, or to fine-tune foundation models. However, this process is computationally and technically demanding. In language processing, in-context learning provides an alternative, where models learn from within prompts, bypassing the need for parameter updates. Yet, in-context learning remains underexplored in medical image analysis. Here, we systematically evaluate the model Generative Pretrained Transformer 4 with Vision capabilities (GPT-4V) on cancer image processing with in-context learning on three cancer histopathology tasks of high importance: Classification of tissue subtypes in colorectal cancer, colon polyp subtyping and breast tumor detection in lymph node sections. Our results show that in-context learning is sufficient to match or even outperform specialized neural networks trained for particular tasks, while only requiring a minimal number of samples. In summary, this study demonstrates that large vision language models trained on non-domain specific data can be applied out-of-the box to solve medical image-processing tasks in histopathology. This democratizes access of generalist AI models to medical experts without technical background especially for areas where annotated data is scarce. Medical image classification remains a challenging process in deep learning. Here, the authors evaluate a large vision language foundation model (GPT-4V) with in-context learning for cancer image processing and show that such models can learn from examples and reach performance similar to specialized neural networks while reducing the gap to current state-of-the art pathology foundation models.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Active Learning Principles for In-Context Learning with Large Language Models
    Margatina, Katerina
    Schick, Timo
    Aletras, Nikolaos
    Dwivedi-Yu, Jane
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 5011 - 5034
  • [2] Learning to Retrieve In-Context Examples for Large Language Models
    Wang, Liang
    Yang, Nan
    Wei, Furu
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 1752 - 1767
  • [3] Adaptive In-Context Learning with Large Language Models for Bundle
    Sun, Zhu
    Feng, Kaidong
    Yang, Jie
    Qu, Xinghua
    Fang, Hui
    Ong, Yew-Soon
    Liu, Wenyuan
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 966 - 976
  • [4] Adapting Large Multimodal Models to Distribution Shifts: The Role of In-Context Learning
    University of New South Wales, Australia
    不详
    不详
    不详
    不详
    不详
    不详
    arXiv,
  • [5] Visual In-Context Learning for Large Vision-Language Models
    Zhou, Yucheng
    Le, Xiang
    Wang, Qianning
    Shen, Jianbing
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 15890 - 15902
  • [6] Are Emergent Abilities in Large Language Models just In-Context Learning?
    Lu, Sheng
    Bigoulaeva, Irina
    Sachdeva, Rachneet
    Madabushi, Harish Tayyar
    Gurevych, Iryna
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 5098 - 5139
  • [7] Steering Large Language Models for Machine Translation with Finetuning and In-Context Learning
    Alves, Duarte M.
    Guerreirol, Nuno M.
    Alves, Joao
    Pombal, Jose
    Rei, Ricardo
    de Souza, Jose G. C.
    Colombo, Pierre
    Martins, Andre F. T.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 11127 - 11148
  • [8] Meta In-Context Learning: Harnessing Large Language Models for Electrical Data Classification
    Zhou, Mi
    Li, Fusheng
    Zhang, Fan
    Zheng, Junhao
    Ma, Qianli
    ENERGIES, 2023, 16 (18)
  • [9] Large Language Models Can be Lazy Learners: Analyze Shortcuts in In-Context Learning
    Tang, Ruixiang
    Kong, Dehan
    Huang, Longtao
    Xue, Hui
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 4645 - 4657
  • [10] Symbol tuning improves in-context learning in language models
    Wei, Jerry
    Hou, Le
    Lampinen, Andrew
    Chen, Xiangning
    Huang, Da
    Tay, Yi
    Chen, Xinyun
    Lu, Yifeng
    Zhou, Denny
    Ma, Tengyu
    Le, Quoc V.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 968 - 979