3D Semantic Novelty Detection via Large-Scale Pre-Trained Models

被引:0
|
作者
Rabino, Paolo [1 ,2 ]
Alliegro, Antonio [1 ]
Tommasi, Tatiana [1 ]
机构
[1] Polytech Univ Turin, Dept Control & Comp Engn, I-10129 Turin, Italy
[2] Italian Inst Technol, I-16163 Genoa, Italy
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Three-dimensional displays; Semantics; Feature extraction; Solid modeling; Point cloud compression; Anomaly detection; Data models; 3D point clouds; semantic novelty detection; out-of-distribution detection; training-free;
D O I
10.1109/ACCESS.2024.3464334
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Shifting deep learning models from lab environments to real-world settings entails preparing them to handle unforeseen conditions, including the chance of encountering novel objects from classes that were not included in their training data. Such occurrences can pose serious threats in various applications. The task of Semantic Novelty detection has attracted significant attention in the last years mainly on 2D images, overlooking the complex 3D nature of the real-world. In this study, we address this gap by examining the geometric structures of objects within 3D point clouds to detect semantic novelty effectively. We advance the field by introducing 3D-SeND, a method that harnesses a large-scale pre-trained model to extract patch-based object representations directly from its intermediate feature representation. These patches are used to characterize each known class precisely. At inference, a normalcy score is obtained by assessing whether a test sample can be reconstructed predominantly from patches of a single known class or from multiple classes. We evaluate 3D-SeND on real-world point cloud samples when the reference known data are synthetic and demonstrate that it excels in both standard and few-shot scenarios. Thanks to its patch-based object representation, it is possible to visualize 3D-SeND's predictions with a valuable explanation of the decision process. Moreover, the inherent training-free nature of 3D-SeND allows for its immediate application to a wide array of real-world tasks, offering a compelling advantage over approaches that require a task-specific learning phase. Our code is available at https://paolotron.github.io/3DSend.github.io.
引用
收藏
页码:135352 / 135361
页数:10
相关论文
共 50 条
  • [41] Text Detoxification using Large Pre-trained Neural Models
    Dale, David
    Voronov, Anton
    Dementieva, Daryna
    Logacheva, Varvara
    Kozlova, Olga
    Semenov, Nikita
    Panchenko, Alexander
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 7979 - 7996
  • [42] Probing Toxic Content in Large Pre-Trained Language Models
    Ousidhoum, Nedjma
    Zhao, Xinran
    Fang, Tianqing
    Song, Yangqiu
    Yeung, Dit-Yan
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4262 - 4274
  • [43] 3D-VisTA: Pre-trained Transformer for 3D Vision and Text Alignment
    Zhu, Ziyu
    Ma, Xiaojian
    Chen, Yixin
    Deng, Zhidong
    Huang, Siyuan
    Li, Qing
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 2899 - 2909
  • [44] Compressing Pre-trained Models of Code into 3 MB
    Shi, Jieke
    Yang, Zhou
    Xu, Bowen
    Kang, Hong Jin
    Lo, David
    PROCEEDINGS OF THE 37TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE 2022, 2022,
  • [45] Parallel Corpus Filtering via Pre-trained Language Models
    DiDi Labs
    arXiv, 2020,
  • [46] Compression of Generative Pre-trained Language Models via Quantization
    Tao, Chaofan
    Hou, Lu
    Zhang, Wei
    Shang, Lifeng
    Jiang, Xin
    Liu, Qun
    Luo, Ping
    Wong, Ngai
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4821 - 4836
  • [47] Multilingual Translation via Grafting Pre-trained Language Models
    Sun, Zewei
    Wang, Mingxuan
    Li, Lei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2735 - 2747
  • [48] ScanComplete: Large-Scale Scene Completion and Semantic Segmentation for 3D Scans
    Dai, Angela
    Ritchie, Daniel
    Bokeloh, Martin
    Reed, Scott
    Sturm, Juergen
    Niessner, Matthias
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 4578 - 4587
  • [49] Performance Evaluation of Pre-trained Models in Sarcasm Detection Task
    Wang, Haiyang
    Song, Xin
    Zhou, Bin
    Wang, Ye
    Gao, Liqun
    Jia, Yan
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2021, PT II, 2021, 13081 : 67 - 75
  • [50] Adapting Pre-trained Language Models to Rumor Detection on Twitter
    Slimi, Hamda
    Bounhas, Ibrahim
    Slimani, Yahya
    JOURNAL OF UNIVERSAL COMPUTER SCIENCE, 2021, 27 (10) : 1128 - 1148