3D Semantic Novelty Detection via Large-Scale Pre-Trained Models

被引:0
|
作者
Rabino, Paolo [1 ,2 ]
Alliegro, Antonio [1 ]
Tommasi, Tatiana [1 ]
机构
[1] Polytech Univ Turin, Dept Control & Comp Engn, I-10129 Turin, Italy
[2] Italian Inst Technol, I-16163 Genoa, Italy
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Three-dimensional displays; Semantics; Feature extraction; Solid modeling; Point cloud compression; Anomaly detection; Data models; 3D point clouds; semantic novelty detection; out-of-distribution detection; training-free;
D O I
10.1109/ACCESS.2024.3464334
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Shifting deep learning models from lab environments to real-world settings entails preparing them to handle unforeseen conditions, including the chance of encountering novel objects from classes that were not included in their training data. Such occurrences can pose serious threats in various applications. The task of Semantic Novelty detection has attracted significant attention in the last years mainly on 2D images, overlooking the complex 3D nature of the real-world. In this study, we address this gap by examining the geometric structures of objects within 3D point clouds to detect semantic novelty effectively. We advance the field by introducing 3D-SeND, a method that harnesses a large-scale pre-trained model to extract patch-based object representations directly from its intermediate feature representation. These patches are used to characterize each known class precisely. At inference, a normalcy score is obtained by assessing whether a test sample can be reconstructed predominantly from patches of a single known class or from multiple classes. We evaluate 3D-SeND on real-world point cloud samples when the reference known data are synthetic and demonstrate that it excels in both standard and few-shot scenarios. Thanks to its patch-based object representation, it is possible to visualize 3D-SeND's predictions with a valuable explanation of the decision process. Moreover, the inherent training-free nature of 3D-SeND allows for its immediate application to a wide array of real-world tasks, offering a compelling advantage over approaches that require a task-specific learning phase. Our code is available at https://paolotron.github.io/3DSend.github.io.
引用
收藏
页码:135352 / 135361
页数:10
相关论文
共 50 条
  • [1] Exploring the Application of Large-Scale Pre-Trained Models on Adverse Weather Removal
    Tan, Zhentao
    Wu, Yue
    Liu, Qiankun
    Chu, Qi
    Lu, Le
    Ye, Jieping
    Yu, Nenghai
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 1683 - 1698
  • [2] Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey
    Xiao Wang
    Guangyao Chen
    Guangwu Qian
    Pengcheng Gao
    Xiao-Yong Wei
    Yaowei Wang
    Yonghong Tian
    Wen Gao
    Machine Intelligence Research, 2023, 20 : 447 - 482
  • [3] Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey
    Wang, Xiao
    Chen, Guangyao
    Qian, Guangwu
    Gao, Pengcheng
    Wei, Xiao-Yong
    Wang, Yaowei
    Tian, Yonghong
    Gao, Wen
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (04) : 447 - 482
  • [4] FASTERMOE: Modeling and Optimizing Training of Large-Scale Dynamic Pre-Trained Models
    He, Jiaao
    Zhai, Jidong
    Antunes, Tiago
    Wang, Haojie
    Luo, Fuwen
    Shi, Shangfeng
    Li, Qin
    PPOPP'22: PROCEEDINGS OF THE 27TH ACM SIGPLAN SYMPOSIUM ON PRINCIPLES AND PRACTICE OF PARALLEL PROGRAMMING, 2022, : 120 - 134
  • [5] Alternating Recurrent Dialog Model with Large-scale Pre-trained Language Models
    Wu, Qingyang
    Zhang, Yichi
    Li, Yu
    Yu, Zhou
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1292 - 1301
  • [6] EXPLORING LARGE SCALE PRE-TRAINED MODELS FOR ROBUST MACHINE ANOMALOUS SOUND DETECTION
    Han, Bing
    Lv, Zhiqiang
    Jiang, Anbai
    Huang, Wen
    Chen, Zhengyang
    Deng, Yufeng
    Ding, Jiawei
    Lu, Cheng
    Zhang, Wei-Qiang
    Fan, Pingyi
    Liu, Jia
    Qian, Yanmin
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 1326 - 1330
  • [7] Leveraging Pre-Trained 3D Object Detection Models For Fast Ground Truth Generation
    Lee, Jungwook
    Walsh, Sean
    Harakeh, Ali
    Waslander, Steven L.
    2018 21ST INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2018, : 2504 - 2510
  • [8] Semantic Programming by Example with Pre-trained Models
    Verbruggen, Gust
    Le, Vu
    Gulwani, Sumit
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2021, 5 (OOPSLA):
  • [9] y-Tuning: an efficient tuning paradigm for large-scale pre-trained models via label representation learning
    Liu, Yitao
    An, Chenxin
    Qiu, Xipeng
    FRONTIERS OF COMPUTER SCIENCE, 2024, 18 (04)
  • [10] CPM-2: Large-scale cost-effective pre-trained language models
    Zhang, Zhengyan
    Gu, Yuxian
    Han, Xu
    Chen, Shengqi
    Xiao, Chaojun
    Sun, Zhenbo
    Yao, Yuan
    Qi, Fanchao
    Guan, Jian
    Ke, Pei
    Cai, Yanzheng
    Zeng, Guoyang
    Tan, Zhixing
    Liu, Zhiyuan
    Huang, Minlie
    Han, Wentao
    Liu, Yang
    Zhu, Xiaoyan
    Sun, Maosong
    AI OPEN, 2021, 2 : 216 - 224