Challenges and Opportunities in Neuro-Symbolic Composition of Foundation Models

被引:0
作者
Jha, Susmit [1 ]
Roy, Anirban [1 ]
Cobb, Adam [1 ]
Berenbeim, Alexander [2 ]
Bastian, Nathaniel D. [2 ]
机构
[1] SRI Int, Comp Sci Lab, Menlo Pk, CA 94025 USA
[2] US Mil Acad, Army Cyber Inst, West Point, NY USA
来源
MILCOM 2023 - 2023 IEEE MILITARY COMMUNICATIONS CONFERENCE | 2023年
关键词
LLMs; Foundation Models; Neuro-symbolic Learning;
D O I
10.1109/MILCOM58377.2023.10356344
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Trustworthy, resilient, and interpretable artificial intelligence (AI) is essential for effective operation of the Internet of Things (IoT) in adversarial environments. Such a robust and interpretable AI is needed to improve tactical coordination through scalability, corroboration, and context-aware intelligence. It is crucial to have robust machine learning (ML) models with characteristics such as low-supervision adaptability, decision explanations, and adaptive inference. Pre-trained large language models (LLMs) and foundation models (FMs) address some of these challenges, but are unpredictable and cannot directly solve complex tasks in mission-critical scenarios. However, their generalization capabilities make them potential building blocks for high-assurance AI/ML systems that compose multiple FMs and LLMs. In this paper, we propose combining neural foundation models (FMs) using symbolic programs that results in a more effective AI for adversarial conditions. Neuro-symbolic composition of FMs to solve complex tasks requires interactive and unambiguous specification of the intent, task decomposition into subtasks that can be solved by individual FMs, program synthesis for composing FMs, and neuro-symbolic inference that schedules inference of different FMs and combines their results. We give examples of such neuro-symbolic programs using foundation models to solve visual question-answering tasks such as out-of-context detection. This position paper identifies the challenges and opportunities in the neuro-symbolic composition of the large language models and foundation models.
引用
收藏
页数:6
相关论文
共 55 条
  • [1] Acharya Manoj, 2022, IJCAI
  • [2] Alayrac Jean-Baptiste, 2022, ADV NEURAL INFORM PR, V35, P23716
  • [3] MORE IS DIFFERENT - BROKEN SYMMETRY AND NATURE OF HIERARCHICAL STRUCTURE OF SCIENCE
    ANDERSON, PW
    [J]. SCIENCE, 1972, 177 (4047) : 393 - &
  • [4] Bomatter P, 2021, Arxiv, DOI arXiv:2104.02215
  • [5] Bommarito M., 2022, arXiv, DOI [10.48550/arXiv.2212.14402, DOI 10.48550/ARXIV.2212.14402]
  • [6] Bulatov A, 2024, Arxiv, DOI arXiv:2304.11062
  • [7] VisualGPT: Data-efficient Adaptation of Pretrained Language Models for Image Captioning
    Chen, Jun
    Guo, Han
    Yi, Kai
    Li, Boyang
    Elhoseiny, Mohamed
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18009 - 18019
  • [8] Chen Mark, 2021, arXiv, DOI DOI 10.48550/ARXIV.2107.03374
  • [9] Context models and out-of-context objects
    Choi, Myung Jin
    Torralba, Antonio
    Willsky, Alan S.
    [J]. PATTERN RECOGNITION LETTERS, 2012, 33 (07) : 853 - 862
  • [10] Chung HW, 2022, Arxiv, DOI arXiv:2210.11416