MetaQA: Enhancing human-centered data search using Generative Pre-trained Transformer (GPT) language model and artificial intelligence

被引:3
|
作者
Li, Diya [1 ]
Zhang, Zhe [1 ,2 ]
机构
[1] Texas A&M Univ, Dept Geog, College Stn, TX 77843 USA
[2] Texas A&M Univ, Dept Elect & Comp Engn, College Stn, TX 77843 USA
来源
PLOS ONE | 2023年 / 18卷 / 11期
关键词
D O I
10.1371/journal.pone.0293034
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Accessing and utilizing geospatial data from various sources is essential for developing scientific research to address complex scientific and societal challenges that require interdisciplinary knowledge. The traditional keyword-based geosearch approach is insufficient due to the uncertainty inherent within spatial information and how it is presented in the data-sharing platform. For instance, the Gulf of Mexico Coastal Ocean Observing System (GCOOS) data search platform stores geoinformation and metadata in a complex tabular. Users can search for data by entering keywords or selecting data from a drop-down manual from the user interface. However, the search results provide limited information about the data product, where detailed descriptions, potential use, and relationship with other data products are still missing. Language models (LMs) have demonstrated great potential in tasks like question answering, sentiment analysis, text classification, and machine translation. However, they struggle when dealing with metadata represented in tabular format. To overcome these challenges, we developed Meta Question Answering System (MetaQA), a novel spatial data search model. MetaQA integrates end-to-end AI models with a generative pre-trained transformer (GPT) to enhance geosearch services. Using GCOOS metadata as a case study, we tested the effectiveness of MetaQA. The results revealed that MetaQA outperforms state-of-the-art question-answering models in handling tabular metadata, underlining its potential for user-inspired geosearch services.
引用
收藏
页数:20
相关论文
共 48 条
  • [21] Revolutionizing teledermatology: Exploring the integration of artificial intelligence, including Generative Pre-trained Transformer chatbots for artificial intelligence-driven anamnesis, diagnosis, and treatment plans
    Shapiro, Jonathan
    Lyakhovitsky, Anna
    CLINICS IN DERMATOLOGY, 2024, 42 (05) : 492 - 497
  • [22] Diagnosing Glaucoma Based on the Ocular Hypertension Treatment Study Dataset Using Chat Generative Pre-Trained Transformer as a Large Language Model
    Raja, Hina
    Huang, Xiaoqin
    Delsoz, Mohammad
    Madadi, Yeganeh
    Poursoroush, Asma
    Munawar, Asim
    Kahook, Malik Y.
    Yousefi, Siamak
    OPHTHALMOLOGY SCIENCE, 2025, 5 (01):
  • [23] ENHANCING SEMANTIC WEB ENTITY MATCHING PROCESS USING TRANSFORMER NEURAL NETWORKS AND PRE-TRAINED LANGUAGE MODELS
    Jabrane, Mourad
    Toulaoui, Abdelfattah
    Hafidi, Imad
    COMPUTING AND INFORMATICS, 2024, 43 (06) : 1397 - 1415
  • [24] Pre-trained artificial intelligence-aided analysis of nanoparticles using the segment anything model
    Monteiro, Gabriel A. A.
    Monteiro, Bruno A. A.
    dos Santos, Jefersson A.
    Wittemann, Alexander
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [25] Pre-trained language model for code-mixed text in Indonesian, Javanese, and English using transformer
    Ahmad Fathan Hidayatullah
    Rosyzie Anna Apong
    Daphne Teck Ching Lai
    Atika Qazi
    Social Network Analysis and Mining, 15 (1)
  • [26] Spatial-Temporal Transformer Networks for Traffic Flow Forecasting Using a Pre-Trained Language Model
    Ma, Ju
    Zhao, Juan
    Hou, Yao
    SENSORS, 2024, 24 (17)
  • [27] Few-Shot Learning for Identification of COVID-19 Symptoms Using Generative Pre-trained Transformer Language Models
    Jiang, Keyuan
    Zhu, Minghao
    Bernard, Gordon R.
    MACHINE LEARNING AND PRINCIPLES AND PRACTICE OF KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT II, 2023, 1753 : 307 - 316
  • [28] Text data augmentation and pre-trained Language Model for enhancing text classification of low-resource languages
    Ziyaden, Atabay
    Yelenov, Amir
    Hajiyev, Fuad
    Rustamov, Samir
    Pak, Alexandr
    PEERJ COMPUTER SCIENCE, 2024, 10
  • [29] The potential of Generative Pre-trained Transformer 4 (GPT-4) to analyse medical notes in three different languages: a retrospective model-evaluation study
    Menezes, Maria Clara Saad
    Hoffmann, Alexander F.
    Tan, Amelia L. M.
    Nalbandyan, Marine
    Omenn, Gilbert S.
    Mazzotti, Diego R.
    Hernandez-Arango, Alejandro
    Visweswaran, Shyam
    Mandl, Kenneth D.
    Bourgeois, Florence
    Lee, James W. K.
    Makmur, Andrew
    Hanauer, David A.
    Semanik, Michael G.
    Kerivan, Lauren T.
    Hill, Terra
    Forero, Julian
    Restrepo, Carlos
    Vigna, Matteo
    Ceriana, Piero
    Abu-el-rub, Noor
    Avillach, Paul
    Bellazzi, Riccardo
    Callaci, Thomas
    Gutierrez-Sacristan, Alba
    Malovini, Alberto
    Mathew, Jomol P.
    Morris, Michele
    Murthy, Venkatesh L.
    Buonocore, Tommaso M.
    Parimbelli, Enea
    Patel, Lav P.
    Saez, Carlos
    Samayamuthu, Malarkodi Jebathilagam
    Thompson, Jeffrey A.
    Tibollo, Valentina
    Xia, Zongqi
    Kohane, Isaac S.
    LANCET DIGITAL HEALTH, 2025, 7 (01): : e35 - e43
  • [30] Comparative Analysis of Generative Pre-Trained Transformer Models in Oncogene-Driven Non-Small Cell Lung Cancer: Introducing the Generative Artificial Intelligence Performance Score
    Hamilton, Zacharie
    Aseem, Aseem
    Chen, Zhengjia
    Naffakh, Noor
    Reizine, Natalie M.
    Weinberg, Frank
    Jain, Shikha
    Kessler, Larry G.
    Gadi, Vijayakrishna K.
    Bun, Christopher
    Nguyen, Ryan H.
    JCO CLINICAL CANCER INFORMATICS, 2024, 8