JARVIS: Joining Adversarial Training With Vision Transformers in Next-Activity Prediction

被引:4
作者
Pasquadibisceglie, Vincenzo [1 ,2 ]
Appice, Annalisa [1 ,2 ]
Castellano, Giovanna [1 ,2 ]
Malerba, Donato [1 ,2 ]
机构
[1] Univ Aldo Moro Bari, Dipartimento Informat, I-70125 Bari, Italy
[2] Consorzio Interuniv Nazl Informat CINI, I-70125 Bari, Italy
关键词
Predictive models; Training; Numerical models; Transformers; Data models; Context modeling; Computer architecture; Adversarial training; attention; computer vision; deep learning; multi-view learning; predictive process monitoring; vision transformers; XAI;
D O I
10.1109/TSC.2023.3331020
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose a novel predictive process monitoring approach, named JARVIS, that is designed to achieve a balance between accuracy and explainability in the task of next-activity prediction. To this aim, JARVIS represents different process executions (traces) as patches of an image and uses this patch-based representation within a multi-view learning scheme combined with Vision Transformers (ViTs). Using multi-view learning we guarantee good accuracy by leveraging the variety of information recorded in event logs as different patches of an image. The use of ViTs enables the integration of explainable elements directly into the framework of a predictive process model trained to forecast the next trace activity from the completed events in a running trace by utilizing self-attention modules that give paired attention values between two picture patches. Attention modules disclose explainable information concerning views of the business process and events of the trace that influenced the prediction. In addition, we explore the effect of ViT adversarial training to mitigate overfitting and improve the accuracy and robustness of predictive process monitoring. Experiments with various benchmark event logs prove the accuracy of JARVIS compared to several current state-of-the-art methods and draw insights from explanations recovered through the attention modules.
引用
收藏
页码:1593 / 1606
页数:14
相关论文
共 43 条
  • [1] Abnar S, 2020, 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), P4190
  • [2] Andriushchenko M., 2020, Advances in Neural Information Processing Systems, V33, P16048
  • [3] Bai T, 2021, PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, P4312
  • [4] Barbon Junior Sylvio, 2021, From Data to Models and Back. 9th International Symposium, DataMod 2020. Revised Selected Papers. Lecture Notes in Computer Science (LNCS 12611), P174, DOI 10.1007/978-3-030-70650-0_11
  • [5] Bergstra J., 2011, NIPS 11 P 24 INT C N, V24, P2546
  • [6] Buijs J., 2014, Flexible evolutionary algorithms for mining structured process models
  • [7] Bukhsh ZA, 2021, Arxiv, DOI arXiv:2104.00721
  • [8] Learning Accurate LSTM Models of Business Processes
    Camargo, Manuel
    Dumas, Marlon
    Gonzalez-Rojas, Oscar
    [J]. BUSINESS PROCESS MANAGEMENT (BPM 2019), 2019, 11675 : 286 - 302
  • [9] A Robust and Versatile Multi-View Learning Framework for the Detection of Deviant Business Process Instances
    Cuzzocrea, Alfredo
    Folino, Francesco
    Guarascio, Massimo
    Pontieri, Luigi
    [J]. INTERNATIONAL JOURNAL OF COOPERATIVE INFORMATION SYSTEMS, 2016, 25 (04)
  • [10] A Multi-view Learning Approach to the Discovery of Deviant Process Instances
    Cuzzocrea, Alfredo
    Folino, Francesco
    Guarascio, Massimo
    Pontieri, Luigi
    [J]. ON THE MOVE TO MEANINGFUL INTERNET SYSTEMS: OTM 2015 CONFERENCES, 2015, 9415 : 146 - 165