Abstractive Text Summarization Using Multimodal Information

被引:1
作者
Rafi, Shaik [1 ]
Das, Ranjita [2 ]
机构
[1] NIT MIZORAM, Dept Comp Sci & Engn, Aizawl 796012, Mizoram, India
[2] NIT AGARTALA, Dept Comp Sci & Engn, Agartala 799046, Tripura, India
来源
2023 10TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING & MACHINE INTELLIGENCE, ISCMI | 2023年
关键词
Abstractive Text Summarization; Multimodality Image Text (MIT); Attention Mechanism; LSTM; Sequence-to-Sequence model;
D O I
10.1109/ISCMI59957.2023.10458505
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Much text generates over the internet through news articles, story writing and blogs. Reading and understanding such an enormous amount of data to the user is problematic, including time and effort. Automatic abstractive text summarization has gained more importance to increase the user's understanding and reduce time. It shortens the given input by preserving the meaning and identifying the context of the whole document to generate meaningful sentences. The research community has proposed different methods for text reduction and generating abstractive summaries. However, problems like semantics and contextual relationship in the summary generation process must be still need to improve. The multimodal abstractive text summarization is a technique that combines text and image information which helps in addressing the semantics and contextual relationship by proposing Multimodality Image Text (MIT) layer that fuses the text-extracted global features by glove embedding and preserves the semantics of the vocabulary and text-related images are used to identify the contextual relationship features from inception v3, which cope in the MIT layer to generate efficient multimodal abstractive text summaries by training and testing with seq-to-seq model. Experiments with the MSMO dataset achieve superior performance on other state-of-art results.
引用
收藏
页码:141 / 145
页数:5
相关论文
共 29 条
  • [1] Bahdanau D, 2016, Arxiv, DOI [arXiv:1409.0473, 10.48550/arXiv.1409.0473]
  • [2] Calixto Iacer, 2017, C EMP METH NAT LANG
  • [3] Chen JQ, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P4046
  • [4] Chen Jingqiang, 2019, Future Gener. Comput. Syst.
  • [5] Chopra Sumit., 2016, Abstractive Sentence summarization with Attentive Recurrent Neural Networks
  • [6] Topic-Based Image Caption Generation
    Dash, Sandeep Kumar
    Acharya, Shantanu
    Pakray, Partha
    Das, Ranjita
    Gelbukh, Alexander
    [J]. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2020, 45 (04) : 3025 - 3034
  • [7] Biomedical-domain pre-trained language model for extractive summarization
    Du, Yongping
    Li, Qingxiao
    Wang, Lulin
    He, Yanqing
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 199 (199)
  • [8] Abstractive summarization: An overview of the state of the art
    Gupta, Som
    Gupta, S. K.
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2019, 121 : 49 - 65
  • [9] He B, 2023, Arxiv, DOI arXiv:2303.07284
  • [10] Kuchaiev O, 2018, Arxiv, DOI arXiv:1703.10722