Deep Learning-Based Short Text Summarization: An Integrated BERT and Transformer Encoder-Decoder Approach

被引:0
作者
Ghanem, Fahd A. [1 ,2 ]
Padma, M. C. [1 ]
Abdulwahab, Hudhaifa M. [3 ]
Alkhatib, Ramez [4 ]
机构
[1] Univ Mysore, PES Coll Engn, Dept Comp Sci & Engn, Mandya 571401, India
[2] Hodeidah Univ, Coll Educ Zabid, Dept Comp Sci, POB 3114, Hodeidah, Yemen
[3] VTU, Ramaiah Inst Technol, Dept Comp Applicat, Bangalore 560054, India
[4] BMB Nord, Res Ctr Borstel, Pk Allee 35, D-23845 Borstel, Germany
关键词
attention mechanism; BERT; deep learning; short text summarization; transformer-based encoder-decoder; Twitter summarization; TWITTER;
D O I
10.3390/computation13040096
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
The field of text summarization has evolved from basic extractive methods that identify key sentences to sophisticated abstractive techniques that generate contextually meaningful summaries. In today's digital landscape, where an immense volume of textual data is produced every day, the need for concise and coherent summaries is more crucial than ever. However, summarizing short texts, particularly from platforms like Twitter, presents unique challenges due to character constraints, informal language, and noise from elements such as hashtags, mentions, and URLs. To overcome these challenges, this paper introduces a deep learning framework for automated short text summarization on Twitter. The proposed approach combines bidirectional encoder representations from transformers (BERT) with a transformer-based encoder-decoder architecture (TEDA), incorporating an attention mechanism to improve contextual understanding. Additionally, long short-term memory (LSTM) networks are integrated within BERT to effectively capture long-range dependencies in tweets and their summaries. This hybrid model ensures that generated summaries remain informative, concise, and contextually relevant while minimizing redundancy. The performance of the proposed framework was assessed using three benchmark Twitter datasets-Hagupit, SHShoot, and Hyderabad Blast-with ROUGE scores serving as the evaluation metric. Experimental results demonstrate that the model surpasses existing approaches in accurately capturing key information from tweets. These findings underscore the framework's effectiveness in automated short text summarization, offering a robust solution for efficiently processing and summarizing large-scale social media content.
引用
收藏
页数:21
相关论文
共 53 条
[1]   Enhancing Persian text summarization through a three-phase fine-tuning and reinforcement learning approach with the mT5 transformer model [J].
Abadi, Vahid Nejad Mahmood ;
Ghasemian, Fahimeh .
SCIENTIFIC REPORTS, 2025, 15 (01)
[2]   Performance Study on Extractive Text Summarization Using BERT Models [J].
Abdel-Salam, Shehab ;
Rafea, Ahmed .
INFORMATION, 2022, 13 (02)
[3]   Effective deep learning approaches for summarization of legal texts [J].
Anand, Deepa ;
Wagh, Rupali .
JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (05) :2141-2150
[4]  
[Anonymous], 2019, Computational natural language learning, DOI [DOI 10.18653/V1/K19-1074, 10.48550/arxiv.1902.09243]
[5]   Arabic abstractive text summarization using RNN-based and transformer-based architectures [J].
Bani-Almarjeh, Mohammad ;
Kurdy, Mohamad-Bassam .
INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (02)
[6]  
Bano Sheher, 2022, 2022 International Conference on Artificial Intelligence of Things (ICAIoT), P1, DOI 10.1109/ICAIoT57170.2022.10121826
[7]   Summarization of scholarly articles using BERT and BiGRU: Deep learning-based extractive approach [J].
Bano, Sheher ;
Khalid, Shah ;
Tairan, Nasser Mansoor ;
Shah, Habib ;
Khattak, Hasan Ali .
JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (09)
[8]   DCBRTS: A Classification-Summarization Approach for Evolving Tweet Streams in Multiobjective Optimization Framework [J].
Bansal, Diksha ;
Saini, Naveen ;
Saha, Sriparna .
IEEE ACCESS, 2021, 9 :148325-148338
[9]   Abstractive text summarization and new large-scale datasets for agglutinative languages Turkish and Hungarian [J].
Baykara, Batuhan ;
Gungor, Tunga .
LANGUAGE RESOURCES AND EVALUATION, 2022, 56 (03) :973-1007
[10]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171