A Novel Real-Time Speech Summarizer System for the Learning of Sustainability

被引:1
|
作者
Wang, Hsiu-Wen [1 ]
Cheng, Ding-Yuan [2 ]
Chen, Chi-Hua [1 ,3 ,4 ]
Wu, Yu-Rou [1 ]
Lo, Chi-Chun [1 ]
Lin, Hui-Fei [4 ]
机构
[1] Natl Chiao Tung Univ, Dept Informat Management & Finance, Hsinchu 300, Taiwan
[2] Hwa Hsia Inst Technol, Dept Informat Management, Zhonghe Dist 235, New Taipei, Taiwan
[3] Chunghwa Telecom Co Ltd, Telecommun Labs, Yangmei City 326, Taoyuan County, Taiwan
[4] Natl Chiao Tung Univ, Dept Commun & Technol, Hsinchu 300, Taiwan
关键词
SERVICE SYSTEM; INFORMATION; DESIGN; EXTRACTION; RETRIEVAL; TEXT;
D O I
10.3390/su7043885
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
As the number of speech and video documents increases on the Internet and portable devices proliferate, speech summarization becomes increasingly essential. Relevant research in this domain has typically focused on broadcasts and news; however, the automatic summarization methods used in the past may not apply to other speech domains (e.g., speech in lectures). Therefore, this study explores the lecture speech domain. The features used in previous research were analyzed and suitable features were selected following experimentation; subsequently, a three-phase real-time speech summarizer for the learning of sustainability (RTSSLS) was proposed. Phase One involved selecting independent features (e.g., centrality, resemblance to the title, sentence length, term frequency, and thematic words) and calculating the independent feature scores; Phase Two involved calculating the dependent features, such as the position compared with the independent feature scores; and Phase Three involved comparing these feature scores to obtain weighted averages of the function-scores, determine the highest-scoring sentence, and provide a summary. In practical results, the accuracies of macro-average and micro-average for the RTSSLS were 70% and 73%, respectively. Therefore, using a RTSSLS can enable users to acquire key speech information for the learning of sustainability.
引用
收藏
页码:3885 / 3899
页数:15
相关论文
共 50 条
  • [41] A Real-Time Text to Audio-Visual Speech Synthesis System
    Wang, Lijuan
    Qian, Xiaojun
    Ma, Lei
    Qian, Yao
    Chen, Yining
    Soong, Frank
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 2338 - +
  • [42] Integration of Speech and Text Processing Modules into a Real-Time Dialogue System
    Ptacek, Jan
    Ircing, Pavel
    Spousta, Miroslav
    Romportl, Jan
    Loose, Zdenek
    Cinkova, Silvie
    Relano Gil, Jose
    Santos, Raul
    TEXT, SPEECH AND DIALOGUE, 2010, 6231 : 552 - +
  • [43] LOW LATENCY PARAMETER GENERATION FOR REAL-TIME SPEECH SYNTHESIS SYSTEM
    Na, Xingyu
    Xie, Xiang
    Kuang, Jingming
    2014 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2014,
  • [44] A real-time Speech-interfaced System for Group Conversation Modeling
    Rocchi, Cesare
    Principi, Emanuele
    Cifani, Simone
    Rotili, Rudy
    Squartini, Stefano
    Piazza, Francesco
    NEURAL NETS WIRN09, 2009, 204 : 70 - 80
  • [46] REAL-TIME SPEECH ENHANCEMENT SYSTEM USING ENVELOPE EXPANSION TECHNIQUE
    CLARKSON, PM
    BAHGAT, S
    ELECTRONICS LETTERS, 1989, 25 (17) : 1186 - 1188
  • [47] DYSFLUENCY CLASSIFICATION IN STUTTERED SPEECH USING DEEP LEARNING FOR REAL-TIME APPLICATIONS
    Jouaiti, Melanie
    Dautenhahn, Kerstin
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6482 - 6486
  • [48] A novel approach for real-time learning and recognition of human activities
    Leon, Olmo
    Cuellar, Manuel P.
    Delgado, Miguel
    JOURNAL OF AMBIENT INTELLIGENCE AND SMART ENVIRONMENTS, 2016, 8 (04) : 423 - 436
  • [49] Real-time speech emotion recognition using deep learning and data augmentation
    Barhoumi, Chawki
    Benayed, Yassine
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 58 (02)
  • [50] Deep-Learning Framework for Efficient Real-Time Speech Enhancement and Dereverberation
    Rosenbaum, Tomer
    Winebrand, Emil
    Cohen, Omer
    Cohen, Israel
    SENSORS, 2025, 25 (03)