SELF-ATTENTION BASED MODEL FOR PUNCTUATION PREDICTION USING WORD AND SPEECH EMBEDDINGS

被引:0
作者
Yi, Jiangyan [1 ]
Tao, Jianhua [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing, Peoples R China
[2] Chinese Acad Sci, CAS Ctr Excellence Brain Sci & Intelligence Techn, Beijing, Peoples R China
[3] Univ Chinese Acad Sci, Beijing, Peoples R China
来源
2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP) | 2019年
基金
中国国家自然科学基金;
关键词
Self-attention; transfer learning; word embedding; speech embedding; punctuation prediction; RECOGNITION; SYSTEM;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This paper proposes to use self-attention based model to predict punctuation marks for word sequences. The model is trained using word and speech embedding features which are obtained from the pre-trainedWord2Vec and Speech2Vec, respectively. Thus, the model can use any kind of textual data and speech data. Experiments are conducted on English IWSLT2011 datasets. The results show that the self-attention based model trained using word and speech embedding features outperforms the previous state-of-the-art single model by up to 7.8% absolute overall F-1-score. The results also show that it obtains performance improvement by up to 4.7% absolute overall F-1-score against the previous best ensemble model.
引用
收藏
页码:7270 / 7274
页数:5
相关论文
共 50 条
  • [21] A Self-Attention Integrated Learning Model for Landing Gear Performance Prediction
    Lin, Lin
    Tong, Changsheng
    Guo, Feng
    Fu, Song
    Lv, Yancheng
    He, Wenhui
    [J]. SENSORS, 2023, 23 (13)
  • [22] An Efficient Link Prediction Model in Dynamic Heterogeneous Information Networks Based on Multiple Self-attention
    Ruan, Beibei
    Zhu, Cui
    [J]. KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT III, 2021, 12817 : 62 - 74
  • [23] Fake news detection and classification using hybrid BiLSTM and self-attention model
    Mohapatra, Asutosh
    Thota, Nithin
    Prakasam, P.
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (13) : 18503 - 18519
  • [24] ON THE USEFULNESS OF SELF-ATTENTION FOR AUTOMATIC SPEECH RECOGNITION WITH TRANSFORMERS
    Zhang, Shucong
    Loweimi, Erfan
    Bell, Peter
    Renals, Steve
    [J]. 2021 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP (SLT), 2021, : 89 - 96
  • [25] SELF-ATTENTION GENERATIVE ADVERSARIAL NETWORK FOR SPEECH ENHANCEMENT
    Huy Phan
    Nguyen, Huy Le
    Chen, Oliver Y.
    Koch, Philipp
    Duong, Ngoc Q. K.
    McLoughlin, Ian
    Mertins, Alfred
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7103 - 7107
  • [26] Fake news detection and classification using hybrid BiLSTM and self-attention model
    Asutosh Mohapatra
    Nithin Thota
    P. Prakasam
    [J]. Multimedia Tools and Applications, 2022, 81 : 18503 - 18519
  • [27] Speech emotion recognition using recurrent neural networks with directional self-attention
    Li, Dongdong
    Liu, Jinlin
    Yang, Zhuo
    Sun, Linyu
    Wang, Zhe
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2021, 173
  • [28] Punctuation Prediction for Unsegmented Transcript Based on Word Vector
    Che, Xiaoyin
    Wang, Cheng
    Yang, Haojin
    Meinel, Christoph
    [J]. LREC 2016 - TENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2016, : 654 - 658
  • [29] Vehicle Interaction Behavior Prediction with Self-Attention
    Li, Linhui
    Sui, Xin
    Lian, Jing
    Yu, Fengning
    Zhou, Yafu
    [J]. SENSORS, 2022, 22 (02)
  • [30] Learning Fragment Self-Attention Embeddings for Image-Text Matching
    Wu, Yiling
    Wang, Shuhui
    Song, Guoli
    Huang, Qingming
    [J]. PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 2088 - 2096