Smart Home Notifications in Croatian Language: A Transformer-Based Approach

被引:0
|
作者
Simunec, Magdalena [1 ]
Soic, Renato [1 ]
机构
[1] Univ Zagreb, Fac Elect Engn & Comp, Unska 3, Zagreb 10000, Croatia
关键词
Natural Language Processing; Natural Language Generation; Croatian; sentence transformer; Internet of Things;
D O I
10.1109/CONTEL58387.2023.10198978
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Natural language generation is one of the essential components in achieving spoken interaction between a human user and a computer system. In this paper, we describe the process of generating notifications in Croatian language by employing an approach that relies on sentence transformer models. The solution described here extracts the key information from raw system events available in scope of the IoT platform and uses them to construct user-friendly sentences. The use-case scenario we present in this paper is related to a smart home environment. In this scenario, the generated notifications are used to construct spoken notifications using a speech synthesis component.
引用
收藏
页数:5
相关论文
共 50 条
  • [21] Transformer-based heart language model with electrocardiogram annotations
    Tudjarski, Stojancho
    Gusev, Marjan
    Kanoulas, Evangelos
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [22] LVBERT: Transformer-Based Model for Latvian Language Understanding
    Znotins, Arturs
    Barzdins, Guntis
    HUMAN LANGUAGE TECHNOLOGIES - THE BALTIC PERSPECTIVE (HLT 2020), 2020, 328 : 111 - 115
  • [23] Is Transformer-Based Attention Agnostic of the Pretraining Language and Task?
    Martin, R. H. J.
    Visser, R.
    Dunaiski, M.
    SOUTH AFRICAN COMPUTER SCIENCE AND INFORMATION SYSTEMS RESEARCH TRENDS, SAICSIT 2024, 2024, 2159 : 95 - 123
  • [24] RadBERT: Adapting Transformer-based Language Models to Radiology
    Yan, An
    McAuley, Julian
    Lu, Xing
    Du, Jiang
    Chang, Eric Y.
    Gentili, Amilcare
    Hsu, Chun-Nan
    RADIOLOGY-ARTIFICIAL INTELLIGENCE, 2022, 4 (04)
  • [25] ParsBERT: Transformer-based Model for Persian Language Understanding
    Mehrdad Farahani
    Mohammad Gharachorloo
    Marzieh Farahani
    Mohammad Manthouri
    Neural Processing Letters, 2021, 53 : 3831 - 3847
  • [26] Applications of transformer-based language models in bioinformatics: a survey
    Zhang, Shuang
    Fan, Rui
    Liu, Yuti
    Chen, Shuang
    Liu, Qiao
    Zeng, Wanwen
    NEURO-ONCOLOGY ADVANCES, 2023, 5 (01)
  • [27] ParsBERT: Transformer-based Model for Persian Language Understanding
    Farahani, Mehrdad
    Gharachorloo, Mohammad
    Farahani, Marzieh
    Manthouri, Mohammad
    NEURAL PROCESSING LETTERS, 2021, 53 (06) : 3831 - 3847
  • [28] TAG: Gradient Attack on Transformer-based Language Models
    Deng, Jieren
    Wang, Yijue
    Li, Ji
    Wang, Chenghong
    Shang, Chao
    Liu, Hang
    Rajasekaran, Sanguthevar
    Ding, Caiwen
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3600 - 3610
  • [29] A Sparse Transformer-Based Approach for Image Captioning
    Lei, Zhou
    Zhou, Congcong
    Chen, Shengbo
    Huang, Yiyong
    Liu, Xianrui
    IEEE Access, 2020, 8 : 213437 - 213446
  • [30] A Sparse Transformer-Based Approach for Image Captioning
    Lei, Zhou
    Zhou, Congcong
    Chen, Shengbo
    Huang, Yiyong
    Liu, Xianrui
    IEEE ACCESS, 2020, 8 : 213437 - 213446