Prosodic focus control in reply speech generation for a spoken dialogue system of information retrieval

被引:2
|
作者
Kiriyama, S [1 ]
Hirose, K [1 ]
Minematsu, N [1 ]
机构
[1] Shizuoka Univ, Fac Informat, Hamamatsu, Shizuoka 4328011, Japan
关键词
D O I
10.1109/WSS.2002.1224393
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A spoken dialogue system of information retrieval on academic documents has been developed with a special attention to reply speech generation. In order to realize speech reply with its prosodic features properly controlled to express dialogue focuses, a scheme was developed to directly generating speech reply from reply content. When developing the system firstly, a priority was placed on the automatic processing, and prosodic focus was controlled by rather simple rules (original rules). Based on the listening test for the reply speech generated using original rules, new rules were then developed. Through the further listening test, the rules were revised and called the revised rules. The validity of the revised rules was verified through an evaluation experiment. It was also indicated that there existed users' preferences on the intonation of the reply speech.
引用
收藏
页码:139 / 142
页数:4
相关论文
共 50 条
  • [21] A multi-modal dialogue system for information navigation and retrieval across spoken document archives with topic hierarchies
    Pan, YC
    Wang, CC
    Hsieh, YC
    Lee, TH
    Lee, YS
    Fu, YS
    Huang, YT
    Lee, LS
    2005 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2005, : 375 - 380
  • [22] The utility of semantic-pragmatic information and dialogue-state for speech recognition in spoken dialogue systems
    Stemmer, G
    Nöth, E
    Niemann, H
    TEXT, SPEECH AND DIALOGUE, PROCEEDINGS, 2000, 1902 : 439 - 444
  • [23] A system for speech driven information retrieval
    Gonzalez-Ferreras, Cesar
    Cardenoso-Payo, Valentin
    2007 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, VOLS 1 AND 2, 2007, : 624 - 628
  • [24] On Appropriateness and Estimation of the Emotion of Synthesized Response Speech in a Spoken Dialogue System
    Kase, Taketo
    Nose, Takashi
    Ito, Akinori
    HCI INTERNATIONAL 2015 - POSTERS' EXTENDED ABSTRACTS, PT I, 2015, 528 : 747 - 752
  • [25] Using Contextual Information in the Input Interface of a Spoken Dialogue System
    Lopez-Cozar, Ramon
    Espejo, Gonzalo
    Abalos, Nieves
    Griol, David
    WORKSHOP PROCEEDINGS OF THE 7TH INTERNATIONAL CONFERENCE ON INTELLIGENT ENVIRONMENTS, 2011, 10 : 429 - 440
  • [26] A Spoken Dialogue System for Electronic Program Guide Information Access
    Kim, Seokhwan
    Lee, Cheongjae
    Jung, Sangkeun
    Lee, Gary Geunbae
    2007 RO-MAN: 16TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, VOLS 1-3, 2007, : 178 - 181
  • [27] Speech interface and information retrieval for Medical Information system
    Hsu, CY
    Chen, B
    FASEB JOURNAL, 2001, 15 (04): : A485 - A485
  • [28] Estimating sentence final tone labels using dialogue-act information for text-to-speech synthesis within a spoken dialogue system
    Hojo N.
    Ijima Y.
    Sugiyama H.
    Transactions of the Japanese Society for Artificial Intelligence, 2020, 35 (04) : 1 - 11
  • [29] On the use of automatic speech recognition for spoken information retrieval from video databases
    Salgado-Garza, LR
    Nolazco-Flores, JA
    PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS AND APPLICATIONS, 2004, 3287 : 381 - 385
  • [30] Towards the Integration of Automatic Speech Recognition and Information Retrieval for Spoken Query Processing
    Moreno-Daniel, A.
    Wilpon, J.
    Juang, B. -H.
    Parthasarathy, S.
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 2154 - 2157