Privacy Enhanced Speech Emotion Communication using Deep Learning Aided Edge Computing

被引:9
作者
Ali, Hafiz Shehbaz [1 ]
ul Hassan, Fakhar [2 ]
Latif, Siddique [3 ]
Manzoor, Habib Ullah [4 ]
Qadir, Junaid [2 ]
机构
[1] EmulationAI, Brisbane, Qld, Australia
[2] Informat Technol Univ ITU, Lahore, Punjab, Pakistan
[3] Univ Southern Queensland, Brisbane, Qld, Australia
[4] Univ Engn & Technol, Lahore, Pakistan
来源
2021 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS) | 2021年
关键词
emotion communication system; speech emotion recognition; privacy enhanced features; deep learning; edge computing; REPRESENTATIONS; HEALTH;
D O I
10.1109/ICCWorkshops50388.2021.9473669
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Speech emotion sensing in communication networks has a wide range of applications in real life. In these applications, voice data are transmitted from the user to the central server for storage, processing, and decision making. However, speech data contain vulnerable information that can be used maliciously without the user's consent by an eavesdropping adversary. In this work, we present a privacy-enhanced emotion communication system for preserving the user personal information in emotion-sensing applications. We propose the use of an adversarial learning framework that can be deployed at the edge to unlearn the users' private information in the speech representations. These privacy-enhanced representations can be transmitted to the central server for decision making. We evaluate the proposed model on multiple speech emotion datasets and show that the proposed model can hide users' specific demographic information and improve the robustness of emotion identification without significantly impacting performance. To the best of our knowledge, this is the first work on a privacy-preserving framework for emotion sensing in the communication network.
引用
收藏
页数:5
相关论文
共 34 条
  • [1] Aloufi R, 2021, Arxiv, DOI arXiv:2011.02930
  • [2] [Anonymous], 2017, ARXIV170202823
  • [3] [Anonymous], 2014, P INTERSPEECH 2014
  • [4] Burkhardt F., 2006, 9 INT C SPOKEN LANGU
  • [5] Burkhardt F, 2005, 9 EUR C SPEECH COMM, P1517, DOI DOI 10.21437/INTERSPEECH.2005-446
  • [6] IEMOCAP: interactive emotional dyadic motion capture database
    Busso, Carlos
    Bulut, Murtaza
    Lee, Chi-Chun
    Kazemzadeh, Abe
    Mower, Emily
    Kim, Samuel
    Chang, Jeannette N.
    Lee, Sungbok
    Narayanan, Shrikanth S.
    [J]. LANGUAGE RESOURCES AND EVALUATION, 2008, 42 (04) : 335 - 359
  • [7] Chen M., 2017, BIG DATA COGNIT COMP, V1, P1, DOI DOI 10.3390/BDCC1010002
  • [8] Costantini G, 2014, LREC 2014 - NINTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, P3501
  • [9] Eyben F., 2013, P 21 ACM INT C MULT, P835, DOI DOI 10.1145/2502081.2502224
  • [10] Ganin Y, 2015, PR MACH LEARN RES, V37, P1180