UrduSER: A comprehensive dataset for speech emotion recognition in Urdu language

被引:0
作者
Akhtar, Muhammad Zaheer [1 ,2 ]
Jahangir, Rashid [2 ]
Ain, QuratUl [1 ]
Nauman, Muhammad Asif [3 ]
Uddin, Mueen [4 ]
Ullah, Syed Sajid [5 ]
机构
[1] Islamia Univ Bahawalpur, Dept Informat Technol, Bahawalpur 63100, Pakistan
[2] COMSATS Univ Islamabad, Dept Comp Sci, Vehari Campus, Islamabad 61100, Pakistan
[3] Riphah Int Univ, Riphah Sch Comp & Innovat, Lahore, Pakistan
[4] Univ Doha Sci & Technol, Coll Comp & IT, Doha 24449, Qatar
[5] Univ Agder UiA, Dept Informat & Commun Technol, N-4898 Grimstad, Norway
来源
DATA IN BRIEF | 2025年 / 60卷
关键词
Speech emotion recognition; Signal processing; Deep learning; Urdu language; Dataset;
D O I
10.1016/j.dib.2025.111627
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Speech Emotion Recognition (SER) is a rapidly evolving field of research that aims to identify and categorize emotional states through speech signal analysis. As SER holds considerable socio <not sign>cultural and business significance, researchers are increasingly exploring machine learning and deep learning techniques to advance this technology. A well-suited dataset is a crucial resource for SER studies in a specific language. However, despite being the 10th most spoken language globally, Urdu lacks SER datasets, creating a significant research gap. The available Urdu SER datasets are insufficient due to their limited scope, including a narrow range of emotions, small datasets, and a limited number of dialogs, which restricts their usability in real-world scenarios. To fill the gap in existing Urdu speech datasets, an Urdu Speech Emotion Recognition Dataset (UrduSER) is developed. This comprehensive dataset consists of 3500 speech signals from 10 professional actors, with a balanced mix of males and females, and diverse age ranges. The speech signals were sourced from a vast collection of Pakistani Urdu drama serials and telefilms available on YouTube. Seven emotional states are covered in the dataset: Angry, Fear, Boredom, Disgust, Happy, Neutral, and Sad. A notable strength of this dataset is the diversity of the dialogs, with each utterance containing almost unique content, in contrast to existing datasets that often feature repetitive samples of predefined dialogs spoken by research volunteers in a laboratory environment. To ensure balance and symmetry, the dataset consists of 500 samples for each emotional class, with 50 samples per actor per emotion. An accompanying Excel file provides a detailed metadata index for each audio sample, including file name, duraand the Urdu dialogue script. This comprehensive metadata index enables researchers and developers to efficiently access, organize, and utilize the UrduSER dataset. The UrduSER dataset underwent a rigorous validation process, integrating expert validation to confirm its validity, reliability, and over (c) 2025 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY license ( http://creativecommons.org/licenses/by/4.0/ )
引用
收藏
页数:14
相关论文
共 17 条
[1]   Speech Emotion Recognition through Hybrid Features and Convolutional Neural Network [J].
Alluhaidan, Ala Saleh ;
Saidani, Oumaima ;
Jahangir, Rashid ;
Nauman, Muhammad Asif ;
Neffati, Omnia Saidani .
APPLIED SCIENCES-BASEL, 2023, 13 (08)
[2]   An Urdu speech corpus for emotion recognition [J].
Asghar, Awais ;
Sohaib, Sarmad ;
Iftikhar, Saman ;
Sha, Muhammad ;
Fatima, Kiran .
PEERJ COMPUTER SCIENCE, 2022, 8
[3]  
Burkhardt Felix, 2005, Interspeech, P1517, DOI [10.21437/Interspeech.2005-446, DOI 10.21437/INTERSPEECH.2005-446]
[4]   IEMOCAP: interactive emotional dyadic motion capture database [J].
Busso, Carlos ;
Bulut, Murtaza ;
Lee, Chi-Chun ;
Kazemzadeh, Abe ;
Mower, Emily ;
Kim, Samuel ;
Chang, Jeannette N. ;
Lee, Sungbok ;
Narayanan, Shrikanth S. .
LANGUAGE RESOURCES AND EVALUATION, 2008, 42 (04) :335-359
[5]  
Jackson P., 2014, Surrey audio-visual expressed emotion (SAVEE) database
[6]   Convolutional neural network-based cross-corpus speech emotion recognition with data augmentation and features fusion [J].
Jahangir, Rashid ;
Teh, Ying Wah ;
Mujtaba, Ghulam ;
Alroobaea, Roobaea ;
Shaikh, Zahid Hussain ;
Ali, Ihsan .
MACHINE VISION AND APPLICATIONS, 2022, 33 (03)
[7]  
Jahangir R, 2021, MULTIMED TOOLS APPL, V80, P23745, DOI 10.1007/s11042-020-09874-7
[8]   Deep Learning Approaches for Bimodal Speech Emotion Recognition: Advancements, Challenges, and a Multi-Learning Model [J].
Kakuba, Samuel ;
Poulose, Alwin ;
Han, Dong Seog .
IEEE ACCESS, 2023, 11 :113769-113789
[9]   Emotion recognition and artificial intelligence: A systematic review (2014-2023) and research recommendations [J].
Khare, Smith K. ;
Blanes-Vidal, Victoria ;
Nadimi, Esmaeil S. ;
Acharya, U. Rajendra .
INFORMATION FUSION, 2024, 102
[10]   Survey of Deep Representation Learning for Speech Emotion Recognition [J].
Latif, Siddique ;
Rana, Rajib ;
Khalifa, Sara ;
Jurdak, Raja ;
Qadir, Junaid ;
Schuller, Bjorn .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (02) :1634-1654