Добірка наукової літератури з теми "Emotional speech database"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся зі списками актуальних статей, книг, дисертацій, тез та інших наукових джерел на тему "Emotional speech database".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Статті в журналах з теми "Emotional speech database":
Tank, Vishal P., and S. K. Hadia. "Creation of speech corpus for emotion analysis in Gujarati language and its evaluation by various speech parameters." International Journal of Electrical and Computer Engineering (IJECE) 10, no. 5 (October 1, 2020): 4752. http://dx.doi.org/10.11591/ijece.v10i5.pp4752-4758.
Byun, Sung-Woo, and Seok-Pil Lee. "A Study on a Speech Emotion Recognition System with Effective Acoustic Features Using Deep Learning Algorithms." Applied Sciences 11, no. 4 (February 21, 2021): 1890. http://dx.doi.org/10.3390/app11041890.
손남호, Hwang Hyosung, and Ho-Young Lee. "Emotional Speech Database and the Acoustic Analysis of Emotional Speech." EONEOHAG ll, no. 72 (August 2015): 175–99. http://dx.doi.org/10.17290/jlsk.2015..72.175.
Vicsi, Klára, and Dávid Sztahó. "Recognition of Emotions on the Basis of Different Levels of Speech Segments." Journal of Advanced Computational Intelligence and Intelligent Informatics 16, no. 2 (March 20, 2012): 335–40. http://dx.doi.org/10.20965/jaciii.2012.p0335.
Quan, Changqin, Bin Zhang, Xiao Sun, and Fuji Ren. "A combined cepstral distance method for emotional speech recognition." International Journal of Advanced Robotic Systems 14, no. 4 (July 1, 2017): 172988141771983. http://dx.doi.org/10.1177/1729881417719836.
Shahin, Ismail. "Employing Emotion Cues to Verify Speakers in Emotional Talking Environments." Journal of Intelligent Systems 25, no. 1 (January 1, 2016): 3–17. http://dx.doi.org/10.1515/jisys-2014-0118.
Caballero-Morales, Santiago-Omar. "Recognition of Emotions in Mexican Spanish Speech: An Approach Based on Acoustic Modelling of Emotion-Specific Vowels." Scientific World Journal 2013 (2013): 1–13. http://dx.doi.org/10.1155/2013/162093.
Sultana, Sadia, M. Shahidur Rahman, M. Reza Selim, and M. Zafar Iqbal. "SUST Bangla Emotional Speech Corpus (SUBESCO): An audio-only emotional speech corpus for Bangla." PLOS ONE 16, no. 4 (April 30, 2021): e0250173. http://dx.doi.org/10.1371/journal.pone.0250173.
Keshtiari, Niloofar, Michael Kuhlmann, Moharram Eslami, and Gisela Klann-Delius. "Recognizing emotional speech in Persian: A validated database of Persian emotional speech (Persian ESD)." Behavior Research Methods 47, no. 1 (May 23, 2014): 275–94. http://dx.doi.org/10.3758/s13428-014-0467-x.
Werner, S., and G. N. Petrenko. "Speech Emotion Recognition: Humans vs Machines." Discourse 5, no. 5 (December 18, 2019): 136–52. http://dx.doi.org/10.32603/2412-8562-2019-5-5-136-152.
Дисертації з теми "Emotional speech database":
Sun, Rui. "The evaluation of the stability of acoustic features in affective conveyance across multiple emotional databases." Diss., Georgia Institute of Technology, 2013. http://hdl.handle.net/1853/49041.
CHENG, KUAN-JUNG, and 程冠融. "Cross-Lingual Speech Emotion Recognition Based on Speech Recognition Technology in An Emotional Speech Database in Mandarin, Taiwanese, and Hakka." Thesis, 2019. http://ndltd.ncl.edu.tw/handle/6c4m2x.
國立雲林科技大學
資訊管理系
107
With the development of artificial intelligence, machine learning and deep learning, there are considerable breakthroughs in recognition techniques such as image recognition and speech recognition. Especially in the speech recognition technology, whether it is everyone's smart phone, or the current popular smart speakers, these products are equipped with voice assistants to provide a convenient voice interactive interface. Allowing machines to understand human emotions helps to increase interaction with machines, so speech emotion recognition is an important issue. An important research direction in speech emotion recognition technology is cross-corpus and cross-lingual speech emotion recognition. Most existing speech emotion recognition researches focus on using the same corpus to train and test the speech emotion recognition system. In the context of cross-corpus and cross-lingual, the effectiveness of such systems is significantly reduced. In order to solve this problem, this study uses the cascaded normalization approach proposed by previous research to eliminate the difference as much as possible, and observe whether the extreme learning machine can improve the rate of cross-lingual speech emotion recognition in An Emotional Speech Database In Mandarin, Taiwanese, and Hakka. Besides An Emotional Speech Database In Mandarin, Taiwanese, and Hakka, we add the Berlin Database of Emotional Speech (Emo-DB) to conduct experiments of cross-corpus speech emotion recognition.
Lu, Jhih-Jheng, and 陸至正. "Construction and Testing of a Mandarin Emotional Speech Database and Its Application." Thesis, 2004. http://ndltd.ncl.edu.tw/handle/54267308290823890882.
大同大學
資訊工程學系(所)
92
Automatic emotional speech recognition is a hot topic in signal processing. In this thesis, we build a Mandarin emotional speech database which includes anger, happiness, sadness, boredom, and neutral emotion utterances. We extract the Mel-frequency cepstrum coefficients from each speech as the emotion feature vector. We use K-nearest neighbor method to be our classifier, and obtained 74.6% recognition accuracy. We also proposed a modified K-nearest neighbor method for emotion evaluation. For training the hearing-impaired people to speak naturally, we design an emotion radar chart to present the intensity of each emotion. With the techniques stated above, we implement a computer-assisted speech training system.
Manamela, Phuti John. "The automatic recognition of emotions in speech." Thesis, 2020. http://hdl.handle.net/10386/3347.
Speech emotion recognition (SER) refers to a technology that enables machines to detect and recognise human emotions from spoken phrases. In the literature, numerous attempts have been made to develop systems that can recognise human emotions from their voice, however, not much work has been done in the context of South African indigenous languages. The aim of this study was to develop an SER system that can classify and recognise six basic human emotions (i.e., sadness, fear, anger, disgust, happiness, and neutral) from speech spoken in Sepedi language (one of South Africa’s official languages). One of the major challenges encountered, in this study, was the lack of a proper corpus of emotional speech. Therefore, three different Sepedi emotional speech corpora consisting of acted speech data have been developed. These include a RecordedSepedi corpus collected from recruited native speakers (9 participants), a TV broadcast corpus collected from professional Sepedi actors, and an Extended-Sepedi corpus which is a combination of Recorded-Sepedi and TV broadcast emotional speech corpora. Features were extracted from the speech corpora and a data file was constructed. This file was used to train four machine learning (ML) algorithms (i.e., SVM, KNN, MLP and Auto-WEKA) based on 10 folds validation method. Three experiments were then performed on the developed speech corpora and the performance of the algorithms was compared. The best results were achieved when Auto-WEKA was applied in all the experiments. We may have expected good results for the TV broadcast speech corpus since it was collected from professional actors, however, the results showed differently. From the findings of this study, one can conclude that there are no precise or exact techniques for the development of SER systems, it is a matter of experimenting and finding the best technique for the study at hand. The study has also highlighted the scarcity of SER resources for South African indigenous languages. The quality of the dataset plays a vital role in the performance of SER systems.
National research foundation (NRF) and Telkom Center of Excellence (CoE)
Ferro, Adelino Rafael Mendes. "Speech emotion recognition through statistical classification." Master's thesis, 2017. http://hdl.handle.net/10400.14/22817.
The purpose of this dissertation is to discuss speech emotion recognition. It was created a validated acted Portuguese emotional speech database, named European Portuguese Emotional Discourse Database (EPEDD), and statistical classification algorithms have been applied on it. EPEDD is an acted database, featuring 12 utterances (2 single-words, 5 short sentences and 5 long sentences) per actor and per emotion, 8 actors, both genders equally represented, and 9 emotions (anger, joy, disgust, excitement, fear, apathy, surprise, sadness and neutral), based on Lövheim’s emotion model. We had 40% of the database evaluated by unexperienced evaluators, enabling us to produce a validated one, filtering 60% of the evaluated utterances. The full database contains 718 instances, while the validated one contains 116 instances. The average acting quality of the original database was evaluated, in a scale from 1 to 5, as 2,3. The validated database is composed by emotional utterances that have their emotions recognized on average at a 69,6% rate, by unexperienced judges. Anger had the highest recognition rate at 79,7%, while disgust had the lowest recognition rate at 40,5%. Feature extraction and statistical classification algorithms were performed respectively applying Opensmile and Weka software. Statistical classification algorithms operated in the full database and in the validated one, best results being obtained by SVMs, respectively the emotion recognition rates being 48,7% and 44,0%. Apathy had the highest recognition rate: 79.0%, while excitement had the lowest emotion recognition rate: 32.9%.
Частини книг з теми "Emotional speech database":
Gajšek, Rok, Vitomir Štruc, Boštjan Vesnicer, Anja Podlesek, Luka Komidar, and France Mihelič. "Analysis and Assessment of AvID: Multi-Modal Emotional Database." In Text, Speech and Dialogue, 266–73. Berlin, Heidelberg: Springer Berlin Heidelberg, 2009. http://dx.doi.org/10.1007/978-3-642-04208-9_38.
Justin, Tadej, Vitomir Štruc, Janez Žibert, and France Mihelič. "Development and Evaluation of the Emotional Slovenian Speech Database - EmoLUKS." In Text, Speech, and Dialogue, 351–59. Cham: Springer International Publishing, 2015. http://dx.doi.org/10.1007/978-3-319-24033-6_40.
Geethashree, A., and D. J. Ravi. "Kannada Emotional Speech Database: Design, Development and Evaluation." In Proceedings of International Conference on Cognition and Recognition, 135–43. Singapore: Springer Singapore, 2017. http://dx.doi.org/10.1007/978-981-10-5146-3_14.
Sapiński, Tomasz, Dorota Kamińska, Adam Pelikant, Cagri Ozcinar, Egils Avots, and Gholamreza Anbarjafari. "Multimodal Database of Emotional Speech, Video and Gestures." In Pattern Recognition and Information Forensics, 153–63. Cham: Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-030-05792-3_15.
Staroniewicz, Piotr, and Wojciech Majewski. "Polish Emotional Speech Database – Recording and Preliminary Validation." In Cross-Modal Analysis of Speech, Gestures, Gaze and Facial Expressions, 42–49. Berlin, Heidelberg: Springer Berlin Heidelberg, 2009. http://dx.doi.org/10.1007/978-3-642-03320-9_5.
Jokić, Ivan, Stevan Jokić, Vlado Delić, and Zoran Perić. "Impact of Emotional Speech to Automatic Speaker Recognition - Experiments on GEES Speech Database." In Speech and Computer, 268–75. Cham: Springer International Publishing, 2014. http://dx.doi.org/10.1007/978-3-319-11581-8_33.
Navas, Eva, Inmaculada Hernáez, Amaia Castelruiz, and Iker Luengo. "Obtaining and Evaluating an Emotional Database for Prosody Modelling in Standard Basque." In Text, Speech and Dialogue, 393–400. Berlin, Heidelberg: Springer Berlin Heidelberg, 2004. http://dx.doi.org/10.1007/978-3-540-30120-2_50.
Pérez-Espinosa, Humberto, Carlos Aleberto Reyes-García, and Luis Villaseñor-Pineda. "EmoWisconsin: An Emotional Children Speech Database in Mexican Spanish." In Affective Computing and Intelligent Interaction, 62–71. Berlin, Heidelberg: Springer Berlin Heidelberg, 2011. http://dx.doi.org/10.1007/978-3-642-24571-8_7.
Gahlawat, Mukta, Amita Malik, and Poonam Bansal. "Phonetic Transcription Comparison for Emotional Database for Speech Synthesis." In Advances in Intelligent Systems and Computing, 187–94. Singapore: Springer Singapore, 2017. http://dx.doi.org/10.1007/978-981-10-6626-9_21.
Atassi, Hicham, Maria Teresa Riviello, Zdeněk Smékal, Amir Hussain, and Anna Esposito. "Emotional Vocal Expressions Recognition Using the COST 2102 Italian Database of Emotional Speech." In Development of Multimodal Interfaces: Active Listening and Synchrony, 255–67. Berlin, Heidelberg: Springer Berlin Heidelberg, 2010. http://dx.doi.org/10.1007/978-3-642-12397-9_21.
Тези доповідей конференцій з теми "Emotional speech database":
Bansal, Sweeta, and Amita Dev. "Emotional hindi speech database." In 2013 International Conference Oriental COCOSDA held jointly with 2013 Conference on Asian Spoken Language Research and Evaluation (O-COCOSDA/CASLRE). IEEE, 2013. http://dx.doi.org/10.1109/icsda.2013.6709867.
Oflazoglu, Caglar, and Serdar Yildirim. "Turkish emotional speech database." In 2011 IEEE 19th Signal Processing and Communications Applications Conference (SIU). IEEE, 2011. http://dx.doi.org/10.1109/siu.2011.5929860.
Burkhardt, Felix, A. Paeschke, M. Rolfes, Walter F. Sendlmeier, and Benjamin Weiss. "A database of German emotional speech." In Interspeech 2005. ISCA: ISCA, 2005. http://dx.doi.org/10.21437/interspeech.2005-446.
Zen, Heiga, Tadashi Kitamura, Murtaza Bulut, Shrikanth Narayanan, Ryosuke Tsuzuki, and Keiichi Tokuda. "Constructing emotional speech synthesizers with limited speech database." In Interspeech 2004. ISCA: ISCA, 2004. http://dx.doi.org/10.21437/interspeech.2004-442.
Sato, Ryota, Ryohei Sasaki, Norisato Suga, and Toshihiro Furukawa. "Creation and Analysis of Emotional Speech Database for Multiple Emotions Recognition." In 2020 23rd Conference of the Oriental COCOSDA International Committee for the Co-ordination and Standardisation of Speech Databases and Assessment Techniques (O-COCOSDA). IEEE, 2020. http://dx.doi.org/10.1109/o-cocosda50338.2020.9295041.
Grimm, Michael, Kristian Kroschel, and Shrikanth Narayanan. "The Vera am Mittag German audio-visual emotional speech database." In 2008 IEEE International Conference on Multimedia and Expo (ICME). IEEE, 2008. http://dx.doi.org/10.1109/icme.2008.4607572.
Ko, Youjung, Insuk Hong, Hyunsoon Shin, and Yoonjoong Kim. "Construction of a database of emotional speech using emotion sounds from movies and dramas." In 2017 International Conference on Information and Communications (ICIC). IEEE, 2017. http://dx.doi.org/10.1109/infoc.2017.8001672.
Mustafa, Mumtaz B., Raja N. Ainon, Roziati Zainuddin, Zuraidah M. Don, and Gerry Knowles. "Assessing the naturalness of malay emotional voice corpora." In 2011 Oriental COCOSDA 2011 - International Conference on Speech Database and Assessments. IEEE, 2011. http://dx.doi.org/10.1109/icsda.2011.6086002.
Pandharipande, Meghna A., Rupayan Chakraborty, and Sunil Kumar Kopparapu. "Methods and challenges for creating an emotional audio-visual database." In 2017 20th Conference of the Oriental Chapter of the International Coordinating Committee on Speech Databases and Speech I/O Systems and Assessment (O-COCOSDA). IEEE, 2017. http://dx.doi.org/10.1109/icsda.2017.8384466.
Li, Runnan, Zhiyong Wu, Jia Jia, Yaohua Bu, Sheng Zhao, and Helen Meng. "Towards Discriminative Representation Learning for Speech Emotion Recognition." In Twenty-Eighth International Joint Conference on Artificial Intelligence {IJCAI-19}. California: International Joint Conferences on Artificial Intelligence Organization, 2019. http://dx.doi.org/10.24963/ijcai.2019/703.