{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,12]],"date-time":"2024-09-12T10:04:34Z","timestamp":1726135474641},"publisher-location":"Cham","reference-count":31,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030862701"},{"type":"electronic","value":"9783030862718"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-86271-8_27","type":"book-chapter","created":{"date-parts":[[2021,9,14]],"date-time":"2021-09-14T19:08:06Z","timestamp":1631646486000},"page":"319-330","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Speech Emotion Recognition by Conventional Machine Learning and Deep Learning"],"prefix":"10.1007","author":[{"given":"Javier","family":"de Lope","sequence":"first","affiliation":[]},{"given":"Enrique","family":"Hern\u00e1ndez","sequence":"additional","affiliation":[]},{"given":"Vanessa","family":"Vargas","sequence":"additional","affiliation":[]},{"given":"Manuel","family":"Gra\u00f1a","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,9,15]]},"reference":[{"issue":"1","key":"27_CR1","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1162\/neco.1995.7.1.108","volume":"7","author":"C Bishop","year":"1995","unstructured":"Bishop, C.: Training with noise is equivalent to Tikhonov regularization. Neural Comput. 7(1), 108\u2013116 (1995)","journal-title":"Neural Comput."},{"key":"27_CR2","unstructured":"Bogert, B., Healy, J., Tukey, J.: The quefrency analysis of time series for echoes: cepstrum, pseudo-autocovariance, cross-cepstrum and saphe cracking. In: Proceeddings of the Symposium on Time Series Analysis, pp. 209\u2013243 (1963)"},{"key":"27_CR3","doi-asserted-by":"crossref","unstructured":"Burkhardt, F., Paeschke, A., Rolfes, M., Sendlmeier, W., Weiss, B.: A database of German emotional speech. In: Proceedings of 9th European Conference on Speech Communication and Technology, pp. 1517\u20131520 (2005)","DOI":"10.21437\/Interspeech.2005-446"},{"issue":"5","key":"27_CR4","first-page":"11","volume":"1","author":"V Chavan","year":"2012","unstructured":"Chavan, V., Gohokar, V.: Speech emotion recognition by using SVM-classifier. Int. J. Eng. Adv. Technol. 1(5), 11\u201315 (2012)","journal-title":"Int. J. Eng. Adv. Technol."},{"key":"27_CR5","doi-asserted-by":"publisher","first-page":"1428","DOI":"10.1109\/PROC.1977.10747","volume":"65","author":"D Chiders","year":"1977","unstructured":"Chiders, D., Skinner, D., Kemerait, R.: The cepstrum: a guide to processing. Proc. IEEE 65, 1428\u20131443 (1977)","journal-title":"Proc. IEEE"},{"issue":"4","key":"27_CR6","doi-asserted-by":"publisher","first-page":"357","DOI":"10.1109\/TASSP.1980.1163420","volume":"28","author":"S Davis","year":"1980","unstructured":"Davis, S., Mermelstein, P.: Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences. IEEE Trans. Acoust. Speech Signal Process. 28(4), 357\u2013366 (1980)","journal-title":"IEEE Trans. Acoust. Speech Signal Process."},{"key":"27_CR7","doi-asserted-by":"crossref","unstructured":"Engberg, I., Hansen, A.: Documentation of the Danish emotional speech database. Technical report, Center for Person Kommunilation, Denmark (1996)","DOI":"10.21437\/Eurospeech.1997-482"},{"key":"27_CR8","doi-asserted-by":"crossref","unstructured":"Engberg, I., Hansen, A., Andersen, O., Dalsgaard, P.: Design, recording and verification of a Danish emotional speech database. In: Proceedings of EuroSpeech, pp. 1695\u20131698 (1997)","DOI":"10.21437\/Eurospeech.1997-482"},{"key":"27_CR9","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-319-70772-3_1","volume-title":"Brain Informatics","author":"Y Gao","year":"2017","unstructured":"Gao, Y., Li, B., Wang, N., Zhu, T.: Speech emotion recognition using local and global features. In: He, Y., et al. (eds.) BI 2017. LNCS (LNAI), vol. 10654, pp. 3\u201313. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-70772-3_1"},{"key":"27_CR10","doi-asserted-by":"crossref","unstructured":"Iqbal, A., Barua, K.: A real-time emotion recognition from speech using gradient boosting. In: Proceedings of International Conference on Electrical, Computer and Communication Engineering, pp. 1\u20135 (2019)","DOI":"10.1109\/ECACE.2019.8679271"},{"key":"27_CR11","doi-asserted-by":"publisher","first-page":"101894","DOI":"10.1016\/j.bspc.2020.101894","volume":"59","author":"D Issa","year":"2020","unstructured":"Issa, D., Faith-Demirci, M., Yazici, A.: Speech emotion recognition with deep convolutional neural networks. Biomed. Signal Process. Control 59, 101894 (2020)","journal-title":"Biomed. Signal Process. Control"},{"key":"27_CR12","first-page":"87","volume":"4","author":"O Kamil","year":"2018","unstructured":"Kamil, O.: Frame blocking and windowing. Speech Signal 4, 87\u201394 (2018)","journal-title":"Speech Signal"},{"issue":"11","key":"27_CR13","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y LeCun","year":"1998","unstructured":"LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278\u20132324 (1998)","journal-title":"Proc. IEEE"},{"key":"27_CR14","doi-asserted-by":"crossref","unstructured":"Lim, W., Jang, D., Lee, T.: Speech emotion recognition using convolutional and recurrent neural networks. In: Proceedings of Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, pp. 1\u20134 (2016)","DOI":"10.1109\/APSIPA.2016.7820699"},{"key":"27_CR15","unstructured":"Lin, Y., Wei, G.: Speech emotion recognition based on HMM and SVM. In: IEEE International Conference on Machine Learning and Cybermetics, pp. 4898\u20134901 (2005)"},{"issue":"5","key":"27_CR16","doi-asserted-by":"publisher","first-page":"e0196391","DOI":"10.1371\/journal.pone.0196391","volume":"13","author":"S Livingstone","year":"2018","unstructured":"Livingstone, S., Russo, F.: The Ryerson audio-visual database of emotional speech and song (RAVDESS): a dynamic, multimodal set of facieal and vocal expressions in North American English. PLoS One 13(5), e0196391 (2018)","journal-title":"PLoS One"},{"key":"27_CR17","doi-asserted-by":"crossref","unstructured":"Lotfidereshgi, R., Gournay, P.: Biologically inspired speech emotion recognition. In: Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 5135\u20135139 (2017)","DOI":"10.1109\/ICASSP.2017.7953135"},{"key":"27_CR18","doi-asserted-by":"crossref","unstructured":"McFee, B., et al.: librosa: Audio and music signal analysis in Python. In: Proceedings of 14th Python in Science Conference, pp. 18\u201325 (2015)","DOI":"10.25080\/Majora-7b98e3ed-003"},{"key":"27_CR19","unstructured":"Mermelstein, P.: Distance measures for speech recognition, psychological and instrumental. In: Chen, C. (ed.) Pattern Recognition and Artificial Intelligence, pp. 374\u2013388. Academic Press (1976)"},{"issue":"4","key":"27_CR20","doi-asserted-by":"publisher","first-page":"1799","DOI":"10.1016\/j.asej.2016.11.001","volume":"9","author":"H Palo","year":"2018","unstructured":"Palo, H., Mohanty, M.: Wavelet based feature combination for recognition of emotion. Ain Shams Eng. J. 9(4), 1799\u20131806 (2018)","journal-title":"Ain Shams Eng. J."},{"key":"27_CR21","first-page":"101","volume":"6","author":"Y Pan","year":"2012","unstructured":"Pan, Y., Sen, P., Shen, L.: Speech emotion recognition using support vector machines. Int. J. Smart Home 6, 101\u2013108 (2012)","journal-title":"Int. J. Smart Home"},{"key":"27_CR22","doi-asserted-by":"crossref","unstructured":"Pandey, S., Shekhawat, H., Prasanna, S.: Deep learning techniques for speech emotion recognition: a review. In: Proceedings of 29th IEEE International Conference on Radioelektronika, pp. 1\u20136 (2019)","DOI":"10.1109\/RADIOELEK.2019.8733432"},{"key":"27_CR23","doi-asserted-by":"crossref","unstructured":"Park, D., et al.: SpecAugment: a simple data augmentation method for automatic speech recognition. In: Proceedings of Interspeech 2019, pp. 2613\u20132617 (2019)","DOI":"10.21437\/Interspeech.2019-2680"},{"key":"27_CR24","first-page":"2825","volume":"12","author":"F Pedregosa","year":"2011","unstructured":"Pedregosa, F., et al.: Scikit-learn: machine learning in Python. J. Mach. Learn. Res. 12, 2825\u20132830 (2011)","journal-title":"J. Mach. Learn. Res."},{"issue":"4","key":"27_CR25","doi-asserted-by":"publisher","first-page":"344","DOI":"10.1511\/2001.4.344","volume":"89","author":"R Plutchik","year":"2001","unstructured":"Plutchik, R.: The nature of emotions: human emotions have deep evolutionary roots. Am. Sci. 89(4), 344\u2013350 (2001)","journal-title":"Am. Sci."},{"issue":"3","key":"27_CR26","doi-asserted-by":"publisher","first-page":"1467","DOI":"10.1007\/s11235-011-9624-z","volume":"52","author":"S Ramakrishnan","year":"2013","unstructured":"Ramakrishnan, S., Emary, I.E.: Speech emotion recognition approaches in human computer interaction. Telecommun. Syst. 52(3), 1467\u20131478 (2013)","journal-title":"Telecommun. Syst."},{"key":"27_CR27","first-page":"99","volume":"15","author":"K Rao","year":"2012","unstructured":"Rao, K., Kumar, T., Anusha, K., Leela, B., Bhavana, I., Gowtham, S.: Emotion recognition from speech. Int. J. Comput. Sci. Inf. Technol. 15, 99\u2013117 (2012)","journal-title":"Int. J. Comput. Sci. Inf. Technol."},{"key":"27_CR28","doi-asserted-by":"crossref","unstructured":"Seehapoch, T., Wongthanavasu, S.: Speech emotion recognition using support vector machines. In: International Conference on Knowledge and Smart Technology (2013)","DOI":"10.1109\/KST.2013.6512793"},{"key":"27_CR29","unstructured":"Stolar, M., Lech, M., Bolia, R., Skinner, M.: Real time speech emotion recognition using RGB image classification and transfer learning. In: Proceedings of 11th IEEE International Conference on Signal Processing and Communication Systems, pp. 1\u20138 (2005)"},{"key":"27_CR30","unstructured":"Tripathi, S., Kumar, A., Ramesh, A., Singh, C., Yenigalla, P.: Focal loss based residual convolutional neural network for speech emotion recognition. arXiv:1906.05682 (2019)"},{"key":"27_CR31","unstructured":"Ververidis, D., Kotropoulos, C.: Automatic speech classification to five emotional states based on gender information. In: 12th IEEE European Signal Processing Conference, pp. 341\u2013344 (2004)"}],"container-title":["Lecture Notes in Computer Science","Hybrid Artificial Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-86271-8_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T15:33:47Z","timestamp":1710344027000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-86271-8_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030862701","9783030862718"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-86271-8_27","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"15 September 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"HAIS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Hybrid Artificial Intelligence Systems","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Bilbao","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Spain","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 September 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 September 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"hais2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2021.haisconference.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"81","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"44","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"11","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"54% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2-3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}