{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T12:40:04Z","timestamp":1732365604688,"version":"3.28.0"},"reference-count":56,"publisher":"Springer Science and Business Media LLC","issue":"28","license":[{"start":{"date-parts":[[2024,7,4]],"date-time":"2024-07-04T00:00:00Z","timestamp":1720051200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,7,4]],"date-time":"2024-07-04T00:00:00Z","timestamp":1720051200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100010607","name":"Universit\u00e0 degli Studi di Perugia","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100010607","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput & Applic"],"published-print":{"date-parts":[[2024,10]]},"abstract":"Abstract<\/jats:title>Inter-species emotional relationships, particularly the symbiotic interaction between humans and dogs, are complex and intriguing. Humans and dogs share fundamental mammalian neural mechanisms including mirror neurons, crucial to empathy and social behavior. Mirror neurons are activated during the execution and observation of actions, indicating inherent connections in social dynamics across species despite variations in emotional expression. This study explores the feasibility of using deep-learning Artificial Intelligence systems to accurately recognize canine emotions in general environments, to assist individuals without specialized knowledge or skills in discerning dog behavior, particularly related to aggression or friendliness. Starting with identifying key challenges in classifying pleasant and unpleasant emotions in dogs, we tested advanced deep-learning techniques and aggregated results to distinguish potentially dangerous human--dog interactions. Knowledge transfer is used to fine-tune different networks, and results are compared on original and transformed sets of frames from the Dog Clips dataset to investigate whether DogFACS action codes detailing relevant dog movements can aid the emotion recognition task. Elaborating on challenges and biases, we emphasize the need for bias mitigation to optimize performance, including different image preprocessing strategies for noise mitigation in dog recognition (i.e., face bounding boxes, segmentation of the face or body, isolating the dog on a white background, blurring the original background). Systematic experimental results demonstrate the system\u2019s capability to accurately detect emotions and effectively identify dangerous situations or signs of discomfort in the presence of humans.<\/jats:p>","DOI":"10.1007\/s00521-024-10042-3","type":"journal-article","created":{"date-parts":[[2024,7,4]],"date-time":"2024-07-04T08:14:55Z","timestamp":1720080895000},"page":"17669-17688","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Advanced techniques for automated emotion recognition in dogs from video data through deep learning"],"prefix":"10.1007","volume":"36","author":[{"ORCID":"http:\/\/orcid.org\/0000-0002-2972-7188","authenticated-orcid":false,"given":"Valentina","family":"Franzoni","sequence":"first","affiliation":[]},{"ORCID":"http:\/\/orcid.org\/0000-0002-1854-2196","authenticated-orcid":false,"given":"Giulio","family":"Biondi","sequence":"additional","affiliation":[]},{"given":"Alfredo","family":"Milani","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,4]]},"reference":[{"key":"10042_CR1","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1146\/annurev.neuro.27.070203.144230","volume":"27","author":"G Rizzolatti","year":"2004","unstructured":"Rizzolatti G, Craighero L (2004) The mirror-neuron system. Annu Rev Neurosci 27:169\u2013192. https:\/\/doi.org\/10.1146\/annurev.neuro.27.070203.144230","journal-title":"Annu Rev Neurosci"},{"key":"10042_CR2","doi-asserted-by":"publisher","unstructured":"Hess U, Fischer A (2013) Emotional mimicry as social regulation. Personality and social psychology review: an official journal of the Society for Personality and Social Psychology, Inc 17(2):142\u2013157. https:\/\/doi.org\/10.1177\/1088868312472607","DOI":"10.1177\/1088868312472607"},{"issue":"1","key":"10042_CR3","doi-asserted-by":"publisher","first-page":"15525","DOI":"10.1038\/s41598-017-15091-4","volume":"7","author":"C Caeiro","year":"2017","unstructured":"Caeiro C, Guo K, Mills D (2017) Dogs and humans respond to emotionally competent stimuli by producing different facial actions. Sci Rep 7(1):15525. https:\/\/doi.org\/10.1038\/s41598-017-15091-4","journal-title":"Sci Rep"},{"issue":"12","key":"10042_CR4","doi-asserted-by":"publisher","DOI":"10.1098\/rsos.150505","volume":"2","author":"E Palagi","year":"2015","unstructured":"Palagi E, Nicotra V, Cordoni G (2015) Rapid mimicry and emotional contagion in domestic dogs. R Soc Open Sci 2(12):150505","journal-title":"R Soc Open Sci"},{"key":"10042_CR5","doi-asserted-by":"publisher","DOI":"10.1155\/2014\/623203","author":"EP Cherniack","year":"2014","unstructured":"Cherniack EP, Cherniack AR (2014) The benefit of pets and animal-assisted therapy to the health of older individuals. Curr Gerontol Geriatr Res. https:\/\/doi.org\/10.1155\/2014\/623203","journal-title":"Curr Gerontol Geriatr Res"},{"key":"10042_CR6","doi-asserted-by":"publisher","DOI":"10.3233\/JAD-160703","author":"S Petersen","year":"2017","unstructured":"Petersen S, Houston S, Qin H, Tague C, Studley J (2017) The Utilization of Robotic Pets in Dementia Care. J Alzheimers Dis. https:\/\/doi.org\/10.3233\/JAD-160703","journal-title":"J Alzheimers Dis"},{"key":"10042_CR7","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-009-0024-4","author":"A Weiss","year":"2009","unstructured":"Weiss A, Wurhofer D, Tscheligi M (2009) \u201cI love this dog\u2019\u2019-children\u2019s emotional attachment to the robotic dog AIBO. Int J Soc Robot. https:\/\/doi.org\/10.1007\/s12369-009-0024-4","journal-title":"Int J Soc Robot"},{"issue":"1\u20132","key":"10042_CR8","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1016\/S1071-5819(03)00018-1","volume":"59","author":"C Breazeal","year":"2003","unstructured":"Breazeal C (2003) Emotion and sociable humanoid robots. Int J Hum-Comput Stud 59(1\u20132):119\u2013155. https:\/\/doi.org\/10.1016\/S1071-5819(03)00018-1","journal-title":"Int J Hum-Comput Stud"},{"issue":"3","key":"10042_CR9","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1016\/S0921-8890(02)00372-X","volume":"42","author":"T Fong","year":"2003","unstructured":"Fong T, Nourbakhsh I, Dautenhahn K (2003) A survey of socially interactive robots. Robot Auton Syst 42(3):143\u2013166. https:\/\/doi.org\/10.1016\/S0921-8890(02)00372-X","journal-title":"Robot Auton Syst"},{"key":"10042_CR10","doi-asserted-by":"publisher","unstructured":"Franzoni V, Milani A, Vallverd\u00fa J (2017) Emotional affordances in human-machine interactive planning and negotiation. In: Proceedings of the International Conference on Web Intelligence. WI \u201917, pp. 924\u2013930. Association for Computing Machinery, New York, NY, USA. https:\/\/doi.org\/10.1145\/3106426.3109421","DOI":"10.1145\/3106426.3109421"},{"key":"10042_CR11","doi-asserted-by":"publisher","unstructured":"Franzoni V, Milani A, Biondi G (2017) Semo: A semantic model for emotion recognition in web objects. In: Proceedings of the International Conference on Web Intelligence. WI \u201917, pp. 953\u2013958. Association for Computing Machinery, New York, NY, USA. https:\/\/doi.org\/10.1145\/3106426.3109417","DOI":"10.1145\/3106426.3109417"},{"key":"10042_CR12","doi-asserted-by":"publisher","unstructured":"Chan SW, Franzoni V, Mengoni P, Milani A (2018) Context-based image semantic similarity for prosthetic knowledge. In: 2018 IEEE First International Conference on Artificial Intelligence and Knowledge Engineering (AIKE), pp. 254\u2013258. https:\/\/doi.org\/10.1109\/AIKE.2018.00057","DOI":"10.1109\/AIKE.2018.00057"},{"key":"10042_CR13","doi-asserted-by":"publisher","unstructured":"Franzoni V, Vallverd\u00f9 J, Milani A (2019) Errors, biases and overconfidence in artificial emotional modeling. In: IEEE\/WIC\/ACM International Conference on Web Intelligence - Companion Volume. WI \u201919 Companion, pp. 86\u201390. Association for Computing Machinery, New York, NY, USA. https:\/\/doi.org\/10.1145\/3358695.3361749","DOI":"10.1145\/3358695.3361749"},{"key":"10042_CR14","doi-asserted-by":"publisher","unstructured":"Holzinger A, R\u00f6cker C, Ziefle M (2015) From smart health to smart hospitals. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). https:\/\/doi.org\/10.1007\/978-3-319-16226-3_1","DOI":"10.1007\/978-3-319-16226-3_1"},{"key":"10042_CR15","doi-asserted-by":"publisher","DOI":"10.1016\/j.jnca.2016.03.014","author":"J Santos","year":"2016","unstructured":"Santos J, Rodrigues JJPC, Silva BMC, Casal J, Saleem K, Denisov V (2016) An IoT-based mobile gateway for intelligent personal assistants on mobile health environments. J Netw Comput Appl. https:\/\/doi.org\/10.1016\/j.jnca.2016.03.014","journal-title":"J Netw Comput Appl"},{"key":"10042_CR16","doi-asserted-by":"crossref","unstructured":"Boyko N, Basystiuk O, Shakhovska N (2018) Performance evaluation and comparison of software for face recognition, based on dlib and opencv library. In: 2018 IEEE Second International Conference on Data Stream Mining & Processing (DSMP), pp. 478\u2013482. IEEE","DOI":"10.1109\/DSMP.2018.8478556"},{"issue":"3","key":"10042_CR17","doi-asserted-by":"publisher","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","volume":"13","author":"S Li","year":"2022","unstructured":"Li S, Deng W (2022) Deep facial expression recognition: A survey. IEEE Trans Affect Comput 13(3):1195\u20131215. https:\/\/doi.org\/10.1109\/TAFFC.2020.2981446","journal-title":"IEEE Trans Affect Comput"},{"key":"10042_CR18","doi-asserted-by":"publisher","unstructured":"Zhao S, Wang S, Soleymani M, Joshi D, Ji Q (2019) Affective computing for large-scale heterogeneous multimedia data: A survey. ACM Trans Multimedia Comput Commun Appl 15(3s) https:\/\/doi.org\/10.1145\/3363560","DOI":"10.1145\/3363560"},{"key":"10042_CR19","doi-asserted-by":"publisher","DOI":"10.3233\/WEB-190397","author":"O Gervasi","year":"2019","unstructured":"Gervasi O, Franzoni V, Riganelli M, Tasso S (2019) Automating facial emotion recognition. Web. Intelligence. https:\/\/doi.org\/10.3233\/WEB-190397","journal-title":"Intelligence"},{"key":"10042_CR20","doi-asserted-by":"publisher","unstructured":"Riganelli M, Franzoni V, Gervasi O, Tasso S (2017) EmEx, a Tool for Automated Emotive Face Recognition Using Convolutional Neural Networks vol. 10406 LNCS. https:\/\/doi.org\/10.1007\/978-3-319-62398-6_49","DOI":"10.1007\/978-3-319-62398-6_49"},{"key":"10042_CR21","doi-asserted-by":"publisher","DOI":"10.1080\/02699939208411068","author":"P Ekman","year":"1992","unstructured":"Ekman P (1992) An Argument for Basic Emotions. Cogn Emot. https:\/\/doi.org\/10.1080\/02699939208411068","journal-title":"Cogn Emot"},{"key":"10042_CR22","doi-asserted-by":"crossref","unstructured":"Ekman P, Friesen WV (1978) Facial action coding system: a technique for the measurement of facial movement","DOI":"10.1037\/t27734-000"},{"key":"10042_CR23","doi-asserted-by":"publisher","unstructured":"Franzoni V, Milani A, Biondi G, Micheli F (2019) A preliminary work on dog emotion recognition. In: IEEE\/WIC\/ACM International Conference on Web Intelligence - Companion Volume. WI \u201919 Companion, pp. 91\u201396. Association for Computing Machinery, New York, NY, USA. https:\/\/doi.org\/10.1145\/3358695.3361750","DOI":"10.1145\/3358695.3361750"},{"issue":"1","key":"10042_CR24","doi-asserted-by":"publisher","first-page":"22611","DOI":"10.1038\/s41598-022-27079-w","volume":"12","author":"T Boneh-Shitrit","year":"2022","unstructured":"Boneh-Shitrit T, Feighelstein M, Bremhorst A, Amir S, Distelfeld T, Dassa Y, Yaroshetsky S, Riemer S, Shimshoni I, Mills DS et al (2022) Explainable automated recognition of emotional states from canine facial expressions: the case of positive anticipation and frustration. Sci Rep 12(1):22611","journal-title":"Sci Rep"},{"issue":"4","key":"10042_CR25","doi-asserted-by":"publisher","first-page":"97","DOI":"10.3390\/fi14040097","volume":"14","author":"K Ferres","year":"2022","unstructured":"Ferres K, Schloesser T, Gloor PA (2022) Predicting dog emotions based on posture analysis using deeplabcut. Future Internet 14(4):97","journal-title":"Future Internet"},{"key":"10042_CR26","unstructured":"Waller BM, Caeiro C, Peirce K, Burrows AM, Kaminski J (2013) Dogfacs: the dog facial action coding system"},{"issue":"3","key":"10042_CR27","doi-asserted-by":"publisher","first-page":"92281","DOI":"10.1371\/journal.pone.0092281","volume":"9","author":"E Dalla Costa","year":"2014","unstructured":"Dalla Costa E, Minero M, Lebelt D, Stucke D, Canali E, Leach MC (2014) Development of the Horse Grimace Scale (HGS) as a Pain Assessment Tool in Horses Undergoing Routine Castration. PLoS ONE 9(3):92281. https:\/\/doi.org\/10.1371\/journal.pone.0092281","journal-title":"PLoS ONE"},{"key":"10042_CR28","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0175839","author":"C H\u00e4ger","year":"2017","unstructured":"H\u00e4ger C, Biernot S, Buettner M, Glage S, Keubler LM, Held N, Bleich EM, Otto K, M\u00fcller CW, Decker S, Talbot SR, Bleich A (2017) The Sheep Grimace Scale as an indicator of post-operative distress and pain in laboratory sheep. PLoS ONE. https:\/\/doi.org\/10.1371\/journal.pone.0175839","journal-title":"PLoS ONE"},{"key":"10042_CR29","doi-asserted-by":"publisher","DOI":"10.1038\/nmeth.1455","author":"DJ Langford","year":"2010","unstructured":"Langford DJ, Bailey AL, Chanda ML, Clarke SE, Drummond TE, Echols S, Glick S, Ingrao J, Klassen-Ross T, Lacroix-Fralish ML, Matsumiya L, Sorge RE, Sotocinal SG, Tabaka JM, Wong D, Van Den Maagdenberg AMJM, Ferrari MD, Craig KD, Mogil JS (2010) Coding of facial expressions of pain in the laboratory mouse. Nat Methods. https:\/\/doi.org\/10.1038\/nmeth.1455","journal-title":"Nat Methods"},{"key":"10042_CR30","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1186\/1744-8069-7-55","volume":"7","author":"SG Sotocinal","year":"2011","unstructured":"Sotocinal SG, Sorge RE, Zaloum A, Tuttle AH, Martin LJ, Wieskopf JS, Mapplebeck JCS, Wei P, Zhan S, Zhang S, McDougall JJ, King OD, Mogil JS (2011) The Rat Grimace Scale: a partially automated method for quantifying pain in the laboratory rat via facial expressions. Mol Pain 7:55. https:\/\/doi.org\/10.1186\/1744-8069-7-55","journal-title":"Mol Pain"},{"key":"10042_CR31","doi-asserted-by":"publisher","unstructured":"Mota-Rojas D, Marcet-Rius M, Ogi A, Hern\u00e1ndez-\u00c1valos I, Mariti C, Mart\u00ednez-Burnes J, Mora-Medina P, Casas A, Dom\u00ednguez A, Reyes B, Gazzano A (2021) Current Advances in Assessment of Dog\u2019s Emotions, Facial Expressions, and Their Use for Clinical Recognition of Pain. Animals 11(11) https:\/\/doi.org\/10.3390\/ani11113334","DOI":"10.3390\/ani11113334"},{"issue":"1","key":"10042_CR32","doi-asserted-by":"publisher","first-page":"0170730","DOI":"10.1371\/journal.pone.0170730","volume":"12","author":"MV Kujala","year":"2017","unstructured":"Kujala MV, Somppi S, Jokela M, Vainio O, Parkkonen L (2017) Human Empathy, Personality and Experience Affect the Emotion Ratings of Dog and Human Facial Expressions. PLoS ONE 12(1):0170730. https:\/\/doi.org\/10.1371\/journal.pone.0170730","journal-title":"PLoS ONE"},{"issue":"5","key":"10042_CR33","doi-asserted-by":"publisher","first-page":"851","DOI":"10.1007\/s10071-012-0510-1","volume":"15","author":"D Custance","year":"2012","unstructured":"Custance D, Mayer J (2012) Empathic-like responding by domestic dogs (Canis familiaris) to distress in humans: an exploratory study. Anim Cogn 15(5):851\u2013859. https:\/\/doi.org\/10.1007\/s10071-012-0510-1","journal-title":"Anim Cogn"},{"issue":"6","key":"10042_CR34","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky A, Sutskever I, Hinton GE (2017) Imagenet classification with deep convolutional neural networks. Commun ACM 60(6):84\u201390","journal-title":"Commun ACM"},{"key":"10042_CR35","doi-asserted-by":"publisher","first-page":"1281","DOI":"10.1038\/s41593-018-0209-y","volume":"21","author":"A Mathis","year":"2018","unstructured":"Mathis A, Mamidanna P, Cury KM, Abe T, Murthy VN, Mathis MW, Bethge M (2018) Deeplabcut: markerless pose estimation of user-defined body parts with deep learning. Nat Neurosci 21:1281\u20131289. https:\/\/doi.org\/10.1038\/s41593-018-0209-y","journal-title":"Nat Neurosci"},{"key":"10042_CR36","doi-asserted-by":"publisher","unstructured":"Franzoni V, Biondi G, Perri D, Gervasi O (2020) Enhancing Mouth-Based Emotion Recognition Using Transfer Learning. Sensors 20(18). https:\/\/doi.org\/10.3390\/s20185222","DOI":"10.3390\/s20185222"},{"issue":"1","key":"10042_CR37","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1109\/79.911197","volume":"18","author":"R Cowie","year":"2001","unstructured":"Cowie R, Douglas-Cowie E, Tsapatsoulis N, Votsis G, Kollias S, Fellenz W, Taylor JG (2001) Emotion recognition in human-computer interaction. IEEE Signal Process Mag 18(1):32\u201380. https:\/\/doi.org\/10.1109\/79.911197","journal-title":"IEEE Signal Process Mag"},{"key":"10042_CR38","doi-asserted-by":"publisher","unstructured":"Mirsamadi S, Barsoum E, Zhang C (2017) Automatic speech emotion recognition using recurrent neural networks with local attention. In: 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2227\u20132231. https:\/\/doi.org\/10.1109\/ICASSP.2017.7952552","DOI":"10.1109\/ICASSP.2017.7952552"},{"issue":"10","key":"10042_CR39","doi-asserted-by":"publisher","first-page":"1175","DOI":"10.1109\/34.954607","volume":"23","author":"RW Picard","year":"2001","unstructured":"Picard RW, Vyzas E, Healey J (2001) Toward machine emotional intelligence: analysis of affective physiological state. IEEE Trans Pattern Anal Mach Intell 23(10):1175\u20131191. https:\/\/doi.org\/10.1109\/34.954607","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"10042_CR40","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1016\/J.NEUNET.2017.02.013","volume":"92","author":"HM Fayek","year":"2017","unstructured":"Fayek HM, Lech M, Cavedon L (2017) Evaluating deep learning architectures for Speech Emotion Recognition. Neural Netw 92:60\u201368. https:\/\/doi.org\/10.1016\/J.NEUNET.2017.02.013","journal-title":"Neural Netw"},{"key":"10042_CR41","doi-asserted-by":"publisher","unstructured":"Fayek HM, Lech M, Cavedon L (2015) Towards real-time speech emotion recognition using deep neural networks. In: 2015 9th International Conference on Signal Processing and Communication Systems (ICSPCS), pp. 1\u20135. https:\/\/doi.org\/10.1109\/ICSPCS.2015.7391796","DOI":"10.1109\/ICSPCS.2015.7391796"},{"issue":"4","key":"10042_CR42","doi-asserted-by":"publisher","first-page":"363","DOI":"10.25046\/aj030437","volume":"3","author":"M Lech","year":"2018","unstructured":"Lech M, Stolar M, Bolia R, Skinner M (2018) Amplitude-frequency analysis of emotional speech using transfer learning and classification of spectrogram images. Adv Sci, Technol Eng Syst J 3(4):363\u2013371. https:\/\/doi.org\/10.25046\/aj030437","journal-title":"Adv Sci, Technol Eng Syst J"},{"key":"10042_CR43","doi-asserted-by":"publisher","unstructured":"Prasomphan S (2015) Detecting human emotion via speech recognition by using speech spectrogram. In: 2015 IEEE International Conference on Data Science and Advanced Analytics (DSAA), pp. 1\u201310. https:\/\/doi.org\/10.1109\/DSAA.2015.7344793","DOI":"10.1109\/DSAA.2015.7344793"},{"issue":"47","key":"10042_CR44","doi-asserted-by":"publisher","first-page":"36063","DOI":"10.1007\/s11042-020-09428-x","volume":"79","author":"V Franzoni","year":"2020","unstructured":"Franzoni V, Biondi G, Milani A (2020) Emotional sounds of crowds: spectrogram-based analysis using deep learning. Multimed Tools Appl 79(47):36063\u201336075. https:\/\/doi.org\/10.1007\/s11042-020-09428-x","journal-title":"Multimed Tools Appl"},{"key":"10042_CR45","unstructured":"Tureckova A (2017) GitHub - tureckova\/Doggie-smile: Computer Vision for Faces - Final project \u2014 github.com. GitHub"},{"issue":"2","key":"10042_CR46","doi-asserted-by":"publisher","first-page":"17","DOI":"10.13164\/mendel.2020.2.017","volume":"26","author":"A Tureckova","year":"2020","unstructured":"Tureckova A, Holik T, Kominkova Oplatkova Z (2020) Dog face detection using yolo network. MENDEL 26(2):17\u201322. https:\/\/doi.org\/10.13164\/mendel.2020.2.017","journal-title":"MENDEL"},{"key":"10042_CR47","doi-asserted-by":"publisher","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779\u2013788. https:\/\/doi.org\/10.1109\/CVPR.2016.91","DOI":"10.1109\/CVPR.2016.91"},{"key":"10042_CR48","doi-asserted-by":"publisher","unstructured":"Hu R, Doll\u00e1r P, He K, Darrell T, Girshick R (2018) Learning to segment every thing. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4233\u20134241. https:\/\/doi.org\/10.1109\/CVPR.2018.00445","DOI":"10.1109\/CVPR.2018.00445"},{"key":"10042_CR49","doi-asserted-by":"crossref","unstructured":"Kirillov A, Mintun E, Ravi N, Mao H, Rolland C, Gustafson L, Xiao T, Whitehead S, Berg AC, Lo W-Y, Doll\u00e1r P, Girshick R (2023) Segment Anything","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"10042_CR50","doi-asserted-by":"publisher","unstructured":"Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. https:\/\/doi.org\/10.1109\/CVPR.2009.5206848","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"10042_CR51","doi-asserted-by":"publisher","unstructured":"Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv. https:\/\/doi.org\/10.48550\/ARXIV.1704.04861. https:\/\/arxiv.org\/abs\/1704.04861","DOI":"10.48550\/ARXIV.1704.04861"},{"key":"10042_CR52","doi-asserted-by":"crossref","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: International Conference on Learning Representations","DOI":"10.1109\/ICCV.2015.314"},{"key":"10042_CR53","doi-asserted-by":"publisher","unstructured":"Chollet F (2017) Xception: Deep learning with depthwise separable convolutions. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1800\u20131807. https:\/\/doi.org\/10.1109\/CVPR.2017.195","DOI":"10.1109\/CVPR.2017.195"},{"key":"10042_CR54","doi-asserted-by":"crossref","unstructured":"Szegedy C, Ioffe S, Vanhoucke V, Alemi AA (2017) Inception-v4, inception-resnet and the impact of residual connections on learning. In: Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence. AAAI\u201917, pp. 4278\u20134284. AAAI Press, ???","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"10042_CR55","doi-asserted-by":"crossref","unstructured":"Bottou L (2012) Stochastic gradient descent tricks. Neural Networks: Tricks of the Trade: Second Edition, 421\u2013436","DOI":"10.1007\/978-3-642-35289-8_25"},{"key":"10042_CR56","doi-asserted-by":"crossref","unstructured":"Lin T, Wang Y, Liu X, Qiu X (2022) A survey of transformers. AI Open","DOI":"10.1016\/j.aiopen.2022.10.001"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-10042-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-024-10042-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-10042-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T12:06:45Z","timestamp":1732363605000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-024-10042-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,4]]},"references-count":56,"journal-issue":{"issue":"28","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["10042"],"URL":"https:\/\/doi.org\/10.1007\/s00521-024-10042-3","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"type":"print","value":"0941-0643"},{"type":"electronic","value":"1433-3058"}],"subject":[],"published":{"date-parts":[[2024,7,4]]},"assertion":[{"value":"8 August 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 June 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The authors declare no ethical issue. In particular, authors make sure to respect third parties\u2019 rights such as copyright and moral rights.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval and informed consent"}}]}}