{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,6]],"date-time":"2024-09-06T08:46:30Z","timestamp":1725612390861},"reference-count":72,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2018,1,30]],"date-time":"2018-01-30T00:00:00Z","timestamp":1517270400000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61622115","61472281"],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Program for Professor of Special Appointment (Eastern Scholar) at Shanghai Institutions of Higher Learning","award":["GZ2015005"]},{"name":"Engineering Research Center of Industrial Vision Perception & Intelligent Computing","award":["17DZ2251600"]},{"name":"Key Research and Development Project of Jiangxi Provincial Department of Science and Technology","award":["20171BBE50065"]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2019,5]]},"DOI":"10.1007\/s11042-018-5662-9","type":"journal-article","created":{"date-parts":[[2018,1,29]],"date-time":"2018-01-29T23:31:00Z","timestamp":1517268660000},"page":"13331-13350","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":24,"title":["Multi-modal learning for affective content analysis in movies"],"prefix":"10.1007","volume":"78","author":[{"given":"Yun","family":"Yi","sequence":"first","affiliation":[]},{"given":"Hanli","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2018,1,30]]},"reference":[{"issue":"9","key":"5662_CR1","doi-asserted-by":"publisher","first-page":"11,809","DOI":"10.1007\/s11042-016-3618-5","volume":"76","author":"E Acar","year":"2017","unstructured":"Acar E, Hopfgartner F, Albayrak S (2017) A comprehensive study on mid-level representation and ensemble learning for emotional analysis of video material. Multimed Tools Appl 76(9):11,809\u201311,837","journal-title":"Multimed Tools Appl"},{"key":"5662_CR2","unstructured":"Anastasia T, Leontios H (2016) AUTH-SGP in MediaEval 2016 emotional impact of movies task. In: MediaEval 2016 Workshop"},{"issue":"2","key":"5662_CR3","doi-asserted-by":"publisher","first-page":"411","DOI":"10.1002\/mrm.20965","volume":"56","author":"V Arsigny","year":"2006","unstructured":"Arsigny V, Fillard P, Pennec X, Ayache N (2006) Log-Euclidean metrics for fast and simple calculus on diffusion tensors. Magn Resonan Med 56(2):411\u2013421","journal-title":"Magn Resonan Med"},{"key":"5662_CR4","doi-asserted-by":"crossref","unstructured":"Baecchi C, Uricchio T, Bertini M, Del Bimbo A (2017) Deep sentiment features of context and faces for affective video analysis. In: ICMR\u201917, pp 72\u201377","DOI":"10.1145\/3078971.3079027"},{"issue":"1","key":"5662_CR5","doi-asserted-by":"publisher","first-page":"43","DOI":"10.1109\/TAFFC.2015.2396531","volume":"6","author":"Y Baveye","year":"2015","unstructured":"Baveye Y, Dellandrea E, Chamaret C, Chen L (2015) LIRIS-ACCEDE: a video database for affective content analysis. IEEE Trans Affect Comput 6(1):43\u201355","journal-title":"IEEE Trans Affect Comput"},{"key":"5662_CR6","doi-asserted-by":"crossref","unstructured":"Baveye Y, Chamaret C, Dellandr\u00e9a E, Chen L (2017) Affective video content analysis: a multidisciplinary insight. IEEE Trans Affect Comput","DOI":"10.1109\/TAFFC.2017.2661284"},{"key":"5662_CR7","doi-asserted-by":"crossref","unstructured":"Bosch A, Zisserman A, Munoz X (2007) Image classification using random forests and ferns. In: ICCV\u201907, pp 1\u20138","DOI":"10.1109\/ICCV.2007.4409066"},{"issue":"4","key":"5662_CR8","doi-asserted-by":"publisher","first-page":"636","DOI":"10.1109\/TCSVT.2012.2211935","volume":"23","author":"L Canini","year":"2013","unstructured":"Canini L, Benini S, Leonardi R (2013) Affective recommendation of movies based on selected connotative features. IEEE Trans Circuits Syst Video Technol 23 (4):636\u2013647","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"5662_CR9","unstructured":"Chakraborty R, Maurya AK, Pandharipande M, Hassan E, Ghosh H, Kopparapu SK (2015) TCS-ILAB-MediaEval 2015: affective impact of movies and violent scene detection. In: MediaEval 2015 Workshop"},{"issue":"3","key":"5662_CR10","doi-asserted-by":"publisher","first-page":"27:1","DOI":"10.1145\/1961189.1961199","volume":"2","author":"CC Chang","year":"2011","unstructured":"Chang CC, Lin CJ (2011) LIBSVM: a library for support vector machines. ACM Trans Intell Syst Technol 2(3):27:1\u201327:27","journal-title":"ACM Trans Intell Syst Technol"},{"key":"5662_CR11","unstructured":"Chen S, Jin Q (2016) RUC at MediaEval 2016 emotional impact of movies task: fusion of multimodal features. In: MediaEval 2016 Workshop"},{"key":"5662_CR12","unstructured":"Dai Q, Zhao RW, Wu Z, Wang X, Gu Z, Wu W, Jiang YG (2015) Fudan-Huawei at MediaEval 2015: detecting violent scenes and affective impact in movies with deep learning. In: MediaEval 2015 Workshop"},{"key":"5662_CR13","doi-asserted-by":"crossref","unstructured":"Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: CVPR\u201905, pp 886\u2013893","DOI":"10.1109\/CVPR.2005.177"},{"key":"5662_CR14","doi-asserted-by":"crossref","unstructured":"Dalal N, Triggs B, Schmid C (2006) Human detection using oriented histograms of flow and appearance. In: ECCV\u201906, pp 428\u2013441","DOI":"10.1007\/11744047_33"},{"key":"5662_CR15","unstructured":"Dellandr\u00e9a E, Chen L, Baveye Y, Sj\u00f6berg MV, Chamaret C et al (2016) The mediaeval 2016 emotional impact of movies task. In: MediaEval 2016 Workshop"},{"key":"5662_CR16","doi-asserted-by":"crossref","unstructured":"Eggink J, Bland D (2012) A large scale experiment for mood-based classification of tv programmes. In: ICME\u201912, pp 140\u2013145","DOI":"10.1109\/ICME.2012.68"},{"key":"5662_CR17","unstructured":"Ellis DPW (2005) PLP and RASTA (and MFCC, and inversion) in Matlab. http:\/\/www.ee.columbia.edu\/dpwe\/~resources\/matlab\/rastamat\/ , online web resource"},{"key":"5662_CR18","doi-asserted-by":"crossref","unstructured":"Eyben F, Weninger F, Gross F, Schuller B (2013) Recent developments in opensmile, the munich open-source multimedia feature extractor. In: ACM MM\u201913, pp 835\u2013838","DOI":"10.1145\/2502081.2502224"},{"key":"5662_CR19","first-page":"1871","volume":"9","author":"RE Fan","year":"2008","unstructured":"Fan RE, Chang KW, Hsieh CJ, Wang XR, Lin CJ (2008) LIBLINEAR: a library for large linear classification. J Mach Learn Res 9:1871\u20131874","journal-title":"J Mach Learn Res"},{"key":"5662_CR20","unstructured":"Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: AISTATS\u201910, pp 249\u2013256"},{"issue":"1","key":"5662_CR21","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1109\/TMM.2004.840618","volume":"7","author":"A Hanjalic","year":"2005","unstructured":"Hanjalic A, Xu LQ (2005) Affective video content representation and modeling. IEEE Trans Multimed 7(1):143\u2013154","journal-title":"IEEE Trans Multimed"},{"key":"5662_CR22","first-page":"3323","volume":"13","author":"CH Ho","year":"2012","unstructured":"Ho CH, Lin CJ (2012) Large-scale linear support vector regression. J Mach Learn Res 13:3323\u20133348","journal-title":"J Mach Learn Res"},{"key":"5662_CR23","unstructured":"Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: ICML\u201915, pp 448\u2013456"},{"issue":"6","key":"5662_CR24","doi-asserted-by":"publisher","first-page":"523","DOI":"10.1109\/TMM.2010.2051871","volume":"12","author":"G Irie","year":"2010","unstructured":"Irie G, Satou T, Kojima A, Yamasaki T, Aizawa K (2010) Affective audio-visual words and latent topic driving model for realizing movie affective scene classification. IEEE Trans Multimed 12(6):523\u2013535","journal-title":"IEEE Trans Multimed"},{"key":"5662_CR25","unstructured":"Jan A, Gaus YFBA, Meng H, Zhang F (2016) BUL in MediaEval 2016 emotional impact of movies task. In: MediaEval 2016 Workshop"},{"key":"5662_CR26","doi-asserted-by":"crossref","unstructured":"Jia Y, Shelhamer E, Donahue J, Karayev S, Long J, Girshick R, Guadarrama S, Darrell T (2014) Caffe: convolutional architecture for fast feature embedding. In: ACM MM\u201914, pp 675\u2013678","DOI":"10.1145\/2647868.2654889"},{"key":"5662_CR27","doi-asserted-by":"crossref","unstructured":"Jiang YG, Xu B, Xue X (2014) Predicting emotions in user-generated videos. In: AAAI\u201914, pp 73\u201379","DOI":"10.1609\/aaai.v28i1.8724"},{"key":"5662_CR28","unstructured":"Lam V, Phan S, Le DD, Satoh S, Duong DA (2015) NII-UIT at MediaEval 2015 affective impact of movies task. In: MediaEval 2015 Workshop"},{"key":"5662_CR29","doi-asserted-by":"crossref","unstructured":"Laptev I, Marszalek M, Schmid C, Rozenfeld B (2008) Learning realistic human actions from movies. In: CVPR\u201908, pp 1\u20138","DOI":"10.1109\/CVPR.2008.4587756"},{"issue":"22","key":"5662_CR30","doi-asserted-by":"publisher","first-page":"14,399","DOI":"10.1007\/s11042-016-3608-7","volume":"75","author":"C Li","year":"2016","unstructured":"Li C, Feng Z, Xu C (2016) Error-correcting output codes for multi-label emotion classification. Multimed Tools Appl 75(22):14,399\u201314,416","journal-title":"Multimed Tools Appl"},{"issue":"2","key":"5662_CR31","first-page":"561","volume":"9","author":"CJ Lin","year":"2007","unstructured":"Lin CJ, Weng RC, Keerthi SS (2007) Trust region newton method for large-scale logistic regression. J Mach Learn Res 9(2):561\u2013568","journal-title":"J Mach Learn Res"},{"key":"5662_CR32","unstructured":"Liu Y, Gu Z, Zhang Y, Liu Y (2016) Mining emotional features of movies. In: MediaEval 2016 Workshop"},{"issue":"2","key":"5662_CR33","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1023\/B:VISI.0000029664.99615.94","volume":"60","author":"DG Lowe","year":"2004","unstructured":"Lowe DG (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60(2):91\u2013110","journal-title":"Int J Comput Vis"},{"issue":"4","key":"5662_CR34","doi-asserted-by":"publisher","first-page":"1706","DOI":"10.1109\/TIP.2014.2307478","volume":"23","author":"J Ma","year":"2014","unstructured":"Ma J, Zhao J, Tian J, Yuille AL, Tu Z (2014) Robust point matching via vector field consensus. IEEE Trans Image Process 23(4):1706\u20131721","journal-title":"IEEE Trans Image Process"},{"key":"5662_CR35","unstructured":"Ma Y, Ye Z, Xu M (2016) THU-HCSI at MediaEval 2016: emotional impact of movies task. In: MediaEval 2016 workshop"},{"key":"5662_CR36","unstructured":"Marin Vlastelica P, Hayrapetyan S, Tapaswi M, Stiefelhagen R (2015) KIT at MediaEval 2015\u2013evaluating visual cues for affective impact of movies task. In: MediaEval 2015 workshop"},{"key":"5662_CR37","unstructured":"Mironica I, Ionescu B, Sj\u00f6berg M, Schedl M, Skowron M (2015) RFA at MediaEval 2015 affective impact of movies task: a multimodal approach. In: MediaEval 2015 workshop"},{"key":"5662_CR38","first-page":"2825","volume":"12","author":"F Pedregosa","year":"2011","unstructured":"Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, Blondel M, Prettenhofer P, Weiss R, Dubourg V, Vanderplas J, Passos A, Cournapeau D, Brucher M, Perrot M, Duchesnay E (2011) Scikit-learn: machine learning in python. J Mach Learn Res 12:2825\u20132830","journal-title":"J Mach Learn Res"},{"key":"5662_CR39","doi-asserted-by":"publisher","first-page":"104","DOI":"10.1016\/j.neunet.2014.10.005","volume":"63","author":"S Poria","year":"2015","unstructured":"Poria S, Cambria E, Hussain A, Huang GB (2015) Towards an intelligent framework for multimodal affective data analysis. Neural Netw 63:104\u2013116","journal-title":"Neural Netw"},{"issue":"3","key":"5662_CR40","doi-asserted-by":"publisher","first-page":"222","DOI":"10.1007\/s11263-013-0636-x","volume":"105","author":"J Sanchez","year":"2013","unstructured":"Sanchez J, Perronnin F, Mensink T, Verbeek J (2013) Image classification with the fisher vector: theory and practice. Int J Comput Vis 105(3):222\u2013245","journal-title":"Int J Comput Vis"},{"key":"5662_CR41","doi-asserted-by":"crossref","unstructured":"Sang J, Xu C (2012) Right buddy makes the difference: an early exploration of social relation analysis in multimedia applications. In: ACM MM\u201912, pp 19\u201328","DOI":"10.1145\/2393347.2393358"},{"issue":"3","key":"5662_CR42","doi-asserted-by":"publisher","first-page":"883","DOI":"10.1109\/TMM.2012.2188782","volume":"14","author":"J Sang","year":"2012","unstructured":"Sang J, Xu C, Liu J (2012) User-aware image tag refinement via ternary semantic analysis. IEEE Trans Multimed 14(3):883\u2013895","journal-title":"IEEE Trans Multimed"},{"key":"5662_CR43","doi-asserted-by":"crossref","unstructured":"Schuller B, Steidl S, Batliner A, Burkhardt F, Devillers L, M\u00fcller CA, Narayanan SS (2010) The INTERSPEECH 2010 paralinguistic challenge. In: INTERSPEECH\u201910","DOI":"10.21437\/Interspeech.2010-739"},{"key":"5662_CR44","unstructured":"Seddati O, Kulah E, Pironkov G, Dupont S, Mahmoudi S, Dutoit T (2015) UMons at MediaEval 2015 affective impact of movies task including violent scenes detection. In: MediaEval 2015 workshop"},{"key":"5662_CR45","unstructured":"Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. In: NIPS\u201914, pp 568\u2013576"},{"key":"5662_CR46","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv: 14091556"},{"key":"5662_CR47","unstructured":"Sj\u00f6berg M, Baveye Y, Wang H, Quang VL, Ionescu B, Dellandr\u00e9a E, Schedl M, Demarty CH, Chen L (2015) The MediaEval 2015 affective impact of movies task. In: MediaEval 2015 workshop"},{"key":"5662_CR48","doi-asserted-by":"crossref","unstructured":"Snoek CG, Worring M, Smeulders AW (2005) Early versus late fusion in semantic video analysis. In: ACM MM\u201905, pp 399\u2013402","DOI":"10.1145\/1101149.1101236"},{"key":"5662_CR49","unstructured":"Soomro K, Zamir AR, Shah M (2012) UCF101: a dataset of 101 human actions classes from videos in the wild. CRCV-TR-12-01"},{"key":"5662_CR50","doi-asserted-by":"crossref","unstructured":"Sun K, Yu J (2007) Video affective content representation and recognition using video affective tree and hidden markov models. In: ACII\u201907, pp 594\u2013605","DOI":"10.1007\/978-3-540-74889-2_52"},{"key":"5662_CR51","doi-asserted-by":"crossref","unstructured":"Sural S, Qian G, Pramanik S (2002) Segmentation and histogram generation using the HSV color space for image retrieval. In: ICIP\u201902, pp 589\u2013592","DOI":"10.1109\/ICIP.2002.1040019"},{"key":"5662_CR52","doi-asserted-by":"crossref","unstructured":"Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: CVPR\u201916, pp 2818\u20132826","DOI":"10.1109\/CVPR.2016.308"},{"issue":"1","key":"5662_CR53","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1007\/s11042-010-0702-0","volume":"61","author":"RMA Teixeira","year":"2012","unstructured":"Teixeira RMA, Yamasaki T, Aizawa K (2012) Determination of emotional content of video clips by low-level audiovisual features. Multimed Tools Appl 61(1):21\u201349","journal-title":"Multimed Tools Appl"},{"key":"5662_CR54","doi-asserted-by":"crossref","unstructured":"Tieleman T (2008) Training restricted boltzmann machines using approximations to the likelihood gradient. In: ICML\u201908, pp 1064\u20131071","DOI":"10.1145\/1390156.1390290"},{"key":"5662_CR55","unstructured":"Trigeorgis G, Coutinho E, Ringeval F, Marchi E, Zafeiriou S, Schuller B (2015) The ICL-TUM-PASSAU approach for the MediaEval 2015 affective impact of movies task. In: MediaEval 2015 Workshop"},{"key":"5662_CR56","doi-asserted-by":"crossref","unstructured":"Vedaldi A, Fulkerson B (2010) Vlfeat: an open and portable library of computer vision algorithms. In: ACM MM\u201910, pp 1469\u20131472","DOI":"10.1145\/1873951.1874249"},{"key":"5662_CR57","unstructured":"Verma GK, Tiwary US (2016) Affect representation and recognition in 3d continuous valence\u2013arousal\u2013dominance space. Multimed Tools Appl 1\u201325"},{"issue":"6","key":"5662_CR58","doi-asserted-by":"publisher","first-page":"689","DOI":"10.1109\/TCSVT.2006.873781","volume":"16","author":"HL Wang","year":"2006","unstructured":"Wang HL, Cheong LF (2006) Affective understanding in film. IEEE Trans Circ Syst Video Technol 16(6):689\u2013704","journal-title":"IEEE Trans Circ Syst Video Technol"},{"issue":"4","key":"5662_CR59","doi-asserted-by":"publisher","first-page":"410","DOI":"10.1109\/TAFFC.2015.2432791","volume":"6","author":"S Wang","year":"2015","unstructured":"Wang S, Ji Q (2015) Video affective content analysis: a survey of state-of-the-art methods. IEEE Trans Affect Comput 6(4):410\u2013430","journal-title":"IEEE Trans Affect Comput"},{"key":"5662_CR60","doi-asserted-by":"crossref","unstructured":"Wang H, Kl\u00e4ser A, Schmid C, Liu CL (2011) Action recognition by dense trajectories. In: CVPR\u201911, pp 3169\u20133176","DOI":"10.1109\/CVPR.2011.5995407"},{"key":"5662_CR61","doi-asserted-by":"crossref","unstructured":"Wang H, Yi Y, Wu J (2015) Human action recognition with trajectory based covariance descriptor in unconstrained videos. In: ACM MM\u201915, pp 1175\u20131178","DOI":"10.1145\/2733373.2806310"},{"key":"5662_CR62","doi-asserted-by":"crossref","unstructured":"Wang L, Xiong Y, Wang Z, Qiao Y, Lin D, Tang X, Van Gool L (2016) Temporal segment networks: towards good practices for deep action recognition. In: ECCV\u201916, pp 20\u201336","DOI":"10.1007\/978-3-319-46484-8_2"},{"issue":"2","key":"5662_CR63","doi-asserted-by":"publisher","first-page":"757","DOI":"10.1007\/s11042-012-1046-8","volume":"70","author":"M Xu","year":"2014","unstructured":"Xu M, Wang J, He X, Jin J S, Luo S, Lu H (2014) A three-level framework for affective content analysis and its case studies. Multimed Tools Appl 70 (2):757\u2013779","journal-title":"Multimed Tools Appl"},{"key":"5662_CR64","unstructured":"Yi Y, Wang H (2017) Motion keypoint trajectory and covariance descriptor for human action recognition. Vis Comput 1\u201313"},{"key":"5662_CR65","unstructured":"Yi Y, Wang H, Zhang B, Yu J (2015) MIC-TJU in MediaEval 2015 affective impact of movies task. In: MediaEval 2015 workshop"},{"issue":"1","key":"5662_CR66","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1007\/s10994-010-5221-8","volume":"85","author":"HF Yu","year":"2011","unstructured":"Yu HF, Huang FL, Lin CJ (2011) Dual coordinate descent methods for logistic regression and maximum entropy models. Mach Learn 85(1):41\u201375","journal-title":"Mach Learn"},{"issue":"5","key":"5662_CR67","doi-asserted-by":"publisher","first-page":"2019","DOI":"10.1109\/TIP.2014.2311377","volume":"23","author":"J Yu","year":"2014","unstructured":"Yu J, Rui Y, Tao D (2014) Click prediction for web image reranking using multimodal sparse coding. IEEE Trans Image Process 23(5):2019\u201332","journal-title":"IEEE Trans Image Process"},{"issue":"12","key":"5662_CR68","doi-asserted-by":"publisher","first-page":"4014","DOI":"10.1109\/TCYB.2016.2591583","volume":"47","author":"J Yu","year":"2017","unstructured":"Yu J, Yang X, Gao F, Tao D (2017) Deep multimodal distance metric learning using click constraints for image ranking. IEEE Trans Cybern 47(12):4014\u20134024","journal-title":"IEEE Trans Cybern"},{"key":"5662_CR69","doi-asserted-by":"crossref","unstructured":"Yu Z, Yu J, Fan J, Tao D (2017) Multi-modal factorized bilinear pooling with co-attention learning for visual question answering. In: ICCV\u201917, pp 1 \u2013 10","DOI":"10.1109\/ICCV.2017.202"},{"key":"5662_CR70","doi-asserted-by":"crossref","unstructured":"Zach C, Pock T, Bischof H (2007) A duality based approach for realtime T V \u2212 L 1 optical flow. In: Joint pattern recognition symposium, pp 214\u2013223","DOI":"10.1007\/978-3-540-74936-3_22"},{"key":"5662_CR71","unstructured":"Zhang S, Tian Q, Jiang S, Huang Q, Gao W (2008) Affective MTV analysis based on arousal and valence features. In: ICME\u201908, pp 1369\u20131372"},{"key":"5662_CR72","doi-asserted-by":"crossref","unstructured":"Zhang S, Tian Q, Huang Q, Gao W, Li S (2009) Utilizing affective analysis for efficient movie browsing. In: ICIP\u201909, pp 1853\u20131856","DOI":"10.1109\/ICIP.2009.5413590"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s11042-018-5662-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-018-5662-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-018-5662-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,8,13]],"date-time":"2022-08-13T12:37:03Z","timestamp":1660394223000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s11042-018-5662-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,1,30]]},"references-count":72,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2019,5]]}},"alternative-id":["5662"],"URL":"https:\/\/doi.org\/10.1007\/s11042-018-5662-9","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2018,1,30]]},"assertion":[{"value":"30 July 2017","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 December 2017","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 January 2018","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 January 2018","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}