{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,6,17]],"date-time":"2024-06-17T02:13:37Z","timestamp":1718590417577},"reference-count":35,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2024,5,2]],"date-time":"2024-05-02T00:00:00Z","timestamp":1714608000000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100002855","name":"Ministry of Science and Technology of the People's Republic of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100002855","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2019YFB1406201"],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62071434"],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Array"],"published-print":{"date-parts":[[2024,7]]},"DOI":"10.1016\/j.array.2024.100348","type":"journal-article","created":{"date-parts":[[2024,5,5]],"date-time":"2024-05-05T14:05:26Z","timestamp":1714917926000},"page":"100348","update-policy":"http:\/\/dx.doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"title":["D2MNet for music generation joint driven by facial expressions and dance movements"],"prefix":"10.1016","volume":"22","author":[{"given":"Jiang","family":"Huang","sequence":"first","affiliation":[]},{"given":"Xianglin","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Lifang","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Zhulin","family":"Tao","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"Issue 23","key":"10.1016\/j.array.2024.100348_bib1","doi-asserted-by":"crossref","first-page":"5192","DOI":"10.1016\/j.cub.2021.09.043","article-title":"Visual and auditory brain areas share a representational structure that supports emotion perception","volume":"31","author":"Sievers","year":"2021","journal-title":"Curr Biol"},{"key":"10.1016\/j.array.2024.100348_bib2","first-page":"3933","article-title":"Adversarial incomplete multi-view clustering","author":"Xu","year":"2019","journal-title":"IJCAI"},{"issue":"2","key":"10.1016\/j.array.2024.100348_bib3","doi-asserted-by":"crossref","first-page":"1456","DOI":"10.1109\/TII.2022.3206343","article-title":"Uncertainty-aware multi-view deep learning for Internet of Things applications","volume":"19","author":"Xu","year":"2023","journal-title":"IEEE Trans Ind Inf"},{"key":"10.1016\/j.array.2024.100348_bib4","series-title":"IEEE conference on computer vision and pattern recognition","first-page":"7574","article-title":"Audio to body dynamics[C]","author":"Shlizerman","year":"2018"},{"key":"10.1016\/j.array.2024.100348_bib5","series-title":"ACM Multimedia","first-page":"1598","article-title":"Dance with melody: an LSTM-autoencoder approach to music-oriented dance synthesis[C]","author":"Tang","year":"2018"},{"key":"10.1016\/j.array.2024.100348_bib6","first-page":"32","article-title":"Dancing to music","author":"Lee","year":"2019","journal-title":"Adv Neural Inf Process Syst"},{"key":"10.1016\/j.array.2024.100348_bib7","series-title":"Aaai","article-title":"Reliable Conflictive multi-view learning","author":"Xu","year":"2024"},{"key":"10.1016\/j.array.2024.100348_bib8","article-title":"Dance2music: automatic dance-driven music generation","author":"Aggarwal","year":"2021","journal-title":"arXiv preprint arXiv:2107.06252"},{"key":"10.1016\/j.array.2024.100348_bib9","series-title":"Proceedings of the IEEE conference on computer vision and pattern recognition","first-page":"3497","article-title":"Learning individual styles of conversational gesture[C]","author":"Ginosar","year":"2019"},{"key":"10.1016\/j.array.2024.100348_bib10","first-page":"29","article-title":"Conditional image generation with pixelcnn decoders","author":"Van den Oord","year":"2016","journal-title":"Adv Neural Inf Process Syst"},{"key":"10.1016\/j.array.2024.100348_bib11","first-page":"758","article-title":"Foley music: learning to generate music from videos[C]","author":"Gan","year":"2020","journal-title":"ECCV"},{"key":"10.1016\/j.array.2024.100348_bib12","series-title":"Acmmm","first-page":"2037","article-title":"Video background music generation with controllable music transformer[C]","author":"Di","year":"2021"},{"key":"10.1016\/j.array.2024.100348_bib13","first-page":"182","article-title":"Quantized GAN for complex music generation from dance videos[C]","author":"Zhu","year":"2022","journal-title":"ECCV"},{"issue":"1","key":"10.1016\/j.array.2024.100348_bib14","first-page":"119","article-title":"Electrocardiogram signal classification using VGGNet: a neural network based classification model","volume":"15","author":"Goswami","year":"2023","journal-title":"Int J Inf Technol"},{"issue":"6","key":"10.1016\/j.array.2024.100348_bib15","doi-asserted-by":"crossref","first-page":"1756","DOI":"10.1111\/tgis.12681","article-title":"A complex junction recognition method based on GoogLeNet model","volume":"24","author":"Li","year":"2020","journal-title":"Trans GIS"},{"issue":"12","key":"10.1016\/j.array.2024.100348_bib16","doi-asserted-by":"crossref","first-page":"5349","DOI":"10.1109\/TNNLS.2020.2966319","article-title":"Why resnet works? residuals generalize","volume":"31","author":"He","year":"2020","journal-title":"IEEE Transact Neural Networks Learn Syst"},{"key":"10.1016\/j.array.2024.100348_bib17","first-page":"657","article-title":"Evaluation of impressions and movements related to negative emotional expressions in dance[C]","author":"Shikanai","year":"2015","journal-title":"ICCAS. IEEE"},{"key":"10.1016\/j.array.2024.100348_bib18","series-title":"The Annual Meeting of the Cognitive Science Society","first-page":"1026","article-title":"Relations between body motion and emotion: analysis based on Laban Movement Analysis[C]","author":"Morita","year":"2013"},{"issue":"3\u20134","key":"10.1016\/j.array.2024.100348_bib19","doi-asserted-by":"crossref","first-page":"311","DOI":"10.1002\/cav.1714","article-title":"Continuous body emotion recognition system during theater performances","volume":"27","author":"Senecal","year":"2016","journal-title":"Comput Animat Virtual Worlds"},{"key":"10.1016\/j.array.2024.100348_bib20","series-title":"A review of automatic music generation based on performance RNN[J]","author":"Cao","year":"2023"},{"key":"10.1016\/j.array.2024.100348_bib21","article-title":"Music transformer[J]","author":"Huang","year":"2018","journal-title":"arXiv preprint arXiv:1809.04281"},{"key":"10.1016\/j.array.2024.100348_bib22","article-title":"Wavenet: a generative model for raw audio","author":"Oord","year":"2016","journal-title":"arXiv preprint arXiv:1609.03499"},{"issue":"4","key":"10.1016\/j.array.2024.100348_bib23","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3197517.3201371","article-title":"Visual rhythm and beat","volume":"37","author":"Davis","year":"2018","journal-title":"ACM Trans Graph"},{"key":"10.1016\/j.array.2024.100348_bib24","doi-asserted-by":"crossref","first-page":"124928","DOI":"10.1109\/ACCESS.2020.3007956","article-title":"Dance emotion recognition based on laban motion analysis using convolutional neural network and long short-term memory","volume":"8","author":"Wang","year":"2020","journal-title":"IEEE Access"},{"key":"10.1016\/j.array.2024.100348_bib25","first-page":"29","article-title":"Soundnet: learning sound representations from unlabeled video","author":"Aytar","year":"2016","journal-title":"Adv Neural Inf Process Syst"},{"key":"10.1016\/j.array.2024.100348_bib26","series-title":"Proceedings of the IEEE conference on computer vision and pattern recognition","first-page":"580","article-title":"Rich feature hierarchies for accurate object detection and semantic segmentation[C]","author":"Girshick","year":"2014"},{"key":"10.1016\/j.array.2024.100348_bib27","series-title":"ICCV","first-page":"13401","article-title":"Ai choreographer: music conditioned 3d dance generation with aist++[C]","author":"Li","year":"2021"},{"key":"10.1016\/j.array.2024.100348_bib28","series-title":"Proceedings of the IEEE\/CVF international conference on computer vision","first-page":"5442","article-title":"AMASS: Archive of motion capture as surface shapes[C]","author":"Mahmood","year":"2019"},{"issue":"7","key":"10.1016\/j.array.2024.100348_bib29","doi-asserted-by":"crossref","first-page":"1325","DOI":"10.1109\/TPAMI.2013.248","article-title":"Human3. 6m: large scale datasets and predictive methods for 3d human sensing in natural environments","volume":"36","author":"Ionescu","year":"2013","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"17","key":"10.1016\/j.array.2024.100348_bib30","first-page":"26","article-title":"GrooveNet: real-time music-driven dance movement generation using artificial neural networks","volume":"8","author":"Alemi","year":"2017","journal-title":"Networks"},{"issue":"2","key":"10.1016\/j.array.2024.100348_bib31","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3485664","article-title":"Music2dance: Dancenet for music-driven dance generation","volume":"18","author":"Zhuang","year":"2022","journal-title":"ACM Trans Multimed Comput Commun Appl"},{"key":"10.1016\/j.array.2024.100348_bib32","series-title":"Proceedings of the IEEE conference on computer vision and pattern recognition","first-page":"7291","article-title":"Realtime multi-person 2d pose estimation using part affinity fields[C]","author":"Cao","year":"2017"},{"issue":"1","key":"10.1016\/j.array.2024.100348_bib33","doi-asserted-by":"crossref","first-page":"51","DOI":"10.1080\/09298210701653344","article-title":"Beat tracking by dynamic programming","volume":"36","author":"Ellis","year":"2007","journal-title":"J N Music Res"},{"key":"10.1016\/j.array.2024.100348_bib34","series-title":"2017 IEEE international conference on acoustics, speech and signal processing (ICASSP)","first-page":"776","article-title":"Audio set: an ontology and human-labeled dataset for audio events[C]","author":"Gemmeke","year":"2017"},{"issue":"5","key":"10.1016\/j.array.2024.100348_bib35","first-page":"5","article-title":"Music audio Sentiment classification based on CNN-LSTM","volume":"52","author":"Chen","year":"2019","journal-title":"Commun Technol"}],"container-title":["Array"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2590005624000146?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2590005624000146?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2024,6,17]],"date-time":"2024-06-17T01:32:05Z","timestamp":1718587925000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S2590005624000146"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7]]},"references-count":35,"alternative-id":["S2590005624000146"],"URL":"https:\/\/doi.org\/10.1016\/j.array.2024.100348","relation":{},"ISSN":["2590-0056"],"issn-type":[{"value":"2590-0056","type":"print"}],"subject":[],"published":{"date-parts":[[2024,7]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"D2MNet for music generation joint driven by facial expressions and dance movements","name":"articletitle","label":"Article Title"},{"value":"Array","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.array.2024.100348","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2024 The Authors. Published by Elsevier Inc.","name":"copyright","label":"Copyright"}],"article-number":"100348"}}