{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,7,9]],"date-time":"2024-07-09T09:34:54Z","timestamp":1720517694555},"reference-count":34,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2024,3,1]],"date-time":"2024-03-01T00:00:00Z","timestamp":1709251200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2024,3,1]],"date-time":"2024-03-01T00:00:00Z","timestamp":1709251200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2024,3,1]],"date-time":"2024-03-01T00:00:00Z","timestamp":1709251200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2024,3,1]],"date-time":"2024-03-01T00:00:00Z","timestamp":1709251200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2024,3,1]],"date-time":"2024-03-01T00:00:00Z","timestamp":1709251200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,3,1]],"date-time":"2024-03-01T00:00:00Z","timestamp":1709251200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Information Fusion"],"published-print":{"date-parts":[[2024,3]]},"DOI":"10.1016\/j.inffus.2023.102090","type":"journal-article","created":{"date-parts":[[2023,10,21]],"date-time":"2023-10-21T19:25:11Z","timestamp":1697916311000},"page":"102090","update-policy":"http:\/\/dx.doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":1,"special_numbering":"C","title":["Virtual lighting environment and real human fusion based on multiview videos"],"prefix":"10.1016","volume":"103","author":[{"given":"Aijia","family":"Zhang","sequence":"first","affiliation":[]},{"ORCID":"http:\/\/orcid.org\/0000-0002-1377-7084","authenticated-orcid":false,"given":"Weiqiang","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Zhiguo","family":"Wan","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Hua","sequence":"additional","affiliation":[]},{"given":"Zisong","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.inffus.2023.102090_b1","doi-asserted-by":"crossref","unstructured":"S. Peng, Y. Zhang, Y. Xu, Q. Wang, Q. Shuai, H. Bao, X. Zhou, Neural body: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 9054\u20139063.","DOI":"10.1109\/CVPR46437.2021.00894"},{"key":"10.1016\/j.inffus.2023.102090_b2","doi-asserted-by":"crossref","unstructured":"S. Peng, J. Dong, Q. Wang, S. Zhang, Q. Shuai, X. Zhou, H. Bao, Animatable neural radiance fields for modeling dynamic human bodies, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 14314\u201314323.","DOI":"10.1109\/ICCV48922.2021.01405"},{"key":"10.1016\/j.inffus.2023.102090_b3","doi-asserted-by":"crossref","unstructured":"Z. Zheng, H. Huang, T. Yu, H. Zhang, Y. Guo, Y. Liu, Structured local radiance fields for human avatar modeling, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 15893\u201315903.","DOI":"10.1109\/CVPR52688.2022.01543"},{"key":"10.1016\/j.inffus.2023.102090_b4","doi-asserted-by":"crossref","unstructured":"Y. Feng, J. Yang, M. Pollefeys, M.J. Black, T. Bolkart, Capturing and Animation of Body and Clothing from Monocular Video, in: SIGGRAPH Asia 2022 Conference Papers, 2022, pp. 1\u20139.","DOI":"10.1145\/3550469.3555423"},{"issue":"6","key":"10.1016\/j.inffus.2023.102090_b5","first-page":"1","article-title":"The relightables: Volumetric performance capture of humans with realistic relighting","volume":"38","author":"Guo","year":"2019","journal-title":"ACM Trans. Graph. (ToG)"},{"issue":"1","key":"10.1016\/j.inffus.2023.102090_b6","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3446328","article-title":"Neural light transport for relighting and view synthesis","volume":"40","author":"Zhang","year":"2021","journal-title":"ACM Trans. Graph."},{"issue":"4","key":"10.1016\/j.inffus.2023.102090_b7","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3450626.3459872","article-title":"Total relighting: learning to relight portraits for background replacement","volume":"40","author":"Pandey","year":"2021","journal-title":"ACM Trans. Graph."},{"issue":"4","key":"10.1016\/j.inffus.2023.102090_b8","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3306346.3323027","article-title":"Deep reflectance fields: high-quality facial reflectance field inference from color gradient illumination","volume":"38","author":"Meka","year":"2019","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.inffus.2023.102090_b9","doi-asserted-by":"crossref","unstructured":"T. Zhou, K. He, D. Wu, T. Xu, Q. Zhang, K. Shao, W. Chen, L. Xu, J. Yu, Relightable Neural Human Assets from Multi-view Gradient Illuminations, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 4315\u20134327.","DOI":"10.1109\/CVPR52729.2023.00420"},{"issue":"6","key":"10.1016\/j.inffus.2023.102090_b10","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3550454.3555442","article-title":"Learning to relight portrait images via a virtual light stage and synthetic-to-real adaptation","volume":"41","author":"Yeh","year":"2022","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.inffus.2023.102090_b11","doi-asserted-by":"crossref","unstructured":"P. Debevec, T. Hawkins, C. Tchou, H.-P. Duiker, W. Sarokin, M. Sagar, Acquiring the reflectance field of a human face, in: Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques, 2000, pp. 145\u2013156.","DOI":"10.1145\/344779.344855"},{"key":"10.1016\/j.inffus.2023.102090_b12","article-title":"Relighting humans: occlusion-aware inverse rendering for full-body human images","volume":"37","author":"Yoshihiro","year":"2018","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.inffus.2023.102090_b13","unstructured":"M. Lagunas, X. Sun, J. Yang, R. Villegas, J. Zhang, Z. Shu, B. Masi\u00e1, D. Gutierrez, Single-image Full-body Human Relighting, in: Eurographics Symposium on Rendering, 2021."},{"key":"10.1016\/j.inffus.2023.102090_b14","series-title":"Computer Vision\u2013ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part XVI","first-page":"388","article-title":"Geometry-aware single-image full-body human relighting","author":"Ji","year":"2022"},{"key":"10.1016\/j.inffus.2023.102090_b15","doi-asserted-by":"crossref","unstructured":"X. Pan, A. Tewari, L. Liu, C. Theobalt, GAN2X: Non-Lambertian Inverse Rendering of Image GANs, in: International Conference on 3D Vision, (3DV), 2022.","DOI":"10.1109\/3DV57658.2022.00081"},{"key":"10.1016\/j.inffus.2023.102090_b16","series-title":"Computer Vision\u2013ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part XIV","first-page":"606","article-title":"Relighting4d: Neural relightable human from videos","author":"Chen","year":"2022"},{"key":"10.1016\/j.inffus.2023.102090_b17","doi-asserted-by":"crossref","unstructured":"J.P. Lewis, M. Cordner, N. Fong, Pose space deformation: a unified approach to shape interpolation and skeleton-driven deformation, in: Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques, 2000, pp. 165\u2013172.","DOI":"10.1145\/344779.344862"},{"issue":"7","key":"10.1016\/j.inffus.2023.102090_b18","doi-asserted-by":"crossref","first-page":"1325","DOI":"10.1109\/TPAMI.2013.248","article-title":"Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments","volume":"36","author":"Ionescu","year":"2013","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.inffus.2023.102090_b19","doi-asserted-by":"crossref","unstructured":"S. Saito, Z. Huang, R. Natsume, S. Morishima, A. Kanazawa, H. Li, Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2019, pp. 2304\u20132314.","DOI":"10.1109\/ICCV.2019.00239"},{"issue":"6","key":"10.1016\/j.inffus.2023.102090_b20","doi-asserted-by":"crossref","first-page":"3170","DOI":"10.1109\/TPAMI.2021.3050505","article-title":"Pamir: Parametric model-conditioned implicit representation for image-based human reconstruction","volume":"44","author":"Zheng","year":"2021","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"6","key":"10.1016\/j.inffus.2023.102090_b21","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/2816795.2818013","article-title":"SMPL: A skinned multi-person linear model","volume":"34","author":"Loper","year":"2015","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"10.1016\/j.inffus.2023.102090_b22","doi-asserted-by":"crossref","unstructured":"T. Wang, B. Zhang, T. Zhang, S. Gu, J. Bao, T. Baltrusaitis, J. Shen, D. Chen, F. Wen, Q. Chen, et al., Rodin: A generative model for sculpting 3d digital avatars using diffusion, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 4563\u20134573.","DOI":"10.1109\/CVPR52729.2023.00443"},{"key":"10.1016\/j.inffus.2023.102090_b23","doi-asserted-by":"crossref","unstructured":"Z. Shu, E. Yumer, S. Hadap, K. Sunkavalli, E. Shechtman, D. Samaras, Neural face editing with intrinsic image disentangling, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 5541\u20135550.","DOI":"10.1109\/CVPR.2017.578"},{"issue":"1","key":"10.1016\/j.inffus.2023.102090_b24","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1145\/3503250","article-title":"Nerf: Representing scenes as neural radiance fields for view synthesis","volume":"65","author":"Mildenhall","year":"2021","journal-title":"Commun. ACM"},{"issue":"6","key":"10.1016\/j.inffus.2023.102090_b25","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3478513.3480500","article-title":"Nerfactor: Neural factorization of shape and reflectance under an unknown illumination","volume":"40","author":"Zhang","year":"2021","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.inffus.2023.102090_b26","doi-asserted-by":"crossref","unstructured":"K. Zhang, F. Luan, Q. Wang, K. Bala, N. Snavely, Physg: Inverse rendering with spherical gaussians for physics-based material editing and relighting, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 5453\u20135462.","DOI":"10.1109\/CVPR46437.2021.00541"},{"key":"10.1016\/j.inffus.2023.102090_b27","doi-asserted-by":"crossref","unstructured":"P.P. Srinivasan, B. Deng, X. Zhang, M. Tancik, B. Mildenhall, J.T. Barron, Nerv: Neural reflectance and visibility fields for relighting and view synthesis, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 7495\u20137504.","DOI":"10.1109\/CVPR46437.2021.00741"},{"key":"10.1016\/j.inffus.2023.102090_b28","doi-asserted-by":"crossref","unstructured":"M. Boss, R. Braun, V. Jampani, J.T. Barron, C. Liu, H. Lensch, Nerd: Neural reflectance decomposition from image collections, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 12684\u201312694.","DOI":"10.1109\/ICCV48922.2021.01245"},{"key":"10.1016\/j.inffus.2023.102090_b29","doi-asserted-by":"crossref","unstructured":"Q. Shuai, C. Geng, Q. Fang, S. Peng, W. Shen, X. Zhou, H. Bao, Novel view synthesis of human interactions from sparse multi-view videos, in: ACM SIGGRAPH 2022 Conference Proceedings, 2022, pp. 1\u201310.","DOI":"10.1145\/3528233.3530704"},{"key":"10.1016\/j.inffus.2023.102090_b30","doi-asserted-by":"crossref","unstructured":"Z. Huang, Y. Xu, C. Lassner, H. Li, T. Tung, Arch: Animatable reconstruction of clothed humans, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 3093\u20133102.","DOI":"10.1109\/CVPR42600.2020.00316"},{"key":"10.1016\/j.inffus.2023.102090_b31","first-page":"12909","article-title":"Loopreg: Self-supervised learning of implicit surface correspondences, pose and shape for 3d human mesh registration","volume":"33","author":"Bhatnagar","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.inffus.2023.102090_b32","doi-asserted-by":"crossref","unstructured":"S. Sengupta, J. Gu, K. Kim, G. Liu, D.W. Jacobs, J. Kautz, Neural inverse rendering of an indoor scene from a single image, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2019, pp. 8598\u20138607.","DOI":"10.1109\/ICCV.2019.00869"},{"key":"10.1016\/j.inffus.2023.102090_b33","unstructured":"J.B. Diederik P. Kingma, Adam: A Method for Stochastic Optimization, in: 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015."},{"issue":"7","key":"10.1016\/j.inffus.2023.102090_b34","doi-asserted-by":"crossref","first-page":"1325","DOI":"10.1109\/TPAMI.2013.248","article-title":"Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments","volume":"36","author":"Ionescu","year":"2013","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."}],"container-title":["Information Fusion"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1566253523004062?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1566253523004062?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2023,12,2]],"date-time":"2023-12-02T11:48:01Z","timestamp":1701517681000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1566253523004062"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3]]},"references-count":34,"alternative-id":["S1566253523004062"],"URL":"https:\/\/doi.org\/10.1016\/j.inffus.2023.102090","relation":{},"ISSN":["1566-2535"],"issn-type":[{"value":"1566-2535","type":"print"}],"subject":[],"published":{"date-parts":[[2024,3]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Virtual lighting environment and real human fusion based on multiview videos","name":"articletitle","label":"Article Title"},{"value":"Information Fusion","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.inffus.2023.102090","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2023 Elsevier B.V. All rights reserved.","name":"copyright","label":"Copyright"}],"article-number":"102090"}}