{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,7,8]],"date-time":"2024-07-08T15:25:07Z","timestamp":1720452307376},"reference-count":86,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,2,14]],"date-time":"2025-02-14T00:00:00Z","timestamp":1739491200000},"content-version":"am","delay-in-days":319,"URL":"http:\/\/www.elsevier.com\/open-access\/userlicense\/1.0\/"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/100000002","name":"National Institutes of Health","doi-asserted-by":"publisher","award":["R21MH123997"],"id":[{"id":"10.13039\/100000002","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100007698","name":"University of Florida","doi-asserted-by":"publisher","award":["1833908"],"id":[{"id":"10.13039\/100007698","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Computers & Graphics"],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.1016\/j.cag.2024.103888","type":"journal-article","created":{"date-parts":[[2024,2,6]],"date-time":"2024-02-06T07:26:45Z","timestamp":1707204405000},"page":"103888","update-policy":"http:\/\/dx.doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":1,"special_numbering":"C","title":["Towards mitigating uncann(eye)ness in face swaps via gaze-centric loss terms"],"prefix":"10.1016","volume":"119","author":[{"ORCID":"http:\/\/orcid.org\/0000-0003-0944-2641","authenticated-orcid":false,"given":"Ethan","family":"Wilson","sequence":"first","affiliation":[]},{"given":"Frederick","family":"Shic","sequence":"additional","affiliation":[]},{"given":"Sophie","family":"J\u00f6rg","sequence":"additional","affiliation":[]},{"given":"Eakta","family":"Jain","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.cag.2024.103888_b1","series-title":"Proceedings of the 2023 symposium on eye tracking research and applications","isbn-type":"print","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3588015.3588416","article-title":"Introducing explicit gaze constraints to face swapping","author":"Wilson","year":"2023","ISBN":"http:\/\/id.crossref.org\/isbn\/9798400701504"},{"issue":"14","key":"10.1016\/j.cag.2024.103888_b2","doi-asserted-by":"crossref","first-page":"4225","DOI":"10.1167\/jov.22.14.4225","article-title":"The uncanniness of face swaps","volume":"22","author":"Wilson","year":"2022","journal-title":"J Vis","ISSN":"http:\/\/id.crossref.org\/issn\/1534-7362","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b3","series-title":"Advances in artificial intelligence, software and systems engineering","isbn-type":"print","doi-asserted-by":"crossref","first-page":"235","DOI":"10.1007\/978-3-030-51328-3_33","article-title":"Deepfakes for the good: A beneficial application of contentious artificial intelligence technology","author":"Caporusso","year":"2021","ISBN":"http:\/\/id.crossref.org\/isbn\/9783030513283"},{"key":"10.1016\/j.cag.2024.103888_b4","series-title":"Proceedings of the AAAI\/ACM conference on AI, ethics, and society","isbn-type":"print","doi-asserted-by":"crossref","first-page":"414","DOI":"10.1145\/3375627.3375849","article-title":"Deepfakes for medical video de-identification: Privacy protection and diagnostic information preservation","author":"Zhu","year":"2020","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450371100"},{"key":"10.1016\/j.cag.2024.103888_b5","series-title":"Proceedings of the 23rd international ACM SIGACCeSS conference on computers and accessibility","isbn-type":"print","first-page":"1","article-title":"American sign language video anonymization to support online participation of deaf and hard of hearing users","author":"Lee","year":"2021","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450383066"},{"issue":"1","key":"10.1016\/j.cag.2024.103888_b6","doi-asserted-by":"crossref","first-page":"32","DOI":"10.1515\/opis-2019-0003","article-title":"\u201cThe word real is no longer real\u201d: Deepfakes, gender, and the challenges of AI-altered video","volume":"3","author":"Wagner","year":"2019","journal-title":"Open Inf Sci","ISSN":"http:\/\/id.crossref.org\/issn\/2451-1781","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b7","doi-asserted-by":"crossref","unstructured":"Meskys\u00a0E, Kalpokiene\u00a0J, Jurcys\u00a0P, Liaudanskas\u00a0A. Regulating deep fakes: Legal and ethical considerations. Rochester, NY; 2019, URL:.","DOI":"10.1093\/jiplp\/jpz167"},{"key":"10.1016\/j.cag.2024.103888_b8","doi-asserted-by":"crossref","first-page":"83144","DOI":"10.1109\/ACCESS.2020.2988660","article-title":"DeepVision: Deepfakes detection using human eye blinking pattern","volume":"8","author":"Jung","year":"2020","journal-title":"IEEE Access","ISSN":"http:\/\/id.crossref.org\/issn\/2169-3536","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b9","series-title":"2020 IEEE international joint conference on biometrics","first-page":"1","article-title":"How do the hearts of deep fakes beat? Deep fake source detection via interpreting residuals with biological signals","author":"Ciftci","year":"2020"},{"key":"10.1016\/j.cag.2024.103888_b10","series-title":"ACM symposium on eye tracking research and applications","isbn-type":"print","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3448017.3457387","article-title":"Where do deep fakes look? Synthetic face detection via gaze tracking","author":"Demir","year":"2021","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450383448"},{"key":"10.1016\/j.cag.2024.103888_b11","series-title":"2020 IEEE international conference on multimedia & expo workshops","first-page":"1","article-title":"Deepfake detection: Current challenges and next steps","author":"Lyu","year":"2020"},{"key":"10.1016\/j.cag.2024.103888_b12","article-title":"Generative adversarial nets","volume":"vol. 27","author":"Goodfellow","year":"2014"},{"key":"10.1016\/j.cag.2024.103888_b13","series-title":"Auto-encoding variational Bayes","author":"Kingma","year":"2022"},{"key":"10.1016\/j.cag.2024.103888_b14","first-page":"6840","article-title":"Denoising diffusion probabilistic models","volume":"vol. 33","author":"Ho","year":"2020"},{"key":"10.1016\/j.cag.2024.103888_b15","series-title":"Analyzing and improving the image quality of StyleGAN","first-page":"8110","author":"Karras","year":"2020"},{"key":"10.1016\/j.cag.2024.103888_b16","article-title":"Generating diverse high-fidelity images with VQ-VAE-2","volume":"vol. 32","author":"Razavi","year":"2019"},{"key":"10.1016\/j.cag.2024.103888_b17","series-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","author":"Radford","year":"2016"},{"key":"10.1016\/j.cag.2024.103888_b18","article-title":"Coupled generative adversarial networks","volume":"vol. 29","author":"Liu","year":"2016"},{"key":"10.1016\/j.cag.2024.103888_b19","series-title":"High-resolution image synthesis with latent diffusion models","first-page":"10684","author":"Rombach","year":"2022"},{"issue":"5","key":"10.1016\/j.cag.2024.103888_b20","doi-asserted-by":"crossref","first-page":"6259","DOI":"10.1007\/s11042-021-11733-y","article-title":"Deepfake generation and detection, a survey","volume":"81","author":"Zhang","year":"2022","journal-title":"Multimedia Tools Appl","ISSN":"http:\/\/id.crossref.org\/issn\/1573-7721","issn-type":"print"},{"issue":"16","key":"10.1016\/j.cag.2024.103888_b21","doi-asserted-by":"crossref","first-page":"3407","DOI":"10.3390\/electronics12163407","article-title":"Digital face manipulation creation and detection: A systematic review","volume":"12","author":"Dang","year":"2023","journal-title":"Electronics","ISSN":"http:\/\/id.crossref.org\/issn\/2079-9292","issn-type":"print"},{"issue":"11","key":"10.1016\/j.cag.2024.103888_b22","doi-asserted-by":"crossref","first-page":"6711","DOI":"10.3390\/app13116711","article-title":"Quick overview of face swap deep fakes","volume":"13","author":"Walczyna","year":"2023","journal-title":"Appl Sci","ISSN":"http:\/\/id.crossref.org\/issn\/2076-3417","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b23","series-title":"Faceswap","author":"\/u\/deepfakes","year":"2023"},{"key":"10.1016\/j.cag.2024.103888_b24","series-title":"FSGAN: Subject agnostic face swapping and reenactment","first-page":"7184","author":"Nirkin","year":"2019"},{"key":"10.1016\/j.cag.2024.103888_b25","series-title":"Advancing high fidelity identity swapping for forgery detection","first-page":"5074","author":"Li","year":"2020"},{"key":"10.1016\/j.cag.2024.103888_b26","series-title":"Proceedings of the 28th ACM international conference on multimedia","isbn-type":"print","doi-asserted-by":"crossref","first-page":"2003","DOI":"10.1145\/3394171.3413630","article-title":"SimSwap: An efficient framework for high fidelity face swapping","author":"Chen","year":"2020","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450379885"},{"key":"10.1016\/j.cag.2024.103888_b27","first-page":"29710","article-title":"BlendGAN: Implicitly GAN blending for arbitrary stylized face generation","volume":"vol. 34","author":"Liu","year":"2021"},{"key":"10.1016\/j.cag.2024.103888_b28","series-title":"Fast face-swap using convolutional neural networks","first-page":"3677","author":"Korshunova","year":"2017"},{"issue":"2","key":"10.1016\/j.cag.2024.103888_b29","doi-asserted-by":"crossref","first-page":"25:1","DOI":"10.1145\/3182644","article-title":"FaceVR: Real-time gaze-aware facial reenactment in virtual reality","volume":"37","author":"Thies","year":"2018","journal-title":"ACM Trans Graph","ISSN":"http:\/\/id.crossref.org\/issn\/0730-0301","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b30","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.109628","article-title":"Deepfacelab: Integrated, flexible and extensible face-swapping framework","volume":"141","author":"Liu","year":"2023","journal-title":"Pattern Recognit","ISSN":"http:\/\/id.crossref.org\/issn\/0031-3203","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b31","series-title":"One shot face swapping on megapixels","first-page":"4834","author":"Zhu","year":"2021"},{"key":"10.1016\/j.cag.2024.103888_b32","series-title":"HifiFace: 3D shape and semantic prior guided high fidelity face swapping, 2","first-page":"1136","author":"Wang","year":"2021"},{"key":"10.1016\/j.cag.2024.103888_b33","series-title":"3D-aware face swapping","first-page":"12705","author":"Li","year":"2023"},{"issue":"6","key":"10.1016\/j.cag.2024.103888_b34","doi-asserted-by":"crossref","first-page":"225:1","DOI":"10.1145\/3414685.3417826","article-title":"Face identity disentanglement via latent space mapping","volume":"39","author":"Nitzan","year":"2020","journal-title":"ACM Trans Graph","ISSN":"http:\/\/id.crossref.org\/issn\/0730-0301","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b35","series-title":"ArcFace: Additive angular margin loss for deep face recognition","first-page":"4690","author":"Deng","year":"2019"},{"key":"10.1016\/j.cag.2024.103888_b36","series-title":"CosFace: Large margin cosine loss for deep face recognition","first-page":"5265","author":"Wang","year":"2018"},{"key":"10.1016\/j.cag.2024.103888_b37","series-title":"Proceedings of the 27th ACM international conference on multimedia","isbn-type":"print","doi-asserted-by":"crossref","first-page":"2052","DOI":"10.1145\/3343031.3350980","article-title":"Cycle in cycle generative adversarial networks for keypoint-guided image generation","author":"Tang","year":"2019","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450368896"},{"key":"10.1016\/j.cag.2024.103888_b38","series-title":"AnonymousNet: Natural face de-identification with measurable privacy","author":"Li","year":"2019"},{"issue":"5","key":"10.1016\/j.cag.2024.103888_b39","doi-asserted-by":"crossref","DOI":"10.1002\/cpe.7554","article-title":"Face image de-identification by feature space adversarial perturbation","volume":"35","author":"Xue","year":"2023","journal-title":"Concurr Comput: Pract Exper","ISSN":"http:\/\/id.crossref.org\/issn\/1532-0634","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b40","series-title":"Natural and effective obfuscation by head inpainting","first-page":"5050","author":"Sun","year":"2018"},{"key":"10.1016\/j.cag.2024.103888_b41","series-title":"Proceedings of the 29th ACM international conference on multimedia","isbn-type":"print","doi-asserted-by":"crossref","first-page":"3182","DOI":"10.1145\/3474085.3475464","article-title":"Effective de-identification generative adversarial network for face anonymization","author":"Kuang","year":"2021","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450386517"},{"key":"10.1016\/j.cag.2024.103888_b42","series-title":"Motion representations for articulated animation","first-page":"13653","author":"Siarohin","year":"2021"},{"key":"10.1016\/j.cag.2024.103888_b43","series-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2015"},{"key":"10.1016\/j.cag.2024.103888_b44","series-title":"Image style transfer using convolutional neural networks","first-page":"2414","author":"Gatys","year":"2016"},{"key":"10.1016\/j.cag.2024.103888_b45","series-title":"Multi-style generative network for real-time transfer","author":"Zhang","year":"2018"},{"key":"10.1016\/j.cag.2024.103888_b46","series-title":"Arbitrary style transfer in real-time with adaptive instance normalization","first-page":"1501","author":"Huang","year":"2017"},{"key":"10.1016\/j.cag.2024.103888_b47","series-title":"High-resolution image synthesis and semantic manipulation with conditional GANs","first-page":"8798","author":"Wang","year":"2018"},{"key":"10.1016\/j.cag.2024.103888_b48","series-title":"2022 IEEE 13th annual ubiquitous computing, electronics & mobile communication conference","first-page":"0547","article-title":"Perception vs. reality: Understanding and evaluating the impact of synthetic image deepfakes over college students","author":"Preu","year":"2022"},{"key":"10.1016\/j.cag.2024.103888_b49","series-title":"FaceForensics++: Learning to detect manipulated facial images","first-page":"1","author":"Rossler","year":"2019"},{"key":"10.1016\/j.cag.2024.103888_b50","series-title":"Proceedings of the 2021 CHI conference on human factors in computing systems","isbn-type":"print","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3411764.3445699","article-title":"Seeing is believing: Exploring perceptual differences in DeepFake videos","author":"Tahir","year":"2021","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450380966"},{"issue":"1","key":"10.1016\/j.cag.2024.103888_b51","doi-asserted-by":"crossref","DOI":"10.1073\/pnas.2110013119","article-title":"Deepfake detection by human crowds, machines, and machine-informed crowds","volume":"119","author":"Groh","year":"2022","journal-title":"Proc Natl Acad Sci"},{"key":"10.1016\/j.cag.2024.103888_b52","series-title":"Human detection of political speech deepfakes across transcripts, audio, and video","author":"Groh","year":"2023"},{"key":"10.1016\/j.cag.2024.103888_b53","series-title":"Computer animation and social agents","isbn-type":"print","doi-asserted-by":"crossref","first-page":"120","DOI":"10.1007\/978-3-030-63426-1_13","article-title":"PEFS: A validated dataset for perceptual experiments on face swap portrait videos","author":"W\u00f6hler","year":"2020","ISBN":"http:\/\/id.crossref.org\/isbn\/9783030634261"},{"key":"10.1016\/j.cag.2024.103888_b54","series-title":"Proceedings of the 2021 CHI conference on human factors in computing systems","isbn-type":"print","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3411764.3445627","article-title":"Towards understanding perceptual differences between genuine and face-swapped videos","author":"W\u00f6hler","year":"2021","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450380966"},{"key":"10.1016\/j.cag.2024.103888_b55","series-title":"Proceedings of the 22nd ACM international conference on intelligent virtual agents","isbn-type":"print","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3514197.3549687","article-title":"Personality analysis of face swaps: Can they be used as avatars?","author":"W\u00f6hler","year":"2022","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450392488"},{"issue":"8","key":"10.1016\/j.cag.2024.103888_b56","doi-asserted-by":"crossref","DOI":"10.1073\/pnas.2120481119","article-title":"AI-synthesized faces are indistinguishable from real faces and more trustworthy","volume":"119","author":"Nightingale","year":"2022","journal-title":"Proc Natl Acad Sci"},{"issue":"4","key":"10.1016\/j.cag.2024.103888_b57","doi-asserted-by":"crossref","first-page":"91:1","DOI":"10.1145\/2185520.2185587","article-title":"Render me real? investigating the effect of render style on the perception of animated virtual humans","volume":"31","author":"McDonnell","year":"2012","journal-title":"ACM Trans Graph","ISSN":"http:\/\/id.crossref.org\/issn\/0730-0301","issn-type":"print"},{"issue":"4","key":"10.1016\/j.cag.2024.103888_b58","doi-asserted-by":"crossref","first-page":"22:1","DOI":"10.1145\/1823738.1823740","article-title":"The saliency of anomalies in animated human characters","volume":"7","author":"Hodgins","year":"2010","journal-title":"ACM Trans Appl Percept","ISSN":"http:\/\/id.crossref.org\/issn\/1544-3558","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b59","series-title":"Proceedings of the ACM symposium on applied perception","isbn-type":"print","doi-asserted-by":"crossref","first-page":"35","DOI":"10.1145\/2492494.2502059","article-title":"Unpleasantness of animated characters corresponds to increased viewer attention to faces","author":"Carter","year":"2013","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450322621"},{"key":"10.1016\/j.cag.2024.103888_b60","series-title":"Proceedings of the 13th ACM SIGGRApH conference on motion, interaction and games","isbn-type":"print","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3424636.3426904","article-title":"Investigating perceptually based models to predict importance of facial blendshapes","author":"Carrigan","year":"2020","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450381710"},{"issue":"3","key":"10.1016\/j.cag.2024.103888_b61","doi-asserted-by":"crossref","first-page":"695","DOI":"10.1016\/j.chb.2008.12.026","article-title":"Too real for comfort? Uncanny responses to computer generated faces","volume":"25","author":"MacDorman","year":"2009","journal-title":"Comput Hum Behav","ISSN":"http:\/\/id.crossref.org\/issn\/0747-5632","issn-type":"print"},{"issue":"4","key":"10.1016\/j.cag.2024.103888_b62","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1109\/MCG.2008.79","article-title":"Overcoming the uncanny valley","volume":"28","author":"Geller","year":"2008","journal-title":"IEEE Comput Graph Appl","ISSN":"http:\/\/id.crossref.org\/issn\/1558-1756","issn-type":"print"},{"issue":"1","key":"10.1016\/j.cag.2024.103888_b63","doi-asserted-by":"crossref","first-page":"129","DOI":"10.1007\/s12369-016-0380-9","article-title":"Measuring the uncanny valley effect","volume":"9","author":"Ho","year":"2017","journal-title":"Int J Soc Robot","ISSN":"http:\/\/id.crossref.org\/issn\/1875-4805","issn-type":"print"},{"issue":"2","key":"10.1016\/j.cag.2024.103888_b64","doi-asserted-by":"crossref","first-page":"98","DOI":"10.1109\/MRA.2012.2192811","article-title":"The uncanny valley [from the field]","volume":"19","author":"Mori","year":"2012","journal-title":"IEEE Robot Autom Mag","ISSN":"http:\/\/id.crossref.org\/issn\/1558-223X","issn-type":"print"},{"issue":"10","key":"10.1016\/j.cag.2024.103888_b65","doi-asserted-by":"crossref","first-page":"968","DOI":"10.1177\/0301006619869134","article-title":"Virtual faces evoke only a weak uncanny valley effect: An empirical investigation with controlled virtual face images","volume":"48","author":"K\u00e4tsyri","year":"2019","journal-title":"Perception","ISSN":"http:\/\/id.crossref.org\/issn\/0301-0066","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b66","doi-asserted-by":"crossref","first-page":"132","DOI":"10.1016\/j.cognition.2017.01.009","article-title":"Categorization-based stranger avoidance does not explain the uncanny valley effect","volume":"161","author":"MacDorman","year":"2017","journal-title":"Cognition","ISSN":"http:\/\/id.crossref.org\/issn\/1873-7838","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b67","doi-asserted-by":"crossref","first-page":"190","DOI":"10.1016\/j.cognition.2015.09.019","article-title":"Reducing consistency in human realism increases the uncanny valley effect; increasing category uncertainty does not","volume":"146","author":"MacDorman","year":"2016","journal-title":"Cognition","ISSN":"http:\/\/id.crossref.org\/issn\/0010-0277","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b68","series-title":"Intelligent virtual agents","isbn-type":"print","doi-asserted-by":"crossref","first-page":"511","DOI":"10.1007\/978-3-642-33197-8_62","article-title":"Evaluation of the uncanny valley in CG characters","author":"Dill","year":"2012","ISBN":"http:\/\/id.crossref.org\/isbn\/9783642331978"},{"issue":"9","key":"10.1016\/j.cag.2024.103888_b69","first-page":"477","article-title":"Motion and the uncanny valley","volume":"7","author":"White","year":"2007","journal-title":"J Vis","ISSN":"http:\/\/id.crossref.org\/issn\/1534-7362","issn-type":"print"},{"issue":"3","key":"10.1016\/j.cag.2024.103888_b70","doi-asserted-by":"crossref","first-page":"271","DOI":"10.1016\/j.cognition.2013.11.001","article-title":"Empirical evaluation of the uncanny valley hypothesis fails to confirm the predicted effect of motion","volume":"130","author":"Piwek","year":"2014","journal-title":"Cognition","ISSN":"http:\/\/id.crossref.org\/issn\/0010-0277","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b71","first-page":"1","article-title":"FakeCatcher: Detection of synthetic portrait videos using biological signals","author":"Ciftci","year":"2020","journal-title":"IEEE Trans Pattern Anal Mach Intell","ISSN":"http:\/\/id.crossref.org\/issn\/1939-3539","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b72","series-title":"2018 IEEE international workshop on information forensics and security","first-page":"1","article-title":"In ictu oculi: Exposing AI created fake videos by detecting eye blinking","author":"Li","year":"2018"},{"issue":"1","key":"10.1016\/j.cag.2024.103888_b73","first-page":"210","article-title":"A new reality: Deepfake technology and the world around us","volume":"48","author":"Mullen","year":"2022","journal-title":"Mitchell Hamline Law Review"},{"issue":"3 Pt 1","key":"10.1016\/j.cag.2024.103888_b74","doi-asserted-by":"crossref","first-page":"857","DOI":"10.2466\/pms.1978.47.3.857","article-title":"Eyes as the center of focus in the visual examination of human faces","volume":"47","author":"Janik","year":"1978","journal-title":"Perceptual Motor Skills","ISSN":"http:\/\/id.crossref.org\/issn\/0031-5125","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b75","series-title":"Proceedings of the 2020 international conference on multimodal interaction","isbn-type":"print","doi-asserted-by":"crossref","first-page":"519","DOI":"10.1145\/3382507.3418857","article-title":"The eyes know it: Fakeet- an eye-tracking database to understand deepfake perception","author":"Gupta","year":"2020","ISBN":"http:\/\/id.crossref.org\/isbn\/9781450375818"},{"issue":"4","key":"10.1016\/j.cag.2024.103888_b76","doi-asserted-by":"crossref","first-page":"173","DOI":"10.1111\/cgf.14062","article-title":"High-resolution neural face swapping for visual effects","volume":"39","author":"Naruniec","year":"2020","journal-title":"Comput Graph Forum","ISSN":"http:\/\/id.crossref.org\/issn\/1467-8659","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b77","series-title":"Questions and answers in attitude surveys: experiments on question form, wording, and context","isbn-type":"print","author":"Schuman","year":"1996","ISBN":"http:\/\/id.crossref.org\/isbn\/9780761903598"},{"key":"10.1016\/j.cag.2024.103888_b78","series-title":"International encyclopedia of statistical science","isbn-type":"print","doi-asserted-by":"crossref","first-page":"1658","DOI":"10.1007\/978-3-642-04898-2_616","article-title":"Wilcoxon-signed-rank test","author":"Rey","year":"2011","ISBN":"http:\/\/id.crossref.org\/isbn\/9783642048982"},{"key":"10.1016\/j.cag.2024.103888_b79","series-title":"How far are we from solving the 2D & 3D face alignment problem? (and a dataset of 230,000 3D facial landmarks)","first-page":"1021","author":"Bulat","year":"2017"},{"issue":"4","key":"10.1016\/j.cag.2024.103888_b80","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: From error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans Image Process","ISSN":"http:\/\/id.crossref.org\/issn\/1941-0042","issn-type":"print"},{"issue":"1","key":"10.1016\/j.cag.2024.103888_b81","doi-asserted-by":"crossref","first-page":"47","DOI":"10.1109\/TCI.2016.2644865","article-title":"Loss functions for image restoration with neural networks","volume":"3","author":"Zhao","year":"2017","journal-title":"IEEE Trans Comput Imaging","ISSN":"http:\/\/id.crossref.org\/issn\/2333-9403","issn-type":"print"},{"key":"10.1016\/j.cag.2024.103888_b82","series-title":"2023 8th international conference on frontiers of signal processing","first-page":"98","article-title":"L2CS-Net : Fine-grained gaze estimation in unconstrained environments","author":"Abdelrahman","year":"2023"},{"key":"10.1016\/j.cag.2024.103888_b83","series-title":"Deep learning face attributes in the wild","first-page":"3730","author":"Liu","year":"2015"},{"key":"10.1016\/j.cag.2024.103888_b84","series-title":"The corsini encyclopedia of psychology","isbn-type":"print","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1002\/9780470479216.corpsy0524","article-title":"Mann-Whitney U test","author":"McKnight","year":"2010","ISBN":"http:\/\/id.crossref.org\/isbn\/9780470479216"},{"key":"10.1016\/j.cag.2024.103888_b85","series-title":"Live face de-identification in video","first-page":"9378","author":"Gafni","year":"2019"},{"key":"10.1016\/j.cag.2024.103888_b86","series-title":"Practical digital disguises: Leveraging face swaps to protect patient privacy","author":"Wilson","year":"2022"}],"container-title":["Computers & Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0097849324000153?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0097849324000153?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2024,5,16]],"date-time":"2024-05-16T08:09:08Z","timestamp":1715846948000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0097849324000153"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4]]},"references-count":86,"alternative-id":["S0097849324000153"],"URL":"https:\/\/doi.org\/10.1016\/j.cag.2024.103888","relation":{},"ISSN":["0097-8493"],"issn-type":[{"value":"0097-8493","type":"print"}],"subject":[],"published":{"date-parts":[[2024,4]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Towards mitigating uncann(eye)ness in face swaps via gaze-centric loss terms","name":"articletitle","label":"Article Title"},{"value":"Computers & Graphics","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.cag.2024.103888","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2024 Elsevier Ltd. All rights reserved.","name":"copyright","label":"Copyright"}],"article-number":"103888"}}