{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,10]],"date-time":"2025-04-10T22:44:34Z","timestamp":1744325074887,"version":"3.28.0"},"reference-count":37,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,10,1]],"date-time":"2023-10-01T00:00:00Z","timestamp":1696118400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,10,1]],"date-time":"2023-10-01T00:00:00Z","timestamp":1696118400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100006785","name":"Google","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100006785","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,10,1]]},"DOI":"10.1109\/iccv51070.2023.00719","type":"proceedings-article","created":{"date-parts":[[2024,1,15]],"date-time":"2024-01-15T15:55:59Z","timestamp":1705334159000},"page":"7788-7798","source":"Crossref","is-referenced-by-count":15,"title":["The Power of Sound (TPoS): Audio Reactive Video Generation with Stable Diffusion"],"prefix":"10.1109","author":[{"given":"Yujin","family":"Jeong","sequence":"first","affiliation":[{"name":"Korea University,Department of Computer Science and Engineering,Seoul,Korea,02841"}]},{"given":"Wonjeong","family":"Ryoo","sequence":"additional","affiliation":[{"name":"Korea University,Department of Artificial Intelligence,Seoul,Korea,02841"}]},{"given":"Seunghyun","family":"Lee","sequence":"additional","affiliation":[{"name":"Korea University,Department of Artificial Intelligence,Seoul,Korea,02841"}]},{"given":"Dabin","family":"Seo","sequence":"additional","affiliation":[{"name":"Korea University,Department of Computer Science and Engineering,Seoul,Korea,02841"}]},{"given":"Wonmin","family":"Byeon","sequence":"additional","affiliation":[{"name":"NVIDIA Research,Santa Clara,USA,95050"}]},{"given":"Sangpil","family":"Kim","sequence":"additional","affiliation":[{"name":"Korea University,Department of Artificial Intelligence,Seoul,Korea,02841"}]},{"given":"Jinkyu","family":"Kim","sequence":"additional","affiliation":[{"name":"Korea University,Department of Computer Science and Engineering,Seoul,Korea,02841"}]}],"member":"263","reference":[{"doi-asserted-by":"publisher","key":"ref1","DOI":"10.1145\/3592450"},{"year":"2022","author":"Brack","article-title":"The stable artist: Steering semantics in diffusion latent space","key":"ref2"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.1109\/CVPR.2017.502"},{"doi-asserted-by":"publisher","key":"ref4","DOI":"10.1007\/978-3-030-58583-9_42"},{"doi-asserted-by":"publisher","key":"ref5","DOI":"10.1109\/ICASSP40776.2020.9053174"},{"doi-asserted-by":"publisher","key":"ref6","DOI":"10.1109\/CVPR.2009.5206848"},{"year":"2019","author":"Gross","article-title":"Automatic realistic music video generation from segments of youtube videos","key":"ref7"},{"doi-asserted-by":"publisher","key":"ref8","DOI":"10.1109\/CVPR.2016.90"},{"year":"2022","author":"Hertz","article-title":"Prompt-to-prompt image editing with cross attention control","key":"ref9"},{"year":"2022","author":"Ho","article-title":"Imagen video: High definition video generation with diffusion models","key":"ref10"},{"year":"2022","author":"Ho","article-title":"Video diffusion models","key":"ref11"},{"doi-asserted-by":"publisher","key":"ref12","DOI":"10.1162\/neco.1997.9.8.1735"},{"issue":"4","key":"ref13","first-page":"10","article-title":"Tr\u00e4umerai: Dreaming music with stylegan","volume":"2","author":"Jeong","year":"2021"},{"doi-asserted-by":"publisher","key":"ref14","DOI":"10.1145\/3528233.3530745"},{"doi-asserted-by":"publisher","key":"ref15","DOI":"10.1109\/CVPR46437.2021.01386"},{"key":"ref16","first-page":"852","article-title":"Alias-free generative adversarial networks","volume":"34","author":"Karras","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"year":"2017","author":"Kay","article-title":"The kinetics human action video dataset","key":"ref17"},{"key":"ref18","first-page":"14042","article-title":"Ccvs: context-aware controllable video synthesis","volume":"34","author":"Le Moing","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"doi-asserted-by":"publisher","key":"ref19","DOI":"10.1007\/978-3-031-19790-1_3"},{"doi-asserted-by":"publisher","key":"ref20","DOI":"10.1109\/CVPR52688.2022.00337"},{"doi-asserted-by":"publisher","key":"ref21","DOI":"10.1109\/CVPR52688.2022.00338"},{"doi-asserted-by":"publisher","key":"ref22","DOI":"10.1007\/978-3-031-19790-1_26"},{"year":"2023","author":"Molad","article-title":"Dreamix: Video diffusion models are general video editors","key":"ref23"},{"year":"2018","author":"van den Oord","article-title":"Representation learning with contrastive predictive coding","key":"ref24"},{"doi-asserted-by":"publisher","key":"ref25","DOI":"10.21437\/Interspeech.2019-2680"},{"doi-asserted-by":"publisher","key":"ref26","DOI":"10.1145\/3394171.3413532"},{"key":"ref27","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"International conference on machine learning","author":"Radford"},{"year":"2022","author":"Ramesh","article-title":"Hierarchical text-conditional image generation with clip latents","key":"ref28"},{"doi-asserted-by":"publisher","key":"ref29","DOI":"10.1109\/CVPR52688.2022.01042"},{"doi-asserted-by":"publisher","key":"ref30","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref31","article-title":"Improved techniques for training gans","volume":"29","author":"Salimans","year":"2016","journal-title":"Advances in neural information processing systems"},{"year":"2022","author":"Singer","article-title":"Make-a-video: Text-to-video generation without text-video data","key":"ref32"},{"doi-asserted-by":"publisher","key":"ref33","DOI":"10.1109\/ICCV48922.2021.01388"},{"doi-asserted-by":"publisher","key":"ref34","DOI":"10.1109\/CVPR52688.2022.00361"},{"year":"2018","author":"Unterthiner","article-title":"Towards accurate generative models of video: A new metric & challenges","key":"ref35"},{"year":"2022","author":"Villegas","article-title":"Phenaki: Variable length video generation from open domain textual description","key":"ref36"},{"doi-asserted-by":"publisher","key":"ref37","DOI":"10.1109\/CVPR52729.2023.00189"}],"event":{"name":"2023 IEEE\/CVF International Conference on Computer Vision (ICCV)","start":{"date-parts":[[2023,10,1]]},"location":"Paris, France","end":{"date-parts":[[2023,10,6]]}},"container-title":["2023 IEEE\/CVF International Conference on Computer Vision (ICCV)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10376473\/10376477\/10377529.pdf?arnumber=10377529","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,17]],"date-time":"2024-01-17T20:37:48Z","timestamp":1705523868000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10377529\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,1]]},"references-count":37,"URL":"https:\/\/doi.org\/10.1109\/iccv51070.2023.00719","relation":{},"subject":[],"published":{"date-parts":[[2023,10,1]]}}}