{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T22:52:29Z","timestamp":1740178349199,"version":"3.37.3"},"reference-count":45,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"4","license":[{"start":{"date-parts":[[2020,8,1]],"date-time":"2020-08-01T00:00:00Z","timestamp":1596240000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,8,1]],"date-time":"2020-08-01T00:00:00Z","timestamp":1596240000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,8,1]],"date-time":"2020-08-01T00:00:00Z","timestamp":1596240000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100004663","name":"Ministry of Science and Technology, Taiwan","doi-asserted-by":"publisher","award":["MOST109-2634-F-002-037"],"id":[{"id":"10.13039\/501100004663","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Emerg. Top. Comput. Intell."],"published-print":{"date-parts":[[2020,8]]},"DOI":"10.1109\/tetci.2020.2968599","type":"journal-article","created":{"date-parts":[[2020,5,25]],"date-time":"2020-05-25T22:39:50Z","timestamp":1590446390000},"page":"440-449","source":"Crossref","is-referenced-by-count":5,"title":["Static2Dynamic: Video Inference From a Deep Glimpse"],"prefix":"10.1109","volume":"4","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4672-8454","authenticated-orcid":false,"given":"Yu-Ying","family":"Yeh","sequence":"first","affiliation":[]},{"given":"Yen-Cheng","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7715-8306","authenticated-orcid":false,"given":"Wei-Chen","family":"Chiu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2333-157X","authenticated-orcid":false,"given":"Yu-Chiang Frank","family":"Wang","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref38","article-title":"Auto-encoding variational bayes","author":"kingma","year":"2014","journal-title":"Proc Intl Conf on Learning Representations"},{"key":"ref33","first-page":"2225","article-title":"Video prediction with appearance and motion conditions","author":"jang","year":"2018","journal-title":"Proc 7th Int Conf Machine Learning"},{"key":"ref32","first-page":"91","article-title":"Probabilistic future frame synthesis via cross convolutional networks","author":"xue","year":"2016","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref31","first-page":"753","article-title":"Contextvp: Fully context-aware video prediction","author":"byeon","year":"2018","journal-title":"Proc Eur Conf Comput Vision"},{"key":"ref30","first-page":"716","article-title":"Folded recurrent neural networks for future video prediction","author":"oliu","year":"2018","journal-title":"Proc Eur Conf Comput Vision"},{"key":"ref37","first-page":"1558","article-title":"Autoencoding beyond pixels using a learned similarity metric","author":"larsen","year":"2016","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref36","first-page":"387","article-title":"Learning to forecast and re_ne residual motion for image-to-video generation","author":"zhao","year":"2018","journal-title":"Proc Eur Conf Comput Vision"},{"key":"ref35","first-page":"201","article-title":"Pose guided human video generation","author":"yang","year":"2018","journal-title":"Proc Eur Conf Comput Vision"},{"key":"ref34","first-page":"600","article-title":"Flow-grounded spatial-temporal video prediction from still images","author":"li","year":"2018","journal-title":"Proc Eur Conf Comput Vision"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2007.60"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2004.1334462"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-006-0001-4"},{"key":"ref12","first-page":"613","article-title":"Generating videos with scene dynamics","author":"vondrick","year":"2016","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.308"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00165"},{"key":"ref15","first-page":"366","article-title":"Deep video generation, prediction and completion of human action sequences","author":"cai","year":"2018","journal-title":"Proc Eur Conf Comput Vision"},{"key":"ref16","first-page":"452","article-title":"Probabilistic video generation using holistic attribute control","author":"he","year":"2018","journal-title":"Proc Eur Conf Comput Vision"},{"key":"ref17","article-title":"Deep multi-scale video prediction beyond mean square error","author":"mathieu","year":"2016","journal-title":"Proc Intl Conf on Learning Representations"},{"key":"ref18","first-page":"64","article-title":"Unsupervised learning for physical interaction through video prediction","author":"finn","year":"2016","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref19","article-title":"Decomposing motion and content for natural video sequence prediction","author":"villegas","year":"2017","journal-title":"Proc Intl Conf on Learning Representations"},{"key":"ref28","first-page":"6038","article-title":"Hierarchical long-term video prediction without supervision","volume":"80","author":"wichers","year":"2018","journal-title":"Proc 35th Int Conf Mach Learn"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.37"},{"key":"ref27","first-page":"1174","article-title":"Stochastic video generation with a learned prior","volume":"80","author":"denton","year":"2018","journal-title":"Proc 35th Int Conf Mach Learn"},{"key":"ref3","first-page":"670","article-title":"Video frame interpolation via adaptive convolution","author":"niklaus","year":"2017","journal-title":"Proc IEEE Conf Comput Vision Pattern Recognit"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00059"},{"key":"ref29","first-page":"170","article-title":"Dyan: A dynamical atoms-based network for video prediction","author":"liu","year":"2018","journal-title":"Proc Eur Conf Comput Vision"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00183"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.478"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00938"},{"key":"ref2","first-page":"434","article-title":"Learning image matching by simply watching video","author":"long","year":"2016","journal-title":"Proc Eur Conf Comput Vision"},{"key":"ref9","first-page":"249","article-title":"A temporally-aware interpolation network for video frame inpainting","author":"sun","year":"2018","journal-title":"Proc Asian Conf Comput Vision"},{"article-title":"Conditional generative adversarial nets","year":"2014","author":"mirza","key":"ref1"},{"key":"ref20","article-title":"Deep predictive coding networks for video prediction and unsupervised learning","author":"lotter","year":"2017","journal-title":"Proc Intl Conf on Learning Representations"},{"key":"ref45","article-title":"Empirical evaluation of gated recurrent neural networks on sequence modeling","author":"chung","year":"2014","journal-title":"NIPS 2014 Deep Learning Workshop"},{"key":"ref22","first-page":"1744","article-title":"Dual motion gan for future-ow embedded video prediction","author":"liang","year":"2017","journal-title":"Proc IEEE Int Conf Comput Vision"},{"key":"ref21","first-page":"3560","article-title":"Learning to generate long-term future via hierarchical prediction","volume":"70","author":"villegas","year":"2017","journal-title":"Proc 34th Int Conf Mach Learn"},{"article-title":"Openface: A general-purpose face recognition library with mobile applications","year":"2016","author":"amos","key":"ref42"},{"key":"ref24","article-title":"Stochastic variational video prediction","author":"babaeizadeh","year":"2018","journal-title":"Proc Intl Conf on Learning Representations"},{"key":"ref41","article-title":"The mug facial expression database","author":"aifanti","year":"2010","journal-title":"Proc IEEE 11th Int Workshop Image Anal Multimedia Interactive Serv WIAMIS 10"},{"key":"ref23","first-page":"4414","article-title":"Unsupervised learning of disentangled representations from video","author":"denton","year":"2017","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref44","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2015","journal-title":"Proc 3rd Int Conf Learn Representations"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00158"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1049\/el:20080522"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00684"}],"container-title":["IEEE Transactions on Emerging Topics in Computational Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7433297\/9145895\/09099414.pdf?arnumber=9099414","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T12:43:14Z","timestamp":1651063394000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9099414\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,8]]},"references-count":45,"journal-issue":{"issue":"4"},"URL":"https:\/\/doi.org\/10.1109\/tetci.2020.2968599","relation":{},"ISSN":["2471-285X"],"issn-type":[{"type":"electronic","value":"2471-285X"}],"subject":[],"published":{"date-parts":[[2020,8]]}}}