{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T00:55:02Z","timestamp":1740099302913,"version":"3.37.3"},"publisher-location":"New York, NY, USA","reference-count":16,"publisher":"ACM","license":[{"start":{"date-parts":[[2019,4,15]],"date-time":"2019-04-15T00:00:00Z","timestamp":1555286400000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100011002","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61761136011 and 61532019"],"id":[{"id":"10.13039\/501100011002","id-type":"DOI","asserted-by":"publisher"}]},{"name":"College of Engineering and Applied Sciences of CU Boulder","award":["ASIRT grant"]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/M027287\/1 and EP\/P020909\/1."],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2019,4,15]]},"DOI":"10.1145\/3313149.3313369","type":"proceedings-article","created":{"date-parts":[[2019,5,10]],"date-time":"2019-05-10T19:01:58Z","timestamp":1557514918000},"page":"16-18","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Limit reachability for model-free reinforcement learning of \u03c9-regular objectives"],"prefix":"10.1145","author":[{"given":"Ernst Moritz","family":"Hahn","sequence":"first","affiliation":[{"name":"Queen's University Belfast"}]},{"given":"Mateo","family":"Perez","sequence":"additional","affiliation":[{"name":"University of Colorado Boulder"}]},{"given":"Sven","family":"Schewe","sequence":"additional","affiliation":[{"name":"University of Liverpool"}]},{"given":"Fabio","family":"Somenzi","sequence":"additional","affiliation":[{"name":"University of Colorado Boulder"}]},{"given":"Ashutosh","family":"Trivedi","sequence":"additional","affiliation":[{"name":"University of Colorado Boulder"}]},{"given":"Dominik","family":"Wojtczak","sequence":"additional","affiliation":[{"name":"University of Liverpool"}]}],"member":"320","published-online":{"date-parts":[[2019,4,15]]},"reference":[{"volume-title":"Principles of Model Checking","author":"Baier C.","key":"e_1_3_2_1_1_1","unstructured":"C. Baier and J.-P. Katoen . 2008. Principles of Model Checking . MIT Press . C. Baier and J.-P. Katoen. 2008. Principles of Model Checking. MIT Press."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"T. Br\u00e1zdil K. Chatterjee M. Chmel\u00edk V. Forejt J. K\u0159et\u00ednsk\u00fd M. Kwiatkowska D. Parker and M. Ujma. 2014. Verification of Markov Decision Processes Using Learning Algorithms. In ATVA. 98--114. T. Br\u00e1zdil K. Chatterjee M. Chmel\u00edk V. Forejt J. K\u0159et\u00ednsk\u00fd M. Kwiatkowska D. Parker and M. Ujma. 2014. Verification of Markov Decision Processes Using Learning Algorithms. In ATVA. 98--114.","DOI":"10.1007\/978-3-319-11936-6_8"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/210332.210339"},{"key":"e_1_3_2_1_4_1","unstructured":"E. M. Hahn G. Li S. Schewe A. Turrini and L. Zhang. 2015. Lazy Probabilistic Model Checking without Determinisation. In Concurrency Theory (CONCUR). 354--367. E. M. Hahn G. Li S. Schewe A. Turrini and L. Zhang. 2015. Lazy Probabilistic Model Checking without Determinisation. In Concurrency Theory (CONCUR). 354--367."},{"key":"e_1_3_2_1_5_1","volume-title":"Omega-Regular Objectives in Model-Free Reinforcement Learning. In TACAS 2019","author":"Hahn E. M.","year":"1810","unstructured":"E. M. Hahn , M. Perez , S. Schewe , F. Somenzi , A. Trivedi , and D. Wojtczak . 2019 . Omega-Regular Objectives in Model-Free Reinforcement Learning. In TACAS 2019 . To appear. ArXiv 1810 .00950. E. M. Hahn, M. Perez, S. Schewe, F. Somenzi, A. Trivedi, and D. Wojtczak. 2019. Omega-Regular Objectives in Model-Free Reinforcement Learning. In TACAS 2019. To appear. ArXiv 1810.00950."},{"key":"e_1_3_2_1_6_1","first-page":"1","article-title":"End-to-end Training of Deep Visuomotor Policies","volume":"17","author":"Levine S.","year":"2016","unstructured":"S. Levine , C. Finn , T. Darrell , and P. Abbeel . 2016 . End-to-end Training of Deep Visuomotor Policies . J. Mach. Learn. Res. 17 , 1 (Jan. 2016), 1334--1373. http:\/\/dl.acm.org\/citation.cfm?id=2946645.2946684 S. Levine, C. Finn, T. Darrell, and P. Abbeel. 2016. End-to-end Training of Deep Visuomotor Policies. J. Mach. Learn. Res. 17, 1 (Jan. 2016), 1334--1373. http:\/\/dl.acm.org\/citation.cfm?id=2946645.2946684","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_2_1_7_1","unstructured":"T. P. Lillicrap J. J. Hunt A. Pritzel N. Heess T. Erez Y. Tassa D. Silver and D. Wierstra. 2015. Continuous control with deep reinforcement learning. CoRR abs\/1509.02971 (2015). http:\/\/arxiv.org\/abs\/1509.02971 T. P. Lillicrap J. J. Hunt A. Pritzel N. Heess T. Erez Y. Tassa D. Silver and D. Wierstra. 2015. Continuous control with deep reinforcement learning. CoRR abs\/1509.02971 (2015). http:\/\/arxiv.org\/abs\/1509.02971"},{"key":"e_1_3_2_1_8_1","unstructured":"R. Liu J. Lehman P. Molino F. Petroski Such E. Frank A. Sergeev and J. Yosinski. 2018. An Intriguing Failing of Convolutional Neural Networks and the CoordConv Solution. ArXiv e-prints 1807.03247 (July 2018). R. Liu J. Lehman P. Molino F. Petroski Such E. Frank A. Sergeev and J. Yosinski. 2018. An Intriguing Failing of Convolutional Neural Networks and the CoordConv Solution. ArXiv e-prints 1807.03247 (July 2018)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"crossref","unstructured":"V. Mnih etal 2015. Human-level control through reinforcement learning. Nature 518 (Feb. 2015) 529--533. V. Mnih et al. 2015. Human-level control through reinforcement learning. Nature 518 (Feb. 2015) 529--533.","DOI":"10.1038\/nature14236"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1007\/11564096_32"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/1143844.1143955"},{"key":"e_1_3_2_1_12_1","volume-title":"Reinforcement Learnging: An Introduction","author":"Sutton R. S.","year":"2018","unstructured":"R. S. Sutton and A. G. Barto . 2018 . Reinforcement Learnging: An Introduction ( second ed.). MIT Press . R. S. Sutton and A. G. Barto. 2018. Reinforcement Learnging: An Introduction (second ed.). MIT Press."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/203330.203343"},{"key":"e_1_3_2_1_14_1","unstructured":"Uber 2018. http:\/\/fortune.com\/2018\/05\/08\/uber-autopilot-death-reason. date accessed: 11-05-2018. Uber 2018. http:\/\/fortune.com\/2018\/05\/08\/uber-autopilot-death-reason. date accessed: 11-05-2018."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/SFCS.1985.12"},{"key":"e_1_3_2_1_16_1","unstructured":"Waymo 2018. https:\/\/en.wikipedia.org\/wiki\/Waymo#Limitations. date accessed: 11-05-2018. Waymo 2018. https:\/\/en.wikipedia.org\/wiki\/Waymo#Limitations. date accessed: 11-05-2018."}],"event":{"name":"CPS-IoT Week '19: Cyber-Physical Systems and Internet of Things Week 2019","acronym":"CPS-IoT Week '19","location":"Montreal Quebec Canada"},"container-title":["Proceedings of the Fifth International Workshop on Symbolic-Numeric methods for Reasoning about CPS and IoT"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3313149.3313369","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,13]],"date-time":"2023-01-13T04:44:40Z","timestamp":1673585080000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3313149.3313369"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,4,15]]},"references-count":16,"alternative-id":["10.1145\/3313149.3313369","10.1145\/3313149"],"URL":"https:\/\/doi.org\/10.1145\/3313149.3313369","relation":{},"subject":[],"published":{"date-parts":[[2019,4,15]]},"assertion":[{"value":"2019-04-15","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}