{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,12,27]],"date-time":"2024-12-27T05:16:57Z","timestamp":1735276617175,"version":"3.32.0"},"reference-count":34,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T00:00:00Z","timestamp":1728864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T00:00:00Z","timestamp":1728864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,10,14]]},"DOI":"10.1109\/iros58592.2024.10801876","type":"proceedings-article","created":{"date-parts":[[2024,12,25]],"date-time":"2024-12-25T19:17:39Z","timestamp":1735154259000},"page":"12726-12733","source":"Crossref","is-referenced-by-count":0,"title":["Adaptive Stochastic Nonlinear Model Predictive Control with Look-ahead Deep Reinforcement Learning for Autonomous Vehicle Motion Control"],"prefix":"10.1109","author":[{"given":"Baha","family":"Zarrouki","sequence":"first","affiliation":[{"name":"Technical University of Munich,Chair of Automotive Technology"}]},{"given":"Chenyang","family":"Wang","sequence":"additional","affiliation":[{"name":"Technical University of Munich,Professorship of Autonomous Vehicle Systems, TUM School of Engineering and Design,Garching,Germany,85748"}]},{"given":"Johannes","family":"Betz","sequence":"additional","affiliation":[{"name":"Technical University of Munich,Professorship of Autonomous Vehicle Systems, TUM School of Engineering and Design,Garching,Germany,85748"}]}],"member":"263","reference":[{"key":"ref1","volume-title":"Model predictive control: theory, computation, and design","volume":"2","author":"Rawlings","year":"2017"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.23919\/ECC.2018.8550249"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/j.compchemeng.2020.106844"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2012.6425919"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ACC.2014.6858851"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.23919\/acc60939.2024.10645032"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/tnn.1998.712192"},{"article-title":"Learning to adapt in dynamic, real-world environments through meta-reinforcement learning","year":"2018","author":"Nagabandi","key":"ref8"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1126\/scirobotics.adg1462"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.23919\/ACC53348.2022.9867643"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1146\/annurev-control-090419-075625"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1016\/j.ifacol.2020.12.1196"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CDC45484.2021.9683750"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2018.8593882"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2021.3049335"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2019.2926677"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TCST.2021.3069825"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICCA54724.2022.9831975"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CDC40024.2019.9030270"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"issue":"1","key":"ref21","first-page":"1334","article-title":"End-to-end training of deep visuomotor policies","volume":"17","author":"Levine","year":"2016","journal-title":"The Journal of Machine Learning Research"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1016\/j.automatica.2021.109597"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/MCS.2023.3291885"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1016\/j.ifacol.2021.08.563"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2023.106211"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.23919\/ECC54610.2021.9655042"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ACC.2013.6580560"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1016\/j.automatica.2018.06.026"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1016\/j.ifacol.2020.12.359"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.23919\/ACC.2017.7963836"},{"article-title":"Edgar: An autonomous driving research platform\u2013 from feature development to real-world application","year":"2023","author":"Karle","key":"ref31"},{"article-title":"Proximal policy optimization algorithms","year":"2017","author":"Schulman","key":"ref32"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1007\/s12532-021-00208-8"},{"issue":"268","key":"ref34","first-page":"1","article-title":"Stable-baselines3: Reliable reinforcement learning implementations","volume-title":"Journal of Machine Learning Research","volume":"22","author":"Raffin","year":"2021"}],"event":{"name":"2024 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)","start":{"date-parts":[[2024,10,14]]},"location":"Abu Dhabi, United Arab Emirates","end":{"date-parts":[[2024,10,18]]}},"container-title":["2024 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10801246\/10801290\/10801876.pdf?arnumber=10801876","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,26]],"date-time":"2024-12-26T07:14:58Z","timestamp":1735197298000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10801876\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,14]]},"references-count":34,"URL":"https:\/\/doi.org\/10.1109\/iros58592.2024.10801876","relation":{},"subject":[],"published":{"date-parts":[[2024,10,14]]}}}