R. Muhammad-awais-abbas, . Milman-ieee, E. Mikael, and . Smieee, Obstacle Avoidance in Real Time with Nonlinear Model Predictive Control of Autonomous Vehicles, 2014.

P. Abbeel and A. Y. Ng, Apprenticeship learning via inverse reinforcement learning, Twenty-rst international conference on Machine learning-ICML '04, 2004.

P. Abbeel, D. Dolgov, Y. Andrew, S. Ng, and . Thrun, Apprenticeship Learning for Motion Planning, with Application to Parking Lot Navigation, 2008.

F. Altché and A. Fortelle, An LSTM Network for Highway Trajectory Prediction, p.353359, 2017.

J. Bagnell, D. Bradley, D. Silver, B. Sofman, and A. Stentz, Learning for autonomous navigation, IEEE Robotics and Automation Magazine, vol.17, issue.2, p.7484, 2010.

H. Bai, S. Cai, N. Ye, D. Hsu, and W. Lee, Intention-Aware Online POMDP Planning for Autonomous Driving in a Crowd, 2015.

T. Bandyopadhyay, E. Kok-sung-won, D. Frazzoli, and . Hsu, Wee Sun Lee, and Daniela Rus. Intention-Aware Motion Planning, 2013.

D. Barnes, W. Maddern, and I. Posner, Find your own way: Weakly-supervised segmentation of path proposals for urban autonomy, Proceedings-IEEE International Conference on Robotics and Automation, p.203210, 2017.

P. Raunak, D. J. Bhattacharyya, B. Phillips, J. Wulfe, A. Morton et al., Multi-Agent Imitation Learning for Driving Simulation, 2018.

M. Bojarski, D. D. Testa, D. Dworakowski, B. Firner, B. Flepp et al., End to End Learning for Self-Driving Cars, 2016.

S. Brechtel, T. Gindele, and R. Dillmann, Probabilistic MDP-behavior planning for cars, IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, 2011.

S. Brechtel, T. Gindele, and R. Dillmann, Probabilistic Decision-Making under Uncertainty for Autonomous Driving using Continuous POMDPs, 2014.

L. Cardamone, D. Loiacono, and P. L. Lanzi, Evolving Competitive Car Controllers for Racing Games with Neuroevolution, 2009.

C. Chen, A. Se, A. Kornhauser, and J. Xiao, DeepDriving: Learning Aordance for Direct Perception in Autonomous Driving, 2015.

S. Yu-fan-chen, M. Liu, J. Liu, J. P. Miller, and . How, Motion Planning with Diusion Maps, 2016.

M. Yu-fan-chen, M. Everett, J. P. Liu, and . How, Socially Aware Motion Planning with Deep Reinforcement Learning, 2017.

F. Codevilla, M. Müller, A. Dosovitskiy, A. López, and V. Koltun, End-to-end Driving via Conditional Imitation Learning, 2017.

R. Coulom, Apprentissage par renforcement utilisant des réseaux de neurones, avec des applications au contrôle moteur, 2002.

K. Driggs, -. , and R. Bajcsy, Identifying Modes of Intent from Driver Behaviors in Dynamic Environments, IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, 2015.

Y. Du, D. Hsu, H. Kurniawati, . Wee-sun-lee, C. W. Sylvie et al., A POMDP Approach to Robot Motion Planning under Uncertainty, 2010.

M. Hesham, M. N. Eraqi, J. Moustafa, and . Honer, End-to-End Deep Learning for Steering Autonomous Vehicles Considering Temporal Dependencies, 2017.

C. Finn and S. Levine, Deep visual foresight for planning robot motion, Proceedings-IEEE International Conference on Robotics and Automation, 2017.

J. Forbes, T. Huang, K. Kanazawa, and S. Russell, The BATmobile: Towards a Bayesian Automated Taxi, 1995.

L. Fridman, B. Jenik, and J. Terwilliger, DeepTrac: Driving Fast through Dense Trac with Deep Reinforcement Learning, vol.18, 2018.

E. Galceran, A. G. Cunningham, R. M. Eustice, and E. Olson, Multipolicy decision-making for autonomous driving via changepoint-based behavior prediction: Theory and experiment, Autonomous Robots, 2017.

J. Garcia and F. Fernandez, Safe exploration of state and action spaces in reinforcement learning, Journal of Articial Intelligence Research, vol.45, p.515564, 2012.

T. Gindele, S. Brechtel, and R. Dillmann, Learning driver behavior models from trac observations for decision making and planning. IEEE Intelligent Transportation Systems Magazine, vol.7, p.6979, 2015.

M. Gómez-plaza, T. Martinez-marin, S. S. Prieto, and D. Luna, Integration of cell-mapping and reinforcement-learning techniques for motion planning of car-like robots, IEEE Transactions on Instrumentation and Measurement, 2009.

D. Ha and J. Schmidhuber, World Models, 2018.

R. Hadsell, P. Sermanet, J. Ben, A. Erkan, M. Scoer et al., Learning long-range vision for autonomous o-road driving, Journal of Field Robotics, vol.26, issue.2, p.120144, 2009.

T. Hester and P. Stone, TEXPLORE: Real-Time Sample-Ecient Reinforcement Learning for Robots, 2006.

S. Hoermann, M. Bach, and K. Dietmayer, Dynamic Occupancy Grid Prediction for Urban Autonomous Driving: A Deep Learning Approach with Fully Automatic Labeling, 2017.

H. Sandy, D. Huang, P. Held, A. D. Abbeel, and . Dragan, Enabling Robots to Communicate their Objectives, 2017.

D. Isele, A. Cosgun, K. Subramanian, and K. Fujimura, Navigating Intersections with Autonomous Vehicles using Deep Reinforcement Learning, 2017.

A. Kendall, J. Hawke, D. Janz, P. Mazur, and D. Reda, Learning to Drive in a Day, 2018.

S. Koppula, Learning a CNN-based End-to-End Controller for a Formula Racecar, p.26, 2017.

J. Koutník, G. Cuccu, J. Schmidhuber, and F. Gomez, Evolving LargeScale Neural Networks for Vision-Based TORCS, 2013.

H. Kretzschmar, M. Kuderer, and W. Burgard, Learning to predict trajectories of cooperatively navigating agents, Proceedings-IEEE International Conference on Robotics and Automation, 2014.

H. Kretzschmar, M. Spies, C. Sprunk, and W. Burgard, Socially Compliant Mobile Robot Navigation via Inverse Reinforcement Learning, 2016.

M. Kuderer, S. Gulati, and W. Burgard, Learning Driving Styles for Autonomous Vehicles from Demonstration, 2015.

A. Kueer, J. Morton, T. Wheeler, and M. Kochenderfer, Imitating Driver Behavior with Generative Adversarial Networks, 2017.

S. Lee and S. Seo, A Learning-Based Framework for Handling Dilemmas in Urban Automated Driving, 2017.

S. Levine and V. Koltun, Continuous Inverse Optimal Control with Locally Optimal Examples. International Conference on Machine Learning (ICML), p.4148, 2012.

S. Levine, Z. Popovic, and V. Koltun, Nonlinear Inverse Reinforcement Learning with Gaussian Processes, Advances in Neural Information Processing Systems, p.1927, 2011.

N. Li, D. Oyler, M. Zhang, Y. Yildiz, I. Kolmanovsky et al., Game-Theoretic Modeling of Driver and Vehicle Interactions for Verication and Validation of Autonomous Vehicle Control Systems, p.113, 2017.

A. Liniger, A. Domahidi, and M. Morari, Optimization-based autonomous racing of 1:43 scale RC cars, Optimal Control Applications and Methods, 2014.

J. Liu, P. Hou, L. Mu, Y. Yu, and C. Huang, Elements of Eective Deep Reinforcement Learning towards Tactical Driving Decision Making, 2018.

D. Loiacono, A. Prete, P. L. Lanzi, and L. Cardamone, Learning to overtake in TORCS using simple reinforcement learning, 2010 IEEE World Congress on Computational Intelligence, WCCI 2010-2010 IEEE Congress on Evolutionary Computation, CEC 2010, 2010.

L. Manuelli and P. Florence, Reinforcement Learning for Autonomous Driving Obstacle Avoidance using LIDAR, 2015.

V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness et al.,

M. Mukadam, A. Cosgun, A. Nakhaei, and K. Fujimura, Tactical Decision Making For Lane Changing With Deep Reinforcement Learning, p.110, 2018.

D. Osipychev, D. Tran, W. Sheng, G. Chowdhary, and R. Zeng, Proactive MDP-based collision avoidance algorithm for autonomous cars, 2015 IEEE International Conference on Cyber Technology in Automation, Control and Intelligent Systems, 2015.

X. Pan, Y. You, Z. Wang, and C. Lu, Virtual to Real Reinforcement Learning for Autonomous Driving, 2017.

C. Paxton, V. Raman, D. Gregory, M. Hager, and . Kobilarov, Combining Neural Networks and Tree Search for Task and Motion Planning in Challenging Environments, 2017.

M. Pfeier, U. Schwesinger, H. Sommer, E. Galceran, and R. Siegwart, Predicting actions to act predictably: Cooperative partial motion planning with maximum entropy models, IEEE International Conference on Intelligent Robots and Systems, 2016.

M. Pfeier, G. Paolo, H. Sommer, J. Nieto, R. Siegwart et al., A Data-driven Model for Interaction-aware Pedestrian Motion Prediction in Object Cluttered Environments, 2017.

M. Pfeier, M. Schaeuble, J. Nieto, R. Siegwart, and C. Cadena, From perception to decision: A data-driven approach to end-to-end motion planning for autonomous ground robots, Proceedings-IEEE International Conference on Robotics and Automation, 2017.

. Mogens-graf-plessen, Automating Vehicles by Deep Reinforcement Learning using Task Separation with Hill Climbing, p.116, 2017.

A. Dean and . Pomerleau, ALVINN, an autonomous land vehicle in a neural network, 1989.

D. Larry, A. E. Pyeatt, and . Howe, Learning to Race: Experiments with a Simulated Race Car, 1998.

C. R. Qi, H. Su, K. Mo, and L. J. Guibas, PointNet: Deep Learning on Point Sets for 3D Classication and Segmentation, Proceedings-2016 4th International Conference on 3D Vision, 3DV 2016, p.601610, 2016.

X. Qian, F. Altché, P. Bender, C. Stiller, and A. Fortelle, Optimal trajectory planning for autonomous driving integrating logical constraints: An MIQP perspective, IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, p.205210, 2016.
URL : https://hal.archives-ouvertes.fr/hal-01342358

E. Rehder and M. Naumann, Niels Ole Salscheider, and Christoph Stiller. Cooperative Motion Planning for Non-Holonomic Agents with Value Iteration Networks, 2017.

E. Rehder, J. Quehl, and C. Stiller, Driving Like a Human: Imitation Learning for Path Planning using Convolutional Neural Networks. (October), 2017.

E. Rehder, F. Wirth, M. Lauer, and C. Stiller, Pedestrian Prediction by Planning using Deep Neural Networks, 2017.

M. Rezagholizadeh and . Haidar, Semi-supervised Regression with Generative Adversarial Networks for End to End Learning in Autonomous Driving, p.19, 2018.

N. Rhinehart, R. Mcallister, and S. Levine, Deep Imitative Models for Flexible Inference, Planning, and Control, vol.112, 2018.

C. Richter, J. Ware, and N. Roy, High-Speed Autonomous Navigation of Unknown Environments using Learned Probabilities of Collision, 2014.

M. Riedmiller, M. Montemerlo, and H. Dahlkamp, Learning to drive a real car in 20 minutes, Proceedings of the Frontiers in the Convergence of Bioscience and Information Technologies, FBIT 2007, p.645650, 2007.

S. Ross, J. Georey, J. Gordon, and . Bagnell, A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning, vol.15, 2011.

D. Sadigh, S. Sastry, A. Sanjit, A. Seshia, and . Dragan, Planning for Autonomous Cars that Leverage Eects on Human Actions, Proceedings of Robotics: Science and Systems, 2016.

A. E. Sallab, M. Abdou, E. Perot, and S. Yogamani, End-to-End Deep Reinforcement Learning for Lane Keeping Assist, p.19, 2016.

A. E. Sallab, M. Abdou, E. Perot, and S. Yogamani, Deep Reinforcement Learning framework for Autonomous Driving, 2017.

A. Se and J. Xiao, Learning from Maps: Visual Common Sense for Autonomous Driving, vol.2, 2016.

S. Shalev, -. Shwartz, and A. Shashua, On the Sample Complexity of End-to-end Training vs. Semantic Abstraction Training, vol.14, 2016.

S. Shalev-shwartz, N. Ben-zrihem, A. Cohen, and A. Shashua, Long-term Planning by Short-term Prediction, vol.19, 2016.

S. Shalev-shwartz, S. Shammah, and A. Shashua, Safe, Multi-Agent, Reinforcement Learning for Autonomous Driving, 2016.

S. Shalev-shwartz, S. Shammah, and A. Shashua, On a Formal Model of Safe and Scalable Self-driving Cars, 2017.

T. Shankar, K. Santosha, P. Dwivedy, and . Guha, Reinforcement Learning via Recurrent Convolutional Neural Networks, 2016.

S. Sharifzadeh, I. Chiotellis, R. Triebel, and D. Cremers, Learning to Drive using Inverse Reinforcement Learning and Deep Q-Networks, p.17, 2016.

W. Song, G. Xiong, and H. Chen, Intention-Aware Autonomous Driving Decision-Making in an Uncontrolled Intersection, Mathematical Problems in Engineering, 2016.

C. Stiller and J. Ziegler, 3D perception and planning for self-driving and cooperative automobiles, International Multi-Conference on Systems, Signals and Devices, SSD 2012-Summary Proceedings, 2012.

Z. N. Sunberg, C. J. Ho, and M. J. Kochenderfer, The value of inferring the internal state of trac participants for autonomous freeway driving, Proceedings of the American Control Conference, p.30043010, 2017.

D. Richard-s-sutton, S. Precup, and . Singh, Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning, Articial Intelligence, vol.112, p.181211, 1999.

A. Tamar, Y. Wu, G. Thomas, S. Levine, and P. Abbeel, Value Iteration Networks, 2016.

H. Tehrani, Q. Huy-do, M. Egawa, K. Muto, K. Yoneda et al., General behavior and motion model for automated lane change, IEEE Intelligent Vehicles Symposium, Proceedings, 2015.

G. Trehard, E. Pollard, B. Bradai, and F. Nashashibi, On line mapping and global positioning for autonomous driving in urban environment based on evidential SLAM, IEEE Intelligent Vehicles Symposium, Proceedings, volume 2015August, p.814819, 2015.
URL : https://hal.archives-ouvertes.fr/hal-01149504

S. Ulbrich and M. Maurer, Probabilistic online POMDP decision making for lane changes in fully automated driving, IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, 2013. ISBN 9781479929146

S. Ulbrich and M. Maurer, Situation Assessment in Tactical Lane Change Behavior Planning for Automated Vehicles, IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, 2015. ISBN 9781467365956

J. Van-den, . Heuvel, A. Marco, W. Wiering, and . Kosters, Temporal exploration for reinforcement learning in continuous action spaces, 2013.

P. Wang and C. Chan, Formulation of Deep Reinforcement Learning Architecture Toward Autonomous Driving for On-Ramp Merge, 2017.

J. Wei, J. M. Dolan, J. M. Snider, and B. B. Litkouhi, A Point-based Markov Decision Process for Robust Single-Lane Autonomous Driving Behavior under Uncertainties. Signicance, p.121, 2010.

T. A. Wheeler, P. Robbel, and M. J. Kochenderfer, A Probabilistic Framework for Microscopic Trac Propagation, IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, 2015.

G. Williams, N. Wagener, B. Goldfain, P. Drews, B. James-m-rehg et al., Information Theoretic MPC for Model-Based Reinforcement Learning. Robotics and Automation (ICRA, 2017 IEEE International Conference on, p.17141721, 2017.

G. Williams, B. Goldfain, P. Drews, K. Saigol, . James-m-rehg et al., Robust Sampling Based Model Predictive Control with Sparse Objective Information, 2018.

K. Hollins-wray, S. J. Witwicki, and S. Zilberstein, Online decision-making for scalable autonomous systems, IJCAI International Joint Conference on Articial Intelligence, p.47684774, 2017.

M. Wulfmeier, D. Z. Wang, and I. Posner, Watch This: Scalable CostFunction Learning for Path Planning in Urban Environments, 2016.

H. Xu, Y. Gao, F. Yu, and T. Darrell, End-to-end Learning of Driving Models from Large-scale Video Datasets, 2016.

A. Yu, R. Palefsky-smith, and R. Bedi, Deep Reinforcement Learning for Simulated Autonomous Vehicle Control, 2016.

A. L. Brian-d-ziebart, . Maas, K. Anind, J. Dey, and . Bagnell, Navigate like a cabbie: Probabilistic reasoning from observed context-aware behavior, 10Th International Conference on Ubiquitous Computing, p.322331, 2008.

N. Brian-d-ziebart, G. Ratli, C. Gallagher, K. Mertz, A. Peterson et al., Planningbased Prediction for Pedestrians, p.39313936, 2009.

J. Ziegler, P. Bender, T. Dang, and C. Stiller, Trajectory planning for Bertha-A local, continuous method, IEEE Intelligent Vehicles Symposium, Proceedings, 2014.