M. Bojarski, D. D. Testa, D. Dworakowski, B. Firner, B. Flepp et al., End to end learning for self-driving cars, 2016.

C. Chen, A. Seff, A. Kornhauser, and J. Xiao, Deepdriving: Learning affordance for direct perception in autonomous driving, Proceedings of the IEEE International Conference on Computer Vision, pp.2722-2730, 2015.

D. Chen, B. Zhou, and V. Koltun, Learning by Cheating, 2019.

F. Codevilla, M. Miiller, A. López, V. Koltun, and A. Dosovitskiy, End-to-end driving via conditional imitation learning, 2018 IEEE International Conference on Robotics and Automation (ICRA), pp.1-9, 2018.

F. Codevilla, E. Santana, A. M. Lpez, and A. Gaidon, Exploring the limitations of behavior cloning for autonomous driving, 2019.

W. Dabney, G. Ostrovski, D. Silver, and R. Munos, Implicit quantile networks for distributional reinforcement learning, 2018.

A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez, and V. Koltun, CARLA: An open urban driving simulator, Proceedings of the 1st Annual Conference on Robot Learning, pp.1-16, 2017.

L. Espeholt, H. Soyer, R. Munos, K. Simonyan, V. Mnih et al., Impala: Scalable distributed deep-rl with importance weighted actor-learner architectures, 2018.

M. Fortunato, M. G. Azar, B. Piot, J. Menick, I. Osband et al., Noisy networks for exploration, 2017.

D. Gordon, A. Kadian, D. Parikh, J. Hoffman, and D. Batra, Splitnet: Sim2sim and task2task transfer for embodied visual navigation, 2019.

K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, 2015.

M. Hessel, J. Modayil, H. Van-hasselt, T. Schaul, G. Ostrovski et al., Rainbow: Combining improvements in deep reinforcement learning, 2018.

D. Horgan, J. Quan, D. Budden, G. Barth-maron, and M. Hessel, Hado Van Hasselt, and David Silver. Distributed prioritized experience replay, 2018.

S. Ioffe and C. Szegedy, Batch normalization: Accelerating deep network training by reducing internal covariate shift, 2015.

M. Jaderberg, V. Mnih, W. M. Czarnecki, T. Schaul, J. Z. Leibo et al., Reinforcement learning with unsupervised auxiliary tasks, 2016.

M. Jaritz, R. De-charette, M. Toromanoff, E. Perot, and F. Nashashibi, End-to-end race driving with deep reinforcement learning, IEEE International Conference on Robotics and Automation (ICRA), pp.2070-2075, 2018.
URL : https://hal.archives-ouvertes.fr/hal-01848067

S. Kapturowski, G. Ostrovski, J. Quan, R. Munos, and W. Dabney, Recurrent experience replay in distributed reinforcement learning, 2019.

A. Kendall, J. Hawke, D. Janz, P. Mazur, D. Reda et al., Learning to drive in a day, 2019 International Conference on Robotics and Automation (ICRA), pp.8248-8254, 2019.

P. Diederik, M. Kingma, and . Welling, Auto-encoding variational bayes, 2013.

X. Liang, T. Wang, L. Yang, and E. Xing, Cirl: Controllable imitative reinforcement learning for vision-based self-driving, Proceedings of the European Conference on Computer Vision (ECCV), pp.584-599, 2018.

P. Timothy, J. J. Lillicrap, A. Hunt, N. Pritzel, T. Heess et al., Continuous control with deep reinforcement learning, 2015.

L. Liu, H. Jiang, P. He, W. Chen, X. Liu et al., On the variance of the adaptive learning rate and beyond, 2019.

A. Mehta, A. Subramanian, and A. Subramanian, Learning end-to-end autonomous driving using guided auxiliary supervision, 2018.

V. Mnih, A. P. Badia, M. Mirza, A. Graves, T. Lillicrap et al., Asynchronous methods for deep reinforcement learning, International conference on machine learning, pp.1928-1937, 2016.

V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness et al., Human-level control through deep reinforcement learning, Nature, vol.518, issue.7540, p.529, 2015.

X. Pan, X. Chen, Q. Cai, J. Canny, and F. Yu, Semantic predictive control for explainable and efficient policy learning, 2019 International Conference on Robotics and Automation (ICRA), pp.3203-3209, 2019.

A. Paszke, A. Chaurasia, S. Kim, and E. Culurciello, Enet: A deep neural network architecture for real-time semantic segmentation, 2016.

A. Dean and . Pomerleau, Alvinn: An autonomous land vehicle in a neural network, Advances in neural information processing systems, pp.305-313, 1989.

O. Ronneberger, P. Fischer, and T. Brox, U-net: Convolutional networks for biomedical image segmentation, 2015.

G. Ros, V. Koltun, F. Codevilla, and A. Lopez, The CARLA Autonomous Driving Challenge, 2019.

A. Sauer, N. Savinov, and A. Geiger, Conditional affordance learning for driving in urban environments, 2018.

K. Simonyan and A. Zisserman, Very deep convolutional networks for large-scale image recognition, 2014.

M. Toromanoff and E. Wirbel, Is Deep Reinforcement Learning Really Superhuman on Atari? Leveling the playing field, 2019.

M. Toromanoff, E. Wirbel, F. Wilhelm, C. Vejarano, X. Perrotton et al., End to end vehicle lateral control using a single fisheye camera, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp.3613-3619, 2018.
URL : https://hal.archives-ouvertes.fr/hal-01861697

Z. Wang, T. Schaul, M. Hessel, M. Hado-van-hasselt, N. Lanctot et al., Dueling network architectures for deep reinforcement learning, 2015.

B. Wymann, C. Dimitrakakis, A. Sumner, E. Espié, and C. Guionneau, TORCS: The open racing car simulator, 2015.

B. Zej, O. Osinski, and A. Jakubowski, Simulation-based reinforcement learning for real-world autonomous driving, Christopher Galias, Silviu Homoceanu, and Henryk Michalewski, 2019.