P. Anderson, Q. Wu, D. Teney, J. Bruce, M. Johnson et al., Vision-and-language navigation: Interpreting visuallygrounded navigation instructions in real environments, In: CVPR, p.4, 2018.

C. Beattie, J. Z. Leibo, D. Teplyashin, T. Ward, M. Wainwright et al., , p.4, 2016.

G. Brockman, V. Cheung, L. Pettersson, J. Schneider, J. Schulman et al., , p.5, 2016.

S. Brodeur, E. Perez, A. Anand, F. Golemo, L. Celotti et al., HoME: a Household Multimodal Environment, In: ICLR, vol.3, p.4, 2018.
URL : https://hal.archives-ouvertes.fr/hal-01653037

J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, Gated Feedback Recurrent Neural Networks. In: ICML, p.4, 2015.

J. Civera, D. Galvez-lopez, L. Riazuelo, J. D. Tards, and J. M. Montiel, Towards semantic slam using a monocular camera, IROS, p.7, 2011.

A. Das, G. Gkioxari, S. Lee, D. Parikh, and D. Batra, Neural Modular Control for Embodied Question Answering, p.1, 2018.

P. Dhariwal, C. Hesse, O. Klimov, A. Nichol, M. Plappert et al., Openai baselines, p.5, 2017.

A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez, and V. Koltun, CARLA: An open urban driving simulator, CoRL, p.5, 2017.

L. Espeholt, H. Soyer, R. Munos, K. Simonyan, V. Mnih et al., IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures, In: ICML, vol.4, issue.8, 2018.

P. Henderson, R. Islam, P. Bachman, J. Pineau, D. Precup et al., Deep reinforcement learning that matters, p.6, 2018.

S. Hochreiter and J. Schmidhuber, Long Short-Term Memory, Neural Computation, vol.9, issue.8, p.4, 1997.

M. Jaderberg, V. Mnih, W. M. Czarnecki, T. Schaul, J. Z. Leibo et al., Reinforcement learning with unsupervised auxiliary tasks, p.4, 2017.

B. Kayalibay, A. Mirchev, M. Soelch, P. Van-der-smagt, and J. Bayer, Navigation and planning in latent maps, FAIM workshop "Prediction and Generative Modeling in Reinforcement Learning", p.4, 2018.

M. Kempka, M. Wydmuch, G. Runc, J. Toczek, and W. Jaskowski, ViZDoom: A Doom-based AI research platform for visual reinforcement learning, IEEE Conference on Computatonal Intelligence and Games, vol.3, p.7, 2017.

E. Kolve, R. Mottaghi, D. Gordon, Y. Zhu, A. Gupta et al., AI2-THOR: An Interactive 3D Environment for Visual AI, vol.3, p.4, 2017.

I. Kostrikov, Pytorch implementations of reinforcement learning algorithms, p.9, 2018.

G. Lample and D. S. Chaplot, Playing FPS games with deep reinforcement learning, p.4, 2017.

Y. Lecun, L. Eon-bottou, Y. Bengio, and P. Haaner, Gradient-Based Learning Applied to Document Recognition, IEEE, vol.86, issue.11, p.4, 1998.

S. Levine, C. Finn, T. Darrell, and P. Abbeel, End-to-End Training of Deep Visuomotor Policies, Journal of Machine Learning Research, vol.17, p.1, 2016.

T. P. Lillicrap, J. J. Hunt, A. Pritzel, N. Heess, T. Erez et al., Continuous Control with Deep Reinforcement Learning, vol.1, 2015.

M. Savva, *. , A. Kadian, *. , O. Maksymets et al., Habitat: A platform for embodied ai research. arXiv, p.4, 2019.

P. Mirowski, R. Pascanu, F. Viola, H. Soyer, A. J. Ballard et al., Learning to Navigate in Complex Environments, vol.1, p.4, 2017.

V. Mnih, A. P. Badia, M. Mirza, A. Graves, T. P. Lillicrap et al., Asynchronous Methods for Deep Reinforcement Learning, 2016.

V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness et al., Human-level control through deep reinforcement learning, Nature, vol.4, p.10, 2015.

P. Moritz, R. Nishihara, S. Wang, A. Tumanov, R. Liaw et al., Ray: A Distributed Framework for Emerging AI Applications, USENIX Symposium on Operating Systems Design and Implementation, p.5, 2017.

M. Müller, V. Casser, J. Lahoud, N. Smith, and B. Ghanem, Sim4CV: A Photo-Realistic Simulator for Computer Vision Applications. arXiv, vol.5, p.4, 2017.

E. Parisotto and R. Salakhutdinov, Neural map: Structured memory for deep reinforcement learning, p.8, 2017.

M. Savva, A. X. Chang, A. Dosovitskiy, T. Funkhouser, and V. Koltun, MINOS: Multimodal Indoor Simulator for Navigation in Complex Environments, vol.3, p.4, 2017.

M. Schaarschmidt, A. Kuhnle, and K. Fricke, Tensorforce: A tensorflow library for applied reinforcement learning, p.5, 2017.

J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, Proximal policy optimization algorithms, 2017.

D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre et al., Mastering the game of Go with deep neural networks and tree search, Nature, vol.529, issue.7587, p.4, 2016.

A. Stooke and P. Abbeel, Accelerated methods for deep reinforcement learning, p.4, 2018.

L. Van-der-maaten and G. Hinton, Visualizing Data using t-SNE, Journal of Machine Learning Research, vol.9, p.13, 2008.

J. X. Wang, Z. Kurth-nelson, D. Tirumala, H. Soyer, J. Z. Leibo et al., Learning to reinforcement learn, p.4, 2016.
URL : https://hal.archives-ouvertes.fr/hal-02613179

R. J. Williams, Simple statistical gradient following algorithms for connectionist reinforcement learning, Machine Learning, p.10, 1992.

Y. Wu, E. Mansimov, S. Liao, R. Grosse, and J. Ba, Scalable trust-region method for deep reinforcement learning using Kronecker-factored approximation, NIPS, issue.2, 2017.

F. Xia, R. Zamir, A. He, Z. Y. Sax, A. Malik et al., Gibson env: realworld perception for embodied agents, p.4, 2018.

C. Zhang, O. Vinyals, R. Munos, and S. Bengio, A study on overfitting in deep reinforcement learning. arxiv, p.6, 2018.