Reinforcement Learning: An Introduction (Adaptive Computation and Machine Learning). A Bradford Book, 1998. ,
Human-level control through deep reinforcement learning, Nature, vol.518, pp.529-533, 2015. ,
Unifying count-based exploration and intrinsic motivation, Advances in Neural Information Processing Systems, pp.1471-1479, 2016. ,
Curriculum learning, Proceedings of the 26th annual international conference on machine learning, pp.41-48, 2009. ,
Transfer Learning for Reinforcement Learning Domains : A Survey, Journal of Machine Learning Research, vol.10, pp.1633-1685, 2009. ,
Cognitive Developmental Robotics : A Survey, IEEE Transactions on Autonomous Mental Development, vol.1, issue.1, pp.1-44, 2009. ,
Learning like a baby: a survey of artificial intelligence approaches, The Knowledge Engineering Review, vol.26, issue.02, pp.209-236, 2011. ,
The arcade learning environment: An evaluation platform for general agents, International Joint Conference on Artificial Intelligence, vol.47, pp.253-279, 2013. ,
OpenAI Gym, 2016. ,
Open dynamics engine, 2005. ,
Mujoco: A physics engine for modelbased control, IEEE/RSJ International Conference on Intelligent Robots and Systems, pp.5026-5033, 2012. ,
Relative Entropy Policy Search, Association for the Advancement of Artificial Intelligence. Atlanta, pp.1607-1612, 2010. ,
Simple statistical gradient-following algorithms for connectionist reinforcement learning, Machine learning, vol.8, issue.3-4, pp.229-256, 1992. ,
Completely derandomized selfadaptation in evolution strategies, Evolutionary computation, vol.9, issue.2, pp.159-195, 2001. ,
Actor-Critic Algorithms, Neural Information Processing Systems, vol.13, pp.1008-1014, 1999. ,
Deep Reinforcement Learning that Matters, 2017. ,
, Continuous control with deep reinforcement learning, 2015.
Reinforcement learning in feedback control, Machine Learning, vol.84, pp.137-169, 2011. ,
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift, 2015. ,
Neural Fitted Actor-Critic, European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, 2016. ,
URL : https://hal.archives-ouvertes.fr/hal-01350651
Neural fitted Q iteration-First experiences with a data efficient neural Reinforcement Learning method, Lecture Notes in Computer Science, vol.3720, pp.317-328, 2005. ,
Reinforcement Learning in Continuous State and Action Spaces, Reinforcement Learning, pp.207-251, 2012. ,
Apprentissage par renforcement développemental, 2018. ,
Dropout: A Simple Way to Prevent Neural Networks from Overfitting, Journal of Machine Learning Research, vol.15, pp.1929-1958, 2014. ,
Catastrophic forgetting in connectionist networks, Trends in Cognitive Sciences, vol.3, issue.4, pp.128-135, 1999. ,
Overcoming catastrophic forgetting in neural networks, Proceedings of the National Academy of Sciences, p.201611835, 2017. ,
How can we define intrinsic motivation, 8th Conf on Epigenetic Robotics, pp.93-101, 2008. ,
URL : https://hal.archives-ouvertes.fr/inria-00420175
Learning to control a 6-degree-of-freedom walking robot, International Conference on Computer as a Tool, pp.698-705, 2007. ,
Synthesis and stabilization of complex behaviors through online trajectory optimization, IEEE/RSJ International Conference on Intelligent Robots and Systems, pp.4906-4913, 2012. ,
Adam: a Method for Stochastic Optimization, International Conference on Learning Representations, pp.1-13, 2015. ,
Bootstrapping Q-Learning for Robotics from Neuro-Evolution Results, IEEE Transactions on Cognitive and Developmental Systems, 2017. ,
URL : https://hal.archives-ouvertes.fr/hal-01494744
Why does unsupervised pre-training help deep learning?, Journal of Machine Learning Research, vol.11, pp.625-660, 2010. ,
, Progressive Neural Networks, 2016.
Pathnet: Evolution channels gradient descent in super neural networks, 2017. ,