Skip to Main content Skip to Navigation
Journal articles

Adaptive reinforcement learning with active state-specific exploration for engagement maximization during simulated child-robot interaction

Abstract : Using assistive robots for educational applications requires robots to be able to adapt their behavior specifically for each child with whom they interact. Among relevant signals, non-verbal cues such as the child's gaze can provide the robot with important information about the child's current engagement in the task, and whether the robot should continue its current behavior or not. Here we propose a reinforcement learning algorithm extended with active state-specific exploration and show its applicability to child engagement maximization as well as more classical tasks such as maze navigation. We first demonstrate its adaptive nature on a continuous maze problem as an enhancement of the classic grid world. There, parame-terized actions enable the agent to learn single moves until the end of a corridor, similarly to "options" but without explicit hierarchical representations. We then apply the algorithm to a series of simulated scenarios, such as an extended Tower of Hanoi where the robot should find the appropriate speed of movement for the interacting child, and to a pointing task where the robot should find the child-specific appropriate level of expressivity of action. We show that the algorithm enables to cope with both global and local non-stationarities in the state space while preserving a stable behavior in other stationary portions of the state space. Altogether, these results suggest a promising way to enable robot learning based on non-verbal cues and the high degree of non-stationarities that can occur during interaction with children.
Document type :
Journal articles
Complete list of metadata

Cited literature [32 references]  Display  Hide  Download

https://hal.archives-ouvertes.fr/hal-02324073
Contributor : Mehdi Khamassi <>
Submitted on : Monday, October 21, 2019 - 6:46:35 PM
Last modification on : Wednesday, May 19, 2021 - 11:58:12 AM
Long-term archiving on: : Wednesday, January 22, 2020 - 7:16:33 PM

File

Velentzas2018_Paladyn.pdf
Publication funded by an institution

Licence


Distributed under a Creative Commons Attribution - NoDerivatives 4.0 International License

Identifiers

Citation

George Velentzas, Theodore Tsitsimis, Iñaki Rañó, Costas Tzafestas, Mehdi Khamassi. Adaptive reinforcement learning with active state-specific exploration for engagement maximization during simulated child-robot interaction. Paladyn: Journal of Behavioral Robotics, De Gruyter, 2018, 9 (1), pp.235-253. ⟨10.1515/pjbr-2018-0016⟩. ⟨hal-02324073⟩

Share

Metrics