Understanding structure of concurrent actionsMoodley, P., Rosman, B. and Hong, X. ORCID: https://orcid.org/0000-0002-6832-2298 (2019) Understanding structure of concurrent actions. In: AI-2019: The Thirty-ninth SGAI International Conference, 17-19 Dec 2019, Cambridge, UK, pp. 78-90.
It is advisable to refer to the publisher's version if you intend to cite from this work. See Guidance on citing. Official URL: https://doi.org/10.1007/978-3-030-34885-4_6 Abstract/SummaryWhereas most work in reinforcement learning (RL) ignores the structure or relationships between actions, in this paper we show that exploiting structure in the action space can improve sample efficiency during exploration. To show this we focus on concurrent action spaces where the RL agent selects multiple actions per timestep. Concurrent action spaces are challenging to learn in especially if the number of actions is large as this can lead to a combinatorial explosion of the action space. This paper proposes two methods: a first approach uses implicit structure to perform high-level action elimination using task-invariant actions; a second approach looks for more explicit structure in the form of action clusters. Both methods are context-free, focusing only on an analysis of the action space and show a significant improvement in policy convergence times.
Download Statistics DownloadsDownloads per month over past year Deposit Details University Staff: Request a correction | Centaur Editors: Update this record |