skip to main content
10.1145/3501409.3501550acmotherconferencesArticle/Chapter ViewAbstractPublication PageseitceConference Proceedingsconference-collections
research-article

Concatenated Dynamic Reinforcement Learning for Multi-staged Tasks (MST)

Published: 31 December 2021 Publication History

Abstract

There exists a class of complex tasks that can be distinctly separated into multiple concatenated simpler tasks [1-2]. This class of complex and continuous motion processes can then be considered as multiple motion sub-processes, and a switch that brings these sub-processes together in accordance to an input function. This paper validates the potential value of this approach by proposing a method to train a complex task through a series of sub-tasks, and then concatenating them through a switch. Through the classic simulation physical task of OpenAI Gym[3]: cartpole, the paper demonstrates the viability and value of this approach. An example complex task of moving the pendulum back and forth between to points was used. A set of reward and switching functions were developed to achieve the complex task though the use of Deep Q-network (DQN) [4]. Results show that the proposed approach achieves the goal within a limited number of training cycles that otherwise cannot be met by other traditional means. Alternative reinforcement learning methods would result in more input parameters, larger deep learning networks, and larger accumulative training cycles.

References

[1]
Guanwen Ding, Yubin Liu, Xizhe Zang, Xuehe Zhang, Jie Zhao. A Task-Learning Strategy for Robotic Assembly Tasks from Human Demonstrations. Sensors, 2020
[2]
W Lee, N Vaughan, D Kim. Task Allocation into a Foraging Task with a Series of Subtasks in Swarm Robotic System. IEEE, 2020
[3]
G Brockman, V Cheung, L Pettersson, J Schneider, J Schulman, J Tang, W Zaremba. OpenAI Gym.
[4]
Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, Martin Riedmiller. Playing Atari with Deep Reinforcement Learning. NIPS Deep Learning Workshop, 2013
[5]
Richard S. Sutton, Andrew G. Barto. Reinforcement Learning: An Introduction. Bradford Books; second edition, 2018
[6]
Panin, A. and Shvechikov, P., Practical Reinforcement Learning. Coursera and National Research University Higher School of Economics, 2017
[7]
Bellman, R.E., 1957. A Markov decision process. Journal of Mathematical Mechanics, 6, pp. 679--684.
[8]
Dimitri P. Bertsekas. Abstract Dynamic Programming. Athena Scientific; 2nd edition, 2018
[9]
B Jang, M Kim, G Harerimana, JW Kim. Q-learning Algorithms: A Comprehensive Classification and Applications. IEEE, 2019
[10]
Volodymyr Mnih*, Koray Kavukcuoglu*, David Silver1*, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves, Martin Riedmiller, Andreas K. Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan Wierstra, Shane Legg, Demis Hassabis. Human-level Control Through Deep Reinforcement Learning. Nature, 2015
[11]
S Gadgil, Y Xin, C Xu. Solving The Lunar Lander Problem under Uncertainty using Reinforcement Learning. IEEE, 2020
[12]
K.P. Modi, F. Sahin, E. Saber. An application of human robot interaction: development of a ping-pong playing robotic arm. IEEE, 2005
[13]
Vijay R. Konda, John N. Tsitsiklis. On Actor-Critic Algorithms. Society for Industrial and Applied Mathematics, 2003
[14]
Volodymyr Mnih, Adrià Puigdomènech Badia, Mehdi Mirza, Alex Graves, Tim Harley, Timothy P. Lillicrap, David Silver, Koray Kavukcuoglu. Asynchronous Methods for Deep Reinforcement Learning. ICML, 2016
[15]
Timothy P. Lillicrap*, Jonathan J. Hunt*, Alexander Pritzel, Nicolas Heess. Continuous Control with Deep Reinforcement Learning. ICLR, 2016
[16]
Hado van Hasselt, Arthur Guez, David Silver. Deep Reinforcement Learning with Double Q-learning. AAAI, 2016
[17]
Chelsea Finn, Pieter Abbeel, Sergey Levine. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. ICML, 2017
[18]
John D. Co-Reyes, Yingjie Miao, Daiyi Peng, Esteban Real, Sergey Levine, Quoc V. Le, Honglak Lee, Aleksandra Faust*. Evolving Reinforcement Learning Algorithms. ICLR, 2021

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
EITCE '21: Proceedings of the 2021 5th International Conference on Electronic Information Technology and Computer Engineering
October 2021
1723 pages
ISBN:9781450384322
DOI:10.1145/3501409
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 31 December 2021

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Concatenate
  2. DQN
  3. MST
  4. Multi-stage task
  5. OpenAI Gym
  6. Reinforcement learning
  7. Reward
  8. inverted pendulum

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

EITCE 2021

Acceptance Rates

EITCE '21 Paper Acceptance Rate 294 of 531 submissions, 55%;
Overall Acceptance Rate 508 of 972 submissions, 52%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)2
  • Downloads (Last 6 weeks)0
Reflects downloads up to 06 Nov 2024

Other Metrics

Citations

Cited By

View all

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media