Abstract
A combined reinforcement learning (RL) and motion planning framework is proposed in this paper to address a multi-class in-rack test tube rearrangement problem. The RL works at the task level to plan a sequence of swap actions while ignoring the details of robotic motion. The motion planning works at the motion level to plan detailed robotic pick-and-place motion based on the action sequences obtained at the task level. The task level and motion level are combined in a closed loop with the help of a condition set maintained for each rack slot, which allows the framework to perform replanning efficiently and thus effectively find solutions in the presence of failures. In particular, for RL, the framework leverages a distributed deep Q learning structure with the dueling double deep Q network (D3QN) to learn policies from training data amplified by an A⋆-based post-processing technique. The D3QN and distributed learning help increase training efficiency. The post-processing helps salvage failed action sequences and remove redundancy, thus making training data more effective. Simulations and real-world studies are carried out in the experiments to understand the performance of the proposed framework. The results verify the advantages of the RL and post-processing, and show that the closed-loop combination improves robustness. The framework is ready to incorporate various sensory feedbacks, as demonstrated by real-world studies.
Get full access to this article
View all access options for this article.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
