A friction-model-based framework for reinforcement learning of robotic tasks in non-rigid environments
Document typeConference report
PublisherInstitute of Electrical and Electronics Engineers (IEEE)
Rights accessOpen Access
Learning motion tasks in a real environment with deformable objects requires not only a Reinforcement Learning (RL) algorithm, but also a good motion characterization, a preferably compliant robot controller, and an agent giving feedback for the rewards/costs in the RL algorithm. In this paper, we unify all these parts in a simple but effective way to properly learn safety-critical robotic tasks such as wrapping a scarf around the neck (so far, of a mannequin). We found that a suitable compliant controller ought to have a good Inverse Dynamic Model (IDM) of the robot. However, most approaches to build such a model do not consider the possibility of having hystheresis on the friction, which is the case for robots such as the Barrett WAM. For this reason, in order to improve the available IDM, we derived an analytical model of friction in the seven robot joints, whose parameters can be automatically tuned for each particular robot. This permits compliantly tracking diverse trajectories in the whole workspace. By using such friction-aware controller, Dynamic Movement Primitives (DMP) as motion characterization and visual/force feedback within the RL algorithm, experimental results demonstrate that the robot is consistently capable of learning such safety-critical tasks.
CitationColomé, A., Planells, A., Torras, C. A friction-model-based framework for reinforcement learning of robotic tasks in non-rigid environments. A: IEEE International Conference on Robotics and Automation. "2015 IEEE International Conference on Robotics and Automation (ICRA 2015): Seattle, Washington, USA, 26-30 May 2015". Seattle, WA: Institute of Electrical and Electronics Engineers (IEEE), 2015, p. 5649-5654.