Mostra el registre d'ítem simple
Safe robot execution in model-based reinforcement learning
dc.contributor.author | Martínez Martínez, David |
dc.contributor.author | Alenyà Ribas, Guillem |
dc.contributor.author | Torras, Carme |
dc.contributor.other | Institut de Robòtica i Informàtica Industrial |
dc.date.accessioned | 2016-04-06T17:46:56Z |
dc.date.available | 2016-04-06T17:46:56Z |
dc.date.issued | 2015 |
dc.identifier.citation | Martínez, D., Alenyà, G., Torras, C. Safe robot execution in model-based reinforcement learning. A: IEEE/RSJ International Conference on Intelligent Robots and Systems. "2015 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2015, Hamburg, Germany, September 28-October 2, 2015". Hamburg: Institute of Electrical and Electronics Engineers (IEEE), 2015, p. 6422-6427. |
dc.identifier.isbn | 978-1-4799-9994-1 |
dc.identifier.uri | http://hdl.handle.net/2117/85331 |
dc.description.abstract | Task learning in robotics requires repeatedly executing the same actions in different states to learn the model of the task. However, in real-world domains, there are usually sequences of actions that, if executed, may produce unrecoverable errors (e.g. breaking an object). Robots should avoid repeating such errors when learning, and thus explore the state space in a more intelligent way. This requires identifying dangerous action effects to avoid including such actions in the generated plans, while at the same time enforcing that the learned models are complete enough for the planner not to fall into dead-ends. We thus propose a new learning method that allows a robot to reason about dead-ends and their causes. Some such causes may be dangerous action effects (i.e., leading to unrecoverable errors if the action were executed in the given state) so that the method allows the robot to skip the exploration of risky actions and guarantees the safety of planned actions. If a plan might lead to a dead-end (e.g., one that includes a dangerous action effect), the robot tries to find an alternative safe plan and, if not found, it actively asks a teacher whether the risky action should be executed. This method permits learning safe policies as well as minimizing unrecoverable errors during the learning process. Experimental validation of the approach is provided in two different scenarios: a robotic task and a simulated problem from the international planning competition. Our approach greatly increases success ratios in problems where previous approaches had high probabilities of failing. |
dc.format.extent | 6 p. |
dc.language.iso | eng |
dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/3.0/es/ |
dc.subject | Àrees temàtiques de la UPC::Informàtica::Intel·ligència artificial |
dc.subject.other | learning (artificial intelligence) |
dc.subject.other | manipulators |
dc.subject.other | planning (artificial intelligence). |
dc.title | Safe robot execution in model-based reinforcement learning |
dc.type | Conference report |
dc.contributor.group | Universitat Politècnica de Catalunya. ROBiri - Grup de Robòtica de l'IRI |
dc.identifier.doi | 10.1109/IROS.2015.7354295 |
dc.description.peerreviewed | Peer Reviewed |
dc.subject.inspec | Classificació INSPEC::Cybernetics::Artificial intelligence::Learning (artificial intelligence) |
dc.relation.publisherversion | http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=7354295 |
dc.rights.access | Open Access |
local.identifier.drac | 17420139 |
dc.description.version | Postprint (author's final draft) |
local.citation.author | Martínez, D.; Alenyà, G.; Torras, C. |
local.citation.contributor | IEEE/RSJ International Conference on Intelligent Robots and Systems |
local.citation.pubplace | Hamburg |
local.citation.publicationName | 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2015, Hamburg, Germany, September 28-October 2, 2015 |
local.citation.startingPage | 6422 |
local.citation.endingPage | 6427 |