Mostra el registre d'ítem simple

dc.contributor.authorAgostini, Alejandro Gabriel
dc.contributor.authorCelaya Llover, Enric
dc.contributor.otherInstitut de Robòtica i Informàtica Industrial
dc.date.accessioned2010-11-22T19:18:29Z
dc.date.available2010-11-22T19:18:29Z
dc.date.created2010
dc.date.issued2010
dc.identifier.citationAgostini, A.G.; Celaya, E. Reinforcement learning for robot control using probability density estimations. A: International Conference on Informatics in Control, Automation and Robotics. "7Th International Conference on Informatics in Control, Automation and Robotics". Funchal: INSTICC Press. Institute for Systems and Technologies of Information, Control and Communication, 2010, p. 160-168.
dc.identifier.urihttp://hdl.handle.net/2117/10368
dc.description.abstractThe successful application of Reinforcement Learning (RL) techniques to robot control is limited by the fact that, in most robotic tasks, the state and action spaces are continuous, multidimensional, and in essence, too large for conventional RL algorithms to work. The well known curse of dimensionality makes infeasible using a tabular representation of the value function, which is the classical approach that provides convergence guarantees. When a function approximation technique is used to generalize among similar states, the convergence of the algorithm is compromised, since updates unavoidably affect an extended region of the domain, that is, some situations are modified in a way that has not been really experienced, and the update may degrade the approximation. We propose a RL algorithm that uses a probability density estimation in the joint space of states, actions and Q-values as a means of function approximation. This allows us to devise an updating approach that, taking into account the local sampling density, avoids an excessive modification of the approximation far from the observed sample.
dc.format.extent9 p.
dc.language.isoeng
dc.publisherINSTICC Press. Institute for Systems and Technologies of Information, Control and Communication
dc.rightsAttribution-NonCommercial-NoDerivs 3.0 Spain
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/3.0/es/
dc.subjectÀrees temàtiques de la UPC::Informàtica::Intel·ligència artificial::Aprenentatge automàtic
dc.subject.lcshMachine learning
dc.subject.othergeneralisation (artificial intelligence) intelligent robots learning (artificial intelligence)
dc.titleReinforcement learning for robot control using probability density estimations
dc.typeConference report
dc.subject.lemacAprenentatge automàtic
dc.contributor.groupUniversitat Politècnica de Catalunya. VIS - Visió Artificial i Sistemes Intel·ligents
dc.description.peerreviewedPeer Reviewed
dc.subject.inspecClassificació INSPEC::Cybernetics::Artificial intelligence::Learning (artificial intelligence)
dc.relation.publisherversionhttp://www.icinco.org/Abstracts/2010/ICINCO_2010_Abstracts.htm
dc.rights.accessRestricted access - publisher's policy
local.identifier.drac4127980
dc.description.versionPostprint (published version)
local.citation.authorAgostini, A.G.; Celaya, E.
local.citation.contributorInternational Conference on Informatics in Control, Automation and Robotics
local.citation.pubplaceFunchal
local.citation.publicationName7Th International Conference on Informatics in Control, Automation and Robotics
local.citation.startingPage160
local.citation.endingPage168


Fitxers d'aquest items

Imatge en miniatura

Aquest ítem apareix a les col·leccions següents

Mostra el registre d'ítem simple