Mostra el registre d'ítem simple

dc.contributor.authorAgostini, Alejandro Gabriel
dc.contributor.authorCelaya Llover, Enric
dc.contributor.otherInstitut de Robòtica i Informàtica Industrial
dc.date.accessioned2015-06-29T18:57:10Z
dc.date.available2015-06-29T18:57:10Z
dc.date.created2014
dc.date.issued2014
dc.identifier.citationAgostini, A.; Celaya, E. "Competitive function approximation for reinforcement learning". 2014.
dc.identifier.urihttp://hdl.handle.net/2117/28454
dc.description.abstractThe application of reinforcement learning to problems with continuous domains requires representing the value function by means of function approximation. We identify two aspects of reinforcement learning that make the function approximation process hard: non-stationarity of the target function and biased sampling. Non-stationarity is the result of the bootstrapping nature of dynamic programming where the value function is estimated using its current approximation. Biased sampling occurs when some regions of the state space are visited too often, causing a reiterated updating with similar values which fade out the occasional updates of infrequently sampled regions. We propose a competitive approach for function approximation where many different local approximators are available at a given input and the one with expectedly best approximation is selected by means of a relevance function. The local nature of the approximators allows their fast adaptation to non-stationary changes and mitigates the biased sampling problem. The coexistence of multiple approximators updated and tried in parallel permits obtaining a good estimation much faster than would be possible with a single approximator. Experiments in different benchmark problems show that the competitive strategy provides a faster and more stable learning than non-competitive approaches.
dc.format.extent32 p.
dc.language.isoeng
dc.relation.ispartofseriesIRI-TR-14-05
dc.rightsAttribution-NonCommercial-NoDerivs 3.0 Spain
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/3.0/es/
dc.subjectÀrees temàtiques de la UPC::Informàtica::Intel·ligència artificial
dc.subject.otherlearning (artificial intelligence)
dc.subject.otherreinforcement learning
dc.subject.othercompetitive strategy
dc.subject.otherGaussian mixture model
dc.titleCompetitive function approximation for reinforcement learning
dc.typeExternal research report
dc.contributor.groupUniversitat Politècnica de Catalunya. VIS - Visió Artificial i Sistemes Intel·ligents
dc.subject.inspecClassificació INSPEC::Cybernetics::Artificial intelligence
dc.rights.accessOpen Access
local.identifier.drac15302509
dc.description.versionPreprint
local.citation.authorAgostini, A.; Celaya, E.
local.citation.publicationNameCompetitive function approximation for reinforcement learning


Fitxers d'aquest items

Thumbnail

Aquest ítem apareix a les col·leccions següents

Mostra el registre d'ítem simple