Mostra el registre d'ítem simple
Modeling robot's world with minimal effort
dc.contributor.author | Villamizar Vergel, Michael Alejandro |
dc.contributor.author | Garrell Zulueta, Anais |
dc.contributor.author | Sanfeliu Cortés, Alberto |
dc.contributor.author | Moreno-Noguer, Francesc |
dc.contributor.other | Institut de Robòtica i Informàtica Industrial |
dc.contributor.other | Universitat Politècnica de Catalunya. Departament d'Enginyeria de Sistemes, Automàtica i Informàtica Industrial |
dc.date.accessioned | 2016-05-17T14:45:15Z |
dc.date.available | 2016-05-17T14:45:15Z |
dc.date.issued | 2015 |
dc.identifier.citation | Villamizar, M.A., Garrell, A., Sanfeliu, A., Moreno-Noguer, F. Modeling robot's world with minimal effort. A: IEEE International Conference on Robotics and Automation. "2015 IEEE International Conference on Robotics and Automation (ICRA 2015): Seattle, Washington, USA, 26-30 May 2015". Seattle, WA: Institute of Electrical and Electronics Engineers (IEEE), 2015, p. 4890-4896. |
dc.identifier.isbn | 978-1-4799-6924-1 |
dc.identifier.uri | http://hdl.handle.net/2117/87122 |
dc.description | © 20xx IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works |
dc.description.abstract | We propose an efficient Human Robot Interaction approach to efficiently model the appearance of all relevant objects in robot’s environment. Given an input video stream recorded while the robot is navigating, the user just needs to annotate a very small number of frames to build specific classifiers for each of the objects of interest. At the core of the method, there are several random ferns classifiers that share the same features and are updated online. The resulting methodology is fast (runs at 8 fps), versatile (it can be applied to unconstrained scenarios), scalable (real experiments show we can model up to 30 different object classes), and minimizes the amount of human intervention by leveraging the uncertainty measures associated to each classifier. We thoroughly validate the approach on synthetic data and on real sequences acquired with a mobile platform in outdoor and challenging scenarios containing a multitude of different objects. We show that the human can, with minimal effort, provide the robot with a detailed model of the objects in the scene. |
dc.format.extent | 7 p. |
dc.language.iso | eng |
dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/3.0/es/ |
dc.subject | Àrees temàtiques de la UPC::Informàtica::Robòtica |
dc.subject.other | Human robot interaction |
dc.title | Modeling robot's world with minimal effort |
dc.type | Conference report |
dc.contributor.group | Universitat Politècnica de Catalunya. VIS - Visió Artificial i Sistemes Intel·ligents |
dc.contributor.group | Universitat Politècnica de Catalunya. ROBiri - Grup de Robòtica de l'IRI |
dc.identifier.doi | 10.1109/ICRA.2015.7139878 |
dc.description.peerreviewed | Peer Reviewed |
dc.subject.inspec | Classificació INSPEC::Automation::Robots::Humanoid robots |
dc.relation.publisherversion | http://ieeexplore.ieee.org/document/7139878/ |
dc.rights.access | Open Access |
local.identifier.drac | 16825595 |
dc.description.version | Postprint (author's final draft) |
dc.relation.projectid | info:eu-repo/grantAgreement/EC/FP7/287617/EU/Aerial Robotics Cooperative Assembly System/ARCAS |
local.citation.author | Villamizar, M.A.; Garrell, A.; Sanfeliu, A.; Moreno-Noguer, F. |
local.citation.contributor | IEEE International Conference on Robotics and Automation |
local.citation.pubplace | Seattle, WA |
local.citation.publicationName | 2015 IEEE International Conference on Robotics and Automation (ICRA 2015): Seattle, Washington, USA, 26-30 May 2015 |
local.citation.startingPage | 4890 |
local.citation.endingPage | 4896 |