Mostra el registre d'ítem simple

dc.contributor.authorMuñoz Ferran, Guillem
dc.contributor.authorBarrado Muxí, Cristina
dc.contributor.authorÇetin, Ender
dc.contributor.authorSalamí San Juan, Esther
dc.contributor.otherUniversitat Politècnica de Catalunya. Departament d'Arquitectura de Computadors
dc.date.accessioned2020-01-24T14:31:37Z
dc.date.available2020-01-24T14:31:37Z
dc.date.issued2019-09-10
dc.identifier.citationMuñoz, G. [et al.]. Deep reinforcement learning for drone delivery. "Drones", 10 Setembre 2019, vol. 3, núm. 3.
dc.identifier.issn2504-446X
dc.identifier.urihttp://hdl.handle.net/2117/175691
dc.descriptionThis work was funded by the Ministry of Science, Innovation and Universities of Spain under Grant No. TRA2016-77012-R.
dc.description.abstractDrones are expected to be used extensively for delivery tasks in the future. In the absence of obstacles, satellite based navigation from departure to the geo-located destination is a simple task. When obstacles are known to be in the path, pilots must build a flight plan to avoid them. However, when they are unknown, there are too many or they are in places that are not fixed positions, then to build a safe flight plan becomes very challenging. Moreover, in a weak satellite signal environment, such as indoors, under trees canopy or in urban canyons, the current drone navigation systems may fail. Artificial intelligence, a research area with increasing activity, can be used to overcome such challenges. Initially focused on robots and now mostly applied to ground vehicles, artificial intelligence begins to be used also to train drones. Reinforcement learning is the branch of artificial intelligence able to train machines. The application of reinforcement learning to drones will provide them with more intelligence, eventually converting drones in fully-autonomous machines. In this work, reinforcement learning is studied for drone delivery. As sensors, the drone only has a stereo-vision front camera, from which depth information is obtained. The drone is trained to fly to a destination in a neighborhood environment that has plenty of obstacles such as trees, cables, cars and houses. The flying area is also delimited by a geo-fence; this is a virtual (non-visible) fence that prevents the drone from entering or leaving a defined area. The drone has to avoid visible obstacles and has to reach a goal. Results show that, in comparison with the previous results, the new algorithms have better results, not only with a better reward, but also with a reduction of its variance. The second contribution is the checkpoints. They consist of saving a trained model every time a better reward is achieved. Results show how checkpoints improve the test results.
dc.language.isoeng
dc.publisherMultidisciplinary Digital Publishing Institute (MDPI)
dc.rightsAttribution 3.0 Spain
dc.rights.urihttp://creativecommons.org/licenses/by/3.0/es/
dc.subjectÀrees temàtiques de la UPC::Aeronàutica i espai::Aeronaus
dc.subject.lcshDrone aircraft
dc.subject.lcshArtificial intelligence
dc.subject.otherDrones
dc.subject.otherDeep learning
dc.subject.otherReinforcement learning
dc.subject.otherQ-learning
dc.subject.otherDQN
dc.subject.otherJNN
dc.titleDeep reinforcement learning for drone delivery
dc.typeArticle
dc.subject.lemacAvions no tripulats
dc.subject.lemacLogística (Indústria) -- Innovacions tecnològiques
dc.subject.lemacGestió de productes
dc.subject.lemacIntel·ligència artificial
dc.contributor.groupUniversitat Politècnica de Catalunya. ICARUS - Intelligent Communications and Avionics for Robust Unmanned Aerial Systems
dc.identifier.doi10.3390/drones3030072
dc.description.peerreviewedPeer Reviewed
dc.relation.publisherversionhttps://www.mdpi.com/2504-446X/3/3/72
dc.rights.accessOpen Access
local.identifier.drac25827522
dc.description.versionPostprint (published version)
local.citation.authorMuñoz, G.; Barrado, C.; Çetin, E.; Salamí, E.
local.citation.publicationNameDrones
local.citation.volume3
local.citation.number3
local.citation.startingPage72


Fitxers d'aquest items

Thumbnail

Aquest ítem apareix a les col·leccions següents

Mostra el registre d'ítem simple