Reinforcement learning based autonomous multi-rotor landing on moving platforms

dc.contributor.authorGoldschmid, Pascal
dc.contributor.authorAhmad, Aamir
dc.date.accessioned2025-05-31T07:04:29Z
dc.date.issued2024
dc.date.updated2025-01-24T18:31:35Z
dc.description.abstractMulti-rotor UAVs suffer from a restricted range and flight duration due to limited battery capacity. Autonomous landing on a 2D moving platform offers the possibility to replenish batteries and offload data, thus increasing the utility of the vehicle. Classical approaches rely on accurate, complex and difficult-to-derive models of the vehicle and the environment. Reinforcement learning (RL) provides an attractive alternative due to its ability to learn a suitable control policy exclusively from data during a training procedure. However, current methods require several hours to train, have limited success rates and depend on hyperparameters that need to be tuned by trial-and-error. We address all these issues in this work. First, we decompose the landing procedure into a sequence of simpler, but similar learning tasks. This is enabled by applying two instances of the same RL based controller trained for 1D motion for controlling the multi-rotor’s movement in both the longitudinal and the lateral directions. Second, we introduce a powerful state space discretization technique that is based on i) kinematic modeling of the moving platform to derive information about the state space topology and ii) structuring the training as a sequential curriculum using transfer learning. Third, we leverage the kinematics model of the moving platform to also derive interpretable hyperparameters for the training process that ensure sufficient maneuverability of the multi-rotor vehicle. The training is performed using the tabular RL method Double Q-Learning . Through extensive simulations we show that the presented method significantly increases the rate of successful landings, while requiring less training time compared to other deep RL approaches. Furthermore, for two comparison scenarios it achieves comparable performance than a cascaded PI controller. Finally, we deploy and demonstrate our algorithm on real hardware. For all evaluation scenarios we provide statistics on the agent’s performance. Source code is openly available at https://github.com/robot-perception-group/rl_multi_rotor_landing .en
dc.description.sponsorshipProjekt DEAL
dc.description.sponsorshipUniversität Stuttgart
dc.identifier.issn1573-7527
dc.identifier.issn0929-5593
dc.identifier.urihttp://nbn-resolving.de/urn:nbn:de:bsz:93-opus-ds-164950de
dc.identifier.urihttps://elib.uni-stuttgart.de/handle/11682/16495
dc.identifier.urihttps://doi.org/10.18419/opus-16476
dc.language.isoen
dc.relation.uridoi:10.1007/s10514-024-10162-8
dc.rightsCC BY
dc.rightsinfo:eu-repo/semantics/openAccess
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subject.ddc620
dc.titleReinforcement learning based autonomous multi-rotor landing on moving platformsen
dc.typearticle
dc.type.versionpublishedVersion
ubs.fakultaetLuft- und Raumfahrttechnik und Geodäsie
ubs.fakultaetFakultätsübergreifend / Sonstige Einrichtung
ubs.institutInstitut für Flugmechanik und Flugregelung
ubs.institutFakultätsübergreifend / Sonstige Einrichtung
ubs.publikation.noppnyesde
ubs.publikation.seiten21
ubs.publikation.sourceAutonomous robots 48 (2024), No. 13
ubs.publikation.typZeitschriftenartikel

Files

Original bundle

Now showing 1 - 1 of 1
Thumbnail Image
Name:
10514_2024_Article_10162.pdf
Size:
2.61 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
3.3 KB
Format:
Item-specific license agreed upon to submission
Description: