Reinforcement Learning based Autonomous Multi-Rotor Landing on Moving Platforms

Multi-rotor UAVs suffer from a restricted range and flight duration due to limited battery capacity. Autonomous landing on a 2D moving platform offers the possibility to replenish batteries and offload data, thus increasing the utility of the vehicle. Classical approaches rely on accurate, complex and difficult-to-derive models of the vehicle and the environment. Reinforcement learning (RL) provides an attractive alternative due to its ability to learn a suitable control policy exclusively from data during a training procedure. However, current methods require several hours to train, have limited success rates and depend on hyperparameters that need to be tuned by trial-and-error. We address all these issues in this work. First, we decompose the landing procedure into a sequence of simpler, but similar learning tasks. This is enabled by applying two instances of the same RL based controller trained for 1D motion for controlling the multi-rotor's movement in both the longitudinal and the lateral directions. Second, we introduce a powerful state space discretization technique that is based on i) kinematic modeling of the moving platform to derive information about the state space topology and ii) structuring the training as a sequential curriculum using transfer learning. Third, we leverage the kinematics model of the moving platform to also derive interpretable hyperparameters for the training process that ensure sufficient maneuverability of the multi-rotor vehicle. The training is performed using the tabular RL method Double Q-Learning. Through extensive simulations we show that the presented method significantly increases the rate of successful landings, while requiring less training time compared to other deep RL approaches. Finally, we deploy and demonstrate our algorithm on real hardware. For all evaluation scenarios we provide statistics on the agent's performance.


Introduction
Classical control approaches suffer from the fact that accurate models of the plant and the environment are required for the controller design.These models are necessary to consider non-linearities of the plant behavior.However, they are subjected to limitations, e.g., with regard to disturbance rejection (Rodriguez-Ramos et al, 2019) and parametric uncertainties (Mo and Farid, 2019).To overcome these problems, reinforcement learning (RL) has been studied for robot control in the last decades, including in the context of UAV control (Mo and Farid, 2019).In model-free, (action-) value based RL methods, an approximation of a value function is learned exclusively through interaction with the environment.On this basis, a policy then selects an optimal action while being in a certain state (Sutton and Barto, 2015).Classical RL approaches such as Q-Learning (Sutton and Barto, 2015) store a tabular representation of that value function approximation.Deep Reinforcement Learning (DRL) methods (Mnih et al, 2013), on the other hand, leverage a deep neural network (DNN) to learn an approximate value function over a continuous state and action space, thus making it a powerful approach for many complex applications.Nevertheless, in both RL and DRL, the training itself is in general sensitive to the formulation of the learning task and in particular to the selection of hyperparameters.This can result in long training times, the necessity to perform a time-consuming hyperparameter search and an unstable convergence behavior.Furthermore, especially in DRL, the neural network acts like a black box, i.e. it is difficult to trace back which training experience caused certain aspects of the learned behaviour.The purpose of this work is to address the aforementioned issues for the task of landing a multi-rotor UAV on a moving platform.Our RL based method aims at i) achieving a high rate of successful landing attempts, ii) requiring short training time, and iii) providing intrepretable ways to compute hyperparameters necessary for training.
To achieve these aims, we leverage the tabular off-policy algorithm Double Q-Learning (Hasselt, 2010), due to the few required hyperparameters which are the decay factor γ and the learning rate α.Using a tabular method does not require defining the DNN's architecture or finding values for additional hyperparameters such as minibatch and buffer size or the number of gradient steps.Furthermore, unlike a NN-based deep learning algorithm, it does not require a powerful GPU for training, making it also suitable for computers with lower performance and less powerconsuming.However, tabular methods only provide discrete state and action spaces and therefore suffer from the "curse of dimensionality" (Lampton and Valasek, 2009).Furthermore, the training performance and control performance is influenced by the sampling rate.When the sampling rate and discretization do not match, the performance of the agent can be low, e. g. due to jittering, where the agent rapidly alternates between discrete states.To solve these problems, our method's novel aspects are outlined below.The first two of these allow us to address the "curse of dimensionality" issue by reducing the complexity of the learning task.The third addresses the problem to find a matching discretization and sampling rate.
1.Under the assumption of a symmetric UAV and decoupled dynamics, common in literature (Wang et al, 2016), our method is able to control the vehicle's motion in longitudinal and lateral direction with two instances of the same RL agent.Thus, the learning task is simplified to 1D movement only.The vertical and yaw movements are controlled by PID controllers.The concept of using independent controllers for different directions of motion is an approach which is often used, e.g. when PID controllers are applied (Wenzel et al, 2011;Araar et al, 2017).2. We introduce a novel state-space discretization approach, motivated by the insight that precise knowledge of the relative position and relative velocity between the UAV and the moving platform is required only when touchdown is imminent.To this end, we leverage a multiresolution technique (Lampton and Valasek, 2009) and augment it with information about the state space topology, derived from a simple kinematics model of the moving platform.This allows us to restructure the learning task of the 1D movement as a sequence of even simpler problems by means of a sequential curriculum, in which learned action values are transferred to the subsequent curriculum step.Furthermore, the discrete state space allows us to accurately track how often a state has been visited during training.3. We leverage the discrete action space to ensure sufficient maneuverability of the UAV to follow the platform.To this end, we derive equations that compute the values of hyperparameters, such as the agent frequency and the maximum value of the roll/pitch angle of the UAV.The intention of these equations is twofold.First, they link the derived values to the maneuverability of the UAV in an interpretable way.Second, they ensure that the discretization of the state space matches the agent frequency.
The aim is to reduce unwanted side effects resulting from the application of a discrete state space, such as jittering.
Sect. 2 presents related work, followed by Sect. 3 describing the proposed approach in detail.Sect. 4 presents the implementation and experimental setup.The results are discussed in Sect.5, with comments on future work given in Sect.6.
2 Related Work

Classical Control Approaches
So far, the problem of landing a multi-rotor aerial vehicle has been tackled for different levels of complexity regarding platform movement.Onedimensional platform movement is considered in (Hu et al, 2015) in the context of maritime applications, where a platform is oscillating vertically.Two-dimensional platform movement is treated in (Wenzel et al, 2011;Ling et al, 2014;Gautam et al, 2015;Vlantis et al, 2015;Araar et al, 2017;Borowczyk et al, 2017;Falanga et al, 2017).Three-dimensional translational movement of the landing platform is covered for docking scenarios involving multi-rotor vehicles in (Zhong et al, 2016) and (Miyazaki et al, 2018).Various control techniques have been applied to enable a multirotor UAV to land in one of these scenarios.In (Hu et al, 2015) an adaptive robust controller is used to control the vehicle during a descend maneuver onto a vertically oscillating platform while considering the ground effect during tracking of a reference trajectory.The authors of (Gautam et al, 2015) apply guidance laws that are based on missile control principles.(Vlantis et al, 2015) uses a model predictive controller to land on an inclined moving platform.(Falanga et al, 2017) relies on a non-linear control approach involving LQR-controllers.However, the most used controller type is a PID controller.Reference (Wenzel et al, 2011) uses four independent PID controllers to follow the platform that has been identified with visual tracking methods.In (Ling et al, 2014), the landing maneuver onto a maritime vessel is structured into different phases.Rendezvous with the vessel, followed by aquiring a visual fiducial marker and descending onto the ship.During all phases, PID controllers are used to control the UAV.Perception and relative pose estimation based methods are the focus of (Araar et al, 2017), where again four PID controllers provide the UAV with the required autonomous motion capability.A PID controller is also applied in (Borowczyk et al, 2017) to handle the final touchdown on a ground vehicle moving up to 50km/h.Also the landing on a 3D moving platform can be solved with PID controllers (Zhong et al, 2016;Miyazaki et al, 2018).Although automatic methods for tuning the gains of a PID controller exist, tuning is often a manual, time-consuming procedure.However, learning-based control methods enable obtaining a suitable control policy exclusively from data through interaction with the environment, making it an attractive and superior approach for controlling an UAV to land on a moving platform.Furthermore, methods such as Q-Learning enable the approximation of an optimal action-value function and thus lead to a (near-) optimal action selection policy (Sutton and Barto, 2015).

Learning-based Control Approaches
The landing problem has been approached with (Deep) Reinforcement Learning for static platforms (Kooi and Babuška, 2021;Shi et al, 2019;Polvara et al, 2018;Polvara et al, 2019) and for moving platforms (Rodriguez-Ramos et al, 2019;Lee et al 2018).The authors of (Kooi and Babuška, 2021) accelerate training by means of a curriculum, where a policy is learned using Proximal Policy Optimization (PPO) to land on an inclined, static platform.Their curriculum is tailored, involving several hyperparameters, whereas in this work we present a structured approach for deriving a curriculum for different scenarios.In (Polvara et al, 2018) the authors assign different control tasks to different agents.A DQN-agent aligns a quadcopter with a marker on the ground, whereas a second agent commands a descending maneuver before a closed-loop controller handles the touchdown.(Polvara et al, 2019) follows a similar approach but allows small pitch and roll maneuvers of the platform.Our approach also leverages separate RL agents, but for controlling the longitudinal and lateral motion.The authors of (Shi et al, 2019) present a deep-learning-based robust non-linear controller to increase the precision of landing and close ground trajectory following.A nominal dynamics model is combined with a DNN to learn the ground effect on aerodynamics and UAV dynamics.(Lee et al, 2018) presents an actor-critic approach for landing using continuous state and action spaces to control the roll and pitch angle of the drone, but no statistics on its performance.In (Rodriguez-Ramos et al, 2019), a DRL-framework is introduced for training in simulation and evaluation in real world.
The Deep Deterministic Policy Gradient (DDPG) algorithm, an actor-critic approach, is used to command the movement of a multi-rotor UAV in longitudinal and lateral direction with continuous states and actions.Detailed data about the agent's success rate in landing in simulation is provided.We use this work as a baseline method and show how we outperform it, providing also statistics about our agent's performance in the real world.However, the baseline method does not provide any systematic and explainable way for deriving hyperparameters used for the learning problem.For our method, we present equations that link values of hyperparameters to problem properties such as the state space discretization and maneuverability of the UAV in an intuitively understandable way.

Problem Statement
The goal is to develop a RL based approach that provides a high rate of successful UAV landings on a horizontally moving platform.A landing trial is considered successful if the UAV touches down on the surface of the moving platform.The approach should require less training time than the baseline method and generalize well enough to be deployed on real hardware.Furthermore, required hyperparameters should be determined in an interpretable way.

Base Notations
Fig. 2: Coordinate frames (blue, e: earth frame, mp: body-fixed frame of moving platform, mr: body-fixed frame of multi-rotor UAV, s: stability frame), forces (red) and attitude angle (green) associated with 1D motion in longitudinal direction.
We first define a fly zone in which the motion of the multi-rotor UAV mr and the moving platform mp are considered, as is illustrated by Fig. 2. Furthermore, for the goal of landing on the moving platform, the RL agent controlling the multi-rotor vehicle has to consider only the relative translational and rotational motion.This is why we express the relative dynamics and kinematics in the stability frame s, to be able to formulate the RL problem from the UAV's and thus the agent's point of view.For this purpose we first denote the Euler angles roll, pitch and yaw in the stability frame with s φ = (ϕ, θ, ψ) T .Each landing trial will begin with the UAV being in hover state, leading to the following initial conditions for rotational movement s φmr,0 = s φmp,0 = 0rad/s, s φ mr,0 = s φ mp,0 = 0rad.The initial conditions for the multi-rotor vehicle's translational motion are expressed in the earth-fixed frame e, it is e ṙmr,0 = 0m/s and e r mr,0 ∈ F. Transforming the translational initial conditions of both, multi-rotor vehicle and moving platform into the stability frame allows to fully express the relative motion as the second order differential equations s φrel = 0 − s φmr and s rrel = s rmp − s rmr.(1) For the remainder of this work we specify the following values that will serve as continuous observations of the environment (see Sec. 3.3.4and Fig. 7) p c denotes the relative position, v c the relative velocity, a c the relative acceleration and φ c the relative orientation between vehicle and platform, expressed in the stability frame.

Motion of the Landing Platform
We introduce the rectilinear periodic movement (RPM) of the platform that is applied during training.With the initial conditions e ṙmp,0 = 0m/s and e r mp,0 = 0m, the translational motion of the platform is expressed as a second order differential equation in the earth-fixed frame e. where r mp denotes the maximum amplitude of the trajectory, v mp the maximum platform velocity and ω mp the angular frequency of the rectilinear movement.The platform is not subjected to any rotational movement.As a consequence, the maximum acceleration of the platform that is required as part of the hyperparameter determination in Sec.3.5 is 3.3 Basis Learning Task

Dynamic Decoupling
The nonlinear dynamics of a multi-rotor vehicle such as a quadcopter can be greatly simplified when the following assumptions are applied.i) Linearization around hover flight, ii) small angles iii) a rigid, symmetric vehicle body, iv) no aerodynamic effects (Wang et al, 2016).Under these conditions, the axes of motion are decoupled.

Markov Decision Process
For the RL task formulation we consider the finite, discrete Markov Decision Process (Sutton and Barto, 2015).The goal of the agent is to learn the optimal action value function Q * to obtain the optimal policy π * = argmax a Q * (s, a), maximizing the sum of discounted rewards Σ T −t−1 k=0 γ k r t+k+1 .

Discrete Action Space
We choose a discrete action space comprising three actions, namely increasing pitch, decreasing pitch and do nothing, denoted by The pitch angle increment is defined by where θ max denotes the maximum pitch angle and n θ the number of intervals intersecting the range [0, θ max ].The set of possible pitch angles that can be taken by the multi-rotor vehicle is then where i θ is used to select a specific element.

Discrete State Space
For the discrete state space we first scale and clip the continuous observations of the environment that are associated with motion in the longitudinal direction to a value range of [−1, 1].For this purpose, we use the function clip(x, x min , x max ) that clips the value of x to the range of [x min , x max ].
where p max , v max , a max are the values used for the normalization of the observations.The reason for the clipping is that our state space discretization technique, a crucial part of the sequential curriculum, is derived assuming a worst case scenario for the platform movement in which the multirotor vehicle is hovering (see Sect. 3.4) and where scaling an observation with its maximum values, p max , v max and a max , would constitute a normalization.However, once the multi-rotor vehicle starts moving too, the scaled observation values could exceed the value range of [−1, 1].Clipping allows the application of the discretization technique also for a moving multi-rotor vehicle.
Next, we define a general discretization function d(x, x 1 , x 2 ) that can be used to map a continuous observation of the environment to a discrete state value.
We apply ( 14) to the normalized observations ( 11)-( 13) to determine the discrete state s = (p d , v d , a d , i θ ) ∈ S, where i θ ∈ {0, . . ., 2n θ }, S = N 3×3×3×2n θ +1 0 and In ( 15)-( 17), the normalized values ±p goal ±, v goal , ±a goal define the boundaries of the discrete states the agent should learn to reach whereas the normalized values of ±p lim ±, v lim , ±a lim denote the limits in the observations the agent should learn not to exceed when controlling the multi-rotor vehicle.

Goal State
We define the goal state s * ∈ S This means the goal state is reached if −p goal ≤ p x ≤ p goal and −v goal ≤ v x ≤ v goal regardless of the values of a d and i θ .

Reward Function
Our reward function r t , inspired by a shaping approach (Rodriguez-Ramos et al, 2018) is given as where r term = r suc if s = s * and r term = r f ail if |p x | > p lim or if the maximum episode duration t max is reached.In all other cases, r term = 0. r suc and r f ail will be defined at the end of this section.Positive rewards r p , r v and r θ are given for a reduction of the relative position, relative velocity and the pitch angle, respectively.The negative reward r dur gives the agent an incentive to start moving.Considering the negative weights w p , w v , w θ , w dur and the agent frequency f ag , derived in Sec.3.5, with which actions are drawn from A, we define one time step as ∆t = 1/f ag and r p , r v , r θ , r dur as The weights are negative so that decreasing the relative position, relative velocity or the pitch angle yields positive reward values.Clipping r p and r v to ±r p,max and ±r v,max , as well as scaling r θ and r dur with v lim is necessary due to their influence on the value of r max .r max denotes the maximum achievable reward in a non-terminal timestep if v x ≤ v lim and a x ≤ a lim and thus complies with the limits set by the motion scenario described in Sect.3.4 during the derivation of the sequential curriculum.To ensure the applicability of the curriculum also in situations in which that compliance is not given is why the clipping is applied in ( 20) and ( 21).r max plays a role in the scaling of Q-values that is performed as part of the knowledge transfer during different curriculum steps.It is defined as follows rv,max = |wv|a lim ∆t, ( 25) With r max derived and the weights w suc and w f ail , we can finally define the success and failure rewards as rsuc = wsucrmax and r f ail = w f ail rmax.(29) All weights of the reward function were determined experimentally.

Double Q-Learning
We use the Double Q-Learning algorithm (Hasselt, 2010) to address the problem of overestimating action values and increasing training stabilty.Inspired by (Even-Dar and Mansour, 2004), we apply a state-action pair specific learning rate that decreases the more often the state-action pair has been visited.
where n c (s t , a t ) denotes the number of visits the state action-pair (s t , a t ) has been experiencing until the timestep t and ω is a decay factor.To keep a certain minimal learning ability, the learning rate is constant once the value of α min has been reached.Sufficient exploration of the state space can be ensured by applying an ϵ-greedy policy for the action selection during training.The exploration rate ϵ is varied according to a schedule for the episode number.

Curriculum and Discretization
Our curriculum and discretization approach builds on (Lampton and Valasek, 2009), where a multiresolution method for state space discretization is presented.There, starting with a coarse discretization, a region around a pre-known learning goal is defined.Once the agent has learned to reach that region, it is further refined by discretizing it with additional states.Then, training starts anew but only in the refined region, repeatedly until a pre-defined resolution around the learning goal has been reached.Consequently, a complex task is reduced to a series of smaller, quicklylearnable sub tasks.In our novel approach, we introduce a similar multiresolution technique combined with transfer learning as part of a sequential curriculum (Narvekar et al, 2020) to accelerate and stabilize training even further.Each new round of training in the multiresolution setting constitutes a different curriculum step.A curriculum step is an instance of the basis learning task introduced in Sect.3.3.This means in particular that each curriculum step has its own Q-table.Since the size of the state and action space is the same throughout all steps, knowledge can be easily transferred to the subsequent curriculum step by copying the Q-table that serves then as a starting point for its training.Fig. 3 illustrates the procedure of the sequential curriculum.However, throughout the curriculum the goal regions become smaller, leading to less spacious maneuvers of the multi-rotor vehicle.Thus, the achievable reward, which depends on relative change in position and velocity, is reduced.In order to achieve a consistent adaptability of the Q-values over all curriculum steps, the initial Qvalues need to be scaled to match the maximum achievable reward per timestep.Therefore, we define where i ≥ 1 is the current curriculum step.Furthermore, adding more curriculum steps with smaller goal regions possibly leads to previously Discrete Relative Position Discrete Relative Velocity Fig. 4: Illustration of the mapping of the normalized observations p x and v x to the discrete states (red -0, yellow -1, blue -2) for a curriculum with three steps (i = 0, i = 1, i = 2).The yellow chequered regions illustrate the size of the goal state when the curriculum step is the last of the sequence.
unconsidered effects, such as overshooting the goal state.This is why in our work, training is not only restricted to the latest curriculum step but performed also on the states of the previous curriculum steps when they are visited by the agent.
Furthermore, the discretization introduced in Sect.3.3.4,which is now associated with a curriculum step, covers only a small part of the continuous observation space with few discrete states.In order to completely capture the predominant environment characteristics of the subspace in which the agent is expected to learn the task, it is necessary for it to have the knowledge of the state space topology (Anderson and Crawford-Hines, 1994).Against the background of the multiresolution approach, this can be rephrased as the question of how to define suitable goal regions, i.e., the size of the goal states.To this end, we again leverage the insight that accurate knowledge about the relative position and velocity is only required when landing is imminent.We begin by choosing an exponential contraction of the size of the unnormalized goal state p c,x,goal,i associated with the i th curriculum step such that where 0 < σ < 1 is a contraction factor, l mp ∈ R[m] is the edge length of the squared moving platform, x max is the boundary of the fly zone F (see Sect. 3.1.2)and n cs is the number of curriculum steps following the initial step.The contraction factor σ plays a role in how easy it is for the agent to reach the goal state.If it is set too low, the agent will receive success rewards only rarely, thus leading to an increased training time.(34) can be solved for n cs because σ, l mp and x max are known parameters.To determine the relationship between the contraction of the positional goal state and the contraction of the velocity goal state, we need a kinematics model that relates these variables.For this purpose, we envision the following worst case scenario for the relative motion (1).Assume that the vehicle hovers right above the platform which is located on the ground in the center of the fly zone F, with v x = 0m/s and ψ rel = 0. Now, the platform starts to constantly accelerate with the maximum possible acceleration a mp,max defined by ( 7) until it has reached x max after a time t 0 = 2x max /a mp,max .This is considered to be the worst case since the platform never slows down, unlike the rectilinear periodic movement ( 6 where β p = β v = β a = 1/3 if the ith curriculum step is the curriculum step that was most recently added to the curriculum sequence, and Thus, for the latest curriculum step the goal values result from scaling the associated limit value with a factor of 1/3, a value that has been found empirically, and for all previous steps from the discretized time applied to (35) -(37).The entire discretization procedure is illustrated by Fig. 4. Introducing a different scaling value only for the latest curriculum step has been empirically found to improve the agent's ability to follow the moving platform.The discretization of the acceleration is the same for all curriculum steps and is defined by a contraction factor σ a which has been empirically set.Finding a suitable value for σ a has been driven by the notion that if this value is chosen too high, the agent will have difficulties reacting to changes in the relative acceleration.This is due to the fact that the goal state would cover an exuberant range in the discretization of the relative acceleration.However, if it is chosen too low the opposite is the case.The agent would only rarely be able to visit the discrete goal state a goal,i , thus unnecessarily foregoing one of only three discrete states.
During the training of the different curriculum steps the following episodic terminal criteria have been applied.On the one hand, the episode terminates with success and the success reward r suc is received if the goal state s * of the latest curriculum step is reached if the agent has been in that curriculum step's discrete states for at least one second without interruption.This is different to all previous curriculum steps where the success reward r suc is received immediately after reaching the goal state of the respective curriculum step as is explained in Sect.3.3.6.On the other hand, the episode is terminated with a failure if the multirotor vehicle leaves the fly-zone F or if the success terminal criterion has not been met after the maximum episode duration t max .The reward received by the agent in the terminal time step for successful or failing termination of the episode is defined by (29).Once the trained agent is deployed, the agent selects the actions based on the Q-table of the latest curriculum step to which the continuous observations can be mapped (see Fig. 3).

Hyperparameter Determination
We leverage the discrete action space (8) to determine the hyperparameters agent frequency f ag and maximum pitch angle θ max in an interpretable way.The purpose is to ensure sufficient maneuverability of the UAV to enable it to follow the platform.For sufficient maneuverability, the UAV needs to possess two core abilities.i) Produce a maximum acceleration bigger than the one the platform is capable of and ii) change direction of acceleration quicker than the moving platform.Against the background of the assumptions explained in Sect.k a denotes the multiple of the platform's maximum acceleration of which the UAV needs to be capable.Fig. 5 illustrates how different values for k a affect the maximum pitch angle θ max for different values of the maximum platform acceleration a mp,max .For the second aspect, we leverage the platform's known frequency of the rectilinear periodic movement.According to (6), the moving platform requires the duration of one period to entirely traverse the available range of acceleration.Leveraging equation ( 9), we can calculate the time required by the copter to do the same as 4n θ ∆t, where ∆t = 1/f ag .Next, we introduce a factor k man which specifies how many times faster the UAV should be able to roam through the entire range of acceleration than the moving platform.Against this background, we obtain the agent frequency as Fig. 6 illustrates how the agent frequency f ag depends on the values of k man and the angular frequency of the platform ω mp executing the rectilinear periodic movement.Both hyperparameters θ max and f ag are eventually based on the maximum acceleration of the moving platform a mp,max as is the discretization of the state space.Therefore, we argue that these hyperparameters pose a matching set of values that is suitable to prevent excessive jittering in the agent's actions.

General
We set up the experiments to showcase the following.
• Empirically show that our method is able to outperform the RL baseline (Rodriguez-Ramos et al, 2019) with regard to the rate of successful landings while requiring a shorter training time.• Empirically show that our method is able to perform successful landings for more complex platform trajectories such as an 8-shape.• Demonstrate our method on real hardware.

Simulation Environment
The environment is built within the physics simulator Gazebo 11, in which the moving platform and the UAV are realized.We use the RotorS simulator (Furrer et al, 2016) to simulate the latter.Furthermore, all required tasks such as computing the observations of the environment, the state space discretization as well as the Double Qlearning algorithm are implemented as nodes in the ROS Noetic framework using Python 3.
The setup and data flow are illustrated by Fig. 7.The associated parameters are given in Table 1.We obtain a c by taking the first order derivative of v c and applying a first order Butterworth-Filter to it.The cut-off frequency is set to 0.3hz.

Initialization
In (Kooi and Babuška, 2021), it is indicated that training can be accelerated when the agent is initialized close to the goal state at the beginning of each episode.For this reason, we use the following normal distribution to determine the UAV's initial position within the fly zone F during the first curriculum step.
We set σ F = p max /3, which will ensure that the UAV is initialized close to the center of the flyzone and thus in proximity to the moving platform more frequently.All subsequent curriculum steps as well as the testing of a fully trained agent are then conducted using a uniform distribution over the entire fly zone.

Training Hardware
Each training is run on a desktop computer with the following specifications.Ubuntu 20, AMD Ryzen threadripper 3960x 24-core processor, 128 GB RAM, 6TB SSD.This allows us to run up to four individual trainings in parallel.Note, that being a tabular method, the Double Q-learning algorithm does not depend on a powerful GPU for training since it does not use a neural network.

Training
We design two training cases, simulation and hardware.Case simulation is similar to the training conditions in the baseline method so that we can compare our approach in simulation.Case hardware is created to match the spatial limitations of our real flying environment, so that we can evaluate our approach on real hardware.
For both cases, we apply the rectilinear periodic movement (RPM) of the platform specified by (6) during training.We consider different scenarios for training regarding the maximum velocity of the platform, which are denoted by "RPM v mp,train ".
For each velocity v mp,train , we train four agents using the same parameters.The purpose is to provide evidence of reproducibility of the training results instead of only presenting a manually selected, best result.We choose the same UAV (Hummingbird) of the RotorS package that is also used in the baseline method.Other notable differences between the baseline and the training cases simulation and hardware are summarized in Table 2. Note, that some of our training cases deal with a higher maximum acceleration of the moving platform than the training case used for the baseline method and are therefore considered as more challenging.For all trainings, we use an initial altitude for the UAV of z init = 4m and a vertical velocity of v z = −0.1m/sso that the UAV is descending during an entire episode.The values used for these variables in the baseline method can not be inferred from the paper.For the first curriculum step, the exploration rate schedule is empirically set to ε = 1 (episode 0 − 800) before it is linearly reduced to ε = 0.01 (episode 800 − 2000).For all later curriculum steps, it is ε = 0.For choosing these values, we could exploit the discrete state-action space where for each stateaction pair the number of visits of the agent can be tracked.The exploration rate schedule presented above is chosen so that most state action pairs have at least received one visit.Other training parameters are presented in Table 3.The training is ended as soon as the agent manages to reach the goal state (18) associated with the latest step in the sequential curriculum in 96% of the last 100 episodes.This value has been empirically set.For all trainings, we use noiseless data to compute the observations fed into the agent.However, we test selected agents for their robustness against noisy observations as part of the evaluation.

Initiation of Motion
During training, the yaw controller ensures ψ rel = 0.However, during evaluation in simulation this sometimes led to the situation that the agent commanding the lateral motion of the UAV occasionally was not able to leave its initial state (see Fig. 8).We hypothesize that the reason for this behavior is that the agent is trained on a platform that moves while considering relative motion in its observations.As a consequence, the agent learns Freq. [hz] Freq. [hz] Pitch / Roll Freq.
[ a policy that, while being in certain states, exclusively relies on the platform movement to achieve a desired state change.However, for evaluation using ψ rel = 0, the agent controlling the lateral motion observes no relative movement if the UAV is hovering and the platform following a rectilinear trajectory in longitudinal direction.The issue can be addressed by setting initial ψ rel ̸ = 0.In this case, the platform's movement shows a component in the values p c,y , v c,y and a c,y that are used as observations for the lateral motion's agent.A change in the states is therefore much more likely, allowing the agent to enter states in which the policy selects an action other than "do nothing".For this reason, we apply ψ rel = π/4 for all experiments.

Cascaded PI Controller Baseline
In order to be able to compare the performance of our approach with a non-learning based control method, we implemented a cascaded PI controller with two cascades as is illustrated by Fig. 9.We apply separate instances for controlling the longitudinal, lateral and vertical movement of the UAV.The properties of the different instances are summarized in Table 4.The outer cascade tracks the setpoint 0m for the relative position s r rel between the multi-rotor vehicle and the moving platform.
For this purpose, it generates a setpoint for the relative velocity s ṙrel that is tracked by the inner cascade.The inner cascade aims to reach the velocity setpoint by computing setpoints for the attitude angle ϕ ref , θ ref or the thrust T ref , respectively.The attitude angle or thrust is then tracked by the multi-rotor vehicle's low level controllers.
In order to achieve comparability between our RL approach and the cascaded PI controller, we limit the velocity setpoints generated by the outer cascade to the value of v max = a mp,max t 0 (see Sect. 3.4) and the attitude angles generated by the inner cascade to the same angles determined using (44) for the training scenario associated with the agent chosen for comparison.Table 4: Properties of the cascaded PI controllers.

Evaluation in Simulation without Noise
In this scenario, all agents trained for case simulation and hardware are evaluated as well in simulation using noiseless observations of the environment, just as in training.Besides a static platform, we use two types of platform trajectories.The first is the rectilinear periodic movement (RPM) specified by ( 6) and the second is an eight-shaped trajectory defined by ermp = rmp [sin(ωmpt), sin(0.5ωmpt),0] T , (47) For all landing attempts, we specify an initial altitude of z init = 2.5m.A landing attempt is ended once the UAV touches the surface of the moving platform or reaches an altitude that is lower than the platform surface, i. e. misses the platform.If the center of the UAV is located above the moving platform at the moment of touchdown, the landing trial is considered successful.The value of z init leads to a duration of a landing attempt which corresponds roughly to the time t max used as maximum episode length during training, see reward function ( 19).The information regarding the training duration of the agents is summarized in Tab. 5 for case simulation and case hardware.
The training durations of the different curriculum steps suggest that the majority of the required knowledge is learned during the first curriculum step.The later curriculum steps required significantly fewer episodes to reach the end condition.This is because the exploration rate is ε = 0. Thus, the agent is only exploiting previously acquired knowledge, which is also supported by the accumulated sum of rewards (Fig. 11).This also implies that the decomposition of the landing procedure into several, similar sub tasks is a suitable approach to solve the problem.The achieved success rates in simulation with noiseless observations of the environment are presented in Fig. 10 for training case simulation and in Table 7 for training case hardware.They were determined over a larger set of RPMs than in the baseline method and indicate a good performance of the approach.The success rates become higher when the platform velocity is lower during evaluation compared to the one applied during training.This is to be expected since the equations presented in Sec.3.5 for hyperparameter determination ensure a sufficient maneuverability up to the velocity of the rectilinear periodic movement used in training.For the training case RPM 0.8, the fourth agent has a comparably poor performance.For the static platform the agent occasionally suffers from the problem of motion initiation as described in Sect.4.6, despite setting ψ rel = π/4rad.For the evaluation case where the platform moves with RPM 0.4, the reason is an oscillating movement, which causes the agent to occasionally overshoot  , 109, 112, 119 102, 87, 87, 91 92, 83, 88, 90 89, 92, 107, 110 1 7, 6, 7, 6 6, 5, 5, 5 5, 4, 6, 6 5, 4, 5, 6 2 7, 6, 7, 6 7, 5, 6, 5 6, 5, 9, 6 5, 6, 5, 7 3 9, 7, 8, 8 8, 6, 7, 7 7, 6, 8, 8 6, 7, 6, 8 4 8, 8, 10, 9 10, 8, 8, 9 8, 8, 8   the platform.A similar problem arises for agent four of training case hardware where the agent occasionally expects a maneuver of the platform during the landing procedure.As a consequence, this agent achieves a success rate of only 82% for a static platform.However, for higher platform velocities, the success rate improves significantly.

Selection of Agents for further Evaluation
We select the first agent of training case simulation which is performing best over all evaluation scenarios and was trained on a rectilinear periodic movement with a platform velocity of v mp = 1.6m/s.It is denoted RPM 1.6/1.Its reward curve is depicted in Fig. 11.We compare its results with the RL baseline method and the cascaded PI controller in Table 8.The comparison shows that our approach is able to outperform the RL baseline method.For the RPM 0.4 evaluation scenario with noiseless observations we achieve a success rate of 99%, which is +8% better than the baseline.For the RPM 1.2 evaluation scenario, our method is successful in 99% of the landing trials, increasing the RL baseline's success rate by +26%.Note that the maximum radius of the rectilinear periodic movement is r mp =∼ 2.5m in the Cascaded PI controller 100 100 100 100 100 100 Table 6: Success rates in percent achieved in simulation with noiseless observations by the cascaded PI controller.It has been evaluated for six different scenarios of platform movement, as is indicated by the column titles.RL baseline and r mp = 2m in our approach.However, the value used for our approach poses a more difficult challenge since the acceleration acting on the platform is higher, due to the same maximum platform velocity, see Table 2. Furthermore, our method requires ∼ 80% less time to train and 53% less episodes.with the cascaded PI controller shows that our method is capable of similar performance than the cascaded PI controller in terms of success rates achieved in the two comparison scenarios.In order to achieve comparability in the first place, we limited the control efforts of the longitudinal and lateral controller's outer and inner cascade (see Table 4) to values also derived for the training case RPM 1.6.We obtain u o = v max = a mp,max t 0 = 3.39m/s as described in Sect.3.4 and • by means of (44), respectively.As is further shown by Table 6 the cascaded PI controllers achieve a success rate of 100% for all types of platform movement.However, the advantage of our learning based method is that it does not require a tuning procedure as is often conducted manually for cascaded PI controllers and therefore time consuming.This is compensated by a smoother flight behaviour enabled by the use of a continuous value range for the control effort of the PI controller.
For training case hardware we select agent 3 for further evaluation, denoted RPM 0.4/3.Its reward curve is also depicted in Fig. 11.It is able to achieve a success rate of 99% for the evaluation scenario with a static platform, 100% in case of a platform movement of RPM 0.2, 99% for RPM 0.4 and 97% for the eight-shaped trajectory of the platform.It required 2343 episodes to train which took 123min.As already in case simulation, the cascaded PI controller achieves a comparable performance across all types of platform movement as is shown by Table 7.The values for the limits of the control efforts of the longitudinal and lateral controller's outer and inner cascade have been calculated as u o = v max = a mp,max t 0 = 0.8m/s and

Evaluation in Simulation with Noise
We evaluate the selected agent of case simulation and case hardware as well as the cascaded PI controller for robustness against noise.For this purpose, we define a set of values σ noise = σ px , σ py , σ pz , σ vx , σ vy , σ vz specifying a level of zero mean Gaussian noise that is added to the noiseless observations in simulation.The noise level corresponds to the noise present in an EKF-based pose estimation of a multi-rotor UAV recorded during real flight experiments.
σ noise = {0.1m,0.1m, 0.1m, 0.25m/s, 0.25m/s, 0.25m/s} (49) We evaluate the landing performance again for static, periodic and eight-shaped trajectories of the landing platform in Table 9 for training case simulation and in Table 10 for training case hardware, both also showing the results of the cascaded PI controller.For the selected agent of training case simulation, adding the realistic noise σ noise leads to a slightly reduced performance.However, the achieved success rates are still higher than the agent's performance in the RL baseline without noise.For the evaluation scenario with RPM 0.4, our success rate is +4% higher.With RPM 1.2, it is +20% higher.The cascaded PI controllers are unaffected by the specified noise level.They enable a success rate of 100% across all scenarios of platform movement.
For the selected agent of training case hardware, the drop in performance is slightly more pronounced.The reason is that the fly zone and platform specified for this training case are significantly smaller than for the case simulation.As a consequence, the size of the discrete states is also significantly reduced, whereas the noise level stays the same.Thus, noise affects the UAV more, since it is more likely that the agent takes a suboptimal action due to an observation that was biased by noise.The cascaded PI controller is unaffectd by noise also in the scenario of case hardware, achieving success rates between 99% and 100% across the different types of platform movement.

General
Unlike the baseline method, we do not evaluate our approach on real hardware in single flights only.Instead, we provide statistics on the agent's performance for different evaluation scenarios to illustrate the sim-to-real-gap.For this purpose, the selected agent of case hardware was deployed on a quadcopter, see Figs. 1 and 12, that has a mass of m uav,RL = 0.72kg and a diameter of d uav,RL = 0.28cm and deviates from the UAV used in simulation (mass 1%, diameter 18%).The cascaded PI controller was tested on a similar quadcopter, the only difference being a different structure carrying the Vicon markers, resulting in a slightly different mass of m uav,P I = 0.696kg.The diameter is the same.Both quadcopters are equipped with a Raspberry Pi 4B providing a ROS interface and a LibrePilot Revolution flight controller to enable the tracking of the attitude angles commanded by the respective controllers via ROS.Furthermore, it also controls the commanded descend velocity.A motion capture system (Vicon) provides almost noiseless values of the position and velocity of the moving platform and the quadcopter.However, we do not use these information directly to compute the observations (2)-( 5).The reason is that the motion capture system occasionally has wrong detections of markers.Table 8: Comparison of our approach with the baseline methods.The agent selected for the comparison is agent RPM 1.6/1.Our method significantly improves the success rate in the two reference evaluation scenarios while requiring significantly less time and fewer episodes to train compared to the RL baseline.Furthermore, the fly zone used in our approach covers a larger area.In the reference evaluation scenarios it achieves a comparable performance than a cascaded PI controller.
This can result in abrupt jumps in the orientation estimate of the UAV.Since the flight controller would immediately react, this could lead to a dangerous condition in our restricted indoor environment.To avoid any dangerous flight condition, we obtain the position and velocity of the UAV from an EKF-based state estimation method running on the flight controller.For this purpose, we provide the flight controller with a fake GPS signal (using Vicon) with a frequency of 10hz.It is then fused with other noisy sensor data (accelerometer, gyroscope, magnetometer, barometer).The EKF is robust to short periods of wrong orientation estimation.All tested algorithms are run off-board and the setpoint values for the attitude angles generated by the respective controllers are sent to the flight controller via ROS using a WLAN connection.Due to hardware limitations regarding the moving platform, we only evaluate the agent for a static platform and the rectilinear periodic movement.Generally, the Vicon system allows for a state estimate of the copter that has a lower noise level than the one specified by (49).However, the velocity of the platform is determined purely by means of the Vicon system.The rough surface of the ground caused vibrations of the platform that induced an unrealistic high level of noise in the Vicon system's readings of the platform velocity.
For this reason, it is filtered using a low-pass filter with a cut-off frequency of 10hz.

Evaluation of the RL Approach
Fig. 13 shows the trajectory of the UAV and moving platform for a landing trial where the platform was executing a rectilinear periodic movement with a maximum velocity of 0.4m/s.The depicted x and y component of the multi-rotor vehicle's trajectory are based on the state estimate calculated by the EKF.All other presented trajectory values are based on the readings of the Vicon system.Table 11 contains the success rates achieved in the experiments with our RL algorithm deployed on real hardware.The starting positions of the UAV were manually selected and as uniformly distributed as possible in an average altitude of 2.17m.Whereas for the static platform a success rate of 96% could be reached, there is    a noticeable drop in performance for the evaluation scenarios in which the platform is performing the rectilinear periodic movement.We argue that these can be attributed to the following main effects.
1.The deviation in the size of the UAV and the different mass plays a role.Whereas the difference in mass is seems negligible, the distribution of mass and thus the inertia of the multi-rotor vehicle is different.Furthermore, during the real flight experiments no significant jittering in the agent's actions could be observed.This substantiates the approach of calculating values for the maximum pitch angle and agent frequency presented in Sect.3.5.

Evaluation of the Cascaded PI Controller
Table 12 shows the success rates achieved with the cascaded PI controller.Again, the starting positions of the UAV were manually selected and as uniformly distributed as possible in an average altitude of 2.09m.The results show that there is a notable drop in performance for the scenarios RPM 0.2 and RPM 0.4.This was not expected considering the strong results obtained in simulation.Careful investigation of the log files revealed several reasons that could play a role.
1.The multi-rotor vehicle used for testing the cascaded PI controller differed from the one used for the evaluation of the RL controller.
Although being identical in terms of size and components, it is 0.02kg lighter due to a different structure carrying the markers for the Vicon system.2. In 81% of the landing trials a loss of state estimate messages that were sent by the multirotor vehicle and received by the ground station running the cascaded PI controller occurred.This happened probably due to issues of ROS queing messages or limited bandwidth of the WLAN module, which was changed.The cascaded PI controller is intended and tuned to run at 100hz, like in the Gazebo simulation, however, had to handle low frequencies of approximately 60hz instead for different lengths of time.As a consequence, the cascaded PI controller's cycles were executed with a non-constant and often reduced frequency.However, drops in the frequency of the received state estimate were also present for the scenario of a static platform where the cascaded PI controller was able to achieve high success rates.The necessity to transmit attitude setpoints with high frequency can be a disadvantage in situations in which the control method is run off-board and the resulting commands are sent to the multi-rotor vehicle via WLAN.In this regard, our RL based approach is beneficial as it is designed to be run at considerably lower frequencies as is illustrated by Table 12: Success rates in percent achieved in real world experiments by the cascaded PI controller structure for case hardware.It has been evaluated over different numbers of landing trials for three different scenarios of platform movement, as is indicated by the column titles.For the cascaded PI controller, violations of the boundaries of the fly zone are not counted as failures.
control is handed over to the tested controller again.This maneuver can increase the momentum of the copter which then has to be handled by the limited attitude angles of the tested controller.Leaving the virtual fly box happened frequently during testing of the cascaded PI controller where the initial position of a landing trial was often located outside the fly box, probably due to issues with the altitude estimation.For this reason, leaving the fly box is not counted as a failed attempt for the cascaded PI controller in Table 12.
Besides the aforementioned issues, an additional problem of the cascaded PI controller became obvious during the real world experiments.Without a differential branch it cannot react to the current change in the control errors of the outer and inner cascade.During tuning in simulation a cascaded PI controller turned out to be successful.However, during the real world experiments it could be noted that the cascaded PI controller tends to overshoot the platform in the scenarios RPM 0.2 and RPM 0.4.Here, the RL controller is advantageous because it has learned how to handle the platform's change in the direction of motion.

Ablation Study
In the context of our approach and in addition to the presented method where the landing task was learned by means of a sequential curriculum leveraging transfer learning between curriculum steps, there are two other ways of how learning the task could be achieved.First, the sequential curriculum is conducted without transfer learning.This means that whenever a new curriculum step is added to the learning sequence the Q-table associated with the latest curriculum step is initialized with zeros.We denote this ablation case Sequential curriculum without transfer learning.And second, we abandon curriculum learning completely.
Instead, we create the full sequence of learning steps right in the beginning with all associated Qtables initialized with zeros.We then conduct one training on the full sequence without adding an additional step after each round of training.This constitutes a scenario in which several separate Q-learning tasks are trained simultaneously.For this reason, we call this ablation case Simultaneous training of sequence.For the training and evaluation, we chose the scenario RPM 1.6.All learning related parameters are the same as presented in Sect.4.5.We compare these additional two methods with the agent selected for further evaluation in Sect.5.2 and name this scenario Sequential curriculum with transfer learning.Table 13 summarizes the results.It is clear that ablation case Sequential curriculum without transfer learning performs worst across all evaluation scenarios in terms of both training duration as well as number of episodes required.Our rationale here is the following.In each new curriculum step the task has to be learned from scratch.However, since in all curriculum steps added to the sequence after the initial step the exploration rate is set to ε = 0, it is even more difficult to learn optimal behaviour.Furthermore, in the early phase after adding a new curriculum step its behaviour is bad due to few updates of the state-action-pairs, therefore acting as a source of disturbance to the previous curriculum steps.For this reason, in order to reach a situation in which the Q-values of the latest curriculum step as well as the Q-values of the previous curriculum steps have sufficiently adapted (96% success rate over the last 100 episodes), a significant amount of time is required.The discrepancy between the success rate necessary to end the training and the success rate achieved during evaluation can be explained as follows.As outlined in Sect.3.4, an episode is terminated with success if the agent spends one second in the latest curriculum step before reaching its goal state.However, during evaluation the Table 13: Results of the ablation study.The columns two to seven show the success rates achieved in the different evaluation scenarios.The last column contains the training information of the respective ablation case.multi-rotor vehicle is required to constantly stay above the moving platform, not only for one second.This leaves room for erratic behaviour after this one second period which can be observed for this ablation case.Even though centering the multi-rotor vehicle above the platform is learned in principle, maintaining that centered state for a longer time period is not possible which leads to low success rates.Conceivably, the situation could be improved by choosing a longer time period for maintaining the goal state before the success criterion is met.
Allowing training on all curriculum steps that were initialized with zeros simultaneously creates the possibility to adapt all Q-values of all the separate Q-tables associated with the different steps in the sequence during one training.This solves the problem where the latest untrained curriculum step acts as a disturbance to the previous curriculum steps.As a result, the training time is considerably shorter for the ablation case Simultaneous training of sequence.Also, the erratic behaviour is reduced, leading to higher success rates across all evaluation scenarios.However, at the end of the training the number of unvisited state-action-pairs is quite high, 497 out of 2835 state-action pairs.This indicates that the training is not yet optimal, although the terminal criterion to end the training has been met.This might be solved with a prolongation of the exploration phase.The best performance in terms of success rates as well as required training time was achieved by the third ablation case Sequential curriculum with transfer learning which constitutes the full method.Here, the erratic behaviour is not present anymore.We hypothize that this is due to the transfer of knowledge between curriculum steps.Not only does this significantly reduce the number of unvisited state-action-pairs (0 out of 2385, a curriculum step adds 567 new state-action-pairs to the sequence) but also the knowledge does not have to be learned from scratch in each new curriculum step but is rather only refined.

Conclusion and Future Work
In this work, we presented a RL based method for autonomous multi-rotor landing.Our method splits the overall task into simpler 1-D tasks, then formulates each of those as a sequential curriculum in combination with a knowledge transfer between curriculum steps.Through rigorous experiments, we demonstrate significantly shorter training time (∼ −80%) and higher success rates (up to +26%) than the DRL-based actor-critic method presented in (Rodriguez-Ramos et al, 2019).For two comparison scenarios, our method achieves a performance comparable to a cascaded PI controller.We present statistics of the performance of our approach on real hardware and show interpretable ways to set hyperparameters.In the future, we plan to test other tabular RL methods such as nstep Q(σ) learning or the use of eligible traces.Furthermore, extending the approach to control the vertical movement and yaw angle of the multirotor vehicle is of interest as well as addressing complex landing problems, such as on an inclined or on a flying platform.

Fig. 1 :
Fig. 1: Our multi-rotor vehicle landing on a platform moving on rails.

Fig. 3 :
Fig. 3: Illustration of the sequential curriculum for n cs +1 curriculum steps.Each curriculum step consists of an instance of the basis learning task.

Fig. 5 :
Fig. 5: Maximum pitch angle θ max for different values of k a and a mp,max .

Fig. 6 :
Fig. 6: Agent frequency f ag for different values of k man and ω mp,max .

Fig. 10 :
Fig.10: Success rates of four agents trained for the case simulation with same parameters on a platform's rectilinear periodic movement with v mp,train = 0.8m/s (RPM 0.8 -red), v mp,train = 1.2m/s (RPM 1.2blue) and v mp,train = 1.6m/s (RPM 1.6 -green).The agents are evaluated on different types of platform movement, indicated by the values of the abscissa.Error bars illustrating the mean success rate and standard deviation are depicted in grey.The success rates have been determined for 150 landing trials for each evaluation scenario.

Fig. 11 :
Fig. 11: Accumulated reward achieved by agent RPM 1.6/1 of case simulation (top) and agent RPM 0.4/3 of case hardware (bottom) during training.Red lines indicate the end of a curriculum step.

Fig. 12 :
Fig. 12: Multi-rotor vehicle and autonomous platform moving on rails that were used for the flight experiments in the real world.The squared platform has an edge length of 0.5m.

Fig. 13 :
Fig. 13: Example trajectory of the multi-rotor vehicle and moving platform during the real flight experiment.The platform's position is determined using the Vicon system, the multi-rotor vehicle's position results from the state estimate of the flight controller.
Overview of the framework's structure.Red are components relying on Gazebo.Green are components designed for converting and extracting / merging data.Yellow are components dealing with the control of the UAV.

Table 1 :
Parameters of the training environment.

Table 2 :
Differences between the training parameters of our method and the baseline.

Table 3 :
Training parameters of our approach that are associated with the Double Q-Learning algorithm, the reward function or the discretization of the state space.Structure of the cascaded PI controller.
Fig.8: Illustration of the problem of motion initiation.After initialization, commanding a yaw angle of ψ rel = π/4 allows the lateral agent to enter a state associated with another action than "do nothing" due to the state change induced by the longitudinal platform movement.The platform movement is now reflected in p y , v y , a y that are the observations fed into the lateral agent.

Table 5 :
Training time in minutes required for the different curriculum steps by four agents trained with identical parameters for the different training cases.The values are rounded to the nearest integer.The training scenarios are identified by the platform's rectilinear periodic movement with a maximum velocity of v mp,train = 0.8 (RPM 0.8), v mp,train = 1.2 (RPM 1.2) and v mp,train = 1.6 (RPM 1.6) for case simulation and v mp,train = 0.4 (RPM 0.4) for case hardware.

Table 7 :
Success rates in percent achieved in simulation with noiseless observations by four agents trained with same parameters for case hardware and the cascaded PI controller.They have been evaluated for four different scenarios of platform movement, as is indicated by the column titles.

Table 9 :
Success rates in percent achieved in simulation with noisy observations by the best performing agent of training case simulation and the cascaded PI controller.They have been evaluated over 150 landing trials for six different scenarios of platform movement, as is indicated by the column titles.

Table 10 :
Success rates in percent achieved in simulation with noisy observations by the selected agent of training case hardware and the cascaded PI controller.They have been evaluated over 150 landing trials for four different scenarios of platform movement, as is indicated by the column titles.

Table 11 :
Success rates in percent achieved in real flights with ground effect by the selected agent of training case hardware.They have been evaluated over at least 20 landing trials for three different scenarios of platform movement, as is indicated by the column titles.