Skip to content

Closed-Loop Q-learning Control of a Small Unmanned Aircraft

Research output: Contribution to conferencePaper

Original languageEnglish
DOIs
DatePublished - 5 Jan 2020

Abstract

Fixed-wing unmanned aerial vehicles (UAVs) using conventional flight controllers are limited in their manoeuvrability, as agile flight manoeuvres require the exploitation of the nonlinear post-stall flight regime. This paper extends previous work which used a Deep Q-Network (DQN) to generate open-loop trajectories for a perching manoeuvre. This paper proposes a closed-loop DQN controller, where live inference of the actuator actions is performed on a small unmanned aircraft during flight. DQN models are trained and evaluated in simulation before being deployed on the vehicle. The training processes uses a numerical flight dynamics model of the aircraft, combined with a baseline DQN implementation, to generate a series of trained models. Models with both fixed and varying start conditions at the start of each learning episode were evaluated, to identify which approach is best for generating robust controllers. Real-world closed-loop control is demonstrated through a series of flight tests, with varying wind conditions. The closed-loop controller was shown to have superior performance compared with open-loop mode, achieving a greater mean reward when performing the manoeuvre. Further, it is found that incorporating wind into the training process, as well as real-world effects such as noise and latency, is essential for further development.

Documents

DOI

View research connections

Related faculties, schools or groups