Deep learning based approaches have shown impressive performance in robotics applications in recent years. Nevertheless, the demand for huge training data and the lack of the explainability and safety guarantee of the trained policy still limit its usage in safety-critical applications such as autonomous cars, surgical robots, etc.
In this thesis, we focus on developing the navigation methods for the racing drones, which need to fly through complex tracks formed by gates at high speed. We use a modularized approach to divide the whole system into perception, planning, and control module. The focus of this thesis is on the perception module which takes the camera image and drone states as input and generates high-level navigation commands such as the navigation direction and speed. As for the planning and control module, we leverage well-studied state-of-the-art methods for path planning and drone control. This modularization choice combines the advantages of the learning-based methods for a robust perception module and the model-based methods for precise planning and control.
An expert policy that automatically maneuvers the drone through the track is proposed in this thesis based on the accurate positions of the drone and the gates which are available in the simulator. While the expert policy generates navigation command only based on the accurate positions of the drone and the gates, informative auxiliary data like the camera images and drone states can still be collected at the same time along with the expert policy’s decisions. We then use imitation learning to train a customized neural network to map from the camera images and drone states to the high-level commands including the navigation direction and speed. Moreover, the extensive domain randomization involved during data collection makes the trained policy robust enough and successfully directly transferred to environments that were unseen during training.