The UZH-FPV Drone Racing Dataset

High-speed, Aggressive 6DoF Trajectories for State Estimation and Drone Racing

News

May 2nd
2024
We have a new competition! The competition is hosted at the RSS 24 workshop on Safety Autonomy.
February
9th 2024
We have a new Leader Board! We extended the test dataset
including the sequences in the SplitS track.
December 1st 2023Added new sequences in a SplitS track with peak speeds up to 100 km/h. Check our Nature paper for details.
February 2nd 2022New ground truth is available. Leader board has been updated.
October 26th 2020The results from our IROS 2020 competition are public.

About

We introduce the UZH-FPV Drone Racing dataset, which is the most aggressive visual-inertial odometry dataset to date. Large accelerations, rotations, and apparent motion in vision sensors make aggressive trajectories difficult for state estimation. However, many compelling applications, such as autonomous drone racing, require high-speed state estimation, but existing datasets do not address this. These sequences were recorded with a first-person-view (FPV) drone racing quadrotor fitted with sensors and flown aggressively by an expert pilot. The trajectories include fast laps around a racetrack with drone racing gates, as well as free-form trajectories around obstacles, both indoor and out. We present the camera images and IMU data from a Qualcomm Snapdragon Flight board, ground truth from a Leica Nova MS60 laser tracker, as well as event data from an mDAVIS 346 event camera, and high-resolution RGB images from the pilot’s FPV camera. With this dataset, our goal is to help advance the state of the art in high-speed state estimation.

Publication

When using this work in an academic context, please cite the following publications:

Publication accompanying the Dataset

Are We Ready for Autonomous Drone Racing? The UZH-FPV Drone Racing Dataset

J. Delmerico, T. Cieslewski, H. Rebecq, M. Faessler, D. Scaramuzza

Are We Ready for Autonomous Drone Racing? The UZH-FPV Drone Racing Dataset

IEEE International Conference on Robotics and Automation (ICRA), 2019.

PDF YouTube Code


@InProceedings{Delmerico19icra,
 author = {Jeffrey Delmerico and Titus Cieslewski and Henri Rebecq and Matthias Faessler and Davide Scaramuzza},
 title = {Are We Ready for Autonomous Drone Racing? The {UZH-FPV} Drone Racing Dataset},
 booktitle = {{IEEE} Int. Conf. Robot. Autom. ({ICRA})},
 year = 2019
}

Publication containing the approach used to generate the ground truth

Continuous-Time vs. Discrete-Time Vision-based SLAM: A Comparative Study

G. Cioffi, T. Cieslewski, D. Scaramuzza

Continuous-Time vs. Discrete-Time Vision-based SLAM: A Comparative Study

Robotics and Automation Letters (RA-L), 2022.

PDF Code


@article{cioffi2022continuous,
  title={Continuous-Time vs. Discrete-Time Vision-based SLAM: A Comparative Study},
  author={Cioffi, Giovanni and Cieslewski, Titus and Scaramuzza, Davide},
  journal={{IEEE} Robotics and Automation Letters, ({RA-L})},
  year={2022},
  publisher={IEEE}
}

Publication containing the approach used to record the sequence in the SplitS track

Champion-level Drone Racing using Deep Reinforcement Learning

E. Kaufmann, L. Bauersfeld, A. Loquercio, M. Müller, V. Koltun, D. Scaramuzza

Champion-level Drone Racing using Deep Reinforcement Learning

Nature, 2023.

PDF Video (ours) Video (Nature)


@article{kaufmann2023champion,
  title={Champion-level drone racing using deep reinforcement learning},
  author={Kaufmann, Elia and Bauersfeld, Leonard and Loquercio, Antonio and M{\"u}ller, Matthias and Koltun, Vladlen and Scaramuzza, Davide},
  journal={Nature},
  volume={620},
  number={7976},
  pages={982--987},
  year={2023},
  publisher={Nature Publishing Group UK London}
}

License

This datasets are released under the Creative Commons license (CC BY-NC-SA 3.0), which is free for non-commercial use (including research).

Acknowledgments

This work was supported by the National Centre of Competence in Research Robotics (NCCR) through the Swiss National Science Foundation, the SNSF-ERC Starting Grant, and the DARPA FLA Program.

This work would not have been possible without the assistance of Stefan Gächter, Zoltan Török, and Thomas Mörwald of Leica Geosystems and their support in gathering our data. Additional thanks go to Innovation Park Zürich, and the Fässler family for providing experimental space, iniVation AG and Prof. Tobi Delbruck for their support and guidance with the mDAVIS sensors, and Hanspeter Kunz from the Department of Informatics at the University of Zurich for his support with setting up this website.