LUCOOP: Leibniz University Cooperative Perception and Urban Navigation Dataset

A real-world multi-vehicle multi-modal V2V and V2X dataset

Recently published datasets have been increasingly comprehensive with respect to their variety of simultaneously used sensors, traffic scenarios, environmental conditions, and provided annotations. However, these datasets typically only consider data collected by one independent vehicle. Hence, there is currently a lack of comprehensive, real-world, multi-vehicle datasets fostering research on cooperative applications such as object detection, urban navigation, or multi-agent SLAM. In this paper, we aim to fill this gap by introducing the novel LUCOOP dataset, which provides time-synchronized multi-modal data collected by three interacting measurement vehicles. The driving scenario corresponds to a follow-up setup of multiple rounds in an inner city triangular trajectory. Each vehicle was equipped with a broad sensor suite including at least one LiDAR sensor, one GNSS antenna, and up to three IMUs. Additionally, Ultra-Wide-Band (UWB) sensors were mounted on each vehicle, as well as statically placed along the trajectory enabling both V2V and V2X range measurements. Furthermore, a part of the trajectory was monitored by a total station resulting in a highly accurate reference trajectory. The LUCOOP dataset also includes a precise, dense 3D map point cloud, acquired simultaneously by a mobile mapping system, as well as an LOD2 city model of the measurement area. We provide sensor measurements in a multi-vehicle setup for a trajectory of more than 4 km and a time interval of more than 26 minutes, respectively. Overall, our dataset includes more than 54,000 LiDAR frames, approximately 700,000 IMU measurements, and more than 2.5 hours of 10 Hz GNSS raw measurements along with 1 Hz data from a reference station. Furthermore, we provide more than 6,000 total station measurements over a trajectory of more than 1 km and 1,874 V2V and 267 V2X UWB measurements. Additionally, we offer 3D bounding box annotations for evaluating object detection approaches, as well as highly accurate ground truth poses for each vehicle throughout the measurement campaign.

Data access

Important: Before downloading and using the data, please check the Updates.zip in the "Data and Resources" section at the bottom of this web site. There, you find updated files and annotations as well as update notes.

  • The dataset is available here.
  • Additional information are provided and constantly updated in our README.
  • The corresponding paper is available here.
  • Cite this as: J. Axmann et al., "LUCOOP: Leibniz University Cooperative Perception and Urban Navigation Dataset," 2023 IEEE Intelligent Vehicles Symposium (IV), Anchorage, AK, USA, 2023, pp. 1-8, doi: 10.1109/IV55152.2023.10186693.

Preview

Watch the video Source LOD2 City model: Auszug aus den Geodaten des Landesamtes für Geoinformation und Landesvermessung Niedersachsen, ©2023, www.lgln.de

Alt text

Sensor Setup of the three measurement vehicles

Alt text Sensor setup of all the three vehicles: Each vehicle is equipped with a LiDAR sensor (green), a UWB unit (orange), a GNSS antenna (purple), and a Microstrain IMU (red). Additionally, each vehicle has its unique feature: Vehicle 1 has an additional LiDAR at the trailer hitch (green) and a prism for the tracking of the total station (dark red hexagon). Vehicle 2 provides an iMAR iPRENA (yellow) and iMAR FSAS (blue) IMU, where the platform containing the IMUs is mounted inside the car (dashed box). Vehicle 3 carries the RIEGL MMS (pink). Along with the sensors and platforms, the right-handed body frame of each vehicle is also indicated.

3D map point cloud

Alt text High resolution 3D map point cloud: Different locations and details along the trajectory. Colors according to reflectance values.

Measurement scenario

Alt text Driven trajectory and locations of the static sensors: The blue hexagons indicate the positions of the static UWB sensors, the orange star represents the location of the total station, and the orange shaded area illustrates the coverage of the total station. The route of the three measurement vehicles is shown in purple. Background map: OpenStreetMap copyright

Watch the video Source LOD2 City model: Auszug aus den Geodaten des Landesamtes für Geoinformation und Landesvermessung Niedersachsen, ©2023, www.lgln.de

Number of annotations per class (final)

Alt text

Watch the video Source LOD2 City model: Auszug aus den Geodaten des Landesamtes für Geoinformation und Landesvermessung Niedersachsen, ©2023, www.lgln.de

Data structure

Alt text

Data format

Alt text

Gallery

<!--

Alt text Van 1

Alt text Van 2

Alt text Van 3

-->

Alt text From left to right: Van 1, van 2, van 3.

Alt text From left to right: Tracking of the prism on van 1 by means of the MS60 total station, the detected prism from the view point of the MS60 total station, PulsON 440 Ultra Wide Band (UWB) sensors, RIEGL VMX-250 Mobile Mapping System.

Acknowledgement

This measurement campaign could not have been carried out without the help of many contributors. At this point, we thank Yuehan Jiang (Institute for Autonomous Cyber-Physical Systems, Hamburg), Franziska Altemeier, Ingo Neumann, Sören Vogel, Frederic Hake (all Geodetic Institute, Hannover), Colin Fischer (Institute of Cartography and Geoinformatics, Hannover), Thomas Maschke, Tobias Kersten, Nina Fletling (all Institut für Erdmessung, Hannover), Jörg Blankenbach (Geodetic Institute, Aachen), Florian Alpen (Hydromapper GmbH), Allison Kealy (Victorian Department of Environment, Land, Water and Planning, Melbourne), Günther Retscher, Jelena Gabela (both Department of Geodesy and Geoin- formation, Wien), Wenchao Li (Solinnov Pty Ltd), Adrian Bingham (Applied Artificial Intelligence Institute, Burwood), and the student assistants Manuel Kramer, Khaled Ahmed, Leonard Göttert, Dennis Mußgnug, Chengqi Zhou, and We- icheng Zhang. Thanks to the Landesamt für Geoinformation and Landesermessung Niedersachsen (LGLN)/Zentrale Stelle SAPOS® for providing the virtual reference station data, in- frastructure and reliable high quality. This project is supported by the German Research Foundation (DFG), as part of the Research Training Group i.c.sens, GRK 2159, ”Integrity and Collaboration in Dynamic Sensor Networks”.

Data and Resources

Cite this as

Axmann, Jeldrik, Moftizadeh, Rozhin, Su, Jingyao, Tennstedt, Benjamin, Zou, Qianqian, Yuan, Yunshuang, Ernst, Dominik, Alkhatib, Hamza, Brenner, Claus, Schön, Steffen (2023). Dataset: LUCOOP: Leibniz University Cooperative Perception and Urban Navigation Dataset. https://doi.org/10.25835/75o9yrc0

DOI retrieved: January 16, 2023

Additional Info

Field Value
Imported on February 22, 2023
Last update November 28, 2024
License CC-BY-NC-3.0
Source https://data.uni-hannover.de/dataset/lucoop-leibniz-university-cooperative-perception-and-urban-navigation-dataset
Author Axmann, Jeldrik
Given Name Jeldrik
Family Name Axmann
More Authors
Moftizadeh, Rozhin
Su, Jingyao
Tennstedt, Benjamin
Zou, Qianqian
Yuan, Yunshuang
Ernst, Dominik
Alkhatib, Hamza
Brenner, Claus
Schön, Steffen
Author Email Axmann, Jeldrik
Maintainer Rozhin Moftizadeh (moftizadeh@gih.uni-hannover.de)
Maintainer Email Rozhin Moftizadeh (moftizadeh@gih.uni-hannover.de)
Source Creation 09 September, 2022, 14:43 PM (UTC+0000)
Source Modified 26 March, 2024, 10:07 AM (UTC+0000)