A real-world multi-vehicle multi-modal V2V and V2X dataset
Recently published datasets have been increasingly
comprehensive with respect to their variety of simultaneously
used sensors, traffic scenarios, environmental conditions, and
provided annotations. However, these datasets typically only
consider data collected by one independent vehicle. Hence, there
is currently a lack of comprehensive, real-world, multi-vehicle
datasets fostering research on cooperative applications such as
object detection, urban navigation, or multi-agent SLAM. In
this paper, we aim to fill this gap by introducing the novel
LUCOOP dataset, which provides time-synchronized multi-modal
data collected by three interacting measurement vehicles. The
driving scenario corresponds to a follow-up setup of multiple
rounds in an inner city triangular trajectory. Each vehicle was
equipped with a broad sensor suite including at least one LiDAR
sensor, one GNSS antenna, and up to three IMUs. Additionally,
Ultra-Wide-Band (UWB) sensors were mounted on each vehicle,
as well as statically placed along the trajectory enabling both
V2V and V2X range measurements. Furthermore, a part of the
trajectory was monitored by a total station resulting in a highly
accurate reference trajectory. The LUCOOP dataset also includes
a precise, dense 3D map point cloud, acquired simultaneously by
a mobile mapping system, as well as an LOD2 city model of
the measurement area. We provide sensor measurements in a
multi-vehicle setup for a trajectory of more than 4 km and a
time interval of more than 26 minutes, respectively. Overall, our
dataset includes more than 54,000 LiDAR frames, approximately
700,000 IMU measurements, and more than 2.5 hours of 10 Hz
GNSS raw measurements along with 1 Hz data from a reference
station. Furthermore, we provide more than 6,000 total station
measurements over a trajectory of more than 1 km and 1,874
V2V and 267 V2X UWB measurements. Additionally, we offer
3D bounding box annotations for evaluating object detection
approaches, as well as highly accurate ground truth poses for
each vehicle throughout the measurement campaign.
Data access
Important: Before downloading and using the data, please check the Updates.zip in the "Data and Resources" section at the bottom of this web site. There, you find updated files and annotations as well as update notes.
- The dataset is available here.
- Additional information are provided and constantly updated in our README.
- The corresponding paper is available here.
- Cite this as: J. Axmann et al., "LUCOOP: Leibniz University Cooperative Perception and Urban Navigation Dataset," 2023 IEEE Intelligent Vehicles Symposium (IV), Anchorage, AK, USA, 2023, pp. 1-8, doi: 10.1109/IV55152.2023.10186693.
Preview
Source LOD2 City model: Auszug aus den Geodaten des Landesamtes für Geoinformation und Landesvermessung Niedersachsen, ©2023, www.lgln.de
Sensor Setup of the three measurement vehicles
Sensor setup of all the three vehicles: Each vehicle is equipped with a LiDAR sensor (green), a UWB unit (orange), a GNSS antenna (purple), and a Microstrain IMU (red). Additionally, each vehicle has its unique feature: Vehicle 1 has an additional LiDAR at the trailer hitch (green) and a prism for the tracking of the total station (dark red hexagon). Vehicle 2 provides an iMAR iPRENA (yellow) and iMAR FSAS (blue) IMU, where the platform containing the IMUs is mounted inside the car (dashed box). Vehicle 3 carries the RIEGL MMS (pink). Along with the sensors and platforms, the right-handed body frame of each vehicle is also indicated.
3D map point cloud
High resolution 3D map point cloud: Different locations and details along the trajectory. Colors according to reflectance values.
Measurement scenario
Driven trajectory and locations of the static sensors: The blue hexagons indicate the positions of the static UWB
sensors, the orange star represents the location of the total station, and the orange shaded area illustrates the coverage of
the total station. The route of the three measurement vehicles is shown in purple. Background map: OpenStreetMap copyright
Source LOD2 City model: Auszug aus den Geodaten des Landesamtes für Geoinformation und Landesvermessung Niedersachsen, ©2023, www.lgln.de
Number of annotations per class (final)
Source LOD2 City model: Auszug aus den Geodaten des Landesamtes für Geoinformation und Landesvermessung Niedersachsen, ©2023, www.lgln.de
Data structure
Data format
Gallery
<!--
Van 1
Van 2
Van 3
-->
From left to right: Van 1, van 2, van 3.
From left to right: Tracking of the prism on van 1 by means of the MS60
total station, the detected prism from the view point of the MS60 total station, PulsON 440 Ultra Wide Band (UWB) sensors, RIEGL VMX-250 Mobile Mapping System.
Acknowledgement
This measurement campaign could not have been carried
out without the help of many contributors. At this point, we
thank Yuehan Jiang (Institute for Autonomous Cyber-Physical
Systems, Hamburg), Franziska Altemeier, Ingo Neumann,
Sören Vogel, Frederic Hake (all Geodetic Institute, Hannover),
Colin Fischer (Institute of Cartography and Geoinformatics,
Hannover), Thomas Maschke, Tobias Kersten, Nina Fletling
(all Institut für Erdmessung, Hannover), Jörg Blankenbach
(Geodetic Institute, Aachen), Florian Alpen (Hydromapper
GmbH), Allison Kealy (Victorian Department of Environment,
Land, Water and Planning, Melbourne), Günther Retscher,
Jelena Gabela (both Department of Geodesy and Geoin-
formation, Wien), Wenchao Li (Solinnov Pty Ltd), Adrian
Bingham (Applied Artificial Intelligence Institute, Burwood),
and the student assistants Manuel Kramer, Khaled Ahmed,
Leonard Göttert, Dennis Mußgnug, Chengqi Zhou, and We-
icheng Zhang. Thanks to the Landesamt für Geoinformation
and Landesermessung Niedersachsen (LGLN)/Zentrale Stelle
SAPOS® for providing the virtual reference station data, in-
frastructure and reliable high quality. This project is supported
by the German Research Foundation (DFG), as part of the
Research Training Group i.c.sens, GRK 2159, ”Integrity and
Collaboration in Dynamic Sensor Networks”.