LUCOOP: Leibniz University Cooperative Perception and Urban Navigation Dataset




Date made available2022
PublisherForschungsdaten-Repositorium der LUH
Date of data production2023


Recently published datasets have been increasingly comprehensive with respect to their variety of simultaneously used sensors, traffic scenarios, environmental conditions, and provided annotations. However, these datasets typically only consider data collected by one independent vehicle. Hence, there is currently a lack of comprehensive, real-world, multi-vehicle datasets fostering research on cooperative applications such as object detection, urban navigation, or multi-agent SLAM. In this paper, we aim to fill this gap by introducing the novel LUCOOP dataset, which provides time-synchronized multi-modal data collected by three interacting measurement vehicles. The driving scenario corresponds to a follow-up setup of multiple rounds in an inner city triangular trajectory. Each vehicle was equipped with a broad sensor suite including at least one LiDAR sensor, one GNSS antenna, and up to three IMUs. Additionally, Ultra-Wide-Band (UWB) sensors were mounted on each vehicle, as well as statically placed along the trajectory enabling both V2V and V2X range measurements. Furthermore, a part of the trajectory was monitored by a total station resulting in a highly accurate reference trajectory. The LUCOOP dataset also includes a precise, dense 3D map point cloud, acquired simultaneously by a mobile mapping system, as well as an LOD2 city model of the measurement area. We provide sensor measurements in a multi-vehicle setup for a trajectory of more than 4 km and a time interval of more than 26 minutes, respectively. Overall, our dataset includes more than 54,000 LiDAR frames, approximately 700,000 IMU measurements, and more than 2.5 hours of 10 Hz GNSS raw measurements along with 1 Hz data from a reference station. Furthermore, we provide more than 6,000 total station measurements over a trajectory of more than 1 km and 1,874 V2V and 267 V2X UWB measurements. Additionally, we offer 3D bounding box annotations for evaluating object detection approaches, as well as highly accurate ground truth poses for each vehicle throughout the measurement campaign. ## Data access The data set is available at the following link: []( Additional information are provided and constantly updated in our README document: []( ##Preview [![Watch the video](]( Source LOD2 City model: Auszug aus den Geodaten des Landesamtes für Geoinformation und Landesvermessung Niedersachsen, ©2023, []( ![Alt text][id_lgln] [id_lgln]: "LGLN logo" ##Sensor Setup of the three measurement vehicles ![Alt text][id] [id]: "Sensor Setup of the three measurement vehicles" Sensor setup of all the three vehicles: Each vehicle is equipped with a LiDAR sensor (green), a UWB unit (orange), a GNSS antenna (purple), and a Microstrain IMU (red). Additionally, each vehicle has its unique feature: Vehicle 1 has an additional LiDAR at the trailer hitch (green) and a prism for the tracking of the total station (dark red hexagon). Vehicle 2 provides an iMAR iPRENA (yellow) and iMAR FSAS (blue) IMU, where the platform containing the IMUs is mounted inside the car (dashed box). Vehicle 3 carries the RIEGL MMS (pink). Along with the sensors and platforms, the right-handed body frame of each vehicle is also indicated. ##3D map point cloud ![Alt text][idmap] [idmap]: "3D map point cloud" High resolution 3D map point cloud: Different locations and details along the trajectory. Colors according to reflectance values. ##Measurement scenario ![Alt text][id1] [id1]: "Measurement scenario" Driven trajectory and locations of the static sensors: The blue hexagons indicate the positions of the static UWB sensors, the orange star represents the location of the total station, and the orange shaded area illustrates the coverage of the total station. The route of the three measurement vehicles is shown in purple. Background map: [OpenStreetMap copyright]( ##Data structure ![Alt text][id_d] [id_d]: "Data structure" ##Data format ![Alt text][id_f] [id_f]: "Data format" ##Gallery ![Alt text][id6] [id6]: "Measurement vehicles" From left to right: Van 1, van 2, van 3. ![Alt text][id5] [id5]: "" From left to right: Tracking of the prism on van 1 by means of the MS60 total station, the detected prism from the view point of the MS60 total station, PulsON 440 Ultra Wide Band (UWB) sensors, RIEGL VMX-250 Mobile Mapping System. ## Acknowledgement This measurement campaign could not have been carried out without the help of many contributors. At this point, we thank Yuehan Jiang (Institute for Autonomous Cyber-Physical Systems, Hamburg), Franziska Altemeier, Ingo Neumann, Sören Vogel, Frederic Hake (all Geodetic Institute, Hannover), Colin Fischer (Institute of Cartography and Geoinformatics, Hannover), Thomas Maschke, Tobias Kersten, Nina Fletling (all Institut für Erdmessung, Hannover), Jörg Blankenbach (Geodetic Institute, Aachen), Florian Alpen (Hydromapper GmbH), Allison Kealy (Victorian Department of Environment, Land, Water and Planning, Melbourne), Günther Retscher, Jelena Gabela (both Department of Geodesy and Geoin- formation, Wien), Wenchao Li (Solinnov Pty Ltd), Adrian Bingham (Applied Artificial Intelligence Institute, Burwood), and the student assistants Manuel Kramer, Khaled Ahmed, Leonard Göttert, Dennis Mußgnug, Chengqi Zhou, and We- icheng Zhang. Thanks to the Landesamt für Geoinformation and Landesermessung Niedersachsen (LGLN)/Zentrale Stelle SAPOS® for providing the virtual reference station data, in- frastructure and reliable high quality. This project is supported by the German Research Foundation (DFG), as part of the Research Training Group i.c.sens, GRK 2159, ”Integrity and Collaboration in Dynamic Sensor Networks”.