3D occupancy prediction benchmark for autonomous driving scene perception
Top 43.2% on sourcepulse
This repository hosts the CVPR 2023 3D Occupancy Prediction Challenge, providing a benchmark for autonomous driving scene perception. It addresses the limitations of traditional 3D bounding box detection by enabling dense, voxel-wise prediction of scene occupancy and semantics from surround-view images. The target audience includes researchers and engineers in autonomous driving and computer vision.
How It Works
The challenge focuses on predicting the occupancy state (free or occupied) and semantic class for each voxel in a 3D scene, using only camera images as input. This approach allows for a more detailed representation of the environment compared to bounding boxes, capturing complex object shapes and background elements. The benchmark utilizes a voxelized representation derived from the nuScenes dataset, requiring models to perform dense 3D prediction.
Quick Start & Requirements
getting_started
for details..npz
format for each frame.Highlighted Details
Maintenance & Community
contact@opendrivelab.com
.Licensing & Compatibility
Limitations & Caveats
The nuScenes dataset has known issues with z-axis translation, potentially affecting precise 6D localization and point cloud accumulation. Some data exhibits ground stratification. The evaluation uses a mask_camera
to exclude voxels not visible to cameras.
2 years ago
1 week