--- license: artistic-2.0 tags: - visual-grounding - lidar - 3d --- # 3EED: Ground Everything Everywhere in 3D — Dataset Card A cross-platform, multi-modal 3D visual grounding dataset spanning **vehicle**, **drone**, and **quadruped** platforms, with synchronized **RGB**, **LiDAR**, and **language** annotations. This page documents how to obtain and organize the dataset from HuggingFace and how to connect it with the training/evaluation code in the 3EED repository. - Project Page: https://3eed.github.io - Code (Baselines & Evaluation): https://github.com/iris0329/3eed - Paper: https://arxiv.org/ (coming soon) ## 1. What’s Included - Platforms: `vehicle`, `drone`, `quad` (quadruped) - Modalities: LiDAR point clouds, RGB images, language referring expressions, metadata - Splits: train/val files per platform under `splits/` - Task: 3D visual grounding (language → 3D box) ## 2. Download You can download via: - HuggingFace CLI: ```bash pip install -U "huggingface_hub[cli]" huggingface-cli download 3EED/3EED --repo-type dataset --local-dir ./3eed_dataset ```` - Python: ```python from huggingface_hub import snapshot_download snapshot_download(repo_id="3EED/3EED", repo_type="dataset", local_dir="./3eed_dataset") ``` - Git (LFS): ```bash git lfs install git clone https://huggingface.co/datasets/3EED/3EED 3eed_dataset ``` ## 3. Directory Structure - Place or verify the files under `data/3eed/` in your project. A minimal expected layout (paths shown relative to the repo root): ``` data/3eed/ ├── drone/ # Drone platform data │ ├── scene-0001/ │ │ ├── 0000_0/ │ │ │ ├── image.jpg │ │ │ ├── lidar.bin │ │ │ └── meta_info.json │ │ └── ... │ └── ... ├── quad/ # Quadruped platform data │ ├── scene-0001/ │ └── ... ├── waymo/ # Vehicle platform data │ ├── scene-0001/ │ └── ... └── splits/ # Train/val split files ├── drone_train.txt ├── drone_val.txt ├── quad_train.txt ├── quad_val.txt ├── waymo_train.txt └── waymo_val.txt ``` ## 4. Connect to the Codebase - Clone the code repository: ```bash git clone https://github.com/iris0329/3eed cd 3eed ``` - Link or copy the downloaded dataset to `data/3eed/`: ```bash # Example: if your dataset is in ../3eed_dataset ln -s ../3eed_dataset data/3eed ``` Now you can follow the **Installation**, **Custom CUDA Operators**, **Training**, and **Evaluation** sections in the GitHub README: * Train on all platforms: ```bash bash scripts/train_3eed.sh ``` * Train on a single platform: ```bash bash scripts/train_waymo.sh # vehicle bash scripts/train_drone.sh # drone bash scripts/train_quad.sh # quadruped ``` * Evaluate: ```bash bash scripts/val_3eed.sh bash scripts/val_waymo.sh bash scripts/val_drone.sh bash scripts/val_quad.sh ``` Remember to set the correct `--checkpoint_path` inside the evaluation scripts. ## 5. Data Splits We provide official splits under `data/3eed/splits/`: * `*_train.txt`: training scene/frame indices for each platform * `*_val.txt`: validation scene/frame indices for each platform Please keep these files unchanged for fair comparison with the baselines and reported results. ## 6. Usage Tips * Storage: LiDAR+RGB data can be large; ensure sufficient disk space and use Git LFS for partial sync if needed. * IO Throughput: For faster training/evaluation, place frequently used scenes on fast local SSDs or use caching. * Reproducibility: Use the exact environment files and scripts from the code repo; platform unions vs. single-platform runs are controlled by the provided scripts. ## 7. License * Dataset license: **Apache-2.0** (see the header of this page). * The **code repository** uses **Apache-2.0**; refer to the LICENSE in the GitHub repo. If you plan to use, redistribute, or modify the dataset, please review the dataset license and any upstream source licenses (e.g., Waymo Open Dataset, M3ED). ## 8. Citation - If you find 3EED helpful, please cite: ```bibtex @inproceedings{li2025_3eed, title = {3EED: Ground Everything Everywhere in 3D}, author = {Rong Li and Yuhao Dong and Tianshuai Hu and Ao Liang and Youquan Liu and Dongyue Lu and Liang Pan and Lingdong Kong and Junwei Liang and Ziwei Liu}, booktitle = {Advances in Neural Information Processing Systems (NeurIPS) Datasets and Benchmarks Track}, year = {2025} } ``` ## 9. Acknowledgements We acknowledge the following upstream sources which make this dataset possible: * Waymo Open Dataset (vehicle platform) * M3ED (drone and quadruped platforms) For baseline implementations and evaluation code, please refer to the GitHub repository.