uniad

Code exploration of UniAD

https://github.com/baophann/uniad

Science Score: 54.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
    Found CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
  • Academic publication links
    Links to: arxiv.org
  • Academic email domains
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (10.0%) to scientific vocabulary
Last synced: 6 months ago · JSON representation ·

Repository

Code exploration of UniAD

Basic Info
  • Host: GitHub
  • Owner: BaophanN
  • License: apache-2.0
  • Language: Python
  • Default Branch: main
  • Size: 9.59 MB
Statistics
  • Stars: 0
  • Watchers: 1
  • Forks: 0
  • Open Issues: 0
  • Releases: 0
Created about 1 year ago · Last pushed 12 months ago
Metadata Files
Readme Funding License Code of conduct Citation

README.md

# Planning-oriented Autonomous Driving

arXiv | Video | Poster | Slides

https://github.com/OpenDriveLab/UniAD/assets/48089846/bcf685e4-2471-450e-8b77-e028a46bd0f7



teaser

Table of Contents:

  1. Highlights
  2. News
  3. Getting Started
  4. Results and Models
  5. TODO List
  6. License
  7. Citation
  8. 🔥 See Also: GenAD & Vista

Highlights

  • :oncoming_automobile: Planning-oriented philosophy: UniAD is a Unified Autonomous Driving algorithm framework following a planning-oriented philosophy. Instead of standalone modular design and multi-task learning, we cast a series of tasks, including perception, prediction and planning tasks hierarchically.
  • :trophy: SOTA performance: All tasks within UniAD achieve SOTA performance, especially prediction and planning (motion: 0.71m minADE, occ: 63.4% IoU, planning: 0.31% avg.Col)

News

  • Paper Title Change: To avoid confusion with the "goal-point" navigation in Robotics, we change the title from "Goal-oriented" to "Planning-oriented" suggested by Reviewers. Thank you!

  • Planning Metric: Discussion [Ref: https://github.com/OpenDriveLab/UniAD/issues/29]: Clarification and Notice regarding open-loop planning results comparison.

  • 2024/08/27 New feature: Implementation for CARLA and closed-loop evaluation on CARLA Leaderboard 2.0 scenarios are available in Bench2Drive.

  • 2023/08/03 Bugfix [Commit]: Previously, the visualized planning results were in opposition on the x axis, compared to the ground truth. Now it's fixed.

  • 2023/06/12 Bugfix [Ref: https://github.com/OpenDriveLab/UniAD/issues/21]: Previously, the performance of the stage1 model (trackmap) could not be replicated when trained from scratch, due to mistakenly adding `losspasttrajand freezingimgneckandBN. By removinglosspasttrajand unfreezingimg_neckandBN` in training, the reported results could be reproduced (AMOTA: 0.393, stage1trainlog).

  • 2023/04/18 New feature: You can replace BEVFormer with other BEV Encoding methods, e.g., LSS, as long as you provide the bev_embed and bev_pos in track_train and track_inference. Make sure your bevs and ours are of the same shape.

  • 2023/04/18 Base-model checkpoints are released.

  • 2023/03/29 Code & model initial release v1.0.

  • 2023/03/21 🌟🌟 UniAD is accepted by CVPR 2023, as an Award Candidate (12 out of 2360 accepted papers)!

  • 2022/12/21 UniAD paper is available on arXiv.

Getting Started

Results and Pre-trained Models

UniAD is trained in two stages. Pretrained checkpoints of both stages will be released and the results of each model are listed in the following tables.

Stage1: Perception training

We first train the perception modules (i.e., track and map) to obtain a stable weight initlization for the next stage. BEV features are aggregated with 5 frames (queue_length = 5).

| Method | Encoder | Tracking
AMOTA | Mapping
IoU-lane | config | Download | | :---: | :---: | :---: | :---: | :---:|:---:| | UniAD-B | R101 | 0.390 | 0.297 | base-stage1 | base-stage1 |

Stage2: End-to-end training

We optimize all task modules together, including track, map, motion, occupancy and planning. BEV features are aggregated with 3 frames (queue_length = 3).

| Method | Encoder | Tracking
AMOTA | Mapping
IoU-lane | Motion
minADE |Occupancy
IoU-n. | Planning
avg.Col. | config | Download | | :---: | :---: | :---: | :---: | :---:|:---:| :---: | :---: | :---: | | UniAD-B | R101 | 0.363 | 0.313 | 0.705 | 63.7 | 0.29 | base-stage2 | base-stage2 |

Checkpoint Usage

  • Download the checkpoints you need into UniAD/ckpts/ directory.
  • You can evaluate these checkpoints to reproduce the results, following the evaluation section in TRAIN_EVAL.md.
  • You can also initialize your own model with the provided weights. Change the load_from field to path/of/ckpt in the config and follow the train section in TRAIN_EVAL.md to start training.

Model Structure

The overall pipeline of UniAD is controlled by uniad_e2e.py which coordinates all the task modules in UniAD/projects/mmdet3d_plugin/uniad/dense_heads. If you are interested in the implementation of a specific task module, please refer to its corresponding file, e.g., motion_head.

License

All assets and code are under the Apache 2.0 license unless specified otherwise.

Citation

If you find our project useful for your research, please consider citing our paper and codebase with the following BibTeX:

bibtex @inproceedings{hu2023_uniad, title={Planning-oriented Autonomous Driving}, author={Yihan Hu and Jiazhi Yang and Li Chen and Keyu Li and Chonghao Sima and Xizhou Zhu and Siqi Chai and Senyao Du and Tianwei Lin and Wenhai Wang and Lewei Lu and Xiaosong Jia and Qiang Liu and Jifeng Dai and Yu Qiao and Hongyang Li}, booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition}, year={2023}, }

bibtex @misc{contributors2023_uniadrepo, title={Planning-oriented Autonomous Driving}, author={UniAD contributors}, howpublished={\url{https://github.com/OpenDriveLab/UniAD}}, year={2023} }

Related Resources

Awesome - BEVFormer (Ours!) - ST-P3 (Ours!) - FIERY - MOTR

🔥 See Also

We are thrilled to launch our recent line of works: GenAD and Vista, to advance driving world models with the largest driving video dataset collected from the web - OpenDV.

GenAD: Generalized Predictive Model for Autonomous Driving (CVPR'24, Highlight ⭐)

Vista: A Generalizable Driving World Model with High Fidelity and Versatile Controllability 🌏

Owner

  • Name: Bao Phan
  • Login: BaophanN
  • Kind: user

Where there's a will, there's a way.

Citation (CITATION.cff)

cff-version: 1.2.0
message: "If you use this software, please cite it as below."
authors:
  - name: "UniAD Contributors"
title: "Planning-oriented Autonomous Driving"
date-released: 2023-03-26
url: "https://github.com/OpenDriveLab/UniAD"
license: Apache-2.0

GitHub Events

Total
  • Push event: 5
  • Create event: 2
Last Year
  • Push event: 5
  • Create event: 2