birds-eye-view-perception
[IEEE T-PAMI 2023] Awesome BEV perception research and cookbook for all level audience in autonomous diriving
Science Score: 77.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
✓CITATION.cff file
Found CITATION.cff file -
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
✓DOI references
Found 1 DOI reference(s) in README -
✓Academic publication links
Links to: arxiv.org, ieee.org -
✓Committers with academic emails
3 of 12 committers (25.0%) from academic institutions -
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (12.0%) to scientific vocabulary
Keywords
Repository
[IEEE T-PAMI 2023] Awesome BEV perception research and cookbook for all level audience in autonomous diriving
Basic Info
- Host: GitHub
- Owner: OpenDriveLab
- License: apache-2.0
- Language: Python
- Default Branch: master
- Homepage: https://doi.org/10.1109/TPAMI.2023.3333838
- Size: 24.3 MB
Statistics
- Stars: 1,314
- Watchers: 36
- Forks: 115
- Open Issues: 3
- Releases: 1
Topics
Metadata Files
README.md
[!IMPORTANT] 🌟 Stay up to date at opendrivelab.com!
Bird's-eye-view (BEV) Perception: A Survey and Collection
Awesome BEV perception papers and toolbox for achieving state-of-the-arts performance.
Table of Contents
Introduction
This repo is associated with the survey paper "Delving into the Devils of Bird’s-eye-view Perception: A Review, Evaluation and Recipe", which provides an up-to-date literature survey for BEV perception and an open-source BEV toolbox based on PyTorch. We also introduce the BEV algorithm family, including follow-up work on BEV percepton such as VCD, GAPretrain, and FocalDistiller. We hope this repo can not only be a good starting point for new beginners but also help current researchers in the BEV perception community. <!-- In the literature survey, it includes different modalities (camera, lidar and fusion) and tasks (detection and segmentation). As for the toolbox, it provides useful recipe for BEV camera-based 3D object detection, including solid data augmentation strategies, efficient BEV encoder design, loss function family, useful test-time augmentation, ensemble policy, and so on. -->
If you find some work popular enough to be cited below, email us or simply open a PR!
Major Features
Up-to-date Literature Survey for BEV Perception
We summarize important methods in recent years about BEV perception, including different modalities (camera, LIDAR, Fusion) and tasks (Detection, Segmentation, Occupancy). More details of the survey paper list can be found here.Convenient BEVPerception Toolbox
We integrate a bag of tricks in the BEV toolbox that helps us achieve 1st in the camera-based detection track of the Waymo Open Challenge 2022, which can be used independently or as a plug-in for popular deep learning libraries. Moreover, we provide a suitable playground for beginners in this area, including a hands-on tutorial and a small-scale dataset (1/5 WOD in kitti format) to validate ideas. More details can be found here.SOTA BEV Knowledge Distillation Algorithms
We include important follow-up works of BEVFormer/BEVDepth/SOLOFusion in the perspective of knowledge distillation(VCD, GAPretrain, FocalDistiller). More details of each paper can be found in each README.md file under here.
| Method | Expert | Apprentice | | :---: | :---: | :---: | | VCD | Vision-centric multi-modal detector | Camera-only detector | | GAPretrain | Lidar-only detector | Camera-only detector | | FocalDistiller | Camera-only detector | Camera-only detector |
What's New
2023/11/04 Our Survey is accepted by IEEE T-PAMI.
2023/10/26 A new paper VCD is coming soon with official implementation.
2023/09/06 We have a new version of the survey. Check it out!
2023/04/06 Two new papers GAPretrain and FocalDistiller are coming soon with official implementation.
2022/10/13 v0.1 was released.
* Integrate some practical data augmentation methods for BEV camera-based 3D detection in the toolbox.
* Offer a pipeline to process the Waymo dataset (camera-based 3D detection).
* Release a baseline (with config) for the Waymo dataset and also 1/5 of the Waymo dataset in Kitti format.
Literature Survey
The general picture of BEV perception at a glance, where consists of three sub-parts based on the input modality. BEV perception is a general task built on top of a series of fundamental tasks. For better completeness of the whole perception algorithms in autonomous driving, we list other topics as well. More details can be found in the survey paper.
We have summarized important datasets and methods in recent years about BEV perception in academia and also different roadmaps used in industry. * Academic Summary of BEV Perception * Datasets for BEV Perception * BEV Camera * BEV Lidar * BEV Fusion * Industrial Roadmap of BEV Perception
We have also summarized some conventional methods for different tasks. * Conventional Methods Camera 3D Object Detection * Conventional Methods LiDAR Detection * Conventional Methods LiDAR Segmentation * Conventional Methods Sensor Fusion
BEV Toolbox
The BEV toolbox provides useful recipes for BEV camera-based 3D object detection, including solid data augmentation strategies, efficient BEV encoder design, loss function family, useful test-time augmentation, ensemble policy, and so on. Please refer to bev_toolbox/README.md for more details.
BEV Knowledge Distillation Algorithms
The BEV algorithm family includes follow-up works of BEVFormer in different aspects, ranging from plug-and-play tricks to pre-training distillation. All paper summary is under nuscenes_playground along with official implementation, check it out!
* VCD (NeurIPS 2023)
Leveraging Vision-Centric Multi-Modal Expertise for 3D Object Detection. More details can be found in nuScenes_playground/VCD.
* GAPretrain (arXiv)
Geometric-aware Pretraining for Vision-centric 3D Object Detection. More details can be found in nuScenes_playground/GAPretrain.md.
* FocalDistiller (CVPR 2023)
Distilling Focal Knowledge from Imperfect Expert for 3D object Detection. More details can be found in nuScenes_playground/FocalDistiller.
License and Citation
This project is released under the Apache 2.0 license.
If you find this project useful in your research, please consider cite:
BibTeX
@article{li2022bevsurvey,
author={Li, Hongyang and Sima, Chonghao and Dai, Jifeng and Wang, Wenhai and Lu, Lewei and Wang, Huijie and Zeng, Jia and Li, Zhiqi and Yang, Jiazhi and Deng, Hanming and Tian, Hao and Xie, Enze and Xie, Jiangwei and Chen, Li and Li, Tianyu and Li, Yang and Gao, Yulu and Jia, Xiaosong and Liu, Si and Shi, Jianping and Lin, Dahua and Qiao, Yu},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
title={Delving Into the Devils of Bird's-Eye-View Perception: A Review, Evaluation and Recipe},
year={2023},
volume={},
number={},
pages={1-20},
doi={10.1109/TPAMI.2023.3333838}
}
BibTeX
@misc{bevtoolbox2022,
title={{BEVPerceptionx-Survey-Recipe} toolbox for general BEV perception},
author={BEV-Toolbox Contributors},
howpublished={\url{https://github.com/OpenDriveLab/Birds-eye-view-Perception}},
year={2022}
}
Owner
- Name: OpenDriveLab
- Login: OpenDriveLab
- Kind: organization
- Email: contact@opendrivelab.com
- Location: Hong Kong
- Website: https://opendrivelab.com
- Twitter: OpenDriveLab
- Repositories: 2
- Profile: https://github.com/OpenDriveLab
AI for Robotics and Autonomous Driving, affiliated at The University of Hong Kong (HKU).
Citation (CITATION.cff)
cff-version: 1.2.0 message: "If you use this software, please cite it as below." authors: - name: "BEVPerception-Survey-Recipe Contributors" title: "BEVPerception-Survey-Recipe: Awesome BEV perception papers and toolbox for achieving SOTA results" date-released: 2022-10-14 url: "https://github.com/OpenPerceptionX/BEVPerception-Survey-Recipe" license: Apache-2.0
GitHub Events
Total
- Issues event: 1
- Watch event: 122
- Issue comment event: 2
- Push event: 1
- Pull request event: 1
- Fork event: 15
Last Year
- Issues event: 1
- Watch event: 122
- Issue comment event: 2
- Push event: 1
- Pull request event: 1
- Fork event: 15
Committers
Last synced: 10 months ago
Top Committers
| Name | Commits | |
|---|---|---|
| Chonghao Sima | s****c@p****u | 44 |
| cyberknight42 | x****w@s****n | 38 |
| linyan | l****u@1****m | 19 |
| Cyberpunk2077 | 1****2 | 14 |
| increase24 | 4****4@q****m | 11 |
| Linyan Huang | 8****I | 10 |
| simachonghao | P****o@s****g | 7 |
| Hongyang Li | h****0@g****m | 6 |
| faikit | h****g@h****m | 4 |
| lijiawei | l****i | 3 |
| EloiZ | E****Z | 2 |
| junwang | j****g@u****u | 1 |
Committer Domains (Top 20 + Academic)
Issues and Pull Requests
Last synced: 9 months ago
All Time
- Total issues: 19
- Total pull requests: 12
- Average time to close issues: about 2 months
- Average time to close pull requests: about 3 hours
- Total issue authors: 15
- Total pull request authors: 5
- Average comments per issue: 2.32
- Average comments per pull request: 0.0
- Merged pull requests: 12
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 1
- Pull requests: 0
- Average time to close issues: N/A
- Average time to close pull requests: N/A
- Issue authors: 1
- Pull request authors: 0
- Average comments per issue: 0.0
- Average comments per pull request: 0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0
Top Authors
Issue Authors
- exiawsh (2)
- YongtaoGe (2)
- linqx7 (2)
- 1146065545 (2)
- northeastsquare (1)
- flyzenr (1)
- Debrove (1)
- c122-ode (1)
- geojed (1)
- Orange-rightperson (1)
- cuge1995 (1)
- MaekTec (1)
- synsin0 (1)
- kevintsq (1)
Pull Request Authors
- ChonghaoSima (4)
- cyberknight42 (4)
- EloiZ (2)
- Petros626 (1)
- faikit (1)
- HenryJunW (1)