[1]GONG Yan,WANG Naibang,ZHANG Xinyu,et al.BEV perception technologies and development trends for intelligent connected vehicles[J].CAAI Transactions on Intelligent Systems,2026,21(1):41-59.[doi:10.11992/tis.202505027]
Copy
CAAI Transactions on Intelligent Systems[ISSN 1673-4785/CN 23-1538/TP] Volume:
21
Number of periods:
2026 1
Page number:
41-59
Column:
综述
Public date:
2026-03-05
- Title:
-
BEV perception technologies and development trends for intelligent connected vehicles
- Author(s):
-
GONG Yan1; 2; 3; WANG Naibang1; 2; ZHANG Xinyu1; 2; SU Nayu1; 2; 4; ZHAO Hongfei1; 2; YUAN Yun1; 2; LU Jianli1; 2; HU Xiaoxi1; 2; LIU Huaping5
-
1. State Key Laboratory of Intelligent Green Vehicle and Mobility, Tsinghua University, Beijing 100084, China;
2. School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China;
3. the State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin 150001, China;
4. Institute of Electrical Engineering, Yanshan University, Qinhuangdao 066004, China;
5. Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China
-
- Keywords:
-
intelligent connected vehicles; vehicle-infrastructure cooperation; cooperative perception; BEV; autonomous driving; dataset; vehicle-to-everything (V2X); multimodal fusion
- CLC:
-
TP391.41;U463.6;U495
- DOI:
-
10.11992/tis.202505027
- Abstract:
-
Bird’s eye view (BEV) perception has become a fundamental technique for environmental understanding in autonomous driving, due to its unified and interpretable spatial representation. This survey provides a comprehensive review of BEV perception technologies tailored for intelligent connected vehicles. It systematically categorizes existing approaches based on sensor modality and deployment configuration, covering vehicle-side, infrastructure-side, and vehicle-infrastructure cooperative scenarios. The review introduces a multi-dimensional framework encompassing vision-only, LiDAR-only, and multimodal fusion methods, and analyzes representative techniques in terms of their design principles and implementation strategies. In addition, this work presents the first consolidated comparison of BEV-related datasets, detailing their sensor setups, task types, and annotation schemes to support standardized benchmarking. Finally, the survey outlines key challenges—such as open-category recognition, unsupervised learning from large-scale data, and robustness under sensor uncertainty—and explores future directions involving end-to-end autonomous driving, embodied intelligence, and large-model-based cooperative BEV perception systems.