RT-BEV: Enhancing Real-Time BEV Perception for Autonomous Vehicles

  • Liangkai Liu
  • , Jinkyu Lee
  • , Kang G. Shin

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

4 Scopus citations

Abstract

Vision-centric Bird's Eye View (BEV) perception has become popular for enhancing the situational awareness of autonomous vehicles (AVs). It uses multiple cameras to create a 360° view, capturing essential details for the vehicle's navigation and decision-making. However, reducing the end-to-end (e2e) BEV perception latency without sacrificing accuracy is challenging due to the lack of co-optimization of message communication and object detection. Prior work either compresses the dense detection model to reduce computation which can hurt accuracy and assume images are well synchronized, or focuses on worstcase communication delay without considering the characteristics of object detection. To meet this challenge, we propose RT-BEV, the first frame-work designed to co-optimize message communication and object detection to improve real-time e2e BEV perception without sacrificing accuracy. The main insight of RT-BEV lies in generating traffic environment- and context-aware Regions of Interest (ROIs) for AV safety, combined with ROI-aware message communication. RT-BEV features an ROI-aware Camera Synchronizer that adaptively determines message groups and allowable delays based on ROIs' coverage. We also develop a ROIs Generator to model context-aware ROIs and a Feature Split & Merge component to handle variable-sized ROIs effectively. Furthermore, a Time Predictor forecasts timelines for processing ROIs, and a Coordinator jointly optimizes latency and accuracy for the entire e2e pipeline. We have implemented RT-BEV in a ROS-based BEV perception pipeline and evaluated it with the nuScenes dataset. RT-BEV is shown to significantly enhances real-time BEV perception, reducing average e2e latency by 1.5 ×, maintaining high mean Average Precision (mAP), doubling the number of processed frames, and improving the frame efficiency score (FES) by 2.9 × compared to the existing approaches. Moreover, RT-BEV is shown to reduce the worst-case e2e latency by 19.3 ×.

Original languageEnglish
Title of host publicationProceedings - 2024 IEEE Real-Time Systems Symposium, RTSS 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages267-279
Number of pages13
ISBN (Electronic)9798331540265
DOIs
StatePublished - 2024
Event45th IEEE Real-Time Systems Symposium, RTSS 2024 - York, United Kingdom
Duration: 10 Dec 202413 Dec 2024

Publication series

NameProceedings - Real-Time Systems Symposium
ISSN (Print)1052-8725

Conference

Conference45th IEEE Real-Time Systems Symposium, RTSS 2024
Country/TerritoryUnited Kingdom
CityYork
Period10/12/2413/12/24

Keywords

  • BEV perception
  • region of interests (ROIs)

Fingerprint

Dive into the research topics of 'RT-BEV: Enhancing Real-Time BEV Perception for Autonomous Vehicles'. Together they form a unique fingerprint.

Cite this