Distance-Aware Tri-Perspective View for Efficient 3D Perception in Autonomous Driving
| dc.contributor.author | Tang, Yutao | en_US |
| dc.contributor.author | Zhao, Jigang | en_US |
| dc.contributor.author | Qin, Zhengrui | en_US |
| dc.contributor.author | Qiu, Rui | en_US |
| dc.contributor.author | Zhao, Lingying | en_US |
| dc.contributor.author | Ren, Jie | en_US |
| dc.contributor.author | Chen, Guangxi | en_US |
| dc.contributor.editor | Christie, Marc | en_US |
| dc.contributor.editor | Han, Ping-Hsuan | en_US |
| dc.contributor.editor | Lin, Shih-Syun | en_US |
| dc.contributor.editor | Pietroni, Nico | en_US |
| dc.contributor.editor | Schneider, Teseo | en_US |
| dc.contributor.editor | Tsai, Hsin-Ruey | en_US |
| dc.contributor.editor | Wang, Yu-Shuen | en_US |
| dc.contributor.editor | Zhang, Eugene | en_US |
| dc.date.accessioned | 2025-10-07T06:04:01Z | |
| dc.date.available | 2025-10-07T06:04:01Z | |
| dc.date.issued | 2025 | |
| dc.description.abstract | Three-dimensional environmental perception remains a critical bottleneck in autonomous driving, where existing vision-based dense representations face an intractable trade-off between spatial resolution and computational complexity. Current methods, including Bird's Eye View (BEV) and Tri-Perspective View (TPV), apply uniform perception precision across all spatial regions, disregarding the fundamental safety principle that near-field objects demand high-precision detection for collision avoidance while distant objects permit lower initial accuracy. This uniform treatment squanders computational resources and constrains real-time deployment. We introduce Distance-Aware Tri-Perspective View (DA-TPV), a novel framework that allocates computational resources proportional to operational risk. DA-TPV employs a hierarchical dual-plane architecture for each viewing direction: low-resolution planes capture global scene context while high-resolution planes deliver fine-grained perception within safety-critical reaction zones. Through distance-adaptive feature fusion, our method dynamically concentrates processing power where it most directly impacts vehicle safety. Extensive experiments on nuScenes demonstrate that DA-TPV matches or exceeds single high-resolution TPV performance while reducing memory consumption by 26.3% and achieving real-time inference. This work establishes distance-aware perception as a practical paradigm for deploying sophisticated three-dimensional understanding within automotive computational constraints. Code is available at https://github.com/yytang2012/DA-TPVFormer. | en_US |
| dc.description.sectionheaders | Point Clouds & Gaussian Splatting | |
| dc.description.seriesinformation | Pacific Graphics Conference Papers, Posters, and Demos | |
| dc.identifier.doi | 10.2312/pg.20251285 | |
| dc.identifier.isbn | 978-3-03868-295-0 | |
| dc.identifier.pages | 10 pages | |
| dc.identifier.uri | https://doi.org/10.2312/pg.20251285 | |
| dc.identifier.uri | https://diglib.eg.org/handle/10.2312/pg20251285 | |
| dc.publisher | The Eurographics Association | en_US |
| dc.rights | Attribution 4.0 International License | |
| dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | |
| dc.subject | CCS Concepts: Computing methodologies → Computer vision; Artificial intelligence | |
| dc.subject | Computing methodologies → Computer vision | |
| dc.subject | Artificial intelligence | |
| dc.title | Distance-Aware Tri-Perspective View for Efficient 3D Perception in Autonomous Driving | en_US |
Files
Original bundle
1 - 1 of 1