| CARVIEW |
Select Language
HTTP/2 200
server: GitHub.com
content-type: text/html; charset=utf-8
last-modified: Mon, 28 Aug 2023 06:39:49 GMT
access-control-allow-origin: *
strict-transport-security: max-age=31556952
etag: W/"64ec4135-4942"
expires: Tue, 30 Dec 2025 00:03:17 GMT
cache-control: max-age=600
content-encoding: gzip
x-proxy-cache: MISS
x-github-request-id: FCD3:3ABDEF:957B06:A7F303:6953146D
accept-ranges: bytes
age: 0
date: Mon, 29 Dec 2025 23:53:17 GMT
via: 1.1 varnish
x-served-by: cache-bom-vanm7210028-BOM
x-cache: MISS
x-cache-hits: 0
x-timer: S1767052398.772508,VS0,VE209
vary: Accept-Encoding
x-fastly-request-id: adb830595bba3241e906eb097bcef449650d9ea8
content-length: 5242
MapPrior
MapPrior: Bird's-Eye View Map Layout Estimation
ICCV2023
MapPrior: Bird's-Eye View Map Layout Estimation
with Generative Models
ICCV2023
- 1University of Illinois at Urbana-Champaign
- 2Massachusetts Institute of Technology
Abstract
Despite tremendous advancements in bird's-eye view (BEV) perception,
existing models fall short in generating realistic and coherent semantic map layouts,
and they fail to account for uncertainties arising from partial sensor information
(such as occlusion or limited coverage). In this work, we introduce MapPrior,
a novel BEV perception framework that combines a traditional
discriminative BEV perception model with a learned generative model for semantic map layouts.
MapPrior delivers predictions with better accuracy, realism and uncertainty awareness.
We evaluate our model on the large-scale nuScenes benchmark. At the time of submission, MapPrior outperforms the strongest competing method, with significantly improved MMD and ECE scores in camera- and LiDAR-based BEV perception. Furthermore, our method can be used to perpetually generate layouts with unconditional sampling.
We evaluate our model on the large-scale nuScenes benchmark. At the time of submission, MapPrior outperforms the strongest competing method, with significantly improved MMD and ECE scores in camera- and LiDAR-based BEV perception. Furthermore, our method can be used to perpetually generate layouts with unconditional sampling.
Diversity Sampling
Our method can sample multiple results per input with diversity, providing better uncertainty awareness:
perpetual Generation
Our method can be exploited in a progressive manner to generate perpetual traffic layouts.
Map Estimation using Generative Models
MapPrior first makes use of an off-the-shelf perception model to generate an initial noisy estimate from the sensory
input, which uses monocular depth estimation to project camera features to BEV. It then encodes the noisy
estimate into a discrete latent code using a generative encoder and generates various samples through a transformer-based
controlled synthesis. Finally, MapPrior decodes these samples into outputs with a decoder
Quantitative Results
We show our quantitative metrics here. Our MapPrior achieves better accuracy (IoU), realism
(MMD) and uncertainty awareness (ECE) than discriminative BEV perception baselines.