| CARVIEW |
Select Language
HTTP/2 200
server: GitHub.com
content-type: text/html; charset=utf-8
last-modified: Tue, 16 Dec 2025 07:39:00 GMT
access-control-allow-origin: *
strict-transport-security: max-age=31556952
etag: W/"69410c94-39a8"
expires: Tue, 30 Dec 2025 00:30:22 GMT
cache-control: max-age=600
content-encoding: gzip
x-proxy-cache: MISS
x-github-request-id: E83F:36A0B4:968FB4:A91357:69531AC3
accept-ranges: bytes
age: 0
date: Tue, 30 Dec 2025 00:20:22 GMT
via: 1.1 varnish
x-served-by: cache-bom-vanm7210077-BOM
x-cache: MISS
x-cache-hits: 0
x-timer: S1767054022.215062,VS0,VE210
vary: Accept-Encoding
x-fastly-request-id: 39397eb21ff5b20e2610dac58278c790531506e3
content-length: 3573
Prior-Enhanced Gaussian Splatting for Dynamic Scene Reconstruction from Casual Video
Prior-Enhanced Gaussian Splatting for Dynamic Scene Reconstruction from Casual Video
Meng-Li Shih1
Ying-Huan Chen2
Yu-Lun (Alex) Liu2
Brian Curless1
1University of Washington 2National Yang Ming Chiao Tung University
SIGGRAPH ASIA 2025
Abstract
We introduce a fully automatic pipeline for dynamic scene reconstruction from casually captured monocular RGB videos. Rather than designing a new scene representation, we enhance the priors that drive Dynamic Gaussian Splatting. Video segmentation combined with epipolar-error maps yields object-level masks that closely follow thin structures; these masks (i) guide an object-depth loss that sharpens the consistent video depth, and (ii) support skeleton-based sampling plus mask-guided re-identification to produce reliable, comprehensive 2-D tracks. Two additional objectives embed the refined priors in the reconstruction stage: a virtual-view depth loss removes floaters, and a scaffold-projection loss ties motion nodes to the tracks, preserving fine geometry and coherent motion. The resulting system surpasses previous monocular dynamic scene reconstruction methods and delivers visibly superior renderings.Method Overview
Acknowledgements
This work was supported by Lenovo and the UW Reality Lab.
BibTex
@inproceedings{10.1145/3757377.3763910,
author = {Shih, Meng-Li and Chen, Ying-Huan and Liu, Yu-Lun and Curless, Brian},
title = {Prior-Enhanced Gaussian Splatting for Dynamic Scene Reconstruction from Casual Video},
year = {2025},
isbn = {9798400721373},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3757377.3763910},
doi = {10.1145/3757377.3763910},
abstract = {We introduce a fully automatic pipeline for dynamic scene reconstruction from casually captured monocular RGB videos. Rather than designing a new scene representation, we enhance the priors that drive Dynamic Gaussian Splatting. Video segmentation combined with epipolar-error maps yields object-level masks that closely follow thin structures; these masks (i) guide an object-depth loss that sharpens the consistent video depth, and (ii) support skeleton-based sampling plus mask-guided re-identification to produce reliable, comprehensive 2-D tracks. Two additional objectives embed the refined priors in the reconstruction stage: a virtual-view depth loss removes floaters, and a scaffold-projection loss ties motion nodes to the tracks, preserving fine geometry and coherent motion. The resulting system surpasses previous monocular dynamic scene reconstruction methods and delivers visibly superior renderings. Project page: https://priorenhancedgaussian.github.io/},
booktitle = {Proceedings of the SIGGRAPH Asia 2025 Conference Papers},
articleno = {74},
numpages = {13},
location = {
},
series = {SA Conference Papers '25}
}
Copyright © Meng-Li Shih 2025

