| CARVIEW |
Select Language
HTTP/2 200
server: GitHub.com
content-type: text/html; charset=utf-8
last-modified: Wed, 11 Sep 2024 19:44:45 GMT
access-control-allow-origin: *
strict-transport-security: max-age=31556952
etag: W/"66e1f32d-561e"
expires: Mon, 29 Dec 2025 12:20:52 GMT
cache-control: max-age=600
content-encoding: gzip
x-proxy-cache: MISS
x-github-request-id: 7066:318CF6:8B7B02:9CB1C1:69526FCC
accept-ranges: bytes
age: 0
date: Mon, 29 Dec 2025 12:10:52 GMT
via: 1.1 varnish
x-served-by: cache-bom-vanm7210026-BOM
x-cache: MISS
x-cache-hits: 0
x-timer: S1767010252.111246,VS0,VE217
vary: Accept-Encoding
x-fastly-request-id: 7ef04ac7ba691a2e9ae85fa3c92633c1e76fbc11
content-length: 4773
DeformGS: Scene Flow in Highly Deformable Scenes for Deformable Object Manipulation
DeformGS: Scene Flow in Highly Deformable Scenes for Deformable Object Manipulation
Bardienus P. Duisterhof1,
Zhao Mandi2,
Yunchao Yao1,
Jia-Wei Liu4
Jenny Seidenschwarz1,5 Mike Zheng Shou4, Deva Ramanan1, Shuran Song2
Stan Birchfield3, Bowen Wen3, Jeffrey Ichnowski1
Jenny Seidenschwarz1,5 Mike Zheng Shou4, Deva Ramanan1, Shuran Song2
Stan Birchfield3, Bowen Wen3, Jeffrey Ichnowski1
1Carnegie Mellon University,
2Stanford University,
3NVIDIA,
4National University of Singapore
5Technical University of Munich
DeformGS was formerly known as MD-Splatting.
Abstract
Teaching robots to fold, drape, or reposition deformable objects such as cloth will unlock a variety of automation applications.
While remarkable progress has been made for rigid object manipulation, manipulating deformable objects poses unique challenges, including
frequent occlusions, infinite-dimensional state spaces and complex dynamics. Just as object pose estimation and tracking have aided robots
for rigid manipulation, dense 3D tracking (scene flow) of highly deformable objects will enable new applications in robotics while aiding existing approaches,
such as imitation learning or creating digital twins with real2sim transfer. We propose DeformGS, an approach to recover scene flow in highly deformable scenes,
using simultaneous video captures of a dynamic scene from multiple cameras. DeformGS builds on recent advances in Gaussian splatting, a method that learns
the properties of a large number of Gaussians for state-of-the-art and fast novel-view synthesis. DeformGS learns a deformation function to project a set
of Gaussians with canonical properties into world space. The deformation function uses a neural-voxel encoding and a multilayer perceptron (MLP) to infer Gaussian position,
rotation, and a shadow scalar. We enforce physics-inspired regularization terms based on conservation of momentum and isometry, which leads to trajectories with smaller
trajectory errors. We also leverage existing foundation models SAM and XMEM to produce noisy masks, and learn a per-Gaussian mask for better physics-inspired regularization.
DeformGS achieves high-quality 3D tracking on highly deformable scenes with shadows and occlusions. In experiments, DeformGS improves 3D
tracking by an average of 55.8% compared to the state-of-the-art. With sufficient texture, DeformGS achieves a median tracking error of 3.3 mm on a cloth of 1.5 × 1.5 m in area.
Method
DeformGS maps a set of Gausians with canonical properties to world space using a deformation function F. The deformation function takes in the position of a Gaussian x and a queried timestamp t, to infer shadow s, rotation R' and position x'. We use the positions to regularize the deformation function to incentivize more physically plausbile trajectories, and learn a per-Gaussian mask to improve the regularization.
Visual Comparisons
DeformGS
4DGS [Wu et al. 2024]
DeformGS
4DGS [Wu et al. 2024]
DeformGS
4DGS [Wu et al. 2024]
DeformGS
4DGS [Wu et al. 2024]
Quantitative Evaluation
@article{deformgs,
title={{DeformGS}: Scene Flow in Highly Deformable Scenes for Deformable Object Manipulation,
author={Duisterhof, Bardienus P and Mandi, Zhao and Yao, Yunchao and Liu, Jia-Wei and Seidenschwarz, Jenny and Shou, Mike Zheng and Ramanan Deva and Song, Shuran and Birchfield, Stan and Wen, Bowen and Ichnowski, Jeffrey},
journal={WAFR},
year={2024}
}