| CARVIEW |
Select Language
HTTP/2 200
server: GitHub.com
content-type: text/html; charset=utf-8
last-modified: Tue, 30 Jul 2024 21:28:21 GMT
access-control-allow-origin: *
strict-transport-security: max-age=31556952
etag: W/"66a95af5-5c10"
expires: Mon, 29 Dec 2025 02:23:39 GMT
cache-control: max-age=600
content-encoding: gzip
x-proxy-cache: MISS
x-github-request-id: 69FA:3655F2:8323AF:934EF5:6951E3D3
accept-ranges: bytes
age: 0
date: Mon, 29 Dec 2025 02:13:40 GMT
via: 1.1 varnish
x-served-by: cache-bom-vanm7210034-BOM
x-cache: MISS
x-cache-hits: 0
x-timer: S1766974420.800585,VS0,VE215
vary: Accept-Encoding
x-fastly-request-id: c9435c75b788944dd712a134ac2ff7a435a9a0e2
content-length: 4955
3D-LFM: Lifting Foundation Model
CVPR, 2024
3D-LFM: Lifting Foundation Model
1Carnegie Mellon University,
2The University of Adelaide
CVPR, 2024
Teaser
3D pose estimation on random videos from the internet using a single model for all the following categories.
We show 3D pose estimation on various deformable categories from the internet, including OpenAI's SORA videos. No camera information is required, allowing 3D-LFM to work out of the box on many of these categories, provided 2D landmarks are available (in any order or joint connectivity).
Unified 3D-LFM Model
A single model for 30+ object categories.
The 3D-LFM scales to multiple categories (30+ in our experiments), managing diverse landmark configurations through proposed architectural changes. See our paper for more details. Key: Red for ground truth, blue for predictions.