Exporters From Japan
Wholesale exporters from Japan   Company Established 1983
CARVIEW
Select Language

Stablization from handheld captures


Prior 2D method (phase-based Eulerian) fails on a handheld-captured video with camera shake as it assumes stablized capture. Our method benefits from having a 3D representation and separates camear motion from scene motion.


Rendering from different poses


We provide motion-magnified rendering at Tracked Poses, which are estimated from the
shaky handheld capture, and Fixed Pose, which is a static viewpoint.


Frequency selection


We capture two tuning forks with different vibration frequnecies (Left: 64 Hz, Right: 128 Hz).
By temporally filtering the point embeddings, we can selectively amplify different frequencies.


Varying the magnification factor


We visualize the impact of varying the magnification factor.

Comparisons of different magnification strategies

Using Positional Encoding as Point Embedding Function

Position Shift predicts a 3D displacement for the input point before positional encoding, while Encoding Shift predicts a phase shift within each sine wave for the input point during positional encoding. We perform Linear Eulerian magnification by amplifying the temporal variations of the predicted shifts. Position Shift leads to false motions, while Encoding Shift reduces such artifacts.


Using Tri-Plane as Point Embedding Function

Linear - Tri-Plane applies linear Eulerian magnification on tri-plane features.
Phase - Tri-Plane applies phase-based Eulerian magnification on tri-plane features.
Linear - Tri-Plane causes clipped intensities, while Phase - Tri-Plane reduces such artifacts.


Comparisons to video-based magnifications

Observed are Blender renderings from scenes with subtle object motions. Ground Truth are Blender renderings where the true object motions are artificially amplified.
Linear - Video and Phase - Video are obtained by deploying 2D magnification methods on the non-magnified RGB videos rendered by NeRF. In general, 3D methods that perform magnification in the embedding space produce fewer artifacts compared to 2D methods. Furthermore, 2D methods requires rendering at a fixed viewpoint and, for every new rendering, re-running video magnification on the fly. In contrast, magnification on tri-planes is performed just once, and the motion-magnified tri-planes can then be used to render at new viewpoints.


BibTeX

@inproceedings{feng2023motionmag,
    author    = {Feng, Brandon Y. and AlZayer, Hadi and Rubinstein, Michael and Freeman, William T. and Huang, Jia-Bin},
    title     = {Visualizing Subtle Motions with Time-Varying Neural Fields},
    booktitle = {International Conference on Computer Vision (ICCV)},
    year      = {2023},
  }

Template from Nerfies