You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
International Conference on Intelligent Robots and Systems (IROS) 2024Oral!
This repo houses code and data for our work in Touch-GS.
Quick Start and Setup
The pipeline has been tested on Ubuntu 22.04.
Requirements:
CUDA 11+
Python 3.8+
Conda or Mamba (optional)
Dependencies (from Nerfstudio)
Install PyTorch with CUDA (this repo has been tested with CUDA 11.7 and CUDA 11.8) and tiny-cuda-nn.
cuda-toolkit is required for building tiny-cuda-nn.
We have implemented our own GPIS (Gaussian Process Implicit Surface) from scratch here!
Please follow the steps to install the repo there.
Install Our Version of Nerfstudio
You can find more detailed instructions in Nerfstudio's README.
# install nerfstudio
bash install_ns.sh
Getting Touch-GS Data Setup and Training
We have made an end-to-end pipeline that will take care of setting up the data, training, and evaluating our method. Note that we will release the code for running the ablations (which includes the baselines) soon!
To prepare each scene:
Install Python packages
pip install -r requirements.txt
Real Bunny Scene (our method)
bash scripts/train_bunny_real.sh
Mirror Scene
bash scripts/train_mirror_data.sh
Prism Scene
bash scripts/train_block_data.sh
Bunny Blender Scene (processing to a unified format in progress)
Get Renders
The renders on the test set are created in the main script under exp-name-renders within touch-gs-data.
Get Rendered Video
You can get rendered videos with a custom camera path detailed here. This is how we were able to get our videos on our website.
About
[IROS 2024] Touch-GS: Visual-Tactile Supervised 3D Gaussian Splatting