You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This work is based on our arXiv tech report. We proposed a novel deep net architecture for auto-encoding point clouds. The learned representations were amenable to semantic part editting, shape analogies, linear classification and shape interpolations.
Citation
If you find our work useful in your research, please consider citing:
@article{achlioptas2017latent_pc,
title={Learning Representations and Generative Models For 3D Point Clouds},
author={Achlioptas, Panos and Diamanti, Olga and Mitliagkas, Ioannis and Guibas, Leonidas J},
journal={arXiv preprint arXiv:1707.02392},
year={2017}
}
To be able to train your own model you need first to compile the EMD/Chamfer losses. In latent_3d_points/external/structural_losses we have included the cuda implementations of Fan et. al.
cd latent_3d_points/external
with your editor modify the first three lines of the makefile to point to
your nvcc, cudalib and tensorflow library.
make
Data Set
We provide ~57K point-clouds, each sampled from a mesh model of
ShapeNetCore
with (area) uniform sampling. To download them (1.4GB):
cd latent_3d_points/
./download_data.sh
The point-clouds will be stored in latent_3d_points/data/shape_net_core_uniform_samples_2048
Use the function snc_category_to_synth_id, defined in src/in_out/, to map a class name such as "chair" to its synthetic_id: "03001627". Point-clouds of models of the same class are stored under a commonly named folder.