You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We provide our PyTorch implementation of our paper 'Domain Adaptation on Point Clouds via Geometry-Aware Implicits' (IEEE CVPR 2022). By using geometry-awrae implicits representation, our method can align point clouds from different domains in feature space well.
Here we show point clouds from different domains in an image.
Domain Alignment
Class-wise MMD for the task: ModelNet to ScanNet in PointDA-10 dataset. Diagonal shows source-target distances of the same class. Upper and lower triangular matrices indicate distances between different classes in the source and target domain, respectively. Our method maintains class-wise distances well.
Dataset Preprocessing
For generating point clouds from GraspNet, we need to render depth maps firstly. Refer to my repo ObjsDepthRender for more information.
So far, this repo only includes the self-supervised pre-training part. As for domain adaptation, my suggestion is to use GAST which is a sufficient codebase for benchmark comparisons.
Citation
If you find this useful for your research, please cite the following paper.
@InProceedings{Shen_2022_CVPR,
author = {Shen, Yuefan and Yang, Yanchao and Yan, Mi and Wang, He and Zheng, Youyi and Guibas, Leonidas J.},
title = {Domain Adaptation on Point Clouds via Geometry-Aware Implicits},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2022},
pages = {7223-7232}
}
About
Domain Adaptation on Point Clouds via Geometry-Aware Implicits