You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We use HHA representation for depth, which transfer one channel representation into three channels.
The source code of computing HHA representation can be found at https://github.com/s-gupta/rcnn-depth.
Compute superpixel and optical flow.
Compute region correspondence.
We provide Matlab code to establish region correspondence in region_correspondence folder.
Run the demo code run.m to see how it works.
As a result, full_std2p_nyud2.caffemodel is generated in the folder examples/models.
Use predict.py to run our demo.
python predict.py -g [gpu_id] -m [model]
Training
Due to GPU memory limitation, we cache pool4 and pool4_hha at the hard disk to provide more views.
And then, we train the parameters after pool4. That is why we need to generate the full model at test stage 4.
In our experiment, we provide 11 views to train the model.
you can use examples/solve.py to train the model.
python solve.py -g [gpu_id]
Citation
If our work is useful for you, please consider citing:
@inproceedings{yang_cvpr17,
title={STD2P: RGBD Semantic Segmentation Using Spatio-Temporal Data-Driven Pooling},
author={Yang He and Wei-Chen Chiu and Margret Keuper and Mario Fritz},
booktitle={IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
year={2017}
}
About
source code for "STD2P: RGBD Semantic Segmentation Using Spatio-Temporal Data-Driven Pooling (CVPR2017)"