You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The Order Prediction Network (OPN) is a model that performs representation learning using unlabeled videos. Our method leverage temporal coherence as a supervisory signal by formulating representation learning as a sequence sorting task. The experimental results show that our method compares favorably against state-of-the-art methods on action recognition, image classification and object detection tasks. For more details and evaluation results, please check out our project webpage and paper.
Citation
If you find the code and data useful in your research, please cite:
@inproceeding{OPN,
author = {Lee, Hsin-Ying and Huang, Jia-Bin and Singh, Maneesh Kumar and Yang, Ming-Hsuan},
title = {Unsupervised Representation Learning by Sorting Sequence},
booktitle = {IEEE International Conference on Computer Vision},
year = {2017}
}
Requirements
Requirements for Caffe and pycaffe (see: Caffe installation instructions)
Note that the Caffe fork needs to support Batch Normalization to run our code.
Model Unsupervised trained on UCF+HMDB+ACT (for Pascal VOC 2007)
Training
There are a few lines need to be customized in UCF_datalayers.py, including the training data location in L22-L27 and L84-87. The default setting includes all processing like channel splitting and spatial jittering, feel free to comment them out.
Action Recognition
The testing on the UCF-101 and HMDB-51 datasets follows the testing sceme of the original two-stream ConvNets, where we sample 25 RGB frames from each video. From each of the frames we then obtain 10 inputs by cropping and flipping four corners and the center of the frame.