You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The hyper-parameters for other datasets are provided in the paper.
Note:
Config files, including refined version of Glow and MaCow, are provided here.
The argument --batch_steps is used for accumulated gradients to trade speed for memory. The size of each segment of data batch is batch-size / (num_gpus * batch_steps).
For distributed training on multi-GPUs, please use distributed.py or slurm.py, and
refer to the pytorch distributed parallel training tutorial.
We also implement the MaCow model with distributed training supported. To train a new MaCow model, please use the MaCow config files for different datasets.
References
@InProceedings{decoupling2021,
title = {Decoupling Global and Local Representations via Invertible Generative Flows},
author = {Ma, Xuezhe and Kong, Xiang and Zhang, Shanghang and Hovy, Eduard},
booktitle = {Proceedings of the 9th International Conference on Learning Representations (ICLR-2021)},
year = {2021},
month = {May},
}
@incollection{macow2019,
title = {MaCow: Masked Convolutional Generative Flow},
author = {Ma, Xuezhe and Kong, Xiang and Zhang, Shanghang and Hovy, Eduard},
booktitle = {Advances in Neural Information Processing Systems 33, (NeurIPS-2019)},
year = {2019},
publisher = {Curran Associates, Inc.}
}