You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
1University of Illinois at Urbana-Champaign,2Microsoft Azure Cognitive Services Research,* Work done at Microsoft internship and UIUC.
Introduction
This is the PyTorch implementation of SeqBoat 🚤 proposed in our paper. This repository is based on MEGA and the fairseq package v0.9.0.
Updates
[Nov. 26] Added a standalone CIFAR-10 training script of SeqBoat for quickstart!
[Nov. 5] Released training scripts for enwik8 and added a standalone implementation of SeqBoat here!
[Sep. 21] Our paper is accepted by NeurIPS 2023!
[July 18] Released training scripts for LRA and Speech Commands.
Code Overview
The compress and extract operators for Sparse Modular Activation (SMA) are implemented in fairseq/modules/seqboat_utils.py with the functions compress_seq and extract respectively.
We also provide the training and testing scripts for each of the tasks in the experiment directory.
Citation
If you find our work useful, please consider citing:
@inproceedings{ren2023sparse,
title={Sparse Modular Activation for Efficient Sequence Modeling},
author={Liliang Ren and Yang Liu and Shuohang Wang and Yichong Xu and Chenguang Zhu and ChengXiang Zhai},
booktitle={Thirty-seventh Conference on Neural Information Processing Systems},
year={2023},
url={https://openreview.net/forum?id=TfbzX6I14i}
}
License
SeqBoat is under MIT license. The license also applies to model checkpoints.