You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Codebase for inspecting approximate 2 layer MLPs in transformers
The official repository for our paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers" (https://arxiv.org/abs//2310.10837).
Please note that this repository is a cleaned-up version of the internal research repository we use. In case you encounter any problems with it, please don't hesitate to contact me.
If you are interested in a plug-and-play MoE layer
Please check out https://github.com/robertcsordas/moe_layer. The development and support of the layer will happen there. This repository is for reproducing our experiments from the paper.
For plotting, LaTeX is required (to avoid Type 3 fonts and to render symbols). Installation is OS specific.
Usage
The code makes use of Weights and Biases for experiment tracking. In the "sweeps" directory, we provide sweep configurations for all experiments we have performed. The sweeps are officially meant for hyperparameter optimization, but we use them to run multiple configurations of our models.
To reproduce our results, start a sweep for each of the YAML files in the "sweeps" directory. Run wandb agent for each of them in the main directory. This will run all the experiments, and they will be displayed on the W&B dashboard.
Re-creating plots from the paper
Edit config file "paper/config.json". Enter your project name in the field "wandb_project" (e.g. "username/modules").
Run the script of interest within the "paper" directory. For example:
cd paper
python3 moe_ablations.py
About
Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"