You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We provide scripts, data, and model checkpoints to run online finetuning on Atari100k benchmark with cross-tasks (i.e., using offline data).
Download the offline data for cross-tasks and pretrained model checkpoints. Link is here. Make sure you have ~100G of storage available to store the data and model checkpoints. Once downloading is finished, move the zip files into the root directory of this repo and directly unzip these files. You should have folder /data and /pretrained_ckpts under this root directory.
Install and use gcc-7 and g++-7 to compile C files. (if your system is already set up to use them, you can safely ignore this step):
Create a new conda environment by using the following code block (for full reproducibility, we used Anaconda 2022.05 version, which can be donwloaded using wget https://repo.anaconda.com/archive/Anaconda3-2022.05-Linux-x86_64.sh:
We provide a sample script to run cross-task finetuning on Alien with cross-tasks Amidar, BankHeist, MsPacman and WizardOfWor from a offline multitask pretrained model from these cross-tasks. To run the framework, execute the following command (at the root directory of the repo, using the specified setup):
bash script xtra_finetune.sh
Citation
@inproceedings{
xu2023on,
title={On the Feasibility of Cross-Task Transfer with Model-Based Reinforcement Learning},
author={Yifan Xu and Nicklas Hansen and Zirui Wang and Yung-Chieh Chan and Hao Su and Zhuowen Tu},
booktitle={The Eleventh International Conference on Learning Representations },
year={2023},
url={https://openreview.net/forum?id=KB1sc5pNKFv}
}
Note: this codebase is based on an older version of EfficientZero implementation.
About
On the Feasibility of Cross-Task Transfer with Model-Based Reinforcement Learning