You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This is the code for <paper>. Reaction rules are extracted from USPTO-MIT dataset and used as actions for reinforcement learning for molecular applications. Two applications are created:
Drug discovery using a gymnasium-compatible RL simulator (Online RL)
Lead optimization without using similarity-based metrics (Goal-conditioned RL + offline RL)
Requirements
This repo was built using python=3.7.
Common requirements:
The gymnasium environment for molecular discovery is contained in folder molecular_discovery. Folder sb3 contains 4 example <agents> = [ppo/sac/td3/ddpg]. For molecular discovery using stable_baselines3, run
# generate some offline data by rolling out a random policy.
python lead_optimization.dump_data_for_offlineRL --train 100000 --steps 5
# train an offline RL agent
python lead_optimization.python offlineRL.py --steps 5 --model actor-critic --actor-loss PG --cuda 0
About
Gym-compatible simulator for reaction sequences using USPTO dataset and then using RL for molecular generation.