Combining generative models and reinforcement learning has become a promising direction for computational drug design, but it is challenging to train an efficient model that produces candidate molecules with high diversity. Jike Wang and colleagues present a method, using knowledge distillation, to condense a conditional transformer model to make it usable in reinforcement learning while still generating diverse molecules that optimize multiple molecular properties.
- Jike Wang
- Chang-Yu Hsieh
- Tingjun Hou