Deep inverse reinforcement learning for structural evolution of small molecules
B Agyemang, WP Wu, D Addo… - Briefings in …, 2021 - academic.oup.com
Briefings in Bioinformatics, 2021•academic.oup.com
The size and quality of chemical libraries to the drug discovery pipeline are crucial for
developing new drugs or repurposing existing drugs. Existing techniques such as
combinatorial organic synthesis and high-throughput screening usually make the process
extraordinarily tough and complicated since the search space of synthetically feasible drugs
is exorbitantly huge. While reinforcement learning has been mostly exploited in the literature
for generating novel compounds, the requirement of designing a reward function that …
developing new drugs or repurposing existing drugs. Existing techniques such as
combinatorial organic synthesis and high-throughput screening usually make the process
extraordinarily tough and complicated since the search space of synthetically feasible drugs
is exorbitantly huge. While reinforcement learning has been mostly exploited in the literature
for generating novel compounds, the requirement of designing a reward function that …
Abstract
The size and quality of chemical libraries to the drug discovery pipeline are crucial for developing new drugs or repurposing existing drugs. Existing techniques such as combinatorial organic synthesis and high-throughput screening usually make the process extraordinarily tough and complicated since the search space of synthetically feasible drugs is exorbitantly huge. While reinforcement learning has been mostly exploited in the literature for generating novel compounds, the requirement of designing a reward function that succinctly represents the learning objective could prove daunting in certain complex domains. Generative adversarial network-based methods also mostly discard the discriminator after training and could be hard to train. In this study, we propose a framework for training a compound generator and learn a transferable reward function based on the entropy maximization inverse reinforcement learning (IRL) paradigm. We show from our experiments that the IRL route offers a rational alternative for generating chemical compounds in domains where reward function engineering may be less appealing or impossible while data exhibiting the desired objective is readily available.
Oxford University Press
以上显示的是最相近的搜索结果。 查看全部搜索结果