Abstract

The size and quality of chemical libraries to the drug discovery pipeline are crucial for developing new drugs or repurposing existing drugs. Existing techniques such as combinatorial organic synthesis and high-throughput screening usually make the process extraordinarily tough and complicated since the search space of synthetically feasible drugs is exorbitantly huge. While reinforcement learning has been mostly exploited in the literature for generating novel compounds, the requirement of designing a reward function that succinctly represents the learning objective could prove daunting in certain complex domains. Generative adversarial network-based methods also mostly discard the discriminator after training and could be hard to train. In this study, we propose a framework for training a compound generator and learn a transferable reward function based on the entropy maximization inverse reinforcement learning (IRL) paradigm. We show from our experiments that the IRL route offers a rational alternative for generating chemical compounds in domains where reward function engineering may be less appealing or impossible while data exhibiting the desired objective is readily available.

Introduction

Identifying promising leads is crucial to the early stages of drug discovery. Combinatorial organic synthesis and high-throughput screening are well-known methods used to generate new compounds in the domain (drug and compound are used interchangeably in this study). This generation process is typically followed by an expert analysis, which focuses on the desired properties such as solubility, activity, pharmacokinetic profile, toxicity and synthetic accessibility, to ascertain the desirability of a generated compound. Compound generation and modification methods are useful for enriching chemical databases and scaffold hopping [1]. A review of the structural and analog entity evolution patent landscape estimates that the pharmaceutical industry constitutes |$70\%$| of the domain [2].

Indeed, the compound generation task is noted to be hard and complicated [3] considering that there exist |$10^{30}-10^{60}$| synthetically feasible drug-like compounds [4]. With about |$96\%$| [5] of the drug development projects failing due to unforeseen reasons, it is significant to ensure diversity in compounds that are desirable to avoid a fatal collapse of the drug discovery process. As a result of these challenges in the domain, there is a need for improved de novo compound generation methods.

In recent times, the proliferation of data, advances in computer hardware, novel algorithms for studying complex problems and other related factors have contributed significantly to the steady growth of data-driven methods such as deep learning (DL). DL-based approaches have been applied to several domains, such as natural language processing [6, 7], computer vision [8], proteochemometric modeling [9], compound and target representation learning [10, 11] and reaction analysis [12]. Consequently, there has been a growing interest in the literature to use data-driven methods to study the compound generation problem.

Deep reinforcement learning (DRL), generative adversarial networks (GANs) and transfer learning (TL) are some of the approaches that have been used to generate compounds represented using the simplified molecular-input line-entry system (SMILES) [13]. The DRL-based methods model the compound generation task as a sequential decision-making process and use reinforcement learning (RL) algorithms to design generators (agents) that estimate the statistical relationship between actions and outcomes. This statistical knowledge is then leveraged to maximize the outcome, thereby biasing the generator according to the desired chemical space. Motivated by the work in [14], GAN-based methods also model the compound generation task as a sequential decision-making problem but with a discriminator parameterizing the reward function. TL methods train a generator on a large dataset to increase the proportion of valid SMILES strings sampled from the generator before performing a fine-tuning training to bias the generator to the target chemical space.

Regarding DRL-based methods, [1] proposed a recurrent neural network (RNN)-based approach to train generative models for producing analogs of a query compound and compounds that satisfy certain chemical properties, such as activity to the dopamine receptor D2 (DRD2) target. The generator takes as input one-hot encoded representations of the canonical SMILES of a compound, and for each experiment, a corresponding reward function is specified. Also, [15] proposed a stack-augmented RNN-generative model using the REINFORCE algorithm [16] where, unlike [1], the canonical SMILES encoding of a compound is learned using backpropagation. The reward functions in [15] are parameterized by a prediction model that is trained separately from the generator. In both studies of [1] and [15], the generator was pretrained on a large SMILES dataset using a supervised learning approach before applying RL to bias the generator. Similarly, [17] proposed a SMILES- and graph-based compound generation model that adopts the supervised pretraining and subsequent RL biasing approach. Unlike [1] and [15], [17] assign intermediate rewards to valid incomplete SMILES strings. We posit that since an incomplete SMILES string could, in some cases, have meaning (e.g. moieties), assigning intermediate rewards could facilitate learning. While the DRL-based generative models can generate biased compounds, an accurate specification of the reward function is challenging and time-consuming in complex domains. In most interesting de novo compound generation scenarios, compounds meeting multiple objectives may be required, and specifying such multi-objective reward functions leads to the generator (agent) exploiting the straightforward objective and generating compounds with low variety.

In another vein, [3] trained an RNN model on a large dataset using supervised learning and then performed TL to the domain of interest to generate focused compounds. Since the supervised approach used to train the generator is different from the autoregressive sampling technique adopted at test time, such methods are not well suited for multi-step SMILES sampling [18]. This discrepancy is referred to as exposure bias. Additionally, methods such as [3] that maximize the likelihood of the underlying data are susceptible to learning distributions that place masses in low-density areas of the multivariate distribution giving rise to the underlying data.

On the other hand, [19] based on the work of [20] to propose a GAN-based generator that produces compounds matching some desirable metrics. The authors adopted an alternating approach to enable multi-objective RL optimization. As pointed out by [21], the challenges with training GAN cause a high rate of invalid SMILES strings, low diversity and reward function exploitation. In [22], a memory-based generator replaced the generator proposed by [19] in order to mitigate the problems in [19]. In these GAN-based methods, the authors adopted a Monte Carlo tree search (MCTS) method to assign intermediate rewards. However, GAN training could be unstable and the generator could get worse as a result of early saturation [23]. Additionally, the discriminator in the GAN-based models is typically discarded after training.

In this paper, we propose a novel framework for training a compound generator and learning a reward function from data using a sample-based deep inverse reinforcement learning (DIRL) objective [24]. We observe that while it may be daunting or impossible to accurately specify the reward function of some complex drug discovery campaigns to leverage DRL algorithms, sample of compounds that satisfy the desired behavior may be readily available or collated. Therefore, our proposed method offers a solution to develop in silico generators and recover reward function from compound samples. As pointed out by [25], the DIRL objective could lead to stability in training and producing effective generators. Also, unlike the typical GAN case where the discriminator is discarded, the sample-based DIRL objective is capable of training a generator and a reward function. Since the learned reward function succinctly represents the agent’s objective, it could be transferred to related domains (with a possible fine-tuning step). Moreover, since the binary cross-entropy (BCE) loss usually applied to train a discriminator does not apply in the case of the sample-based DIRL objective, this eliminates saturation problems in training the generator. The DIRL approach also mitigates the challenge of imbalance between different RL objectives.

Preliminaries

In this section, we review the concepts of RL and inverse reinforcement learning (IRL) related to this study.

Reinforcement learning

The aim of artificial intelligence (AI) is to develop autonomous agents that can sense their environment and act intelligently. Since learning through interaction is vital for developing intelligent systems, the paradigm of RL has been adopted to study several AI research problems. In RL, an agent receives an observation from its environment, reasons about the received observation to decide on an action to execute in the environment and receives a signal from the environment as to the usefulness of the action executed in the environment. This process continues in a trial-and-error manner over a finite or infinite time horizon for the agent to estimate the statistical relationship between the observations, actions and their results. This statistical relationship is then leveraged by the agent to optimize the expected signal from the environment.

Formally, an RL agent receives a state |$s_t$| from the environment and takes an action |$a_t$| in the environment leading to a scalar reward |$r_{t+1}$| in each time step |$t$|⁠. The agent’s behavior is defined by a policy |$\pi (a_t|s_t)$|⁠, and each action performed by the agent transitions the agent and the environment to a next state |$s_{t+1}$| or a terminal state |$s_T$|⁠. The RL problem is modeled as a Markov decision process (MDP) consisting of

  • a set of states |$\mathcal{S}$|⁠, with a distribution over starting states |$p(s_0)$|⁠;

  • a set of actions |$\mathcal{A}$|⁠;

  • state transition dynamics function |$\mathcal{T}(s_{t+1}|s_t, a_t)$| that maps a state-action pair at a time step |$t$| to a distribution of states at time step |$t+1$|⁠;

  • a reward function |$\mathcal{R}(s_t,a_t)$| that assigns a reward to the agent after taking action |$a_t$| in state |$s_t$|⁠;

  • a discount factor |$\gamma \in [0,1]$| that is used to specify a preference for immediate and long-term rewards. Also, |$\gamma < 1$| ensures that a limit is placed on the time steps considered in the infinite horizon case.

The policy |$\pi $| maps a state to a probability distribution over the action space: |$\pi : S \rightarrow p(\mathcal{A}=a|\mathcal{S})$|⁠. In an episodic MDP of length |$T$|⁠, the sequence of states, actions and rewards is referred to as a trajectory or rollout of the policy |$\pi $|⁠. The return or total reward for a trajectory could then be represented as |$R=\sum _{t=0}^{T-1}\gamma ^tr_{t+1}$|⁠. The goal of an RL agent is to learn a policy |$\pi ^*$| that maximizes its expected total reward from all states:
(1)
Existing RL methods for solving such an MDP could be classified into value function-based and policy search algorithms. Value function methods estimate the expected return of being in a given state. The state-value function |$V^\pi (s)$| estimates the expected return of starting in state |$s$| and following |$\pi $| afterward:
(2)
Given the optimal values of all states, |$V^*(s)=\underset{\pi }{max}V^\pi (s), \forall s\in S$|⁠, the optimal policy |$\pi ^*$| could be determined as
(3)
A closely related concept to the state-value function is the state-action value function |$Q^\pi (s,a)$|⁠. With the state-action value function, the initial action |$a$| is given and the policy |$\pi $| is effective from the next state onward:
(4)
The agent can determine the optimal policy, given |$Q^\pi (s,a)$|⁠, by acting greedily in every state, |$\pi ^*=\underset{\pi }{argmax} Q^\pi (s,a)$|⁠. Also, |$V^\pi (s)=\underset{a}{max}Q^\pi (s,a)$|⁠.

On the other hand, policy search/gradient algorithms directly learn the policy |$\pi $| of the agent instead of indirectly estimating it from value-based methods. Specifically, policy-based methods optimize the policy |$\pi (a|s,\theta )$|⁠, where |$\theta $| is the set of parameters of the model approximating the true policy |$\pi ^*$|⁠. We review the policy gradient (PG) methods used in this study below.

PG optimization

The REINFORCE algorithm [16] is a policy gradient method that estimates the gradient |$g:= \nabla _\theta \mathbb{E}\left [\sum _{t=0}^{T-1}r_{t+1}\right ]$| which has the form
(5)
and updates |$\theta $| in the direction of the estimated gradient. |$\Psi _t$| could be any of the functions defined by [26]. For instance, [15] defined |$\Psi _t=\gamma ^t r(s_T)$| where |$r(s_T)$| denotes the reward at the terminal state. In another vein, [22] used the reward of a fully generated sequence. Due to high variance in using the total reward of the trajectory |$\sum _{t=0}^{T-1}\gamma ^t{r_{t+1}}$|⁠, a baseline version could be adopted to reduce the variance, |$\sum _{t^{\prime}=t}^{T-1}\gamma ^{t^{\prime}-t}r_{t^{\prime}+1}-b(s_t)$|⁠. When a value function approximator is used to estimate the baseline |$b(s_t)$|⁠, the resulting method is referred to as an actor–critic method. A related concept to determine |$\Psi _t$| is realizing that it is less challenging to identify that an action has a better outcome than the default behavior compared with learning the value of the action [27]. This concept gives rise to the advantage function |$A^\pi (s,a) = Q^\pi (s,a) - V^\pi (s)$| where |$V(s)$|⁠, serving as a baseline, is approximated by a function such as a neural network.

Proximal policy optimization

Schulman et al. [28] proposed a robust and data efficient policy gradient algorithm as an alternative to prior RL training methods such as the REINFORCE, the Q-learning [29] and the relatively complicated trust region policy optimization (TRPO) [30]. The proposed method, named proximal policy optimization (PPO), shares some similarities with the TRPO in using the ratio between the new and old policies scaled by the advantages of actions to estimate the policy gradient instead of the logarithm of action probabilities (as seen in Equation (5)). While TRPO uses a constrained optimization objective that requires the conjugate gradient algorithm to avoid large policy updates, PPO uses a clipped objective that forms a pessimistic estimate of the policy’s performance to avoid destructive policy updates. These attributes of the PPO objective enables performing multiple epochs of policy update with the data sampled from the environment. In REINFORCE, such multiple epochs of optimization often leads to destructive policy updates. Thus, PPO is also more sample efficient than the REINFORCE algorithm.

More formally, the PPO objective is expressed as
(6)
where
(7)
and
(8)
where |$\delta _t=r_t+\gamma V(s_{t+1})-V(s_t)$| with hyperparameters |$\gamma $| and |$\lambda $|⁠. The |$clip(\cdot )$| function creates an alternate policy update to the expectation of |$r_t(\theta )A_t$| where the action probability ratios between the old and updated policies are maintained in the range of |$[1-\epsilon , 1+\epsilon ]$|⁠. Taking the minimum of the two possible policy updates ensures a lower-bound update is preferred. This is useful since a large update to the parameters of a highly nonlinear policy function, such as a neural network, often results in a worse policy. In this paper, we set |$\epsilon =0.2$|⁠.

Inverse reinforcement learning

IRL is the problem of learning the reward function of an observed agent, given its policy or behavior, thereby avoiding the manual specification of a reward function [31]. The IRL class of solutions assumes the following MDP|$\setminus R_E$|⁠:

  • a set of states |$\mathcal{S}$|⁠, with a distribution over starting states |$p(s_0)$|⁠;

  • a set of actions |$\mathcal{A}$|⁠;

  • state transition dynamics function |$\mathcal{T}(s_{t+1}|s_t, a_t)$| that maps a state-action pair at a time step |$t$| to a distribution of states at time step |$t+1$|⁠;

  • a set of demonstrated trajectories

    |$\mathcal{D}=\left \{\left <(s_0^i,a_0^i),...,(s_{T-1}^i,a_{T-1}^i)\right>_{i=1}^N\right \}$| from the observed agent or expert;

  • a discount factor |$\gamma \in [0,1]$| may be used to discount future rewards.

The goal of IRL then is to learn the reward function |$R_E$| that best explains the expert demonstrations. It is assumed that the demonstrations are perfectly observed and that the expert follows an optimal policy.

While learning reward functions from data is appealing, the IRL problem is ill-posed since there are many reward functions under which the observed expert behavior is optimal [31, 32]. An instance is a reward function that assigns 0 (or any constant value) for all selected actions; in such a case, any policy is optimal. Other main challenges are the accurate inference, generalizability, correctness of prior knowledge and computational cost with an increase in problem complexity.

To this end, several IRL proposals exist in the literature to mitigate the IRL challenges mentioned above [31]. In recent times, the entropy optimization class of IRL methods has been widely used by researchers due to the maximum entropy (MaxEnt)’s goal to obtain an unbiased distribution of potential reward [24, 33–35]. The intuition is that the solution that maximizes entropy violates the optimization constraints least, and hence least wrong. In the MaxEnt formulation, the probability of the expert’s trajectory is proportional to the exponential of the total reward [33],
(9)
where |$R(\tau )=\sum _{(s,a)\in \tau }r(s,a)$| and |$r(s,a)$| gives the reward for taking action |$a$| in state |$s$|⁠.

Methods

Problem statement

We consider the problem of training a model |$G_\theta (a_t|s_{0:t})$| to generate a set of compounds |$\mathbf{C}=\left \{c_1, c_2,..., c_M|M\in \mathbb{N}\right \}$|⁠, each encoded as a valid SMILES string, such that |$\mathbf{C}$| is generally biased to satisfy a predefined criteria that could be evaluated by a function |$E$|⁠. Considering the sequential nature of a SMILES string, we study this problem in the framework of IRL following the MDP described in Section 2.2; instead of the RL approach mostly adopted in the literature on compound generation, a parameterized reward function |$R_\psi $| has to be learned from |$\mathcal{D}$|⁠, which is a set of SMILES that satisfy the criteria evaluated by |$E$|⁠, to train |$G_\theta $| following the MDP in Section 2.1. Note that in the case of SMILES generation, |$\mathcal{T}(s_{t+1}|s_t, a_t)=1$|⁠.

In this context, the action space |$\mathcal{A}$| is defined as the set of unique tokens that are combined to form a canonical SMILES string and the state space |$\mathcal{S}$| is the set of all possible combinations of these tokens to encode a valid SMILES string, each of length |$l\in [1, T], T\in \mathbb{N}$|⁠. We set |$s_0$| to a fixed state that denotes the beginning of a SMILES string, and |$s_T$| is the terminal state.

Furthermore, we compute |$r(s_{T-1},a_{T-1})=R_\psi (Y_{1:T})$|⁠, where |$Y_{1:T}$| denotes a full generated sequence. Intermediate state rewards are obtained by performing an |$N$|-time MCTS with |$G_\theta $| as the rollout policy to obtain the corresponding rewards. Specifically, given the |$N$|-time MCTS set of rollouts,
(10)
the reward for a state is then calculated as

Proposed approach

The workflow we propose in this study for the structural evolution of compounds is presented in Figure 1a. The framework is described in what follows.

(A) Illustration of the proposed framework for training a small molecule generator and learning a reward function using IRL. The workflow begins with pretraining the generator model using large-scale compound sequence dataset, such as ChEMBL. The pretrained model is used to initialize the agent network in an IRL training scheme where a reward function is learned and the agent is biased to generate desirable compounds. The generated SMILES could be examined by an evaluation function where compounds satisfying a specified threshold could be persisted for further analysis. (B) The general structure of the models used in the study. The agent is represented as an actor–critic architecture that is trained using the PPO [28] algorithm. The actor, critic and reward functions are all RNNs that share the same SMILES encoder. The actor net becomes the SMILES generator at test time.
Fig. 1

(A) Illustration of the proposed framework for training a small molecule generator and learning a reward function using IRL. The workflow begins with pretraining the generator model using large-scale compound sequence dataset, such as ChEMBL. The pretrained model is used to initialize the agent network in an IRL training scheme where a reward function is learned and the agent is biased to generate desirable compounds. The generated SMILES could be examined by an evaluation function where compounds satisfying a specified threshold could be persisted for further analysis. (B) The general structure of the models used in the study. The agent is represented as an actor–critic architecture that is trained using the PPO [28] algorithm. The actor, critic and reward functions are all RNNs that share the same SMILES encoder. The actor net becomes the SMILES generator at test time.

Datasets

The first stage of the workflow is creating a SMILES dataset, similar to most existing GAN and vanilla-RL methods. Possible repositories for facilitating this dataset creation are DrugBank [36], KEGG [37], STITCH [38] and ChEMBL [39]. This dataset is used at the next stage for pretraining the generator. For this purpose, we used the curated ChEMBL dataset of [15], which consists of approximately |$1.5$| million drug-like compounds.

Also, a set of SMILES satisfying the constraints or criteria of interest are collated from an appropriate source as the demonstrations |$\mathcal{D}$| of the IRL phase (for instance, see Section 3.6). To evaluate SMILES during the training of the generator, we assume the availability of an evaluation function |$ E $| that can evaluate the extent to which a given compound satisfies the optimization constraints. Here, |$E$| could be an ML model, a robot that conducts chemical synthesis or a molecular docking program.

In this study, we avoided composing the demonstrations dataset from the data used to train the evaluation function. Since, in practice, the data, rule set or method used for developing any of the evaluation functions mentioned above could differ from the approach for constructing the demonstrations dataset, this independence provides a more realistic context to assess our proposed method.

Generative pretraining of prior model

This stage of the framework entails using the large-scale dataset of the previous step to pretrain a model that would be a prior for initializing the agent network/policy. This pretraining step aims to enable the model to learn the SMILES syntax to attain a high rate of valid generated SMILES strings.

Since the SMILES string is a sequence of characters, we represent the prior model as an RNN with gated recurrent units (GRUs). The architecture of the generator at this stage of the workflow is depicted in Figure 3b. The generator takes as input the output of an encoder that learns the embedding of SMILES characters, as shown in Figure 1b. For each given SMILES string, the generator is trained to predict the next token in the sequence using the cross-entropy loss function.

According to the findings of [40], regular RNNs cannot effectively generate sequences of a context-free language due to the lack of memory. To this end, [40] proposed the use of a Stack-RNN architecture, which equips the standard RNN cell with a stack unit to mitigate this problem. Since the SMILES encoding of a compound is a context-free language, we follow [15] to use the Stack-RNN for generating SMILES as it ensures that tasks such as noting the start and end parts of aromatic moieties and parentheses for branches in the chemical structure are well addressed. We note that while [15] used a single-layer Stack-RNN, we adopt a multi-layer Stack-RNN structure in our study. We reckon that the multi-layer Stack-RNN could facilitate learning better representations at multiple levels of abstraction akin to multi-layer RNNs.

The stack unit enables the model to persist information across different time steps. Three differentiable operations could be performed on the stack at each time step: POP, PUSH and NO-OP. The POP operation deletes the entry at the top of the stack, the PUSH operation updates the top of the stack with a new record and the NO-OP operation leaves the stack unchanged for the next time step. In [40], the entry at the top of stack |$s$| at time step |$t$| is indexed as |$s_t[0]$| and is updated as
(11)
where |$h_t\in \mathbb{R}^{m\times 1}$| is the RNN hidden vector of a sequence at time step |$t$|⁠, |$D\in \mathbb{R}^{w\times m}$|⁠, |$w$| is the dimension of the stack and |$a_t\in \mathbb{R}^3$| is a vector whose elements map to the stack control operators PUSH, POP and NO-OP, respectively. Here, |$a_t$| is computed as
(12)
where |$A\in \mathbb{R}^{3\times m}$|⁠. Intuitively, if |$a_t[PUSH]=1$| then |$s_t[0]$| is updated with |$Dh_t$|⁠; if |$a_t[POP]=1$| then the top of the stack is updated with the second entry of the stack and |$a_t[NO-OP]=1$| leaves the stack unchanged.
Regarding the remaining entries of the stack, thus indexes |$i>0$|⁠, the update rule is
(13)
Subsequently, the RNN cell’s hidden vector |$h_t$| is updated as
(14)
where |$\sigma (x)=1/(1+exp(-x))$|⁠, |$U\in \mathbb{R}^{m\times d}$|⁠, |$R \in \mathbb{R}^{m\times m}$|⁠, |$P\in \mathbb{R}^{m\times w}$|⁠, |$s_{t-1}[0]\in \mathbb{R}^{w\times 1}$|⁠, |$d$| is the dimension of encoded representation of token |$x_t$| and |$m$| is the dimension of the hidden state.
Sample compounds generated using the pretrained model.
Fig. 2

Sample compounds generated using the pretrained model.

IRL phase

As stated earlier, we frame the SMILES string generation task as an MDP problem where the reward function is unknown. Consequently, this MDP does not enable the use of RL algorithms to approximate the solution since RL methods require a reward function. Therefore, we learn the reward function using demonstrations that exhibit the desired behavior.

Specifically, given the set of trajectories |$D=\left \{\tau _1,\tau _2,...\tau _N\right \}$| observed from an expert that acts stochastically under an optimal policy, the density function is specified as
(15)
where |$\tau =\left \{\left <s_0,a_0\right>,..., \left <s_{T-1},a_{T-1}\right>\right \}$| is a trajectory of a generated SMILES string |$Y_{1:T}$| and |$R\psi $| is an unknown reward function parameterized by |$\psi $|⁠. Thus, the expert generates SMILES strings that satisfy the desired criteria with a probability that increases exponentially with the return. Here, the main challenge for this energy-based model is computing the partition function
(16)
Ziebart et al. [33] computed |$Z$| using a dynamic programming (DP) approach to estimate the partition function. However, the DP method is not scalable to large-state space domains, such as the space of synthetically accessible compounds mentioned earlier.
The architectures of the RNN models used in this study. (A) The structure of the reward net (left) and the critic net (right). (B) The structure of the generator net during pretraining to learn the SMILES syntax. The model is trained using the cross-entropy loss. (C) The structure of the generator net during the RL phase; the model autoregressively generates a SMILES string given a fixed start state. The sampling process terminates when an $end$ token is sampled or a length limit is reached.
Fig. 3

The architectures of the RNN models used in this study. (A) The structure of the reward net (left) and the critic net (right). (B) The structure of the generator net during pretraining to learn the SMILES syntax. The model is trained using the cross-entropy loss. (C) The structure of the generator net during the RL phase; the model autoregressively generates a SMILES string given a fixed start state. The sampling process terminates when an |$end$| token is sampled or a length limit is reached.

In [24], the authors proposed a sample-based entropy maximization IRL algorithm. The authors use a sample or background distribution to approximate the partition function and continuously adapt this sample distribution to provide a better estimation of Equation (15). Therefore, if the sample distribution represents the agent’s policy and this distribution is trained using an RL algorithm, given an approximation of |$R_\psi $|⁠, then the RL algorithm guides the sample distribution to a space that provides a better estimation of |$Z$|⁠. We refer to the method discussed by [24] as guided reward learning (GRL) in this study. We adopt this approach to bias a pretrained generator to the desired compound space since the GRL method produces a policy and a learned reward function that could be transferred to other related domains.

Formally, taking the log-likelihood of Equation (15) provides the maximization objective
(17)
(18)
where |$\hat{\mathcal{D}}$| is the set of |$M$| trajectories sampled using the background distribution |$q$|⁠. As depicted in Figure 3a-left, |$R_\psi $| is represented as a two-layer GRU RNN that takes as input the output of a SMILES encoder (Figure 1b) and predicts a scalar reward for a generated valid or invalid SMILES string.

RL is used to train |$q$| to improve the background distribution used in estimating the partition function. The model architecture of |$q$| during the RL training and SMILES generation phases is shown in Figure 3c.

Since we represent |$q$| as an RNN in this study, training the sample distribution with high-variance RL algorithms such as the REINFORCE [16] objective could override the efforts of the pretraining due to the nonlinearity of the model. Therefore, we train the sample distribution using the PPO algorithm [28]. The PPO objective ensures a gradual change in the parameters of the agent policy (sample distribution). The PPO algorithm learns a value function to estimate |$A^\pi (s, a)$| making it an actor–critic method. The architecture of the critic in this study, modeled as an RNN, is illustrated in Figure 3a-right. Like the actor model, the critic takes the SMILES encoder’s outputs as input to predict |$V(s_t)$| at each time step |$t$|⁠.

Lastly, the evaluation function |$E$| of our proposed framework (Figure 1a) is used, at designated periods of training, to evaluate a set of generated SMILES strings. Compounds deemed to satisfy the learning objective could be persisted for further examination in the drug discovery process and the training of the generator could be terminated.

Experiments setup

We performed four experiments to evaluate the approach described in this study. In each experiment, the aim is to ascertain the effectiveness of our proposed approach and how it compares with the case where the reward function is known, as found in the literature. The hardware specifications we used for our experiments are presented in Table 1. The performance of each of the evaluation functions used in the experiments are shown in Table 2.

Table 1

Experiments hardware specifications

ModelNumber of coresRAM (GB)GPUs
Intel Xeon CPU E5-2687W481281 GeForce GTX 1080
Intel Xeon CPU E5-2687W241284 GeForce GTX 1080Ti
ModelNumber of coresRAM (GB)GPUs
Intel Xeon CPU E5-2687W481281 GeForce GTX 1080
Intel Xeon CPU E5-2687W241284 GeForce GTX 1080Ti
Table 1

Experiments hardware specifications

ModelNumber of coresRAM (GB)GPUs
Intel Xeon CPU E5-2687W481281 GeForce GTX 1080
Intel Xeon CPU E5-2687W241284 GeForce GTX 1080Ti
ModelNumber of coresRAM (GB)GPUs
Intel Xeon CPU E5-2687W481281 GeForce GTX 1080
Intel Xeon CPU E5-2687W241284 GeForce GTX 1080Ti
Table 2

Performance of ML models used as evaluation functions in the experiments. The reported values are averages of a 5-fold CV in each case. The standard deviation values are shown in parenthesis. The corresponding experiment(s) that used each ML model is/are specified in parenthesis in the first column

ModelBinary classificationRegression
PrecisionRecallAccuracyAUCRMSE|$R^2$|
RNN-Bin (DRD2)0.971 (0.120)0.970 (0.120)0.985 (0.011)0.996 (0.001)
RNN-Reg (LogP)0.845 (0.508)0.708 (0.359)
XGB-Reg (JAK2 Min & Max)0.646 (0.037)0.691 (0.039)
ModelBinary classificationRegression
PrecisionRecallAccuracyAUCRMSE|$R^2$|
RNN-Bin (DRD2)0.971 (0.120)0.970 (0.120)0.985 (0.011)0.996 (0.001)
RNN-Reg (LogP)0.845 (0.508)0.708 (0.359)
XGB-Reg (JAK2 Min & Max)0.646 (0.037)0.691 (0.039)
Table 2

Performance of ML models used as evaluation functions in the experiments. The reported values are averages of a 5-fold CV in each case. The standard deviation values are shown in parenthesis. The corresponding experiment(s) that used each ML model is/are specified in parenthesis in the first column

ModelBinary classificationRegression
PrecisionRecallAccuracyAUCRMSE|$R^2$|
RNN-Bin (DRD2)0.971 (0.120)0.970 (0.120)0.985 (0.011)0.996 (0.001)
RNN-Reg (LogP)0.845 (0.508)0.708 (0.359)
XGB-Reg (JAK2 Min & Max)0.646 (0.037)0.691 (0.039)
ModelBinary classificationRegression
PrecisionRecallAccuracyAUCRMSE|$R^2$|
RNN-Bin (DRD2)0.971 (0.120)0.970 (0.120)0.985 (0.011)0.996 (0.001)
RNN-Reg (LogP)0.845 (0.508)0.708 (0.359)
XGB-Reg (JAK2 Min & Max)0.646 (0.037)0.691 (0.039)

DRD2 activity

This experiment’s objective was to train the generator to produce compounds that target the DRD2 protein. Hence, we retrieved a DRD2 dataset of |$351~529$| compounds from ExCAPE-DB [41] (https://git.io/JUgpt). This dataset contained |$8323$| positive compounds (binding to DRD2). We then sampled an equal number of the remaining compounds as the negatives (non-binding to DRD2) to create a balanced dataset. The balanced DRD2 dataset of |$16~646$| samples was then used to train a two-layer LSTM RNN, similar to the reward function network shown in Figure 3a-left but with an additional sigmoid endpoint, using five-fold cross-validation with the BCE loss function.

The resulting five models of the CV training then served as the evaluation function |$E$| of this experiment. This evaluation function is referred to as RNN-Bin in Table 2. At test time, the average value of the predicted DRD2 activity probability was assigned as the result of the evaluation of a given compound.

Also, to create the set of demonstrations |$\mathcal{D}$| of this experiment, we used the SVM classifier of [1] to filter the ChEMBL dataset of [15] for compounds with a probability of activity greater than |$0.8$|⁠. This filtering resulted in a dataset of |$7732$| compounds to serve as |$\mathcal{D}$|⁠.

LogP optimization

In this experiment, we trained a generator to produce compounds biased toward having their octanol-water partition coefficient (logP) less than five and greater than one. LogP is one of the elements of Lipinski’s rule of five.

On the evaluation function |$E$| of this experiment, we used the LogP dataset of [15], consisting of |$14~176$| compounds, to train an LSTM RNN, similar to the reward function network shown in Figure 3a-left, using five-fold cross-validation with the mean square error (MSE) loss function. Similar to the DRD2 experiment, the five models serve as the evaluation function of |$E$|⁠. The evaluation function of this experiment is labeled RNN-Reg in Table 2.

We constructed the LogP demonstrations dataset |$\mathcal{D}$| by using the LogP-biased generator of [15] to produce |$10~000$| SMILES strings, of which |$5019$| were unique valid compounds. The |$5019$| compounds then served as the set of demonstrations for this experiment.

JAK2 inhibition

We also performed two experiments on producing compounds for JAK2 modulation. In the first JAK2 experiment, we trained a generator to produce compounds that maximize the negative logarithm of half-maximal inhibitory concentration (pIC|$_{50}$|⁠) values for JAK2. In this instance, we used the public JAK2 dataset of [15], consisting of |$1911$| compounds, to train an XGBoost model in a five-fold cross-validation method with the MSE loss function. The five resulting models then served as the evaluation function |$E$|⁠, similar to the LogP and DRD2 experiments. The JAK2 inhibition evaluation function is referred to as XGB-Reg in Table 2. We used the JAK2-maximization-biased generator of [15] to produce a demonstration set of |$3608$| unique valid compounds out of |$10~000$| generated SMILES.

On the other hand, we performed an experiment to bias the generator toward producing compounds that minimize the pIC|$_{50}$| values for JAK2. JAK2 minimization is useful for reducing off-target effects. While we maintained the evaluation function of the JAK2 maximization experiment in the JAK2 minimization experiment, we replaced the demonstrations set with |$285$| unique valid SMILES, out of |$10000$| generated SMILES, produced by the JAK2-minimization-biased generator of [15]. Also, we did not train a reward function for JAK2 minimization but rather transferred the reward function learned for JAK2 maximization to this experiment. However, we negated each reward obtained from the JAK2 maximization reward function for the minimization case study. This was done to examine the ability to transfer a learned reward function to a related domain.

Model types

As discussed in Section 3.4, we pretrained a two-layer Stack-RNN model with the ChEMBL dataset of [15] for one epoch. The training time was approximately 14 days. This pretrained model served as the initializing model for the following generators.

  1. PPO-GRL: this model type follows our proposed approach. It is trained using the GRL objective at the IRL phase and the PPO algorithm at the RL phase.

  2. PPO-Eval: this model type follows our proposed approach but without the IRL phase. The RL algorithm used is PPO. Since the problem presented in Section 3.1 assumes an evaluation function |$E$| to periodically determine the performance of the generator during training, this model enables us to evaluate an instance where |$E$| is able to serve as a reward function directly, such as ML models in our experiments. We note that other instances of |$E$|⁠, such as molecular docking or a robot performing synthesis, may be expensive to serve as the reward function in RL training.

  3. REINFORCE: this model type uses the proposed SMILES generation method in [15] to train a two-layer Stack-RNN generator following the method and reward functions of [15]. Thus, no IRL is performed. In the DRD2 experiment, we used the reward function of [1]. Also, for the JAK2 and LogP experiments, we used their respective reward functions in [15].

  4. REINFORCE-GRL: likewise, the REINFORCE-GRL model type is trained using the REINFORCE algorithm at the RL phase and the GRL method to learn the reward function. This model facilitates assessment of the significance of the PPO algorithm properties to our proposed DIRL approach.

  5. Stack-RNN-TL: this model type is trained using TL. Specifically, after the pretraining stage, the demonstrations dataset is used to fine-tune the prior model to bias the generator toward the desired chemical space. Unlike the previous model types which are either trained using RL/IRL, this generator is trained using supervised learning (cross-entropy loss function). Since this model type is a possible candidate for scenarios where the reward function is not specified but exemplar data could be provided, TL serves as a useful baseline to compare the PPO-GRL approach.

The training time of each generator in each experiment is shown in Table 3.

Table 3

Results of experiments without applying threshold to generated or dataset samples. PPO-GRL follows our proposed approach, REINFORCE follows the work in [15], PPO-Eval, REINFORCE-GRL, and Stack-RNN-TL are baselines in this study

ObjectiveAlgorithm/ datasetNumber of unique canonical SMILESInternal diversityExternal diversitySolubilityNaturalnessSynthesizabilityDruglikenessApprox. training time (mins)
Demonstrations77320.8970.7780.5360.6380.603
Unbiased20480.9170.3810.6610.5560.6400.602
PPO-GRL32640.8870.1890.8870.6740.5790.27141
DRD2PPO-Eval45380.8780.1830.8850.7220.5210.23633
REINFORCE73930.9040.3250.8020.7030.7180.44342
REINFORCE-GRL5410.9240.5130.8460.7940.2930.304390
Stack-RNN-TL69270.9180.3910.6550.5510.6470.6113064
Demonstrations50190.8800.9000.5530.8030.512
Unbiased20510.9170.3720.6580.5530.6400.603-
PPO-GRL46040.9030.1590.8360.5800.7700.49460
LogPPPO-Eval49750.7330.0600.9820.7310.4490.07550
REINFORCE77040.8970.1010.7740.5800.8270.61856
REINFORCE-GRL72250.9150.3470.6630.5510.6890.62762
Stack-RNN-TL69270.9180.3910.6550.5510.6470.6111480
Demonstrations36080.8050.0830.4850.5600.483
Unbiased20500.9170.3790.6540.5540.6410.604
PPO-GRL59110.9280.6140.5870.6640.4680.58122
JAK2 maxPPO-Eval69370.9170.3860.6570.5540.6440.60438
REINFORCE67680.9160.3510.6080.6080.5290.62730
REINFORCE-GRL70390.9170.3810.6580.5550.6440.607133
Stack-RNN-TL69270.9180.3910.6550.5510.6470.611608
Demonstrations2850.8280.5340.5480.8950.604
Unbiased20500.9170.3790.6540.5540.6410.604-
PPO-GRL34460.9070.2440.4880.5060.7760.66334
JAK2 minPPO-Eval15330.7030.0080.9970.7560.4140.049148
REINFORCE76940.9080.2340.6550.5910.7990.64940
REINFORCE-GRL69530.9170.3760.6620.5470.6570.61347
Stack-RNN-TL69270.9180.3910.6550.5510.6470.61183
ObjectiveAlgorithm/ datasetNumber of unique canonical SMILESInternal diversityExternal diversitySolubilityNaturalnessSynthesizabilityDruglikenessApprox. training time (mins)
Demonstrations77320.8970.7780.5360.6380.603
Unbiased20480.9170.3810.6610.5560.6400.602
PPO-GRL32640.8870.1890.8870.6740.5790.27141
DRD2PPO-Eval45380.8780.1830.8850.7220.5210.23633
REINFORCE73930.9040.3250.8020.7030.7180.44342
REINFORCE-GRL5410.9240.5130.8460.7940.2930.304390
Stack-RNN-TL69270.9180.3910.6550.5510.6470.6113064
Demonstrations50190.8800.9000.5530.8030.512
Unbiased20510.9170.3720.6580.5530.6400.603-
PPO-GRL46040.9030.1590.8360.5800.7700.49460
LogPPPO-Eval49750.7330.0600.9820.7310.4490.07550
REINFORCE77040.8970.1010.7740.5800.8270.61856
REINFORCE-GRL72250.9150.3470.6630.5510.6890.62762
Stack-RNN-TL69270.9180.3910.6550.5510.6470.6111480
Demonstrations36080.8050.0830.4850.5600.483
Unbiased20500.9170.3790.6540.5540.6410.604
PPO-GRL59110.9280.6140.5870.6640.4680.58122
JAK2 maxPPO-Eval69370.9170.3860.6570.5540.6440.60438
REINFORCE67680.9160.3510.6080.6080.5290.62730
REINFORCE-GRL70390.9170.3810.6580.5550.6440.607133
Stack-RNN-TL69270.9180.3910.6550.5510.6470.611608
Demonstrations2850.8280.5340.5480.8950.604
Unbiased20500.9170.3790.6540.5540.6410.604-
PPO-GRL34460.9070.2440.4880.5060.7760.66334
JAK2 minPPO-Eval15330.7030.0080.9970.7560.4140.049148
REINFORCE76940.9080.2340.6550.5910.7990.64940
REINFORCE-GRL69530.9170.3760.6620.5470.6570.61347
Stack-RNN-TL69270.9180.3910.6550.5510.6470.61183
Table 3

Results of experiments without applying threshold to generated or dataset samples. PPO-GRL follows our proposed approach, REINFORCE follows the work in [15], PPO-Eval, REINFORCE-GRL, and Stack-RNN-TL are baselines in this study

ObjectiveAlgorithm/ datasetNumber of unique canonical SMILESInternal diversityExternal diversitySolubilityNaturalnessSynthesizabilityDruglikenessApprox. training time (mins)
Demonstrations77320.8970.7780.5360.6380.603
Unbiased20480.9170.3810.6610.5560.6400.602
PPO-GRL32640.8870.1890.8870.6740.5790.27141
DRD2PPO-Eval45380.8780.1830.8850.7220.5210.23633
REINFORCE73930.9040.3250.8020.7030.7180.44342
REINFORCE-GRL5410.9240.5130.8460.7940.2930.304390
Stack-RNN-TL69270.9180.3910.6550.5510.6470.6113064
Demonstrations50190.8800.9000.5530.8030.512
Unbiased20510.9170.3720.6580.5530.6400.603-
PPO-GRL46040.9030.1590.8360.5800.7700.49460
LogPPPO-Eval49750.7330.0600.9820.7310.4490.07550
REINFORCE77040.8970.1010.7740.5800.8270.61856
REINFORCE-GRL72250.9150.3470.6630.5510.6890.62762
Stack-RNN-TL69270.9180.3910.6550.5510.6470.6111480
Demonstrations36080.8050.0830.4850.5600.483
Unbiased20500.9170.3790.6540.5540.6410.604
PPO-GRL59110.9280.6140.5870.6640.4680.58122
JAK2 maxPPO-Eval69370.9170.3860.6570.5540.6440.60438
REINFORCE67680.9160.3510.6080.6080.5290.62730
REINFORCE-GRL70390.9170.3810.6580.5550.6440.607133
Stack-RNN-TL69270.9180.3910.6550.5510.6470.611608
Demonstrations2850.8280.5340.5480.8950.604
Unbiased20500.9170.3790.6540.5540.6410.604-
PPO-GRL34460.9070.2440.4880.5060.7760.66334
JAK2 minPPO-Eval15330.7030.0080.9970.7560.4140.049148
REINFORCE76940.9080.2340.6550.5910.7990.64940
REINFORCE-GRL69530.9170.3760.6620.5470.6570.61347
Stack-RNN-TL69270.9180.3910.6550.5510.6470.61183
ObjectiveAlgorithm/ datasetNumber of unique canonical SMILESInternal diversityExternal diversitySolubilityNaturalnessSynthesizabilityDruglikenessApprox. training time (mins)
Demonstrations77320.8970.7780.5360.6380.603
Unbiased20480.9170.3810.6610.5560.6400.602
PPO-GRL32640.8870.1890.8870.6740.5790.27141
DRD2PPO-Eval45380.8780.1830.8850.7220.5210.23633
REINFORCE73930.9040.3250.8020.7030.7180.44342
REINFORCE-GRL5410.9240.5130.8460.7940.2930.304390
Stack-RNN-TL69270.9180.3910.6550.5510.6470.6113064
Demonstrations50190.8800.9000.5530.8030.512
Unbiased20510.9170.3720.6580.5530.6400.603-
PPO-GRL46040.9030.1590.8360.5800.7700.49460
LogPPPO-Eval49750.7330.0600.9820.7310.4490.07550
REINFORCE77040.8970.1010.7740.5800.8270.61856
REINFORCE-GRL72250.9150.3470.6630.5510.6890.62762
Stack-RNN-TL69270.9180.3910.6550.5510.6470.6111480
Demonstrations36080.8050.0830.4850.5600.483
Unbiased20500.9170.3790.6540.5540.6410.604
PPO-GRL59110.9280.6140.5870.6640.4680.58122
JAK2 maxPPO-Eval69370.9170.3860.6570.5540.6440.60438
REINFORCE67680.9160.3510.6080.6080.5290.62730
REINFORCE-GRL70390.9170.3810.6580.5550.6440.607133
Stack-RNN-TL69270.9180.3910.6550.5510.6470.611608
Demonstrations2850.8280.5340.5480.8950.604
Unbiased20500.9170.3790.6540.5540.6410.604-
PPO-GRL34460.9070.2440.4880.5060.7760.66334
JAK2 minPPO-Eval15330.7030.0080.9970.7560.4140.049148
REINFORCE76940.9080.2340.6550.5910.7990.64940
REINFORCE-GRL69530.9170.3760.6620.5470.6570.61347
Stack-RNN-TL69270.9180.3910.6550.5510.6470.61183
Column 1: the distribution plots of the evaluation function’s outputs for samples generated from the different model types used in the experiment. Column 2: the convergence plot of the RL and IRL-trained models during training; the y-axis represents the mean value of the experiment’s evaluation function output for samples generated at a point during training. The demo SMILES results correspond to the demonstration files of a given experiment. The unbiased SMILES results correspond to samples generated from the pretrained (unbiased or prior) model. Column 3: the convergence plot of the Stack-RNN-TL model during training. (A) The results of DRD2 experiment. (B) The results of LogP optimization experiment. (C) The results of JAK2 maximization. (D) The results of JAK2 minimization experiment.
Fig. 4

Column 1: the distribution plots of the evaluation function’s outputs for samples generated from the different model types used in the experiment. Column 2: the convergence plot of the RL and IRL-trained models during training; the y-axis represents the mean value of the experiment’s evaluation function output for samples generated at a point during training. The demo SMILES results correspond to the demonstration files of a given experiment. The unbiased SMILES results correspond to samples generated from the pretrained (unbiased or prior) model. Column 3: the convergence plot of the Stack-RNN-TL model during training. (A) The results of DRD2 experiment. (B) The results of LogP optimization experiment. (C) The results of JAK2 maximization. (D) The results of JAK2 minimization experiment.

Metrics

Apart from the Stack-RNN-TL model, all other models in each experiment were trained for a maximum of |$600$| episodes (across all trajectories), with early termination once the threshold of the experiment was reached. The Stack-RNN-TL model was trained for two epochs (due to time constraint) in each experiment. The threshold for each experiment is the average score of the evaluation function on the demonstration set. Also, only the model weights yielding the best score, as evaluated by |$E$| during training, were saved for each model type.

Furthermore, we assessed the performance of all trained generators using the metrics provided by [20] and the internal diversity metric in [21]. In each of these metrics, the best value is 1 and the worst value is 0. We give a brief introduction of the metrics below.

  • Diversity: the diversity metrics measure the relative diversity between the generated compounds and a reference set. Given a compound from the generated set, a value of 1 connotes that the substructures of the compound is diverse from the referenced set, whereas a value of 0 indicates that the compound shares several substructures with the compounds in the reference set. In our study, a random sample of the demonstrations dataset (1000 compounds) constitute the external diversity [20] reference set. On the other hand, all generated compounds constitute the reference set when calculating internal diversity [21]. Intuitively, the internal diversity metric indicates whether the compound generator repeats similar substructures. We used ECFP8 (extended-connectivity fingerprint with diameter 8 represented using 2048 bits) vector of each compound for calculating internal and external diversities.

  • Solubility: measures the likeliness for a compound to mix with water. This is also referred to as the water-octanol partition coefficient.

  • Naturalness: measures how similar a generated compound is to the structure space of natural products (NPs). NPs are small molecules that are synthesized by living organisms and are viewed as good starting points for drug development [42].

  • Synthesizability: measures how a compound lends itself to chemical synthesis (0 indicates hard to make and 1 indicates easy to make) [43].

  • Druglikeness: estimates the plausibility of a generated compound being a drug candidate. The synthesizability and solubility of a compound contribute to the compound’s druglikeness.

Discussion

We generated |$10~000$| samples for each trained generator and used the molecular metrics in Section 3.6.5 to assess the generator’s performance. However, during training, we generated |$200$| samples and maintained an exponential average to monitor performance. Figure 4 shows the density plots of each model’s valid SMILES string and the convergence progression of each model toward the threshold of an experiment, beginning from the score of samples generated by the pretrained model.

Also, Table 3 presents the results of each metric for the valid SMILES samples of each model. Likewise, Table 4 presents the results for a set of compounds filtered from the valid SMILES samples of each model by applying the threshold of each experiment. In the case of the logP results in Table 4, we selected compounds whose values are less than 5 and greater than 1. The added proportion in threshold column in Table 4 presents the quota of a generator’s valid SMILES that were within the threshold of the experiment; the maximum value is 1 and minimum is 0.

Table 4

Results of experiments with applied optimization threshold on generated or dataset samples. PPO-GRL follows our proposed approach, REINFORCE follows the work in [15], PPO-Eval, REINFORCE-GRL, Stack-RNN-TL are baselines in this study

ObjectiveAlgorithm/ datasetNumber of unique canonical SMILESProportion in thresholdInternal diversityExternal diversitySolubilityNaturalnessSynthesizabilityDruglikeness
Demonstrations49410.6350.8960.7930.5240.6420.588
Unbiased5280.2660.9190.4120.6910.5860.5880.609
PPO-GRL24060.7370.8640.1040.9390.7090.5080.181
DRD2PPO-Eval34650.7640.8490.0920.9470.7560.4580.144
REINFORCE50050.6770.8880.2220.8370.7150.7060.392
REINFORCE-GRL2010.3720.9070.2230.9290.8530.2110.194
Stack-RNN-TL17150.2480.9200.4470.6960.5860.5850.608
Demonstrations50071.0000.8800.8980.5530.8030.512
Unbiased19030.9270.9150.3460.6870.5460.6520.605
PPO-GRL45480.9880.9030.1490.8420.5790.7720.493
LogPPPO-Eval49690.9990.7330.0610.9830.7310.4490.075
REINFORCE76380.9910.8970.0980.7770.5790.8290.618
REINFORCE-GRL67960.9410.9140.3170.6860.5460.6980.630
Stack-RNN-TL64790.9350.9160.3640.6780.5450.6550.614
Demonstrations34490.9580.7940.0730.4820.5600.479
Unbiased7170.3540.9160.3600.6390.5720.5650.572
PPO-GRL27680.4680.9250.5430.5780.7000.3760.533
JAK2 MaxPPO-Eval24270.3500.9160.3690.6400.5640.5650.576
REINFORCE29420.4340.9120.2810.5860.6090.4620.605
REINFORCE-GRL25080.3560.9160.3550.6440.5610.5680.577
Stack-RNN-TL23760.3430.9170.3700.6500.5570.5780.580
Demonstrations1450.5180.7850.5390.5060.9080.617
Unbiased1230.0620.8870.0750.6800.4220.8270.655
PPO-GRL6930.2010.880.0610.5340.4530.8520.673
JAK2 MinPPO-Eval100.0070.6180.0001.0000.7280.3870.050
REINFORCE10780.1400.8910.0860.6810.5000.8870.677
REINFORCE-GRL4900.0700.8970.1420.6820.4320.8310.671
Stack-RNN-TL4260.0.0610.8970.1300.6830.4300.8320.662
ObjectiveAlgorithm/ datasetNumber of unique canonical SMILESProportion in thresholdInternal diversityExternal diversitySolubilityNaturalnessSynthesizabilityDruglikeness
Demonstrations49410.6350.8960.7930.5240.6420.588
Unbiased5280.2660.9190.4120.6910.5860.5880.609
PPO-GRL24060.7370.8640.1040.9390.7090.5080.181
DRD2PPO-Eval34650.7640.8490.0920.9470.7560.4580.144
REINFORCE50050.6770.8880.2220.8370.7150.7060.392
REINFORCE-GRL2010.3720.9070.2230.9290.8530.2110.194
Stack-RNN-TL17150.2480.9200.4470.6960.5860.5850.608
Demonstrations50071.0000.8800.8980.5530.8030.512
Unbiased19030.9270.9150.3460.6870.5460.6520.605
PPO-GRL45480.9880.9030.1490.8420.5790.7720.493
LogPPPO-Eval49690.9990.7330.0610.9830.7310.4490.075
REINFORCE76380.9910.8970.0980.7770.5790.8290.618
REINFORCE-GRL67960.9410.9140.3170.6860.5460.6980.630
Stack-RNN-TL64790.9350.9160.3640.6780.5450.6550.614
Demonstrations34490.9580.7940.0730.4820.5600.479
Unbiased7170.3540.9160.3600.6390.5720.5650.572
PPO-GRL27680.4680.9250.5430.5780.7000.3760.533
JAK2 MaxPPO-Eval24270.3500.9160.3690.6400.5640.5650.576
REINFORCE29420.4340.9120.2810.5860.6090.4620.605
REINFORCE-GRL25080.3560.9160.3550.6440.5610.5680.577
Stack-RNN-TL23760.3430.9170.3700.6500.5570.5780.580
Demonstrations1450.5180.7850.5390.5060.9080.617
Unbiased1230.0620.8870.0750.6800.4220.8270.655
PPO-GRL6930.2010.880.0610.5340.4530.8520.673
JAK2 MinPPO-Eval100.0070.6180.0001.0000.7280.3870.050
REINFORCE10780.1400.8910.0860.6810.5000.8870.677
REINFORCE-GRL4900.0700.8970.1420.6820.4320.8310.671
Stack-RNN-TL4260.0.0610.8970.1300.6830.4300.8320.662
Table 4

Results of experiments with applied optimization threshold on generated or dataset samples. PPO-GRL follows our proposed approach, REINFORCE follows the work in [15], PPO-Eval, REINFORCE-GRL, Stack-RNN-TL are baselines in this study

ObjectiveAlgorithm/ datasetNumber of unique canonical SMILESProportion in thresholdInternal diversityExternal diversitySolubilityNaturalnessSynthesizabilityDruglikeness
Demonstrations49410.6350.8960.7930.5240.6420.588
Unbiased5280.2660.9190.4120.6910.5860.5880.609
PPO-GRL24060.7370.8640.1040.9390.7090.5080.181
DRD2PPO-Eval34650.7640.8490.0920.9470.7560.4580.144
REINFORCE50050.6770.8880.2220.8370.7150.7060.392
REINFORCE-GRL2010.3720.9070.2230.9290.8530.2110.194
Stack-RNN-TL17150.2480.9200.4470.6960.5860.5850.608
Demonstrations50071.0000.8800.8980.5530.8030.512
Unbiased19030.9270.9150.3460.6870.5460.6520.605
PPO-GRL45480.9880.9030.1490.8420.5790.7720.493
LogPPPO-Eval49690.9990.7330.0610.9830.7310.4490.075
REINFORCE76380.9910.8970.0980.7770.5790.8290.618
REINFORCE-GRL67960.9410.9140.3170.6860.5460.6980.630
Stack-RNN-TL64790.9350.9160.3640.6780.5450.6550.614
Demonstrations34490.9580.7940.0730.4820.5600.479
Unbiased7170.3540.9160.3600.6390.5720.5650.572
PPO-GRL27680.4680.9250.5430.5780.7000.3760.533
JAK2 MaxPPO-Eval24270.3500.9160.3690.6400.5640.5650.576
REINFORCE29420.4340.9120.2810.5860.6090.4620.605
REINFORCE-GRL25080.3560.9160.3550.6440.5610.5680.577
Stack-RNN-TL23760.3430.9170.3700.6500.5570.5780.580
Demonstrations1450.5180.7850.5390.5060.9080.617
Unbiased1230.0620.8870.0750.6800.4220.8270.655
PPO-GRL6930.2010.880.0610.5340.4530.8520.673
JAK2 MinPPO-Eval100.0070.6180.0001.0000.7280.3870.050
REINFORCE10780.1400.8910.0860.6810.5000.8870.677
REINFORCE-GRL4900.0700.8970.1420.6820.4320.8310.671
Stack-RNN-TL4260.0.0610.8970.1300.6830.4300.8320.662
ObjectiveAlgorithm/ datasetNumber of unique canonical SMILESProportion in thresholdInternal diversityExternal diversitySolubilityNaturalnessSynthesizabilityDruglikeness
Demonstrations49410.6350.8960.7930.5240.6420.588
Unbiased5280.2660.9190.4120.6910.5860.5880.609
PPO-GRL24060.7370.8640.1040.9390.7090.5080.181
DRD2PPO-Eval34650.7640.8490.0920.9470.7560.4580.144
REINFORCE50050.6770.8880.2220.8370.7150.7060.392
REINFORCE-GRL2010.3720.9070.2230.9290.8530.2110.194
Stack-RNN-TL17150.2480.9200.4470.6960.5860.5850.608
Demonstrations50071.0000.8800.8980.5530.8030.512
Unbiased19030.9270.9150.3460.6870.5460.6520.605
PPO-GRL45480.9880.9030.1490.8420.5790.7720.493
LogPPPO-Eval49690.9990.7330.0610.9830.7310.4490.075
REINFORCE76380.9910.8970.0980.7770.5790.8290.618
REINFORCE-GRL67960.9410.9140.3170.6860.5460.6980.630
Stack-RNN-TL64790.9350.9160.3640.6780.5450.6550.614
Demonstrations34490.9580.7940.0730.4820.5600.479
Unbiased7170.3540.9160.3600.6390.5720.5650.572
PPO-GRL27680.4680.9250.5430.5780.7000.3760.533
JAK2 MaxPPO-Eval24270.3500.9160.3690.6400.5640.5650.576
REINFORCE29420.4340.9120.2810.5860.6090.4620.605
REINFORCE-GRL25080.3560.9160.3550.6440.5610.5680.577
Stack-RNN-TL23760.3430.9170.3700.6500.5570.5780.580
Demonstrations1450.5180.7850.5390.5060.9080.617
Unbiased1230.0620.8870.0750.6800.4220.8270.655
PPO-GRL6930.2010.880.0610.5340.4530.8520.673
JAK2 MinPPO-Eval100.0070.6180.0001.0000.7280.3870.050
REINFORCE10780.1400.8910.0860.6810.5000.8870.677
REINFORCE-GRL4900.0700.8970.1420.6820.4320.8310.671
Stack-RNN-TL4260.0.0610.8970.1300.6830.4300.8320.662

Firstly, we observed from the results that the PPO-GRL model focused on generating compounds that either satisfy the demonstration dataset threshold or are toward the threshold after a few episodes during training in each of the experiments. This early convergence of the PPO-GRL generator (in terms of number of episodes) connotes that an appropriate reward function has been retrieved from the demonstration dataset to facilitate the biasing optimization using RL. Also, while the evaluation function provided an appropriate signal for training the PPO-Eval in some cases, the diversity metrics of the PPO-Eval model were typically less than the PPO-GRL model. We also realized in training that generating focused compounds toward a threshold was accompanied by an increase in the number of invalid SMILES strings. This increase in invalid compounds explains the drop in the number of valid SMILES of PPO-GRL in Table 3. Although the REINFORCE model mostly generated a high number of valid SMILES samples, it was less sample efficient. We reckon that the difference in sample efficiency between PPO-GRL and REINFORCE-GRL is as a result of the variance in estimating |$\Psi $| (see Equation (5)). Thus, a stable estimation of |$\Psi $| is significant to the sample-based objective in Equation (18) since a better background distribution could be estimated. The performance of the REINFORCE-GRL, as shown in Figure 4, reifies this high-variance challenge. Also, the Stack-RNN-TL model recorded the same scores for all metrics across the experiments, as seen in Table 3. This performance of the Stack-RNN-TL model connotes that no focused compounds could be generated in each experiment after two epochs of training but mostly took longer time to train as compared with the other models. We also note that the Stack-RNN-TL model produced a higher number of valid canonical SMILES than the unbiased or prior generator after the fine-tuning step.

Concerning the DRD2 experiment, although the PPO-GRL model’s number of valid SMILES strings was more than the number of unique compounds in the unbiased dataset, it performed approximately one-third lesser than the PPO-Eval model. As shown in Figure 4a, this is observed in the mean predicted DRD2 activity of the PPO-Eval model, which reaches the threshold in fewer episodes than the PPO-GRL and REINFORCE models. The PPO-GRL model produced a higher proportion of its valid compounds in the activity threshold than the REINFORCE model, and the generated samples seem to share a relatively lesser number of substructures with the demonstrations set (external diversity) than the PPO-Eval approach as reported in Table 4. Unsurprisingly, due to the variance problem mentioned earlier, the REINFORCE-GRL performed poorly and generated the fewest number of valid compounds with more than half of its produced SMILES being less than the DRD2 activity threshold.

Regarding the logP optimization experiment, most of the compounds sampled from all the generators had logP values that fell within the range of |$1$| and |$5$|⁠. However, while the REINFORCE-GRL and Stack-RNN-TL models recorded an average logP value of approximately |$1.6$|⁠, the PPO-GRL, PPO-Eval and REINFORCE models recorded higher average logP values closer to the demonstration dataset’s average logP value, as shown in Figure 4b. Considering that the PPO-GRL model was trained without the reward function used to train the PPO-Eval and REINFORCE generators, our proposed approach was effective at recovering a proper reward function for biasing the generator. Interestingly, the samples of the PPO-GRL generator recorded better diversity scores and were deemed more druglike than the PPO-Eval generator’s samples, as shown in Table 3.

On the JAK2 maximization experiment, the PPO-Eval method could not reach the threshold. The PPO-GRL generator reached the threshold in less than |$100$| episodes but with fewer valid SMILES strings than the REINFORCE generator. As shown in Table 4, the proportion of compounds in the JAK2 max threshold of the PPO-GRL generated samples, when compared with the other models in the JAK2 maximization experiment, indicates that the PPO-GRL model seems to focus on the objective early in training than the other generators.

On the other hand, the JAK2 minimization experiment provides an insightful view of the PPO-GRL behavior despite its inability, as well as the other models, to reach the threshold in Figure 4d. In Figure 5, we show the size of the demonstrations dataset in each experiment and the number of compounds that satisfy the experiment’s threshold. We think of the proportion of each demonstration dataset satisfying the threshold as vital to the learning objective, and hence, fewer numbers could make learning a useful reward function more challenging. Hence, the size and quality of the demonstration dataset contribute significantly to learning a useful reward function. We suggest this explains the PPO-GRL generator’s inability to reach the JAK2 minimization threshold. Therefore, it is no surprise that the REINFORCE-GRL generator’s best mean pIC|$_{50}$| value was approximately the same as that of the pretrained model’s score, as seen in Figure 4d. We note that the number of unique PPO-GRL generated compounds was almost five times larger than the demonstrations with approximately the same internal variation. It is worth noting that while the REINFORCE approach used the JAK2 minimization reward function of [15], the PPO-GRL method used the negated rewards of the learned JAK2 maximization reward function. This ability to transfer learned reward functions could be a useful technique in drug discovery campaigns.

The size of the demonstrations dataset used in each experiment of this study.
Fig. 5

The size of the demonstrations dataset used in each experiment of this study.

In a nutshell, the preceding results and analysis show that our proposed framework offers a rational approach to train compound generators and learn a reward function, transferable to related domains, in situations where specifying the reward function for RL training is challenging or not possible. While the PPO-Eval model can perform well in some instances, the evaluation function may not be readily available or expensive to serve in the training loop as a reward function in some real-world scenarios such as a robot synthesis or molecular docking evaluation function.

Conclusion

This study reviewed the importance of chemical libraries to the drug discovery process and discussed some notable existing proposals in the literature for evolving chemical structures to satisfy specific objectives. We pointed out that while RL methods could facilitate this process, specifying the reward function in certain drug discovery studies could be challenging, and its absence renders even the most potent RL technique inapplicable. Our study has proposed a reward function learning and a structural evolution model development framework based on the entropy maximization IRL method. The experiments conducted have shown the promise such a direction offers in the face of the large-scale chemical data repositories that have become available in recent times. We conclude that further studies into improving the proposed method could provide a powerful technique to aid drug discovery.

Further Studies

An area for further research could be techniques for reducing the number of invalid SMILES while the generator focuses on the training objective. An approach that could be investigated is a training method that actively teaches the reward function to distinguish between valid and invalid SMILES strings. Also, a study to provide a principled understanding of developing the demonstrations dataset could be an exciting and useful direction. Additionally, considering the impact of transformer models in DL research, the gated transformer proposed by Parisotto et al. [44] to extend the capabilities of transformer models to the RL domain offers the opportunity to develop better compound generators. In particular, our work could be extended by replacing the Stack-RNN used to model the generator with a memory-based gated transformer.

Summary
  • The size and quality of chemical libraries to the drug discovery pipeline are crucial for drug design.

  • Existing reinforcement learning-based de novo drug design methods require a reward function that could be challenging to specify or design.

  • We propose a method for learning a transferable reward function from a given set of compounds.

  • The proposed training approach produces a compound generator and a reward function.

  • Our findings indicate that the proposed method provides a rational alternative for scenarios where the reward function design could be challenging.

Data Availability

The source code and data of this study are available at https://github.com/bbrighttaer/irelease.

Acknowledgments

We would like to thank Zhihua Lei, Kwadwo Boafo Debrah and all reviewers of this study.

Funding

This work was partly supported by SipingSoft Co. Ltd.

Conflict of interest

The authors declare that they have no competing interests.

Brighter Agyemang completed is a research assistant at SipingSoft Co. Ltd and a PhD candidate at the University of Electronic Science and Technology of China (UESTC). His current research interests are reinforcement learning, machine learning, the internet of things and cheminformatics.

Wei-Ping Wu received a PhD in metallurgy from RWTH Aachen, Germany, in 1992. He is a professor at UESTC. His research interests include the application of artificial intelligence in drug design, IoT and mechanical manufacturing.

Daniel Addo is a PhD student at UESTC. His current research interests are machine learning, internet of things and security in deep brain implants.

Michael Y. Kpiebaareh is a PhD candidate at UESTC. His current research interests include recommendation systems for configurable product design engineering and machine learning for mass customization.

Ebenezer Nanor is a research assistant at Sipingsoft Co. Ltd, China. His research interests include the development of oversampling techniques for imbalanced and high-dimensional datasets for drug design.

Charles Roland Haruna received a PhD in computer science and technology at UESTC in 2020. His research interests are distributed data, data cleaning and data privacy.

References

1.

Olivecrona
M
,
Blaschke
T
,
Engkvist
O
, et al.
Molecular de-novo design through deep reinforcement learning
.
J Chem
2017
;
9
(
1
):
1
14
.

2.

Ivanenkov
YA
,
Aladinskiy
VA
,
Bushkov
NA
, et al.
Small-molecule inhibitors of hepatitis c virus (HCV) non-structural protein 5a (ns5a): a patent review (2010–2015)
.
Expert Opin Ther Pat
2017
;
27
(
4
):
401
14
.
PMID: 27967269
.

3.

Segler
MHS
,
Kogej
T
,
Tyrchan
C
, et al.
Generating focused molecule libraries for drug discovery with recurrent neural networks
.
ACS Cent Sci
2018
;
4
(
1
):
120
31
.

4.

Polishchuk
PG
,
Madzhidov
TI
,
Varnek
A
.
Estimation of the size of drug-like chemical space based on GDB-17 data
.
J Comput Aided Mol Des
2013
;
27
(
8
):
675
9
.

5.

Hingorani
AD
,
Kuan
V
,
Finan
C
, et al.
Improving the odds of drug development success through human genomics: modelling study
.
Sci Rep
2019
;
9
(
1
):
1
25
.

6.

Bahdanau
D
,
Cho
K
,
Bengio
Y
.
Neural machine translation by jointly learning to align and translate
.
arXiv
2014
;
1
15
.

7.

Radford
A
,
Wu
J
,
Child
R
, et al.
Language models are unsupervised multitask learners
.
OpenAI Technical Report
, San Francisco: OpenAI Blog,
2018
.

8.

Xu
K
,
Ba
J
,
Kiros
R
, et al.
Show, attend and tell: neural image caption generation with visual attention
. In:
Bach F and Blei D (eds)
.
Proceedings of the 32nd International Conference on Machine Learning
, Vol. 37.
Proceedings of Machine Learning Research
.
Lille, France
:
PMLR
,
2015
,
2048
57
.

9.

Ragoza
M
,
Hochuli
J
,
Idrobo
E
, et al.
Protein-ligand scoring with convolutional neural networks
.
J Chem Inf Model
2017
;
57
(
4
):
942
57
.

10.

Agyemang
B
,
Wu
W-P
,
Kpiebaareh
MY
, et al.
Multi-view self-attention for interpretable drug-target interaction prediction
.
J Biomed Inform
2020
;
110
:
103547
.

11.

Tsubaki
M
,
Tomii
K
,
Sese
J
.
Compound-protein interaction prediction with end-to-end learning of neural networks for graphs and sequences
.
Bioinformatics
2019
;
35
(
2
):
309
18
.

12.

Segler
MHS
,
Waller
MP
.
Modelling chemical reasoning to predict and invent reactions
.
Chem A Eur J
2017
;
23
(
25
):
6118
28
.

13.

Weininger
D
.
SMILES, a chemical language and information system: 1: introduction to methodology and encoding rules
.
J Chem Inf Comput Sci
1988
;
28
(
1
):
31
6
.

14.

Yu
L
,
Zhang
W
,
Wang
J
,
Yu
Y
.
Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence
San Francisco, California, USA
:
AAAI Press
,
2017
.
2852
2858
.

15.

Popova
M
,
Isayev
O
,
Tropsha
A
.
Deep reinforcement learning for de novo drug design
.
Sci Adv
2018
;
4
(
7
):
1
15
.

16.

Williams
RJ
.
Simple statistical gradient-following methods for connectionist reinforcement learning
.
Mach Learn
1992
;
8
:
229
56
.

17.

Shi
F
,
You
S
,
Chang
X
.
Reinforced molecule generation with heterogeneous states
. In:
Wang J, Shim K, Wu X (eds)
.
IEEE International Conference on Data Mining (ICDM)
. New York City, USA:
IEEE
,
2019
,
548
57
.

18.

Schmidt
F
.
Generalization in generation: a closer look at exposure bias
.
CoRR
. New York, USA: Cornell University,
2019
,
1
12
. arXiv: abs/1910.00292.

19.

Sanchez-Lengeling
B
,
Outeiral
C
,
Guimaraes
GL
, et al.
Optimizing distributions over molecular space. An objective-reinforced generative adversarial network for inverse-design chemistry (ORGANIC)
.
ChemRxiv
2017
;
1
18
. https://chemrxiv.org/articles/ORGANIC_1_pdf/5309668.

20.

Guimaraes
GL
,
Sanchez-Lengeling
B
,
Farias
PLC
, et al.
Objective-reinforced generative adversarial networks (ORGAN) for sequence generation models
.
CoRR
2017
.
abs/1705.10843
,
1
10
.

21.

Benhenda
M
.
ChemGAN challenge for drug discovery: can AI reproduce natural chemical diversity
?
arXiv
2017
. abs/1708.08227,
1
6
.

22.

Putin
E
,
Asadulaev
A
,
Ivanenkov
Y
, et al.
Reinforced adversarial neural computer for de novo molecular design
.
J Chem Inf Model
2018
;
58
(
6
):
1194
204
.

23.

Arjovsky
M
,
Bottou
L
. 5th International Conference on Learning Representations, ICLR 2017.
Towards principled methods for training generative adversarial networks
, Toulon, France: ICLR,
2017
.

24.

Finn
C
,
Levine
S
,
Abbeel
P
.
Guided cost learning: deep inverse optimal control via policy optimization
. In:
Proceedings of the 33rd International Conference on Machine Learning (ICML)
, Vol. 48.
JMLR Workshop and Conference Proceedings
. New York, NY, USA:
JMLR
,
2016
,
49
58
.

25.

Finn
C
,
Christiano
PF
,
Abbeel
P
, et al.
A connection between generative adversarial networks, inverse reinforcement learning, and energy-based models
.
arXiv
2016
.
abs/1611.03852
,
1
10
.

26.

Schulman
J
,
Moritz
P
,
Levine
S
, et al.
High-dimensional continuous control using generalized advantage estimation
. In:
4th International Conference on Learning Representations, {ICLR}
, San Juan, Puerto Rico,
2016
.

27.

Arulkumaran
K
,
Deisenroth
MP
,
Brundage
M
, et al.
A brief survey of deep reinforcement learning
.
arXiv
2017
;
1
14
.

28.

Schulman
J
,
Wolski
F
,
Dhariwal
P
, et al.
Proximal policy optimization algorithms
.
CoRR
arXiv,
2017
.
abs/1707.06347
,
1
12
.

29.

Mnih
V
,
Kavukcuoglu
K
,
Silver
D
, et al.
Human-level control through deep reinforcement learning
.
Nature
2015
;
518
(
7540
):
529
33
.

30.

Schulman
J
,
Levine
S
,
Abbeel
P
, et al.
Trust Region Policy Optimization
. In: Francis Bach and David Blei (eds.),
Proceedings of the 32nd International Conference on Machine Learning
USA
:
PMLR
,
2015
,
37
,
1889
97
.

31.

Arora
S
,
Doshi
P
.
A survey of inverse reinforcement learning: challenges, methods and progress
.
arXiv
2018
. abs/1806.06877,
1
47
.

32.

Ng
A
,
Russell
S
.
Algorithms for inverse reinforcement learning
. In:
Proceedings of the Seventeenth International Conference on Machine Learning
,
Burlington, Massachusetts, USA
:
Morgan Kaufmann, Elsevier
,
2000
,
663
70
.

33.

Ziebart
BD
,
Maas
A
,
Bagnell
JA
, et al.
Maximum entropy inverse reinforcement learning
. In:
AAAI Conference on Artificial Intelligence
, Palo Alto, California, USA: AAAI Press,
2008
,
1433
8
.

34.

Wulfmeier
M
,
Ondr
P
.
Maximum entropy deep inverse reinforcement learning
.
arXiv
2016
. abs/1507.04888,
1
10
.

35.

Ho
J
,
Ermon
S
.
Generative adversarial imitation learning
. In:
Lee DD, Sugiyama M, von Luxburg U, et al. (eds)
.
Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems
. Red Hook, New York, United States: Curran Associates, Inc.,
2016
,
4565
73
.

36.

Knox
C
,
Law
V
,
Jewison
T
, et al.
DrugBank 3.0: a comprehensive resource for ‘Omics’ research on drugs
.
Nucleic Acids Res
Oxford University Press,
2010
,
39
,
D1035
D1041
.

37.

Kanehisa
M
,
Goto
S
,
Sato
Y
, et al.
KEGG for integration and interpretation of large-scale molecular data sets
.
Nucleic Acids Res
Oxford University Press,
2011
;
40
(
D1
):
D109
14
.

38.

Szklarczyk
D
,
Santos
A
,
Von Mering
C
, et al.
STITCH 5: augmenting protein–chemical interaction networks with tissue and affinity data
.
Nucleic Acids Res
Oxford University Press,
2016
,
44
,
D380
4
.

39.

Bento
AP
,
Gaulton
A
,
Hersey
A
, et al.
The ChEMBL bioactivity database: an update
.
Nucleic Acids Res
Oxford University Press,
2014
.
42
,
D1083
D1090
.

40.

Joulin
A
,
Mikolov
T
.
Inferring algorithmic patterns with stack-augmented recurrent nets
. In:
Cortes C, Lawrence ND, Lee DD, et al. (eds)
.
Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems
. Red Hook, New York, United States: Curran Associates, Inc.,
2015
,
190
8
.

41.

Sun
J
,
Jeliazkova
N
,
Chupakin
V
, et al.
ExCAPE-DB: an integrated large scale dataset facilitating big data analysis in chemogenomics
.
J Chem
BMC,
2017
,
9
, 17.

42.

Sorokina
M
,
Steinbeck
C
.
Naples: a natural products likeness scorer-web application and database
.
J Chem
2019
;
11
(
1
):
1
7
.

43.

Ertl
P
,
Schuffenhauer
A
.
Estimation of synthetic accessibility score of drug-like molecules based on molecular complexity and fragment contributions
.
J Chem
BMC,
2009
,
1
, 8.

44.

Parisotto
E
,
Song
HF
,
Rae
JW
, et al.
Stabilizing transformers for reinforcement learning
.
CoRR
Cornell University, New York, USA: arXiv,
2019
, abs/1910.06764,
1
20
.

This article is published and distributed under the terms of the Oxford University Press, Standard Journals Publication Model (https://dbpia.nl.go.kr/journals/pages/open_access/funder_policies/chorus/standard_publication_model)

Supplementary data