Computational and Mathematical Methods in Medicine

Volume 2014, Article ID 258627, 8 pages

http://dx.doi.org/10.1155/2014/258627

## Designing Lead Optimisation of MMP-12 Inhibitors

^{1}European Centre for Living Technology, 30124 Venice, Italy^{2}Department of Environmental Sciences, Informatics and Statistics, Ca’ Foscari University of Venice, 30123 Venice, Italy

Received 9 October 2013; Revised 16 December 2013; Accepted 16 December 2013; Published 12 January 2014

Academic Editor: Rudolf Füchslin

Copyright © 2014 Matteo Borrotti et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

The design of new molecules with desired properties is in general a very difficult problem, involving heavy experimentation with high investment of resources and possible negative impact on the environment. The standard approach consists of iteration among formulation, synthesis, and testing cycles, which is a very long and laborious process. In this paper we address the so-called lead optimisation process by developing a new strategy to design experiments and modelling data, namely, the evolutionary model-based design for optimisation (EDO). This approach is developed on a very small set of experimental points, which change in relation to the response of the experimentation according to the principle of evolution and insights gained through statistical models. This new procedure is validated on a data set provided as test environment by Pickett et al. (2011), and the results are analysed and compared to the genetic algorithm optimisation (GAO) as a benchmark. The very good performance of the EDO approach is shown in its capacity to uncover the optimum value using a very limited set of experimental points, avoiding unnecessary experimentation.

#### 1. Introduction

Designing molecules with particular properties is usually a long and complex process, in which the nonlinearity of the model, the high number of variables with a leading role, and the categorical structure of these variables can make difficult modelling, experimentation, and analysis. In new drug discovery, a key phase concerns the generation of small molecules modulators of protein function, under the hypothesis that this activity can affect a particular disease state. Current practices rely on the screening of vast libraries of small molecules (often 1-2 million molecules) in order to identify a molecule that specifically inhibits or activates the protein function, commonly known as the lead molecule. The lead molecule interacts with the required target, but it generally lacks the other attributes needed for a drug candidate such as absorption, distribution, metabolism, and excretion (ADME). In order to achieve these attributes, retaining the interaction capacity with the target protein, the lead molecule must be modified. This transformation of the lead molecule is known as lead optimisation. Lead optimisation research involves long synthesis and testing cycles, analyses of the structure-activity relationships (SAR), and quantitative structure activity relationships (QSAR), which are currently the bottleneck of this process [1]. Under traditional approaches these analyses are conducted by experimentation involving an extremely large number of experimental units, which requires large investments of resources and time to reach the target and measure the possible impact on the environment. Computational approaches for the SAR and QSAR analyses, mostly based upon machine learning techniques, have been proposed over the last few years [2–5]. Search and optimisation algorithms inspired by evolution have been also developed and applied with success to drug discovery process and related activities. In Clark [6], different evolutionary algorithms are presented and discussed, such as genetic algorithms (GAs), evolutionary programming (EPs), and evolution strategies (ESs). In this work, several applications of these computational methods have been derived for a wide range of research. Other bio-inspired algorithms such as Ant Colony Optimisation (ACO) and Artificial Neural Networks (ANNs) have been applied in drug discovery [7]. The evolutionary principle is the basic structure of a new approach proposed for designing experiments in an efficient way [8–12]. In this paper we would like to contribute to the development of this research by proposing a new procedure with the objective of finding the optimal value of MMP-12 conducting a very small number of tests and thus with small investments of resources and limited negative impact on the environment. This new procedure is an evolutionary model-based design of experiments: the search for the optimum value is restricted to relatively few experimental points, chosen with the evolutionary paradigm and the information provided by statistical models. Starting from lead molecules, randomization augmented by expert knowledge is used to choose the initial set of compositions to be tested in the laboratory. After chemical synthesis and in vitro screening of these molecules, the resulting response data are evaluated with respect to their capacity to reach the target. They are then transformed according to the operators involved in the evolutionary search and to the information from statistical models estimated on the data. Successive populations of molecules are analysed, modelled, and transformed to generate compositions that are closer to the optimum value. The procedure will be developed and validated, and its efficiency will be measured by a suitable index. Given the successful performance of the simple genetic algorithm for lead optimisation of MMP-12 inhibitors developed by Pickett et al. [1], we will compare our procedure with GAO on this problem. To allow the comparison we will consider the same number of experimental points considered in Pickett et al. [1]. Results exhibit a better performance of our approach in reaching the optimum value and reducing the number of experiments. The paper is organised as follows: in Section 2 we describe the data set on which we developed the procedure and the key idea of the proposed design. In Section 3 we present the results and make comparison with the GAO approach. Section 4 offers some concluding remarks.

#### 2. Materials and Methods

##### 2.1. Data Set

We build the design for optimisation using a data set presented and analysed by Pickett et al. [1]. These data, available at http://pubs.acs.org, have been constructed by the authors as a test environment on which assessing the effectiveness and the efficiency of new designs for lead optimisations. The data concern a library of 2500 molecules, identified by their chemical compositions (*reagents*) and their experimental response (*activity*). These data represent the whole experimental space. Each data point, coding a particular experiment, is described by two categorical variables, which represent the reagents, each of which can assume 50 different values. The response variable measures the molecular activity of the reaction product. The aim of the analysis is to find the reaction whose product maximises the molecular activity. As a first exploratory analysis of these data we compute a set of descriptive statistics to get some insights into the frequency distribution of the response variable. These data are represented in Figure 1, reporting the histogram and the boxplot. From the exploratory analysis we learn that the maximum value of the molecular activity is 8.00 and the minimum value is 3.40. The mean of the response is equal to 5.28 and the median 5.20; the first quartile is 4.40 and the third quartile is 6.20. These values indicate a right-skewed distribution, and this particular shape is clearly shown in the histogram. In designing experiments for optimisation we will have the objective to find the experimental point with the maximum value of 8.00.

To describe the behaviour of the response in relation to the molecule composition, we built a *heatmap plot* as presented in Figure 2. The two reagents, say reagents A and B, are reported in the axes of the plot with 50 levels each. For each combination of the reagent levels we can read the value of the molecular activity: high activity values are represented by dark blue squares and small activity values by light blue squares. White squares indicate molecules for which the response is not available. From Figure 2 we can see that just two molecule compositions reach the maximum value of 8.00 (A21; B07 and A31; B25, marked by red circle); also we notice that some reagents, B16 and B20, can give rise to molecules with very high activity values.

##### 2.2. Design for Optimisation

An *optimisation* problem is commonly described as follows.

Let be a subset of the Euclidean space and let be the function . Let be a point in ; can affect the response variable , and the response can then be described as . The optimisation problem consists in searching the element in such that for all in .

For the lead optimisation problem that we are addressing in this research the dimension equals 2, since two variables are considered in the dependence relation.

These variables are categorical variables, namely, the reagents, and can take different levels.

With this setting of the problem the experiment is run to provide each experimental point with a response value that is a measure of the activity of the resulting molecule. The experimental data set is then , with being an matrix, where is the size of , and being an vector. This data set represents the evidence for inferring the dependence relation among variables and identifies the design point that gives the optimum value of the response.

The *design* problem for optimisation consists in finding a small set of experimental points that contain the relevant information to reach the optimal value. Our contribution to address this problem is to adopt evolution as a paradigm to build a design approach guiding the evolution with the information achieved by statistical models.

##### 2.3. Evolutionary Design for Optimisation

Building on the evolutionary strategy we introduce a new approach, named evolutionary design for optimisation (EDO), testing a very small set of different experimental points able to find the optimal response value or the region of optimality. This approach evolves an initial design through generations by means of a set of genetic operators (selection, recombination, and mutation) that are built on the information provided by models estimated on the data of each design generation , .

For the lead optimisation problem addressed in this research we build the EDO approach with the objective of achieving the optimum value of the molecule activity testing 140 experimental points as in genetic algorithm optimisation (GAO) strategy introduced by [1]. More specifically, we select an initial population of experimental points , consisting of two sets of compounds: the first set is created by assigning at each level of reagent A a randomly selected level of reagent B; then, the second set is created by assigning at each level of reagent B a randomly selected level of reagent A. Each of these sets include 50 compounds, and then the design consists of different compounds. Each of these experimental points receives a response value . Therefore on the data set , a statistical model is estimated to achieve information on the goodness of these compositions in reaching the target of the optimisation. In this research we developed several Monte Carlo simulation studies comparing different classes of statistical models in their predictive capacity, and we selected the random forest model [13–15] as our best choice. The random forests are regression methods frequently used when the relationship between response and predictors is complex, and the predictors are categorical variables [16].

Following the evolutionary paradigm, we then adopt a selection operator where the probability of each experimental point to be selected for next generation is proportional to the square of the response value, according to the following expression: where is the total number of experimental points tested up to and including the current generation. For this optimisation problem we select 10 compounds, and each selected compound is then recombined in order to create a set of new and not already tested points. We estimate a random forest model and proceed in the following way: for each selected compound we fix the reagent, randomly chosen between A and B, and its corresponding level in the compound, and then we generate all the possible compounds by changing the 50 levels of the other reagent. For all these new generated experimental points we then predict the responses using the estimated random forest model and the compound with the highest estimated response value is considered for the following generation. A mutation operator is then performed (with probability ) by randomly changing one reagent level. The second population of experimental points is then defined. We iterate the procedure across generations until the optimum value is archived or a stopping rule is satisfied.

The EDO procedure is represented in Figure 3 and described as follows.(1)Create a population of compounds ().(2)Conduct experimentation and evaluate the response.(3)Estimate the statistical model (i.e., random forest).(4)Select a compound (according to (1)).(5)Combine the reagents using EDO crossover as follows:(i)select a reagent in a random way;(ii)generate all the possible compounds by changing the 50 levels of the other reagent;(iii)infer the molecules activity with the estimated model (random forest);(iv)select the compound with the highest estimated activity value for next generation.(6)Repeat step’s 4 and 5 until the 10 new compounds are created.(7)Mutate the new compounds with .(8)Conduct experimentation and evaluate the response.(9)If number of generations is equal to stop the algorithm. Otherwise repeat steps from 3 to 8.

The EDO procedure is developed in R code (http://cran.r-project.org/) and uses randomForest package [16]. Random forest model is estimated running 500 trees, and model selection is performed with standard parameterisation of the package.

##### 2.4. Measure of the Design Goodness

To evaluate the design goodness we introduce two criteria. The first criterion is a measure of the distance between the response value of the best experimental point provided by the design and the actual optimal value of the whole system response. In particular, let be the maximum value found by the design, and let and be the known maximum and minimum of on the whole search space. The design goodness for optimisation criterion (DGO) is

This measure ranges in value from 0 to 1. The second criterion of design goodness evaluates the capacity of the approach to find response values in defined regions of optimality. We derive this indicator by counting the number of experimental points with response value greater than a defined threshold. This threshold is identified by the right tail area of the response values distribution measured by the probability values and . The can be expressed as follows: where is the total number of tested compounds, is the number of compounds of the whole experimental space, is the percentile of the response distribution at level, and is the indicator function. The ranges from 0 to 1, where indicates that no compound selected by the design is in the optimal region and indicates that all the selected compounds are in the optimal region.

#### 3. Results

To derive an efficient design for the lead optimisation problem we apply the EDO design and search in the experimental space of 2500 compounds for the optimum value. The initial population of compounds sampled from the whole experimental space (as in Section 2.3) are spread in the response distribution as described in Figure 4.

This result and the evolution of the experimental response achieved by generations are shown in Figure 5 (response values greater than 6) where we notice that EDO finds the optimal value of 8.00 (global optimum of the whole experimental space) at the third generation. We also notice that this approach is able to find a set of very good values close to the optimum.

To evaluate the performance of the procedure we developed a comparison of the EDO approach with the GAO, which is considered as a benchmark for this new approach. Reporting the response values by generations of the GAO in Figure 6 (response values greater than 6) we observe that the simple GA, without the statistical modelling contribution, is not able to find the optimum value testing 140 compounds and conducting 10 generations of experiments. Moreover we notice that most of the GAO response values remain under the threshold of 7.50.

Since the GAO approach has shown very good performance with respect to the traditional approach in lead optimisation, the results achieved under EDO design can be regarded as satisfactory. The EDO design on this set of data discovers the global optimum testing just 120 compounds.

Computing the design goodness criterion presented in (2) we compare the optimal response values achieved by the GAO, , and by EDO, , with the known optimum value of the whole space and derive the following measures of design goodness: confirming the superior performance of the EDO design.

As a principal result the new procedure has been able to discover this value testing just 120 compounds and conducting 3 generations of the algorithm. Furthermore, the comparison in performance between EDO and GAO can be realised considering a region of optimality instead of the single optimal value.

Deriving the measure , as presented in (3) for both optimisation approaches, we obtain that, in the right tail region of the distribution of the responses with , the EDO approach can find 29 of the best compounds, while the GAO is able to discover just 17 of these compounds. We achieve similar results considering the region of the right tail distribution , where EDO approach outperforms the results of GAO finding 21 of the best experimental units. These results are reported in Table 1.

We test the statistical significance of the null hypothesis that EDO and GAO have equal proportion of responses in the best response region. From this statistical test, we compute the one-tailed value to evaluate the improvement of EDO compared to GAO: we obtain value = 0.0991 considering the optimality region and value = 0.1361 considering the optimality region . These statistical tests confirm the improvement of EDO design compared to GAO design.

Studying the evolution of the proportion of the best compounds found in the optimality area and described in Figure 7, we notice that this proportion increases much faster and more intensively for the EDO design than for the GAO design. Selecting the region of optimality (Figure 7(a)) the number of responses from the EDO design that fall in this area increases rapidly reaching in 5 generations 29 of the best responses, instead of the 17 of the GAO design. Similar behaviour can be observed for the region of optimality with size (Figure 7(b)). Finally we derived the frequency distributions of the best response values () comparing the EDO and GAO optimisation procedures, as described in Figure 8. We can observe that the proportion of the compounds achieved with EDO design (blue bars) grows for increasing values of the responses. Moreover for values greater than 7.5 this proportion is very high and is much higher than the proportion achieved with GAO procedure.

In order to study the robustness of EDO design with respect to changes in the initial population, we performed a simulative study where we run our algorithm 100 times with different initial populations.

As a result, 90 of the simulations have been able to find greater or equal response values with respect to the best result obtained in Pickett et al. [1]. This result shows the robustness of the EDO with respect to the choice of the initial population confirming the good performance of the approach.

#### 4. Concluding Remarks

In this research we addressed the lead optimisation problem for drug discovery process by developing a design for experiments which is evolutionary and based on the information provided by statistical models. The motivation of this research is to give a contribution to the study of finding an efficient design that tests a very small set of experimental points instead of the whole space, which due to the high dimensionality of the system or the high number of the variable levels may be very large.

The approach that we derived outperforms the GAO methodology developed by Pickett et al. [1]. In fact selecting 120 experimental points from the whole search space, EDO is able to find the global optimum value. These results suggest that the development of an evolutionary design as in GAO is certainly successful in optimisation problems, but the introduction of statistical models at each step of the evolution as in EDO can improve the optimisation procedure.

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of the paper.

#### Acknowledgments

The authors would like to acknowledge the European Centre for Living Technology (www.ecltech.org) for providing opportunities of presentations and fruitful discussions of the research. Thanks are due to the optimisation team in GlaxoSmithKline (www.gsk.com) led by Dr. Darren Green and Professor Philip Brown from University of Kent (www.kent.ac.uk) for valuable suggestions to this work. The authors gratefully acknowledge the two anonymous reviewers and the editor for their helpful comments.

#### References

- S. D. Pickett, D. V. S. Green, D. L. Hunt, D. A. Pardoe, and I. Hughes, “Automated lead optimization of MMP-12 inhibitors using a genetic algorithm,”
*ACS Medicinal Chemistry Letters*, vol. 2, no. 1, pp. 28–33, 2011. View at Publisher · View at Google Scholar · View at Scopus - A. Z. Dudek, T. Arodz, and J. Gálvez, “Computational methods in developing quantitative structure-activity relationships (QSAR): a review,”
*Combinatorial Chemistry and High Throughput Screening*, vol. 9, no. 3, pp. 213–228, 2006. View at Publisher · View at Google Scholar · View at Scopus - M. Butkiewicz, R. Mueller, D. Selic, E. Dawson, and J. Meiler, “Application of machine learning approaches on quantitative structure activity relationships,” in
*Proceedings of the IEEE Symposium on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB '09)*, pp. 255–262, April 2009. View at Publisher · View at Google Scholar · View at Scopus - J. C. Stalring, L. A. Carlsson, P. Almeida, and S. Boyer, “AZOrange—high performance open source machine learning for QSAR modeling in a graphical programming environment,”
*Journal of Cheminformatics*, vol. 3, no. 28, 2011. View at Google Scholar - R. Cox, D. V. S. Green, C. N. Luscombe, N. Malcolm, and S. D. Pickett, “QSAR workbench: automating QSAR modeling to drive compound design,”
*Journal of Computer-Aided Molecular Design*, vol. 27, no. 4, pp. 321–336, 2013. View at Google Scholar - D. E. Clark,
*Evolutionary Algorithms in Molecular Design*, John Wiley and Sons, 2008. - T. Solmajer and J. Zupan, “Optimization algorithms and natural computing in drug discovery,”
*Drug Discovery Today*, vol. 1, no. 3, pp. 247–252, 2004. View at Publisher · View at Google Scholar · View at Scopus - M. Forlin, I. Poli, D. De March, N. Packard, G. Gazzola, and R. Serra, “Evolutionary experiments for self-assembling amphiphilic systems,”
*Chemometrics and Intelligent Laboratory Systems*, vol. 90, no. 2, pp. 153–160, 2008. View at Publisher · View at Google Scholar · View at Scopus - D. De March, M. Forlin, D. Slanzi, and I. Poli, “An evolutionary predictive approach to design high dimensional experiments,” in
*Proceedings of the Articial Life and Evolutionary Computation (WIVACE '08)*, R. Serra, I. Poli, and M. Villani, Eds., pp. 81–88, World Scientific, 2008. - R. Baragona, F. Battaglia, and I. Poli, “Evolutionary Statistical Procedures,” in
*Statistics and Computing*, Springer, Berlin, Germany, 2011. View at Google Scholar - D. Ferrari, M. Borrotti, and D. De March, “Response improvement in complex experiments by co-information composite likelihood optimisation,”
*Statistics and Computing*, pp. 1–13, 2013. View at Publisher · View at Google Scholar - M. Borrotti and I. Poli, “Nave Bayes ant colony optimisation for experimental design,” in
*Synergies of Soft Computing and Statistics for Intelligent Data Analysis, Advances in Intelligent Systems and Computing*, R. Kruse, M. R. Berthold, C. Moewes et al., Eds., vol. 190, pp. 489–497, Springer, 2013. View at Google Scholar - L. Breiman, “Random forests,”
*Machine Learning*, vol. 45, no. 1, pp. 5–32, 2001. View at Publisher · View at Google Scholar · View at Scopus - A. Verikas, A. Gelzinis, and M. Bacauskiene, “Mining data with random forests: a survey and results of new tests,”
*Pattern Recognition*, vol. 44, no. 2, pp. 330–349, 2011. View at Publisher · View at Google Scholar · View at Scopus - A. L. Boulesteix, S. Janitza, J. Kruppa, and I. R. Knig, “Overview of random forest method- ology and practical guidance with emphasis on computational biology and bioinformatics,”
*Wiley Interdisciplinary Reviews*, vol. 2, no. 6, pp. 493–507, 2012. View at Google Scholar - A. Liaw and M. Wiener, “Classification and regression by random forest,”
*R News*, vol. 2, no. 3, pp. 18–22, 2002. View at Google Scholar