Skip to main content
  • Research article
  • Open access
  • Published:

Robust optimization of SVM hyperparameters in the classification of bioactive compounds

Abstract

Background

Support Vector Machine has become one of the most popular machine learning tools used in virtual screening campaigns aimed at finding new drug candidates. Although it can be extremely effective in finding new potentially active compounds, its application requires the optimization of the hyperparameters with which the assessment is being run, particularly the C and \(\gamma\) values. The optimization requirement in turn, establishes the need to develop fast and effective approaches to the optimization procedure, providing the best predictive power of the constructed model.

Results

In this study, we investigated the Bayesian and random search optimization of Support Vector Machine hyperparameters for classifying bioactive compounds. The effectiveness of these strategies was compared with the most popular optimization procedures—grid search and heuristic choice. We demonstrated that Bayesian optimization not only provides better, more efficient classification but is also much faster—the number of iterations it required for reaching optimal predictive performance was the lowest out of the all tested optimization methods. Moreover, for the Bayesian approach, the choice of parameters in subsequent iterations is directed and justified; therefore, the results obtained by using it are constantly improved and the range of hyperparameters tested provides the best overall performance of Support Vector Machine. Additionally, we showed that a random search optimization of hyperparameters leads to significantly better performance than grid search and heuristic-based approaches.

Conclusions

The Bayesian approach to the optimization of Support Vector Machine parameters was demonstrated to outperform other optimization methods for tasks concerned with the bioactivity assessment of chemical compounds. This strategy not only provides a higher accuracy of classification, but is also much faster and more directed than other approaches for optimization. It appears that, despite its simplicity, random search optimization strategy should be used as a second choice if Bayesian approach application is not feasible.

The improvement of classification accuracy obtained after the application of Bayesian approach to the optimization of Support Vector Machines parameters.

Background

The application of computational methods at various stages of drug design and development has become a vital part of the process. As the methods developed become constantly more effective, despite the aims at optimizing their performance, the focus of the attention shifts away from performance optimization to the minimization of requirements for computational resources. The attainment of both effectiveness and the desired speed has been responsible for the recent extreme popularity of machine learning (ML) methods in computer-aided drug design (CADD) approaches. Machine learning methods are mostly used for virtual screening (VS) tasks, in which they are supposed to identify potentially active compounds in large databases of chemical structures. One of the most widely used ML methods in CADD is the Support Vector Machine (SVM). Although it has a potential of providing very high VS performance, its application requires the optimization of the parameters used during the training process, which was proved to be crucial for obtaining accurate predictions. To date, various approaches have been developed to make SVM faster and more effective. In cheminformatics applications, the most popular optimization strategies are grid search [1, 2] and heuristic choice [3, 4]. Depending on the problem, they are able to provide high classification accuracy—for example Wang et al. obtained 86% of accuracy in the classification of hERG potassium channel inhibitors for the heuristic choice of the SVM parameters  [4]. On the other hand, Hamman et al. [1] were able to evaluate the cytochrome P450 activities with 66–83% of accuracy using grid search method of SVM parameters optimization. The need for optimizing SVM parameters is undeniable, as classification efficiency can change dramatically for various parameters values. A high computational cost of a systematic search over a predefined set of parameters’ values is a trigger for development of new optimization algorithms. In recent years, Bayesian optimization [5, 6] (including gaussian processes [7]) and random search-based selection [8] have become more popular [9, 10]. As those approaches were not explored so far in the field of cheminformatics, we analyze their impact on classification accuracy and, more importantly, the speed and ease of use, that these approaches have lent to the optimization of SVM hyperparameters in the search for bioactive compounds.

Hyperparameters optimization

In the classical ML approach to a classification problem, we are given a training set \(\left\{ (x_i,y_i) \right\} _{i=1}^N\) (with \(x_i\) representing samples’ features, in our case—fingerprint, and \(y_i\) being the class assignment) and we try to build a predictive model based on these data using a training algorithm that sets the parameters w (for example the weight of each fingerprint element) for fixed hyperparameters \(\varvec{\lambda }\) (for example a type of SVM kernel, the regularization strength C or the width of the RBF kernel \(\gamma\)). In other words, given an objective \(\mathcal {S}\) that must be maximized, we are supposed to solve the following problem:

$$\begin{aligned} \underset{w}{\text {maximize }} \mathcal {S}\left( w | \left\{ (x_i,y_i) \right\} _{i=1}^N, \varvec{\lambda }\right) . \end{aligned}$$

While this problem is often easily solvable (for example, in SVM, \(\mathcal {S}\) is a concave function, and thus, we can find the maximum by using a simple steepest ascent algorithm), in general, it is very hard to find an optimal \(\varvec{\lambda }\). This difficulty stems from the very complex shape of the \(\mathcal {S}\) function once we treat \(\varvec{\lambda }\) as its arguments, which results in the joint optimization of the model parameters (w) and the set of hyperparameters (\(\varvec{\lambda }\)):

$$\begin{aligned} \underset{w, \varvec{\lambda }}{\text {maximize }} \mathcal {S}\left( w, \varvec{\lambda }| \left\{ (x_i,y_i) \right\} _{i=1}^N\right) . \end{aligned}$$

A basic method for solving this problem is a grid search-based approach, which simply samples the set of possible \(\varvec{\lambda }\) values in a regular manner. For example, we choose the parameter C for a SVM in a geometrical progression, obtaining the values \(\varvec{\lambda }_1, \ldots, \varvec{\lambda }_k\) and returning the best solution among each of the subproblems:

$$\begin{aligned} w_i = \arg \max _w \mathcal {S}\left( w| \left\{ (x_i,y_i) \right\} _{i=1}^N, \varvec{\lambda }_i\right) . \end{aligned}$$

While such an approach guarantees finding the global optimum for \(k \rightarrow \infty\), it might be extremely computationally expensive, as we need to train k classifiers, each of which can take hours. Instead, we can actually try to solve the optimization problem directly by performing an adaptive process that on one hand tries to maximize the objective function and on the other hand samples the possible \(\varvec{\lambda }\) space intelligently in order to minimize the number of classifier trainings. The main idea behind Bayesian optimization for such a problem is to use all of the information gathered in previous iterations for performing the next step. It is apparent that grid search-based methods violate this assumption as we do not use any knowledge coming out from the results of models trained with other \(\varvec{\lambda }\) values.

We can consider this problem as the process of finding the maximum for \(f(\varvec{\lambda })\), defined as

$$\begin{aligned} f(\varvec{\lambda }) = \max _w \mathcal {S}\left( w| \left\{ (x_i,y_i) \right\} _{i=1}^N, \varvec{\lambda }\right) . \end{aligned}$$

Unfortunately, f is an unknown function and we cannot compute its gradient, Hessian, or any other characteristics that could guide the optimization process. The only action we can perform is to obtain a value for f at a given point. However, doing so is very expensive (because it requires training a classifier); thus, we need a fast (with respect to evaluating the function), derivative-free optimization technique to solve this problem.

For the task under consideration, \(\mathcal {S}\) is the accuracy of the resulting SVM model with the RBF kernel, and \(\varvec{\lambda }=\{ C, \gamma \}\) is the set of two hyperparameters that we must fit to optimize the SVM performance to predict the bioactivity of compounds, which (loosely speaking) is measured by f.

Table 1 Details of the classification experiments performed

Results and discussion

Six SVM optimization approaches were evaluated in the classification experiments of compounds possessing activity towards 21 protein targets, represented by six different fingerprints (Table 1).

Fig. 1
figure 1

Global analysis of classification accuracy obtained for different methods for SVM parameters optimization expressed as the number of experiments in which a particular strategy provided the highest accuracy values.

Classification effectiveness analysis

A global analysis of the classification efficiency revealed that Bayesian optimization definitely outperformed the other methods of SVM parameters’ optimization (Fig. 1). For a particular target and fingerprint, Bayesian approach provided a higher classification accuracy in 80 experiments, a significantly greater number than the other strategies (22 for the runner-up: grid search). On the other hand, the SVMlight and libSVM were definitely the least effective methods of SVM usage; they did not provide the highest accuracy values for any of the target/fingerprint combinations. This result is an obvious consequence of the fact that SVMlight and libSVM are just basic heuristics and their results cannot be comparable with any hyperparameters optimization technique. Interestingly, libSVM achieved much better results than SVMlight even though its heuristic is much simpler.

Fig. 2
figure 2

Analysis of the effectiveness of different SVM optimization strategies with respect to various fingerprints expressed as the number of experiments in which a particular strategy provided the highest accuracy values for a given compounds representation.

The relationships between various methods tested were preserved when the results were analyzed with respect to the various fingerprints (Fig. 2)—the Bayesian optimization always provided the highest classification accuracy (for 13–14 targets for each of the fingerprints analyzed), whereas the ‘global’ second-place method—grid search—was outperformed by ‘small grid’-cv for two fingerprints: MACCSFP and PubchemFP. The runners-up (grid search or ‘small grid’-cv, depending on fingerprint) provided the best predictive power of the model for 3 proteins on average. The ineffectiveness of the SVMlight and libSVM strategies has been already indicated in the 'global' analysis, and with respect to various fingerprints, there was no protein for which those SVM optimization methods provided the highest classification accuracy.

Fig. 3
figure 3

Analysis of effectiveness of different SVM optimization strategies with respect to various targets expressed as the number of experiments in which a particular strategy provided the highest accuracy values for a given protein target.

The situation becomes more complex when separate targets are taken into account (Fig. 3). The Bayesian optimization provided the best results for all considered representations for some proteins (CDK2, H1, ABL); however, in few cases, other optimization approaches for tuning SVM parameters outperformed the Bayesian method (5-HT6—random and grid search, beta1AR—‘small grid’-cv and grid search, beta3AR—grid search, HIVi—grid search, ‘small grid’-cv, random, MAP kinases ERK2—‘small grid’-cv and random search). These results show that a more careful model accuracy approximation is required for some proteins. Because we are interested in maximizing the accuracy on a naive test set, we approximate this set by performing internal cross-validation for each method. This is a well-known technique in ML; however, it might be not reliable for small datasets. Beta1AR, beta3AR, and HIVi are very small datasets in our comparison; thus, it seems probable that the poor results of the Bayesian approach (a poor approximation of the \(\mathcal {S}\) value) were caused by the high internal variance in the dataset rather than because the Bayesian approach was actually worse than the grid search method.

Table 2 A comparison of the number of highest accuracies obtained with the Bayesian optimization and grid search

Because grid search was the second-place method in the majority of the analyses, both for global analysis, and fingerprint- and target-based comparisons, a direct comparison of the number of the highest accuracies obtained for Bayesian optimization and the grid search approach was performed (Table 2). The sum of the number of wins is not equal for the given fingerprint-based or target-based comparison as the draws were also considered.

The comparison of the number of 'wins' for Bayesian optimization over the grid search indicates the superiority of the former approach. In the ‘global’ analysis, the Bayesian optimization strategy gave a higher accuracy for approximately a 3-fold higher number of experiments than the grid search. For the fingerprint-based analysis, the ratio of Bayesian/grid search wins was similar to the best ratio (in favor of Bayesian optimization) obtained for MACCSFP (18 : 4) and the worst (15 : 7 and 15 : 6) for EstateFP and SubFP, respectively. When target-based comparisons were considered, Bayesian optimization outperformed the grid search approach for some targets in all cases (i.e., CDK2, H\(_\text {1}\), ABL); for others, there was only 1 case when the grid search strategy won (i.e., 5-HT\(_\text {2A}\), 5-HT\(_\text {2C}\), M\(_\text {1}\), ERK2, AChE, A\(_\text {1}\), alpha2AR, CB1, D\(_\text {4}\), H\(_\text {3}\), IR), still others were draws (i.e., 5-HT\(_\text {7}\), beta1AR), and in two cases the grid search provided top accuracies (beta3AR, HIVi).

Fig. 4
figure 4

Analysis of the changes in accuracy during the SVM optimization procedure for the subsequent optimization steps.

Examination of optimization steps in time

A time course study of the accuracy values was also conducted. Figure 4 shows analyses for 5-HT\(_\text {2A}\) as an example; the results for the remaining targets are in the Additional files section (Additional file 1). We demonstrated not only that Bayesian optimization required the smallest number of iterations to achieve optimal performance (for all representations the number of iterations was less than 20), but also that in the majority of cases, SVM optimized using a Bayesian approach achieved better performance than all of the other optimization methods. SVMlight and libSVM were not iteratively optimized; therefore, the accuracy/number of iterations function is constant for these approaches. In general, the Bayesian and random search approach were optimized very quickly (in less than 20 iterations), whereas the grid search method required many more iterations before the SVM reached optimal performance: 57 iterations (the lowest number) were required for EstateFP, and 138 (the highest number) for MACCSFP). Figure 4 also shows the rate of the improvement of the accuracy after the application of particular optimization approach, which depended on the representation of the compounds—for EstateFP, it was improvement from 0.8 (grid search) up to 0.82 (Bayesian), but for libSVM and SVMlight the improvement was significantly higher; for these two strategies, the classification accuracy was equal to 0.68. A similar result was obtained for ExtFP—the rate of improvement for the Bayesian optimization strategy compared with the random search approach was approximately 0.02 (from 0.88 to 0.90), but it was higher for the other optimization methods: 0.03 for grid search, 0.05 for libSVM and 0.22 for SVMlight. The pattern was similar for KlekFP and MACCSFP, with differences occurring only in the performance of libSVM. However, for PubchemFP and SubFP, grid search optimization provided the same predictive power for SVM as Bayesian optimization; for the SubFP, there was a selected range of iterations (117–142) when grid search provided slightly better SVM performance (by about 2%) in comparison to Bayesian approach.

Table 3 The AUC values obtained in 5-HT\(_\text {2A}\), ExtFP for curves illustrating changes in the accuracy in time and final optimal accuracy values obtained

In order to provide the comprehensive and global analysis of the changes in accuracy with an increasing number of iterations, the areas under curves (AUC) presented in Fig. 4 (and other curves that are placed in the Additional files section) were calculated. Example analysis for selected target/fingerprint pair (5-HT\(_\text {2A}\), ExtFP) is presented in Table 3; the remaining analyses are in the Additional files section (Additional file 2). The global average AUC, the average AUC for particular fingerprints and targets are presented in Tables  4 and  5. The Tables also include final (for the selected target/fingerprint) and averaged (for the rest of the cases) final accuracy values obtained for a given strategy; the highest AUC/accuracy values for the particular case considered are marked with an asterisk sign. In general, the AUC of a curve indicates the strength of the trained model at any randomly chosen iteration. In other words, the AUC measures how quickly a given strategy converges to a strong model.

Fig. 5
figure 5

Analysis of the number of iterations of the optimization procedure required to achieve the highest accuracy. The figure presents the number of iterations required for a particular optimization strategy to achieve optimal performance for the predictive model.

Table 4 The average AUC values–global, obtained for a particular fingerprint and particular target
Table 5 The average final accuracy values—global, obtained for a particular fingerprint and particular target

The analysis of the results obtained for the example target/fingerprint pair (5-HT\(_\text {2A}\), ExtFP; Table 3) shows that both the highest AUC and final optimal accuracy values were obtained with the Bayesian strategy for SVM optimization. A similar observation was made for the global and fingerprint-based analysis; Bayesian optimization provided the best average AUC and average optimal accuracy for all fingerprints, as well as the global average value of this parameter. Interestingly, although grid search was the second-place method for optimal accuracy, it was actually the random search that outperformed this method in terms of AUC, which could be explained from an analysis of the respective curves. Although the grid search method provided higher final accuracy values, these occurred relatively 'late' (after a series of iterations), high accuracies were obtained almost immediately for random search (Figs. 4, 5). Similarly, the average AUC and optimal accuracy values calculated for various targets were highest for Bayesian optimization in the great majority of cases. HIVi and ERK2 were the only targets for which the averaged AUC obtained with the Bayesian optimization strategy was outperformed by other optimization methods. On the other hand, the group of targets for which the average optimal accuracy values were the highest for methods other than Bayesian optimization was a bit more extensive (i.e., M\(_\text {1}\), ERK2, A\(_\text {1}\), beta3AR, HIVi, IR). However, for most of these targets, the difference between the best average accuracy and that obtained with Bayesian optimization was approximately 3% (however, for example for beta3AR this difference approached to 10%, from 0.879 to 0.972). On the other hand, an improvement of several percentage points was also observed when the average AUC and optimal accuracy obtained with the Bayesian strategy were compared with the strategy that provided the ‘second-best’ accuracy value in the ranking.

The number of iterations required to achieve optimal SVM performance was also analyzed in detail (Fig. 5; Additional file 3). The most striking observation was that all curves corresponding to the Bayesian optimization results were both shifted towards higher accuracy values and were much ‘shorter’, meaning that a significantly lower number of iterations was necessary in total to reach optimal SVM performance. Two relevant points arise from a comparison of Bayesian optimization with the grid search method (which sometimes outperformed Bayesian optimization): obtaining optimal accuracy with the grid search method required many more calculations, and even when grid search yielded higher accuracy values than Bayesian optimization, the difference between the two was approximately 1–2%. This result indicates that even when Bayesian optimization ‘lost’, the results provided by this strategy were still very good and taking into account the calculation speed, it can be successfully applied also in experiments for which it was not indicated to be the best approach. A very interesting observation arising from Fig. 5 is that random search reached the optimal classification effectiveness (as measured by accuracy) in the least number of iterations, below 10 in the majority of cases. EstateFP, ExtFP, MACCSFP and PubchemFP, showed similar tendency with respect to the comparison of Bayesian optimization and the grid search strategy; for an initial number of iterations (40), the accuracy values obtained with the grid search were approximately 20% lower than those obtained with the Bayesian approach. However, as the number of iterations for grid search increased, the accuracy values were also higher, and when the number of iterations reached approximately 100, the grid search results were similar to those obtained with Bayesian optimization. On the other hand, for both KlekFP and SubFP, the initial observations were the same; for a lower number of iterations, Bayesian optimization led to significantly higher accuracy values than the grid search approach, and for a higher number of iterations (over 80 for KlekFP and over 115 for SubFP), grid search provided accuracy values at a similar level to the values obtained with the Bayesian strategy. However, increasing the number of iterations for Bayesian optimization from approximately 10 to 90 for KlekFP and 150 for SubFP did not lead to a significant increase in the accuracy (an almost vertical line corresponding to these numbers of iterations), which was already very high (over 0.85 for KlekFP and over 0.8 for SubFP). Further optimization led to further improvement in accuracy of approximately 2–3%.

The results were also analyzed regarding the changes in the accuracy when additional steps were applied. A panel of example results is shown in Fig. 6 for the cannabinoid CB1/SubFP combination (the remaining targets are in Additional file 4). The black dots show the set of parameters tested in the particular approach, and the black squares represent the set of parameters selected as optimal. This chart shows the advantage of Bayesian optimization in terms of the way of work, and the sequence of selected parameters. The set of tested parameters is fixed for grid search optimization, whereas in case of random search, it is based on the random selection. On the other hand, the selection of parameters for Bayesian optimization is more directed, which also affects the effectiveness of the classification. For grid search, only a small fraction of the parameters tested provided satisfactory predictive power of the model (only approximately 35% of the predictions resulted in an accuracy exceeding 0.7). Surprisingly, a relatively high classification efficiency was obtained with the use of the random search approach—60% of the sets of parameters tested provided predictions with an accuracy over 0.7. However, investigation of the Bayesian optimization approach to parameter selection revealed that the choice of parameters tested was justified, and hence, the results obtained with their use were significantly better than those obtained with the other approaches—75% predictions with accuracy over 0.7.

We conclude that there are three SVM hyperparameters selection approaches worth using for activity prediction for compounds:

  • libSVM heuristic (when only one set of hyperparameters is needed),

  • random search (when we need a strong model quickly, using less than a few dozen iterations),

  • a Bayesian approach (when we want the strongest model and can wait a bit longer).

The SVMlight heuristic as well as the traditional grid search approach have definitely been shown to be significantly worse in terms of the resulting model accuracy as well as time needed to construct such model.

Fig. 6
figure 6

Analysis of the changes in accuracy for different steps during the SVM optimization procedure.

Table 6 The number of active and inactive compounds in the dataset

Experimental

Several compounds datasets were prepared and their proper description using various fingerprints was provided for the planned experiments. The ChEMBL database constituted a source of active and inactive compounds with experimentally verified activity towards selected targets. The following proteins were considered in the study: serotonin receptors 5-HT\(_\text {2A}\) [11], 5-HT\(_\text {2C}\) [12], 5-HT\(_\text {6}\) [13], and 5-HT\(_\text {7}\) [14], cyclin dependent kinase 2 (CDK2) [15], muscarinic receptor M\(_\text {1}\) [16], MAP kinase ERK2 [17], acetylcholinesterase (AChE) [18], adenosine receptor A\(_\text {1}\) [19], alpha-2A adrenergic receptor [20], beta-1 adrenergic receptor (beta1AR) [21], beta-3 adrenergic receptor (beta3AR) [21], cannabinoid CB1 receptor [22], delta opioid receptor (DOR) [23], dopamine receptor D\(_\text {4}\) [24], histamine receptor H\(_\text {1}\) [25], histamine receptor H\(_\text {3}\) [26], HIV integrase (HIVi) [27], insulin receptor (IR) [28], tyrosine kinase ABL [29], and human leukocyte elastase (HLE) [30]. Only molecules whose activities were quantified in \(K_{i}\) or \(IC_{50}\) and that were tested in assays on human, rat-cloned or native receptors were taken into account. The compounds were considered active when the median value of all \(K_{i}\) values provided for a particular instance was lower than 100 nM, and inactive when the median \(K_{i}\) value was greater than 1000 nM. The number of compounds from each group for the selected targets is shown in Table 6. The following fingerprints were used for compounds representation: E-state Fingerprint (EstateFP) [31], Extended Fingerprint (ExtFP) [32], Klekota and Roth Fingerprint (KlekFP) [33], MACCS Fingerprints (MACCSFP) [34], Pubchem Fingerprint (PubchemFP), and Substructure Fingerprint (SubFP), generated with the use of the PaDEL-Descriptor [35]. A brief characterization of the fingerprints is provided in Table 7).

Table 7 Fingerprints used for compounds representation

The following SVM strategies were used:

  • default SVM parameters used in the WEKA package (\(C = 1, \gamma = \tfrac{1}{d}\))—libSVM.

  • default SVM parameters from the SVMlight library (\(C = \frac{1}{\mathrm {\underset{i}{mean}} \Vert x_{i}\Vert ^2},\; \gamma = \tfrac{1}{d}\)).

  • grid search optimization of SVM parameters—\(\log _{10}(C) \in [-2, 5]\), \(\log _{10}(\gamma ) \in [-10, 3]\).

  • SVM parameters optimization in the truncated cross-validation mode (‘small grid’-cv).

  • SVM parameters optimization in the random cross-validation mode—number of iterations: up to 150.

  • Bayesian optimization using BayesOpt [36]—number of iterations: up to 150.

The range of C and \(\gamma\) values tested was as follows: \(\log _{10}(C) \in [-2, 5]\), \(\log _{10}(\gamma ) \in [-10, 3]\) (the result of preliminary grid search experiments). The number of iterations in which random search, 'small grid'-cv and Bayesian optimization experiments were performed fell within the following set: 20, 30, 50, 75, 100, 150.

The predictive power of SVM for different optimization strategies applied was measured by the accuracy:

$$\begin{aligned} \mathrm {Accuracy} (TP, FP, TN, FN)= \frac{\mathrm {TP + TN}}{\mathrm {TP + TN + FP + FN}}, \end{aligned}$$

with TP being the number of true positives (correctly classified actives), TN—the number of true negatives (correctly classified inactives), FP—the number of false positives (inactives wrongly classified as active), and FN—the number of false negatives (actives wrongly classified as inactive).

Conclusions

The paper presents strengths of Bayesian optimization applied for fitting SVM hyperparameters in cheminformatics tasks. Because the importance and necessity of the SVM optimization procedure is undeniable, various approaches to this task have neen developed so far. However, the most popular approaches to SVM optimization are not always very effective, in terms of both the predictive power of the models obtained and the computational requirements. This study demonstrated that Bayesian optimization not only provides better classification accuracy than the other optimization approaches tested but is also much faster and directed—in the majority of cases, the number of iterations required to achieve optimal performance was the lowest out of the all methods tested, and the set of parameters tested provided the best predictions on average. Interestingly, if good classification results are desired to be obtained quickly (using a low number of iterations and without complex algorithms), the random search method in which hyperparameters are randomly selected from a predefined range) leads to very good performance of the SVM for predicting the activity of compounds and can thus be used when Bayesian optimization approach is not feasible.

Consequently, we can formulate the following rule of thumb for tuning SVM’s hyperparameters for the classification of bioactive compounds:

  1. 1.

    If you have no resources for performing hyperparameters optimization, use \(C=1, \gamma = \frac{1}{d}\) (as defined in libSVM).

  2. 2.

    If you have limited resources (up to 20 learning procedures) or limited access to complex optimization software, use a random search for C and \(\gamma\) with distribution defined in the “Methods” section.

  3. 3.

    If you have resources for 20 or more training runs and access to Bayesian optimization softwarea, use a Bayesian optimization of \(C, \gamma\).

In general, there is no scenario in which one should use a grid search approach (it is always preferable to use random search or a Bayesian method) or SVMlight heuristics (it is always better to use libSVM) in the tasks connected with the assessment of compounds bioactivity.

Methods

The objective of the iterative global optimization of a function \(f : \mathcal {L} \rightarrow \mathbb {R}\) is to find the sequence of points

$$\begin{aligned} \varvec{\lambda }_1, \dots , \varvec{\lambda }_n \in \mathcal {L}, \end{aligned}$$

that converges to the optimal \(\hat{\varvec{\lambda }}\), \(f(\hat{\varvec{\lambda }}) = \sup _{\varvec{\lambda }\in \mathcal {L}} f(\varvec{\lambda })\). A good algorithm should find a solution at least over some family of functions \(\mathcal {F}\), not necessarily containing f.

The above-mentioned issue can be viewed as a sequential decision making problem [37] in which at time step i a decision based on all previous points \(\alpha _i(\varvec{\lambda }_{1:i-1}, \bar{f}_{1:i-1})\), where \(\bar{f}_i = f(x_i) + \varepsilon _i\) is made. In other words, we have access to approximations of f values from previous steps. For simplicity, assume that \(\varepsilon _i = 0\) (f is deterministic); however, in general, all methods considered can be used in a stochastic scenario (for example, when randomized cross-validation is used as underlying method for f evaluation).

The goal is to find \(\alpha\) which minimizes \(\delta _n(\alpha ) = f(\hat{\varvec{\lambda }}) - f(\varvec{\lambda }_n)\), meaning that we are interested in

$$\begin{aligned} \arg \min _\alpha \delta _n(\alpha ) = \arg \min _\alpha f(\hat{\varvec{\lambda }}) - f(\varvec{\lambda }_n), \end{aligned}$$
(1)

which could be efficiently solved if f is known.

Approximation of generalization capabilities

In general, we are interested in how well our predictive model behaves on a naive test set. In other words, we are assuming that our data are a finite iid (independent and identically distributed) sample from some underlying joint distribution over samples (compounds) and their binary labels (biological activity) \(\mu\):

$$\begin{aligned} \mathrm {T}=\{(x_i, y_i)_{i=1}^N\} \sim \mu ( \mathcal {X} \times \{-1, +1\}), \end{aligned}$$

where \(\mathcal {X}\) represents a feature space of compounds under investigation. We want to maximize the expected accuracy over all possible compounds from \(\mu\), in other words

$$\begin{aligned} f(\varvec{\lambda }) &= \mathbb {E}_{(x_i,y_i) \sim \mu }[\mathrm {svm}(x_i|\mathrm {T},\varvec{\lambda })=y_i]\\ &= \int [\mathrm {svm}(x_i|\mathrm {T},\varvec{\lambda })=y_i] d\mu , \end{aligned}$$

where [ p ] is a characteristic function returning 1 if and only if p is true, and \(\mathrm {svm}(x_i|\mathrm {T}, \varvec{\lambda })\) is a prediction of \(x_i\)’s label by SVM trained with hyperparameters \(\varvec{\lambda }\) on training set \(\mathrm {T}\).

Clearly, we cannot integrate over an unknown probability distribution, but we can approximate this value using internal cross-validation. In other words, we are using a stochastic approximation

$$\begin{aligned} \bar{f}(\varvec{\lambda }) = \text {CV}_\mathrm {T}(\mathrm {svm}(X_\mathrm {test}|\mathrm {T}_\mathrm {train},\varvec{\lambda }), Y_\mathrm {test}), \end{aligned}$$

where \(\mathrm {CV}_\mathrm {T}(p, y)\) is the mean accuracy of the model of predictions p as compared to the true labels y over splits of set \(\mathrm {T}\) into \(\mathrm {T}_\mathrm {train}\) and \(\mathrm {T}_\mathrm {test}\) (composed of \(X_\mathrm {test}\) data and corresponding labels \(Y_\mathrm {test}\)). Thus we can assume [38] that

$$\begin{aligned} \bar{f}(\varvec{\lambda }) = f(\varvec{\lambda }) + \varepsilon , \end{aligned}$$

where \(\varepsilon\) is a random noise variable (resulting from the approximating error and stochastic nature of cross validation).

Random optimization

First, let us define a random optimization technique as a strategy \(\alpha ^\mathcal {R}(\varvec{\lambda }_{1:i-1}, \bar{f}_{1:i-1}) = \alpha ^\mathcal {R}_i = \varvec{\lambda }_i \sim \varvec{P}(\mathcal {L})\), for some probability distribution over the hyperparameters \(\varvec{P}(\mathcal {L})\). In other words, in each iteration, we sample from \(\varvec{P}(\mathcal {L})\), ignoring all previous samples and their results. Finally, we return the maximum of the values obtained.

It is easily seen that a random search, under the assumption that \(\forall _{\varvec{\lambda }\in \mathcal {L}} \varvec{P}(\alpha ^\mathcal {R}_i = \varvec{\lambda }) > 0\), has a property described in (1). A random search will converge to the optimum [39], if only each set of parameters is possible to generate when taking new sample from our decision making process. In practise, it is only necessary that \(\varvec{P}(f(\alpha ^\mathcal {R}_i) = f(\hat{\varvec{\lambda }})) > 0\). Similarly, if one uses a grid search approach that discretizes \(\mathcal {L}\), then given enough iterations and the assumption that f is continuous, one will converge to the optimal solution. It is important to note that the speed of such a convergence can be extremely low.

The only thing missing is the selection of \(\varvec{P}(\mathcal {L})\). According to many empirical studies showing that meaningful changes in the SVM results as the function of its hyperparameters can be expressed in log-scale of these parameters we use

$$\begin{aligned} \varvec{P}(\varvec{\lambda }= (C, \gamma )) = \tfrac{\log _{10} C - \log _{10} C_\mathrm {min}}{\log _{10} C_\mathrm {max} - \log _{10} C_\mathrm {min}} \cdot \tfrac{\log _{10}\gamma - \log _{10}\gamma _\mathrm {min}}{\log _{10}\gamma _\mathrm {max} - \log _{10}\gamma _\mathrm {min}} \end{aligned}$$

where we are interested in \(\mathcal {L} = [C_\mathrm {min}, C_\mathrm {max}] \times [\gamma _\mathrm {min}, \gamma _\mathrm {max}]\). In other words, we are using a log-uniform distributions independently over C and \(\gamma\).

Grid search

For grid search optimization we select \(\varvec{\lambda }\) in a similar manner to the random search approach, uniformly in the log-scale of the parameters, and given \(M_p\) choices of parameter p

$$\begin{aligned} \varvec{\lambda }_{ij}&= (C_i, \gamma _j) \\&= \left( 10^{\log _{10} C_\mathrm {min} + (i-1) \tfrac{\log _{10} C_\mathrm {max} - \log _{10} C_\mathrm {min}}{M_C-1}}, 10^{\log _{10} \gamma _\mathrm {min} + (j-1) \tfrac{\log _{10} \gamma _\mathrm {max} - \log _{10} \gamma _\mathrm {min}}{M_\gamma -1}} \right) . \end{aligned}$$

We put the linear order of \(\varvec{\lambda }_{ij}\) by raveling the resulting matrix by column, which is the most common practice in most ML libraries. It is worth noting that one could achieve better scores by alternating this ordering to any random permutation; however, in practice, such alternation is rarely performed.

Bayesian optimization

If the exact form of f is known (for example, if f is convex and its derivative is known), then the optimization procedure would be much simpler. Unfortunately, f is a black-box function wih a very complex structure, expensive even to evaluate. However, some simplifying assumptions for f might make a problem solvable. Assume that f can be represented as a sample from a probability distribution over a family of functions \(f \sim \varvec{P}(f), f \in \mathcal {F}\).

We can now express the expectation over the loss function \(\delta _n\):

$$\begin{aligned} \arg \min _\alpha \mathbb {E}_{\varvec{P}(f)} [\delta _n(\alpha )] = \arg \min _\alpha \int _{\mathcal {F}} \delta _n(\alpha ) d \varvec{P}(f). \end{aligned}$$

Given that in step n the values of \(\varvec{\lambda }_i, \bar{f}_i\) for \(i=1, \dots , n-1\) are already known and using the Bayes rule, we can write:

$$\begin{aligned} \varvec{P}(f | x_{1:i}, \bar{f}_{1:i}) &= \frac{\varvec{P}(x_i, \bar{f}_i | f) \varvec{P}(f|x_{1:i-1}, \bar{f}_{1:i-1})}{\varvec{P}(x_i, \bar{f}_i)},\\ &\quad\quad\forall i = 1, \dots , n-1, \end{aligned}$$

thus

$$\begin{aligned}& \arg \min _\alpha \mathbb {E}_{\varvec{P}(f|x_{1:n-1}, \bar{f}_{1:n-1})} [\delta _n(\alpha )] \\ &\quad = \arg \min _\alpha \int _{\mathcal {F}} \delta _n(\alpha ) d \varvec{P}(f|x_{1:n-1}, \bar{f}_{1:n-1}). \end{aligned}$$

This is a very basic equation for general Bayesian optimization techniques. Given additional assumptions about the prior distribution of \(\varvec{P}(f)\), very efficient solutions for the entire process can be provided. In the case considered here, a very common approach exploiting features of the Gaussian processes is employed; thus, we assume that our target function (the generalization capabilities of the SVM with RBF kernel applied to the prediction of the activity of chemical compounds) f is a sample of the stochastic process. A crucial advantage of such a simplification is that we can easily manipulate the distribution over such functions: in particular, the posterior distribution is also a Gaussian process. Consequently, in each iteration a calculated posterior can be used as an informative prior for the next iteration, creating a relatively simple iterative procedure.

The only thing missing is the selection of the loss function, because \(\delta _n(\alpha )\) defined above requires knowledge of the optimal solution. There are many surrogate functions (also called proxies) that might be of use. In our investigations we used one of the most well-known [7] and studied expected improvement [40, 41], which has the following closed form solution under the above assumptions:

$$\begin{aligned} \alpha _\mathrm {EI}(\varvec{\lambda }| \varvec{\lambda }_{1:n-1}, \bar{f}_{1:n-1})&= \sigma (\varvec{\lambda }| \varvec{\lambda }_{1:n-1}, \bar{f}_{1:n-1})(\gamma (\varvec{\lambda }) \phi (\gamma (\varvec{\lambda })) + \mathcal {N}(\gamma (\varvec{\lambda }))), \\ \gamma (\varvec{\lambda })&= \frac{\mu (\varvec{\lambda }| \varvec{\lambda }_{1:n-1}, \bar{f}_{1:n-1}) - f(\varvec{\lambda }_\text {best})}{\sigma (\varvec{\lambda }| \varvec{\lambda }_{1:n-1}, \bar{f}_{1:n-1})} \end{aligned}$$

where \(\mu , \sigma ^2\) are Gaussian process mean and variance predictions, \(\varvec{\lambda }_\text {best}\) is a current best solution \(f(\varvec{\lambda }_\text {best}) = \max _{i=1,\dots ,n-1} f(\varvec{\lambda }_i)\), \(\phi\) is a cumulative density function of the standard normal distribution and \(\mathcal {N}\) is a probability density function of the standard normal distribution. Thus, in each iteration we simply select the point that maximizes the above equation

$$\begin{aligned} \alpha _n ( \varvec{\lambda }_{1:n-1}, \bar{f}_{1:n-1} ) = {\arg \max }_{\varvec{\lambda }} \alpha _\mathrm {EI}(\varvec{\lambda }| \varvec{\lambda }_{1:n-1}, \bar{f}_{1:n-1}). \end{aligned}$$

Endnotes

aFor example BayesOpt http://rmcantin.bitbucket.org/html/

References

  1. Hammann F, Gutmann H, Baumann U, Helma C, Drewe J (2009) Classification of cytochrome P 450 activities using machine learning methods. Mol Pharm 33(1):796–801

    Google Scholar 

  2. Smusz S, Czarnecki WM, Warszycki D, Bojarski AJ (2015) Exploiting uncertainty measures in compounds activity prediction using support vector machines. Bioorganic Med Chem Lett 25(1):100–105

    Article  CAS  Google Scholar 

  3. Lee JH, Lee S, Choi S (2010) In silico classification of adenosine receptor antagonists using Laplacian-modified naïve Bayesian, support vector machine, and recursive partitioning. J Mol Graph Model 28(8):883–890

    Article  CAS  Google Scholar 

  4. Wang M, Yang X-G, Xue Y (2008) Identifying hERG potassium channel inhibitors by machine learning methods. QSAR Comb Sci 27(8):1028–1035

    Article  Google Scholar 

  5. Swersky K, Snoek J, Adams RP (2013) Multi-task bayesian optimization. In: Advances in neural information processing systems, vol 26. Lake Tahoe, pp 2004–2012

  6. Snoek J, Rippel O, Swersky K, Kiros R, Satish N, Sundaram N et al (2015) Scalable bayesian optimization using deep neural networks. arXiv preprint aXiv:1502.05700

  7. Snoek J, Larochelle H, Adams RP (2012) Practical bayesian optimization of machine learning algorithms. In: Advances in neural information processing systems, vol 25. Lake Tahoe, pp 2951–2959

  8. Bergstra J, Bengio Y (2012) Random search for hyper-parameter optimization. J Mach Learn Res 13(1):281–305

    Google Scholar 

  9. Eggensperger K, Feurer M, Hutter F, Bergstra J, Snoek J, Hoos H et al (2013) Towards an empirical foundation for assessing bayesian optimization of hyperparameters. In: NIPS Workshop on Bayesian Optimization in Theory and Practice. Lake Tahoe, pp 1–5

  10. Thornton C, Hutter F, Hoos HH, Leyton-Brown K (2013) Auto-weka: combined selection and hyperparameter optimization of classification algorithms. In: Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, pp 847–855

  11. Sencanski M, Sukalovic V, Shakib K, Soskic V, Dosen-Micovic L, Kostic-Rajacic S (2014) Molecular modelling of 5HT2A receptor—arylpiperazine ligands interactions. Chem Biol Drug Design 83(4):462–471

    Article  CAS  Google Scholar 

  12. Millan MJ (2005) Serotonin 5-HT2C receptors as a target for the treatment of depressive and anxious states: focus on novel therapeutic strategies. Therapie 60(5):441–460

    Article  Google Scholar 

  13. Upton N, Chuang TT, Hunter AJ, Virley DJ (2008) 5-HT6 receptor antagonists as novel cognitive enhancing agents for Alzheimer’s disease. Neurotherapeutics 5(3):458–469

    Article  CAS  Google Scholar 

  14. Roberts AJ, Hedlund PB (2012) The 5-HT(7) receptor in learning and memory. Hippocampus 22(4):762–771

    Article  CAS  Google Scholar 

  15. Chen H, Van Duyne R, Zhang N, Kashanchi F, Zeng C (2009) A novel binding pocket of cyclin-dependent kinase 2. Proteins 74(1):122–132

    Article  CAS  Google Scholar 

  16. Leach K, Simms J, Sexton PM, Christopoulos A (2012) Structure-function studies of muscarinic acetylcholine receptors.  In: Fryer AD, Christopoulos A, Nathanson NM (eds) Handbook of experimental pharmacology, vol 208. Springer, UK, pp 29–48

  17. Zhang F, Strand A, Robbins D, Cobb MH, Goldsmith EJ (1994) Atomic structure of the MAP kinase ERK2 at 2.3 A resolution. Nature 367(6465):704–711

    Article  CAS  Google Scholar 

  18. Soreq H, Seidman S (2001) Acetylcholinesterase–new roles for an old actor. Nat Rev Neurosci 2(4):294–302

    Article  CAS  Google Scholar 

  19. Hocher B (2010) Adenosine A1 receptor antagonists in clinical research and development. Kidney Int 78(5):438–445

    Article  CAS  Google Scholar 

  20. Hein L (2001) The alpha 2-adrenergic receptors: molecular structure and in vivo function. Z Kardiol 90(9):607–612

    Article  CAS  Google Scholar 

  21. Wallukat G (2002) The beta-adrenergic receptors. Herz 27(7):683–690

    Article  Google Scholar 

  22. Pertwee RG (1997) Pharmacology of cannabinoid CB1 and CB2 receptors. Pharmacol Ther 4(2):129–180

    Google Scholar 

  23. Quock RM (1999) The delta-opioid receptor: molecular pharmacology, signal transduction, and the determination of drug efficacy. Pharmacol Rev 51(3):503–532

    CAS  Google Scholar 

  24. Rondou P, Haegeman G, Van Craenenbroeck K (2010) The dopamine D4 receptor: biochemical and signalling properties. Cell Mol Life Sci 67(12):1971–1986

    Article  CAS  Google Scholar 

  25. Thurmond RL, Gelfand EW, Dunford PJ (2008) The role of histamine h1 and h4 receptors in allergic inflammation: the search for new antihistamines. Nat Rev Drug Discov 7(1):41–53

    Article  CAS  Google Scholar 

  26. Passani MB, Lin J-S, Hancock A, Crochet S, Blandina P (2004) The histamine H3 receptor as a novel therapeutic target for cognitive and sleep disorders. Trend Pharmacol Sci 25(12):618–625

    Article  CAS  Google Scholar 

  27. Craigie R (2001) Hiv integrase, a brief overview from chemistry to therapeutics. J Biol Chem 276(26):23213–23216

    Article  CAS  Google Scholar 

  28. Whitehead JP, Clark SF, Ursø B, James DE (2000) Signalling through the insulin receptor. Curr Opin Cell Biol 12(2):222–228

    Article  CAS  Google Scholar 

  29. Lanier LM, Gertler FB (2000) From Abl to actin: Abl tyrosine kinase and associated proteins in growth cone motility. Curr Opin Neurobiol 10(1):80–87

    Article  CAS  Google Scholar 

  30. Lee WL, Downey GP (2001) Leukocyte elastase: physiological functions and role in acute lung injury. Am J Respir Crit Care Med164(5):896–904

    Article  CAS  Google Scholar 

  31. Hall LH, Kier LB (1995) Electrotopological state indices for atom types: a novel combination of electronic, topological, and valence state information. J Chem Inform Model 35(6):1039–1045

    Article  CAS  Google Scholar 

  32. Steinbeck C, Han Y, Kuhn S, Horlacher O, Luttmann E, Willighagen E (2003) The chemistry development kit (CDK): an open-source Java library for chemo- and bioinformatics. J Chem Inform Comp Sci 43(2):493–500

    Article  CAS  Google Scholar 

  33. Klekota J, Roth FP (2008) Chemical substructures that enrich for biological activity. Bioinform Oxf Engl 24(21):2518–2525

  34. Ewing T, Baber JC, Feher M (2006) Novel 2D fingerprints for ligand-based virtual screening. J Chem Inform Model 46(6):2423–2431

    Article  CAS  Google Scholar 

  35. Yap CW (2011) Padel-descriptor: an open source software to calculate molecular descriptors and fingerprints. J Comput Chem 32(7):1466–1474

    Article  CAS  Google Scholar 

  36. Martinez-Cantin R (2014) Bayesopt: a bayesian optimization library for nonlinear optimization, experimental design and bandits. arXiv preprint arXiv:1405.7430

  37. Mockus J (1994) Application of bayesian approach to numerical methods of global and stochastic optimization. J Glob Optim 4(4):347–365

    Article  Google Scholar 

  38. Bishop CM (2006) Pattern recognition and machine learning. Springer, NJ

    Google Scholar 

  39. Auger A, Doerr B (2011) Theory of randomized search heuristics: foundations and recent developments, vol 1. World Scientific, Series on Theoretical Computer Science

    Google Scholar 

  40. Mockus J, Tiesis V, Zilinskas A (1978) The application of bayesian methods for seeking the extremum. Towards Glob Optim 2(117–129):2

    Google Scholar 

  41. Schonlau M, Welch WJ, Jones DR (1998) Global versus local search in constrained optimization of computer models. Lect Notes Monogr Ser 34:11–25

    Article  Google Scholar 

Download references

Author’s contributions

WCz and SP performed the experiments. All the authors analyzed and discussed the results and wrote the manuscript. All authors read and approved the final version of the manuscript.

Acknowledgements

The study was partially supported by a Grant OPUS 2014/13/B/ST6/01792 financed by the Polish National Science Centre (http://www.ncn.gov.pl) and by the Statutory Funds of the Institute of Pharmacology Polish Academy of Sciences. SP and AJB participate in the European Cooperation in Science and Technology (COST) Action CM1207: GPCR-Ligand Interactions, Structures, and Transmembrane Signalling: an European Research Network (GLISTEN).

Compliance with ethical guidelines

Competing interests The authors declare that they have no competing interests.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Andrzej J Bojarski.

Additional files

Additional file 1:

Analysis of the time course of accuracy values during execution of the SVM optimization procedure. The file contains the analysis of the changes in accuracy values with different time for SVM optimizations strategies for all targets tested.

Additional file 2:

AUC values obtained for all target/fingerprint pairs for curves illustrating changes in the accuracy with time and the optimal accuracy values obtained. The file contains the AUC and optimal accuracy values obtained in the experiments. The AUC values were calculated on the basis of curves illustrating changes in the accuracy values in time for different SVM optimization strategies.

Additional file 3:

Analysis of the number of iterations of the optimization procedure required for reaching the highest accuracy for all tested targets. The file presents the number of iterations after which the optimal accuracy values were reached for all targets tested.

Additional file 4:

Analysis of the changes of accuracy for different steps for all tested targets. The file presents the set of parameters tested in the subsequent iterations of the optimization procedure with the analysis of classification accuracy values that they were providing.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Czarnecki, W.M., Podlewska, S. & Bojarski, A.J. Robust optimization of SVM hyperparameters in the classification of bioactive compounds. J Cheminform 7, 38 (2015). https://doi.org/10.1186/s13321-015-0088-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13321-015-0088-0

Keywords