Skip to main content
Advertisement
  • Loading metrics

Neighborhood Regularized Logistic Matrix Factorization for Drug-Target Interaction Prediction

  • Yong Liu ,

    liuyo@i2r.a-star.edu.sg

    Affiliations Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly (LILY), and School of Computer Engineering, Nanyang Technological University, Singapore, Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), Singapore

  • Min Wu,

    Affiliation Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), Singapore

  • Chunyan Miao,

    Affiliation Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly (LILY), and School of Computer Engineering, Nanyang Technological University, Singapore

  • Peilin Zhao,

    Affiliation Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), Singapore

  • Xiao-Li Li

    Affiliation Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), Singapore

Abstract

In pharmaceutical sciences, a crucial step of the drug discovery process is the identification of drug-target interactions. However, only a small portion of the drug-target interactions have been experimentally validated, as the experimental validation is laborious and costly. To improve the drug discovery efficiency, there is a great need for the development of accurate computational approaches that can predict potential drug-target interactions to direct the experimental verification. In this paper, we propose a novel drug-target interaction prediction algorithm, namely neighborhood regularized logistic matrix factorization (NRLMF). Specifically, the proposed NRLMF method focuses on modeling the probability that a drug would interact with a target by logistic matrix factorization, where the properties of drugs and targets are represented by drug-specific and target-specific latent vectors, respectively. Moreover, NRLMF assigns higher importance levels to positive observations (i.e., the observed interacting drug-target pairs) than negative observations (i.e., the unknown pairs). Because the positive observations are already experimentally verified, they are usually more trustworthy. Furthermore, the local structure of the drug-target interaction data has also been exploited via neighborhood regularization to achieve better prediction accuracy. We conducted extensive experiments over four benchmark datasets, and NRLMF demonstrated its effectiveness compared with five state-of-the-art approaches.

Author Summary

This work introduces a computational approach, namely neighborhood regularized logistic matrix factorization (NRLMF), to predicting potential interactions between drugs and targets. The novelty of NRLMF lies in integrating logistic matrix factorization with neighborhood regularization for drug-target interaction prediction. In NRLMF, we model the interaction probability for each drug-target pair using logistic matrix factorization. As the observed interacting drug-target pairs are experimentally verified, they are more trustworthy than the unknown pairs. We propose to assign higher importance levels to interaction pairs and lower importance levels to unknown pairs. In addition, we further improve the prediction accuracy by neighborhood regularization, which considers the neighborhood influences from most similar drugs and most similar targets. To evaluate the performance of NRLMF, we conducted extensive experiments on four benchmark datasets. The experimental results demonstrated that NRLMF usually outperformed five state-of-the-art methods under three different cross-validation settings, in terms of the area under the ROC curve (AUC) and the area under the precision-recall curve (AUPR). In addition, we confirmed the practical prediction ability of NRLMF by mapping with the latest version of four online biological databases, including ChEMBL, DrugBank, KEGG, and Matador.

Introduction

The drug discovery is one of the primary objectives of the pharmaceutical sciences, which is an interdisciplinary research field of fundamental sciences covering biology, chemistry, physics, statistics, etc. In the drug discovery process, the prediction of drug-target interactions (DTIs) is an important step that aims to identify potential new drugs or new targets for existing drugs. Therefore, it can help guide the experimental validation and reduce costs. In recent years, the DTI prediction has attracted vast research attentions and numerous algorithms have been proposed [1]. Existing methods predict DTIs based on a small number of experimentally validated interactions in existing databases, such as ChEMBL [2], DrugBank [3], KEGG DRUG [4], and SuperTarget [5]. Previous studies have shown that a fraction of new interactions between drugs and targets can be predicted based on the experimentally validated DTIs, and the computational methods for identifying DTIs can significantly improve the drug discovery efficiency.

In general, traditional computational methods proposed for DTI prediction can be categorized into two main groups: docking simulation approaches and ligand-based approaches [68]. The docking simulation approaches predict potential DTIs, considering the structural information of target proteins. However, the docking simulation is extensively time-consuming, and the structural information may not be available for some protein families, for example the G-protein coupled receptors (GPCRs). In the ligand-based approaches, potential DTIs are predicted by comparing a candidate ligand with the known ligands of the target proteins. This kind of approaches may not perform well for the targets with a small number of ligands.

Recently, the quick development of machine learning techniques provides effective and efficient ways to predict DTIs. An intuitive idea is to formulate the DTI prediction as a binary classification problem, where the drug-target pairs are treated as instances, and the chemical structures of drugs and the amino acid subsequences of targets are treated as features. Then, classical classification methods can be used, e.g., support vector machines (SVM) [9] and regularized least square (RLS) [10]. For example, in [11], a SVM model was utilized to classify a given drug-target pair into interaction and non-interaction, considering the amino acid sequences of proteins, chemical structures, and the mass spectrometry data. Bleakley and Yamanishi proposed a supervised approach for DTI prediction based on the bipartite local models (BLMs), where SVM was used to build the local models [12]. Xia et al. proposed a semi-supervised DTI prediction approach, namely Laplacian regularized least square (LapRLS), and extended it to incorporate the kernel constructed from the known DTI network [13]. van Laarhoven et al. defined a Gaussian interaction profile (GIP) kernel to represent the interactions between drugs and targets, and they employed RLS with the GIP kernel for DTI prediction problems [14, 15]. Cheng et al. developed three supervised inference methods for DTI prediction based on the complex network theory [16]. Mei et al. integrated BLM method with a neighbor-based interaction-profile inferring (NII) procedure to form a DTI prediction approach called BLM-NII, where the RLS classifier with GIP kernel was used as the local model [17]. Moreover, Yamanishi et al. developed a web server called DINIES, which utilized supervised machine learning techniques, e.g., pairwise kernel learning and distance metric learning, to predict unknown DTIs from different sources of biological data [18]. Ding et al. used a uniform experimental setting to empirically review the advantages and limitations of existing similarity-based learning approaches for DTI prediction [19]. Furthermore, other auxiliary information has also been exploited for DTI prediction. For example, in [20], Li et al. developed a computational framework that integrated literature mining and the protein and drug connectivity information derived from protein interaction networks to build the disease-specific drug-protein connectivity maps. In [21], Chen et al. utilized the data from public datasets to build a semantic linked network connecting drugs and targets. A statistical model was also proposed to evaluate the association of drug-target pairs.

Essentially, the DTI prediction problem is a recommendation task that aims to suggest a list of potential DTIs. Thus, another line of research for DTI prediction is the application of recommendation technologies. In the literature, collaborative filtering (CF) based approaches are the most widely adopted recommendation methods, which can be categorized into two main groups, i.e., memory-based CF and model-based CF approaches [22, 23]. As the most successful model-based CF approach, matrix factorization has been explored for DTI prediction in recent studies. For example, Gönen proposed a kernelized Bayesian matrix factorization (KBMF) method, which combined the kernel-based dimensionality reduction, matrix factorization, and binary classification for DTI prediction [24]. Cobanoglu et al. utilized probabilistic matrix factorization (PMF) [25] to predict unknown DTIs [26]. The accuracy of the PMF based approach was further improved by an active learning strategy. Moreover, Zheng et al. introduced the multiple similarities collaborative matrix factorization (MSCMF) model, which exploited multiple kinds of drug similarities and target similarities to improve the DTI prediction accuracy [27].

In this paper, we propose a novel matrix factorization approach, namely neighborhood regularized logistic matrix factorization (NRLMF), for DTI prediction. The proposed NRLMF method focuses on predicting the probability that a drug would interact with a target. Specifically, the properties of a drug and a target are represented by two latent vectors in the shared low dimensional latent space, respectively. For each drug-target pair, the interaction probability is modeled by a logistic function of the drug-specific and target-specific latent vectors. This is different from the KBMF method [24] that predicts the interaction probability using a standard normal cumulative distribution function of the drug-specific and target-specific latent vectors [28]. In NRLMF, an observed interacting drug-target pair (i.e., positive observation) is treated as c (c ≥ 1) positive examples, while an unknown pair (i.e., negative observation) is treated as a single negative example. As such, NRLMF assigns higher importance levels to positive observations than negatives. Because the positive observations are biologically validated and thus usually more trustworthy. However, the negative observations could contain potential DTIs and are thus unreliable. This differs from previous matrix factorization based DTI prediction methods [24, 26, 27] that treat the interaction and unknown pairs equally.

Additionally, NRLMF also studies the local structure of the interaction data to further improve the DTI prediction accuracy, by exploiting the neighborhood influences from most similar drugs and most similar targets. In particular, NRLMF imposes individual regularization constraints between the latent representations of a drug and its nearest neighbors, which are most similar with the given drug. Similar neighborhood regularization constraints have also been added on the latent representations of targets. Note that this neighborhood regularization method is different from previous approaches that exploit the drug similarities and target similarities using kernels [13, 14, 17, 29] or factorizing the similarity matrices [27]. Moreover, the proposed approach only considers nearest neighbors instead of all similar neighbors as used in previous approaches, avoiding noisy information, thus achieves more accurate results.

The performances of NRLMF were empirically evaluated on four benchmark datasets, compared with five state-of-the-art DTI prediction methods. Experimental results showed that NRLMF usually outperformed other competing methods on all datasets under different experimental settings, in terms of the widely adopted measures, i.e., the area under the ROC curve (AUC) and the area under the precision-recall curve (AUPR). In addition, the practical prediction ability of NRLMF was also confirmed by mapping with the latest version of online biological databases, including ChEMBL [2], DrugBank [30], KEGG [4], and Matador [5].

Materials and Methods

Materials

The performances of DTI prediction algorithms were evaluated on four benchmark datasets, including Nuclear Receptors, G-Protein Coupled Receptors (GPCR), Ion Channels, and Enzymes. These datasets were originally provided by [31] and were publicly available at http://web.kuicr.kyoto-u.ac.jp/supp/yoshi/drugtarget/. Table 1 summarizes the statistics of all four datasets. Each dataset contains three types of information: 1) the observed DTIs, 2) the drug similarities, and 3) the target similarities. Particularly, the observed DTIs were retrieved from public databases KEGG BRITE [32], BRENDA [33], SuperTarget [5], and DrugBank [3]. The drug similarities were computed based on the chemical structures of the compounds derived from the DRUG and COMPOUND sections in the KEGG LIGAND database [32]. For a pair of compounds, the similarity between their chemical structures was measured by the SIMCOMP algorithm [34]. The target similarities, on the other hand, were calculated based on the amino acid sequences of target proteins, which were retrieved from the KEGG GENES database [32]. The normalized Smith-Waterman score was used to compute the sequence similarity between two proteins.

Problem Formalization

In this paper, the set of drugs is denoted by , and the set of targets is denoted by , where m and n are the number of drugs and number of targets, respectively. The interactions between drugs and targets are represented by a binary matrix , where each element yij ∈ {0, 1}. If a drug di has been experimentally verified to interact with a target tj, yij is set to 1; otherwise, yij is set to 0. The non-zero elements in Y are called “interaction pairs” and regarded as positive observations. The zero elements in Y are called “unknown pairs” and regarded as negative observations. We define the set of positive drugs and targets as and , respectively. Then, the set of negative drugs (i.e., new drugs without any known interaction targets) and negative targets (i.e., new targets without any known interaction drugs) are defined as D = DD+ and T = TT+, respectively. In addition, the drug similarities are represented by , where the (i, μ) element is the similarity between di and dμ. The target similarities are described using , where the (j, ν) element is the similarity between tj and tν.

The objective of this study is to first predict the interaction probability of a drug-target pair and subsequently rank the candidate drug-target pairs according to the predicted probabilities in descending order, such that the top-ranked pairs are the most likely to interact.

Logistic Matrix Factorization

The matrix factorization technique has been successfully applied for DTI prediction in previous studies. In this work, we develop the DTI prediction model based on logistic matrix factorization (LMF) [35], which has been demonstrated to be effective for personalized recommendations. The primary idea of applying LMF for DTI prediction is to model the probability that a drug would interact with a target. In particular, both drugs and targets are mapped into a shared latent space, with a low dimensionality r, where r ≪ min(m, n). The properties of a drug di and a target tj are described by two latent vectors and , respectively. Then, the interaction probability pij of a drug-target pair (di, tj) is modeled by the following logistic function: (1) For simplicity, we further denote the latent vectors of all drugs and all targets by and respectively, where ui is the ith row in U and vj is the jth row in V.

In DTI prediction tasks, the observed interacting drug-target pairs have been experimentally verified, thus they are more trustworthy and important than the unknown pairs. Towards a more accurate modeling for DTI prediction, we propose to assign higher importance levels to the interaction pairs than unknown pairs. In particular, each interaction pair is treated as c (c ≥ 1) positive training examples, and each unknown pair is treated as a single negative training example. Here, c is a constant used to control the importance levels of observed interactions and is empirically set to 5 in the experiments. This importance weighting strategy has been demonstrated to be effective for personalized recommendations [3537]. However, to the best of our knowledge, it has not been explored for DTI prediction in previous studies.

By assuming that all the training examples are independent, the probability of the observations is as follows: (2) Note that when yij = 1, c(1 − yij) = 1 − yij, and when yij = 0, cyij = yij. Hence, we can rewrite Eq (2) as follows: (3)

In addition, we also place zero-mean spherical Gaussian priors on the latent vectors of drugs and targets as: (4) where and are parameters controlling the variances of Gaussian distributions, and I denotes the identity matrix. Hence, through a Bayesian inference, we have (5) The log of the posterior distribution is thus derived as follows: (6) where C is a constant term independent of the model parameters (i.e., U and V). The model parameters can then be learned by maximizing the posterior distribution, which is equivalent with minimizing the following objective function: (7) where , , and ‖⋅‖F denotes the Frobenius norm of a matrix. The problem in Eq (7) can be solved using an alternating gradient descent method [35].

Regularized by Neighborhood

Through mapping both drugs and targets into a shared latent space, the LMF model can effectively estimate the global structure of the DTI data. However, LMF ignores the strong neighborhood associations among a small set of closely related drugs or targets. Thus, we propose to exploit the nearest neighborhood of a drug and that of a target to further improve the DTI prediction accuracy. For a drug di, we denote the set of its nearest neighbors by N(di) ∈ D\di, where N(di) is constructed by choosing K1 most similar drugs with di. Then, we construct the set N(tj) ∈ T\tj, which consists of the K1 most similar targets with tj. In the experiments, we empirically set K1 to 5.

In this paper, the drug neighborhood information is represented using an adjacency matrix A, where the (i, μ) element a is defined as follows: (8) Similarly, the adjacency matrix used to describe the target neighborhood information is denoted by B, where its (j, ν) element b is defined as follows: (9) Note that the adjacency matrices A and B are not symmetric.

The primary idea of exploiting the drug neighborhood information for DTI prediction is to minimize the distances between di and its nearest neighbors N(di) in the latent space. This objective can be achieved by minimizing the following objective function: (10) where tr(⋅) is the trace of a matrix, . Dd and are two diagonal matrices, in which the diagonal elements are and respectively. Moreover, we also exploit the neighborhood information of targets for DTI prediction by minimizing the following objective function: (11) where , Dt and are two diagonal matrices, in which the diagonal elements are and . Note that the proposed neighborhood regularization only considers influences from the K1 nearest neighbors of each drug and each target. It is different from the graph Laplacian constraints used in previous studies [38, 39] which consider influences from all similar drugs and targets. Clearly, given a drug-target pair, we leverage their nearest neighbors, instead of all the neighbors that could potentially introduce noisy information, to enhance the prediction accuracy.

NRLMF

The final DTI prediction model can be formulated by considering the drug-target interactions as well as the neighborhood of drugs and targets. By plugging Eqs (10) and (11) into Eq (7), the proposed NRLMF model is formulated as follows: (12) The optimization problem in Eq (12) can be solved by an alternating gradient ascent procedure. Denoting the objective function in Eq (12) by L, the partial gradients with respect to U and V are as follows: (13) where , in which the (i, j) element is pij (see Eq (1)), ⊙ denotes the Hadamard product of two matrices. To accelerate the convergence of the gradient descent optimization methods, we use the AdaGrad algorithm [40] to adaptively choose the gradient step size. The details of the optimization algorithm to the proposed NRLMF model are described in Algorithm 1, where U and V are randomly initialized using a Gaussian distribution with mean 0, standard deviation .

Algorithm 1: NRLMF

Input: Y, Sd, St, c, r, K1, K2, λd, λt, α, β, γ

Output: U, V

1 Initialize U and V randomly, and set φik = 0, ϕjk = 0, ∀1 ≤ im, 1 ≤ jn, and 1 ≤ kr;

2 Construct the adjacency matrices A and B according to Eq (8) and Eq (9) respectively;

3 Compute the neighborhood regularization matrices Ld and Lt according to Eq (10) and Eq (11) respectively;

4 for t = 1, …, max_iter do

5  ; // fix V and compute the gradient with respect to U

6  for i = 1, …, m do

7   for k = 1, …, r do

   // and uik are the (i, k) element in Gd and U respectively.

8    

9    ; // update each element of di’s latent vector

10  ; // fix U and compute the gradient with respect to V

11  for j = 1, …, n do

12   for k = 1, …, r do

   // and vjk are the (j, k) element in Gt and V respectively.

13    

14    ; // update each element of tj’s latent vector

Once the latent vectors U and V have been learned, the probability associated with any unknown drug-target pair (di, tj) can be predicted by Eq (1). However, in the training procedure, the latent vectors of drugs belonging to the negative drug set D and those of the targets belonging to the negative target set T are learned solely based on negative observations (i.e., unknown pairs). As we know, some negative observations may be potential positive DTIs. Due to such uncertainty over negative observations, the learned latent vectors of the negative drugs and targets may not be accurate enough to describe their properties. One remedy for this problem is to replace the latent vector of a negative drug/target using the linear combination of the latent vectors of its nearest neighbors in the positive set. For a drug diD, we denote the set of its K2 nearest neighbors in D+ by N+(di). Similarly, for a target tjT, the set of its K2 nearest neighbors in T+ is denoted by N+(tj). Note that N+(di) and N+(tj) are built using the same criteria as that used to construct the neighborhood in the training procedure. Then, the prediction of the interaction probability of a drug-target pair (ui, vj) is modified as, (14) where (15) Note that Eq (15) shows a general case for smoothing the learned drug-specific and target-specific latent vectors. In the experiments, K2 is empirically set to 5 to simplify the model.

Results

We have performed extensive experiments to evaluate the performance of the proposed NRLMF method.

Experimental Settings

Following previous studies [1315, 19, 24, 27], the performance of the DTI prediction methods were evaluated under five trials of 10-fold cross-validation (CV), and both AUC and AUPR were used as the evaluation metrics. In particular, for each method, we performed 10-fold CV for five times, each time with a different random seed. Then, we calculated an AUC score in each repetition of CV and reported a final AUC score that was the average over the five repetitions. The AUPR score was calculated in the same manner.

The drug-target interaction matrix had m rows for drugs and n columns for targets. We conducted CV under three different settings as follows [19, 27, 41].

  • CVS1: CV on drug-target pairs—random entries in Y (i.e., drug-target pairs) were selected for testing.
  • CVS2: CV on drugs—random rows in Y (i.e., drugs) were blinded for testing.
  • CVS3: CV on targets—random columns in Y (i.e., targets) were blinded for testing.

Under CVS1, in each round, we used 90% of elements in Y as training data and the remaining 10% of elements as test data. Under CVS2, in each round, we used 90% of rows in Y as training data and the remaining 10% of rows as test data. Under CVS3, in each round, we used 90% of columns in Y as training data and the remaining 10% of columns as test data. Note that these three settings CVS1, CVS2, and CVS3 refer to the DTI prediction for 1) new (unknown) pairs, 2) new drugs, and 3) new targets, respectively.

In this paper, we compared the proposed NRLMF method with the following state-of-the-art methods, namely, NetLapRLS [13], KBMF2K [24], BLM-NII [17], WNN-GIP [15], and CMF [27], by testing their prediction capabilities under the above three settings. The settings of the hyper-parameters of each method were as follows. For the matrix factorization based methods, the dimensionality of the latent space r was selected from {50, 100} [27]. In NRLMF, we set λd = λt and chose these two parameters from {2−5, 2−4, ⋯, 21}. The neighborhood regularization parameters α and β of NRLMF were selected from {2−5, 2−4, ⋯, 22} and {2−5, 2−4, ⋯, 20}, respectively, and the optimal learning rate γ was selected from {2−3, 2−2, ⋯, 20}. In KBMF2K, the margin parameter ν was selected from {0, 1}. For CMF, the regularization coefficient λl was chosen from {2−2, ⋯, 21}, while λd and λt were chosen from {2−3, 2−2, ⋯, 25}. For NetLapRLS, we set γd2/γd1 = γp2/γp1, βd = βp, and chose their values from {10−6, 10−5, ⋯, 102}. In BLM-NII, the linear combination weight α was chosen from {0.0, 0.1, ⋯, 1.0}, and the max function was used to integrate the interaction scores predicted independently from the drug side and the target side. For WNN-GIP, the decay value T was chosen from {0.1, 0.2, ⋯, 0.9}. We set the weighting parameters αd = αt and chose their values from {0.0, 0.1, ⋯, 1.0}. For a machine learning methods, the most suitable hyper-parameters on different datasets are usually different. Thus, we need to choose the optimal hyper-parameters for each method on different datasets. In the literature, the most widely used hyper-parameter optimization strategies are grid search and manual search [42]. In this work, we adopted grid search to choose the optimal hyper-parameters for each DTI prediction method on each dataset. As part of future work, we would like to use the random search strategy proposed in [42] to improve the efficiency of hyper-parameter optimization for DTI prediction methods.

Comparisons with the State-of-the-Arts

Table 2 shows the AUC and AUPR values obtained by various methods under the setting CVS1. As shown in Table 2, NRLMF attains the best AUC values over all datasets. The final average AUC obtained by NRLMF is 0.974, which is 2.10% better than the second method BLM-NII. Moreover, NRLMF achieves the highest AUPR over three datasets (i.e., Nuclear Receptor, GPCR, and Enzyme) and obtains the second best AUPR on the Ion Channel dataset, where CMF outperforms NRLMF (0.923 for CMF vs. 0.906 for NRLMF). The average AUPR obtained by NRLMF is 0.819, which is 4.73% higher than that obtained by the second best method CMF. In summary, under the setting CVS1, NRLMF outperforms other competing methods, being statistically significant except two comparison cases with CMF at the significant level of 0.05 using t-test.

thumbnail
Table 2. The AUC and AUPR obtained under the setting CVS1.

https://doi.org/10.1371/journal.pcbi.1004760.t002

The results obtained under the setting CVS2 for new drugs are shown in Table 3. In particular, NRLMF outperforms the other methods over the Nuclear Receptor, GPCR, and Ion Channel datasets, in terms of AUC. On the Enzyme dataset, WNN-GIP achieves a little better AUC than NRLMF (0.882 for WNN-GIP vs. 0.871 for NRLMF). Over all datasets, NRLMF obtains the best average AUC value 0.870. For the AUPR metric, NRLMF achieves the best results on all datasets except the GPCR dataset, where KBMF2K and CMF are slightly better than NRLMF. Overall, NRLMF achieves the best average AUPR 0.403, which is 13.84% higher than the second-best method KBMF2K and 17.84% higher than the third-best method CMF.

thumbnail
Table 3. The AUC and AUPR obtained under the setting CVS2.

https://doi.org/10.1371/journal.pcbi.1004760.t003

In addition, Table 4 summarizes the results obtained under the setting CVS3 for new targets. We observe that WNN-GIP outperforms other methods on the Nuclear Receptor dataset, in terms of AUC and AUPR. On the other three datastes, the proposed NRLMF achieves the best AUC and AUPR values. Over all datasets, WNN-GIP achieves the highest average AUC value 0.940, which is 1.29% better than the second-best method NRLMF. For the AUPR measure, NRLMF achieves the best average AUPR 0.651, which is a 11.09% better than the second-best method WNN-GIP.

thumbnail
Table 4. The AUC and AUPR obtained under the setting CVS3.

https://doi.org/10.1371/journal.pcbi.1004760.t004

The task under the setting CVS1 focuses on predicting the unknown pair (di, tj), where at least one DTI is known for di and tj respectively in the training data. However, the tasks under CVS2 and CVS3 focus on the predictions for new drugs and new targets respectively, where no DTIs are observed for new drugs and new targets in the training data. Therefore, the task under CVS1 is easier than those under CVS2 and CVS3, and the AUC and AUPR values obtained by DTI prediction methods under CVS1 are higher than those obtained under CVS2 and CVS3 as expected. For all CV settings, the proposed NRLMF method achieves the best AUC values in 10 out of 12 scenarios (i.e., 3 CV settings on 4 datasets) via integrating LMF with neighborhood regularization. In the remaining 2 scenarios (i.e., CVS2 on Enzyme dataset and CVS3 on Nuclear Receptor dataset), WNN-GIP attains better AUC values than NRLMF. The results in these 2 scenarios can be interpreted as follows. For instance, under CVS2, the interactions for 10% of the drugs (i.e., the set of negative drugs D) have been blinded in the training phase. The latent vectors of D are learned solely based on negative observations, and thus are not accurate. This may lead to the inaccuracies of the learned latent vectors of targets (see Eq (13)). Especially, for the targets with only one interaction, the accuracies of the learned latent vectors may be drastically reduced. In NRLMF, the latent vectors of negative drugs and targets are smoothed using their nearest neighbors. However, there is no smoothing for the latent vectors of targets with only one interaction (see Eq (15)). As such, the performances of NRLMF under CVS2 may be affected more on the dataset with a higher percentage of targets that have only one interaction. Interestingly, over 4 datasets, the percentage of targets that have only one interaction is 30.77%, 35.79%, 11.27%, and 43.37%, for Nuclear Receptor, GPCR, Ion Channel, and Enzyme, respectively. Enzyme dataset has the highest percentage of targets with only one interaction, and thus the performance of NRLMF on this dataset under CVS2 is likely to be affected most. Similarly, the percentage of drugs with only one interaction is 72.22% for Nuclear Receptor, 47.53% for GPCR, 38.57% for Ion Channel, and 39.78% for Enzyme. Thus, by blinding the interactions of 10% targets (i.e., under CVS3), the performance of NRLMF on Nuclear Receptor dataset is the most likely to be affected. For the AUPR metric, NRLMF attains the best AUPR values in 9 out of 12 scenarios, which is to be expected, since the methods that optimize AUC are not guaranteed to optimize AUPR [43]. In addition, the target sequence similarity St is more reliable and informative than the drug chemical similarity Sd[14]. Hence, the information propagated from the neighbors to the new targets by the regularization term in Eq (11) will be more accurate than those to new drugs by the term in Eq (10). This explains the results well that various methods usually achieve higher AUC and AUPR under CVS3 than CVS2.

Neighborhood Benefits

The proposed NRLMF method incorporates neighborhood information for DTI prediction via the neighborhood regularization in training and the neighborhood smoothing in prediction. Next, we will study how the neighborhood information benefits DTI prediction under the setting CVS1. For the results under CVS2 and CVS3, please refer to the supporting S1S8 Figs for details.

Fig 1 shows the AUC values obtained by NRLMF with respect to different settings of the neighborhood size K1 used for the neighborhood regularization in the training procedure. As shown in Fig 1, the optimal values of K1 are 3, 5, 5, and 5, for four datasets, respectively. Under the setting CVS1, the average AUC of NRLMF is 0.958 when K1 is set as 0 (i.e., without neighborhood regularization in training), while it is increased to 0.974 when K1 is set as 5. Fig 2 illustrates the AUPR values with respect to different settings of K1. We find that NRLMF achieves the best AUPR by setting K1 as 7, 7, 9, and 3, respectively. When K1 = 0, the average AUPR achieved by NRLMF without neighborhood regularization is 0.772, while it is increased to 0.818 by setting K1 = 5. These results highlight that the neighborhood regularization is highly desirable for DTI prediction.

thumbnail
Fig 1. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of K1 under CVS1.

The best competitors on these datasets are (a) BLM-NII, (b) BLM-NII, (c) BLM-NII and CMF, and (d) BLM-NII, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.g001

thumbnail
Fig 2. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of K1 under CVS1.

The best competitors on these datasets are (a) BLM-NII, (b) CMF, (c) CMF, and (d) CMF, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.g002

In addition, we also study the impact of the neighborhood size K2 used for neighborhood smoothing in the prediction procedure. Figs 3 and 4 plot the AUC and AUPR values obtained by NRLMF with respect to different settings of K2. As shown in Fig 3, NRLMF achieves best AUC via setting K2 as 5, 3, 5, and 5, respectively. For AUPR measure, the best results are achieved by setting K2 as 5, 3, 9, and 5, respectively. Over all datasets, when K2 = 0 (i.e., without neighborhood smoothing in prediction), the average AUC and AUPR values obtained by NRLMF are 0.950 and 0.772, respectively, while these values are 0.974 and 0.819 when K2 = 5. These observations demonstrate the effectiveness of nearest neighbors to predict the interaction probability for a given drug-target pair. In addition, when we set K1 and K2 as 5, we can get reasonably good results for both AUC and AUPR, respectively.

thumbnail
Fig 3. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of K2 under CVS1.

The best competitors on these datasets are (a) BLM-NII, (b) BLM-NII, (c) BLM-NII and CMF, and (d) BLM-NII, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.g003

thumbnail
Fig 4. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of K2 under CVS1.

The best competitors on these datasets are (a) BLM-NII, (b) CMF, (c) CMF, and (d) CMF, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.g004

Parameter Sensitivity Analysis for c and r

In this section, we focus on the sensitivity analysis for other two parameters, i.e., the importance levels of observed DTIs c and the dimensionality of the latent space r, under the setting CVS1. As to the performance trend of NRLMF with respect to different settings for c and r under CVS2 and CVS3, please refer to the supporting S9S16 Figs for details.

As shown in Fig 5, when the importance level c is set as 1 (i.e., without importance weighting), NRLMF outperforms other competitors on Nuclear Receptor, GPCR, and Ion Channel datasets, and is comparable with the best competitor on the Enzyme dataset (0.971 for NRLMF vs. 0.978 for the best competitor), in terms of AUC. This again highlights the effectiveness of integrating logistic matrix factorization with neighborhood regularization for DTI prediction. By setting c = 5, NRLMF is able to achieve the optimal AUC values and outperforms all competing methods over all datasets. For the AUPR metric, Fig 6 shows that NRLMF with setting c = 1 outperforms other competitors on the Nuclear Receptor dataset and performs poorer than the best competitor on the remaining three datasets. This is expected, since the methods that optimize AUC are not guaranteed to optimize AUPR [43]. In addition, NRLMF achieves better AUPR under the setting c > 1 than under the setting c = 1, on the GPCR, Ion Channel, and Enzyme datasets. On the Nuclear Receptor dataset, NRLMF attains slightly better AUPR under the setting c = 1 than under the other settings. These observations demonstrate that assigning more importance on the observed interactions can boost the performance of NRLMF. However, when c is large enough, the performance of NRLMF tends to become saturated, where further increasing c has very limited improvement.

thumbnail
Fig 5. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of c under CVS1.

The best competitors on these datasets are (a) BLM-NII, (b) BLM-NII, (c) BLM-NII and CMF, and (d) BLM-NII, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.g005

thumbnail
Fig 6. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of c under CVS1.

The best competitors on these datasets are (a) BLM-NII, (b) CMF, (c) CMF, and (d) CMF, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.g006

The impact of the dimensionality of the latent space r on the performance of NRLMF, in terms of AUC and AUPR, is shown in Figs 7 and 8, respectively. We find that larger r generally achieves better results. The two exceptions are the AUPR measure on Nuclear Receptor and Ion Channel datasets, where r = 30 leads to slightly better results than r = 50. Nevertheless, r = 100 achieves the best results or the second best results measured by AUC and AUPR, on all datasets. Thus, the parameter r is recommended to be set in the range [50, 100], which is consistent with previous studies [27].

thumbnail
Fig 7. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of r under CVS1.

The best competitors on these datasets are (a) BLM-NII, (b) BLM-NII, (c) BLM-NII and CMF, and (d) BLM-NII, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.g007

thumbnail
Fig 8. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of r under CVS1.

The best competitors on these datasets are (a) BLM-NII, (b) CMF, (c) CMF, and (d) CMF, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.g008

Predicting Novel Interactions

In this section, we evaluate the practical ability of NRLMF on predicting novel interactions, which refer to interactions with high probabilities that do not occur in the benchmark datasets. Following similar settings in previous studies [12, 14, 15, 19, 24], four well-known biological databases, i.e., ChEMBL [2], DrugBank [30], KEGG [4], and Matador [5], are used as references to verify whether the predicted new DTIs are true or not.

To conduct this study, we have collected the online profiles associated with the drugs and targets in each benchmark dataset from the online reference databases and parsed the approved drug-target interactions. Over all benchmark datasets, there are 791 drugs and 986 targets, and 1,999 novel interactions have been confirmed in one or more reference databases. The number of confirmed novel interactions in Nuclear Receptor, GPCR, Ion Channel, and Enzyme datasets are 21, 512, 1034, and 432, respectively. For each dataset, the entire dataset is used as training set. The unknown interactions will be ranked based on the interaction probabilities predicted using the optimal parameters learned under CVS1 instead of those learned under other two settings (i.e., CVS2 and CVS3). This is because that our objective is to predict those novel likely drug-target interactions, instead of focusing a new drug or a new target. Then, the predicted novel interactions are the top ranked unknown drug-target interaction pairs.

Table 5 shows the top 30 novel interactions predicted by NRLMF on the GPCR dataset. In this table, the DTIs are bolded to indicate that they exist in one or more of the reference databases. The third column of Table 5 shows the predicted interaction probability of a drug-target pair. For each pair, the databases containing it are listed in the last column of the table, where C is short for ChEMBL, D for DrugBank, K for KEGG, and M for Matador. For example, the highest ranked DTI is (D00283, hsa1814) with predicted probability 0.9181, which exists in the databases ChEMBL, DrugBank, and Metador. We find that 67% of the predictions (20 out of 30) are currently confirmed in at least one of the reference databases. Since these databases are still being updated as new DTIs are found, the fraction of new DTIs correctly predicted by NRLMF may increase in the future. This encouraging result that NRLMF can successfully detect quite a few novel interactions that are not in the GPCR dataset, implies that NRLMF is very effective in predicting new true DTIs from sparse matrices consisted of very few DTIs.

thumbnail
Table 5. The top 30 novel interactions predicted by NRLMF on GPCR dataset.

https://doi.org/10.1371/journal.pcbi.1004760.t005

Finally, Table 6 summarizes the fractions of true DTIs among the top N (N = 10, 30, 50) predictions generated by various DTI methods, using the optimal parameters learned under CVS1. We observe that NRLMF is able to achieve consistently accurate prediction results across all the datasets. For example, the fractions of true DTIs among the top 10 predicted interactions are 50%, 60%, 50%, and 90% for all datasets, respectively. Compared with other methods, NRLMF is able to achieve comparable or even better prediction results across all the datasets. These observations indicate that the proposed algorithm is very effective for finding novel DTIs, thus it may help biologists or clinicians significantly reduce the cost of biological test. For more details about the novel DTI prediction, please refer to the supporting S1S4 Texts, where the top 1000 novel DTIs predicted by NRLMF are provided.

thumbnail
Table 6. The fractions of true DTIs among the predicted top N (N = 10, 30, 50) interactions under CVS1.

https://doi.org/10.1371/journal.pcbi.1004760.t006

Discussion

This paper presents a novel drug-target interaction prediction method, namely neighborhood regularized logistic matrix factorization (NRLMF). The novelty of NRLMF comes from integrating logistic matrix factorization with neighborhood regularization to predict the interaction probability of a given drug-target pair. Specifically, both drugs and targets are mapped into a shared latent space, and the drug-target interactions are modeled using the linear combinations of the drug-specific and target-specific latent vectors. In addition, higher importance level is assigned to the positive observations (i.e., interaction pairs), while lower level is for negative observations (i.e., unknown pairs). Moreover, the neighborhood regularization based on the drug similarities and target similarities is utilized to further improve the prediction ability of the model.

To evaluate the performance of NRLMF, an extensive set of experiments were performed on four benchmark datasets, compared with five state-of-the-art DTI prediction methods. The promising results further validated the empirical efficacy of the proposed algorithm. For example, on average, NRLMF attains the best AUC values under CVS1 and CVS2, and the second best AUC value under CVS3. In terms of AUPR, NRLMF achieves the best averaged AUPR values over all datasets, under all three CV settings. These results indicate that NRLMF outperforms existing state-of-the-art methods in predicting new pairs and new drugs, and is comparable or even better than existing methods in predicting new targets. However, on the dataset with a large fraction of drugs which have only one interaction (e.g., 72.22% on the Nuclear Receptor dataset), WNN-GIP may outperform NRLMF in predicting new targets. On the dataset with a large fraction of targets which have only one interaction (e.g., 43.37% on the Enzyme dataset), WNN-GIP may achieve better results than NRLMF in predicting new drugs. In addition, the high practical predicting ability of NRLMF have also been verified. For example, on the Enzyme dataset, 90% of the top 10 novel DTIs predicted by NRLMF have been confirmed by the latest version of four well-known biological databases, including ChEMBL, DrugBank, KEGG, and Matador.

The optimization problem of NRLMF is solved using an alternating gradient descent optimization algorithm, the time complexity of which is O(iterrmn), where iter denotes the number of iterations. However, the time complexity of the solutions to the other two matrix factorization based DTI prediction methods (i.e., KBMF2K and CMF) are O(iter ⋅ (rm3+rn3+r3)) and O(iter ⋅ (r2 ⋅ (m+n)2+r3 ⋅ (m+n))), respectively. Therefore, NRLMF is more efficient than KBMF2K and CMF. In addition, NRLMF can also be extended to incorporate multiple types of similarities from drugs and targets for DTI prediction. One direction for future work is to couple logistic matrix factorization with the multiple kernel learning techniques [44]. Another potential direction for future work is to exploit boosting technique, e.g., the AdaBPR model in [45], to improve the prediction accuracy of the proposed NRLMF method.

Supporting Information

S1 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of K1 under CVS2.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) KBMF2K, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s001

(EPS)

S2 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of K1 under CVS2.

The best competitors on these datasets are (a) WNN-GIP, (b) KBMF2K, (c) KBMF2K, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s002

(EPS)

S3 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of K1 under CVS3.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) WNN-GIP, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s003

(EPS)

S4 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of K1 under CVS3.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) BLM-NII, and (d) BLM-NII, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s004

(EPS)

S5 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of K2 under CVS2.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) KBMF2K, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s005

(EPS)

S6 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of K2 under CVS2.

The best competitors on these datasets are (a) WNN-GIP, (b) KBMF2K, (c) KBMF2K, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s006

(EPS)

S7 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of K2 under CVS3.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) WNN-GIP, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s007

(EPS)

S8 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of K2 under CVS3.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) BLM-NII, and (d) BLM-NII, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s008

(EPS)

S9 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of c under CVS2.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) KBMF2K, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s009

(EPS)

S10 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of c under CVS2.

The best competitors on these datasets are (a) WNN-GIP, (b) KBMF2K, (c) KBMF2K, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s010

(EPS)

S11 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of c under CVS3.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) WNN-GIP, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s011

(EPS)

S12 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of c under CVS3.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) BLM-NII, and (d) BLM-NII, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s012

(EPS)

S13 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of r under CVS2.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) KBMF2K, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s013

(EPS)

S14 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of r under CVS2.

The best competitors on these datasets are (a) WNN-GIP, (b) KBMF2K, (c) KBMF2K, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s014

(EPS)

S15 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUC with different settings of r under CVS3.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) WNN-GIP, and (d) WNN-GIP, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s015

(EPS)

S16 Fig. Performance trend of NRLMF on the benchmark datasets (a) Nuclear Receptor, (b) GPCR, (c) Ion Channel, and (d) Enzyme, measured by AUPR with different settings of r under CVS3.

The best competitors on these datasets are (a) WNN-GIP, (b) WNN-GIP, (c) BLM-NII, and (d) BLM-NII, respectively.

https://doi.org/10.1371/journal.pcbi.1004760.s016

(EPS)

S1 Text. The top 1000 novel interactions predicted by NRLMF on the Nuclear Receptor Dataset.

https://doi.org/10.1371/journal.pcbi.1004760.s017

(TXT)

S2 Text. The top 1000 novel interactions predicted by NRLMF on the GPCR Dataset.

https://doi.org/10.1371/journal.pcbi.1004760.s018

(TXT)

S3 Text. The top 1000 novel interactions predicted by NRLMF on the Ion Channel Dataset.

https://doi.org/10.1371/journal.pcbi.1004760.s019

(TXT)

S4 Text. The top 1000 novel interactions predicted by NRLMF on the Enzyme Dataset.

https://doi.org/10.1371/journal.pcbi.1004760.s020

(TXT)

Acknowledgments

The authors would like to express thanks to Dr. Sicong Chen for her kindly help.

Author Contributions

Conceived and designed the experiments: YL MW PZ XLL. Performed the experiments: YL. Analyzed the data: YL MW. Contributed reagents/materials/analysis tools: YL CM. Wrote the paper: YL MW CM PZ XLL. Designed the software used in analysis: YL.

References

  1. 1. Dai YF, Zhao XM. A survey on the computational approaches to identify drug targets in the postgenomic era. Biomed Res Int. 2015;2015.
  2. 2. Gaulton A, Bellis LJ, Bento AP, Chambers J, Davies M, Hersey A, et al. ChEMBL: a large-scale bioactivity database for drug discovery. Nucleic Acids Res. 2012;40(D1):D1100–D1107. pmid:21948594
  3. 3. Wishart DS, Knox C, Guo AC, Cheng D, Shrivastava S, Tzur D, et al. DrugBank: a knowledgebase for drugs, drug actions and drug targets. Nucleic Acids Res. 2008;36(suppl 1):D901–D906. pmid:18048412
  4. 4. Kanehisa M, Goto S, Sato Y, Furumichi M, Tanabe M. KEGG for integration and interpretation of large-scale molecular data sets. Nucleic Acids Res. 2012;40(gkr988):D109–D114. pmid:22080510
  5. 5. Günther S, Kuhn M, Dunkel M, Campillos M, Senger C, Petsalaki E, et al. SuperTarget and Matador: resources for exploring drug-target relationships. Nucleic Acids Res. 2008;36(suppl 1):D919–D922. pmid:17942422
  6. 6. Kinnings SL, Liu N, Buchmeier N, Tonge PJ, Xie L, Bourne PE. Drug discovery using chemical systems biology: repositioning the safe medicine Comtan to treat multi-drug and extensively drug resistant tuberculosis. PLoS Comput Biol. 2009;5(7):e1000423. pmid:19578428
  7. 7. Li YY, An J, Jones SJM. A computational approach to finding novel targets for existing drugs. PLoS Comput Biol. 2011;7(9):e1002139–e1002139. pmid:21909252
  8. 8. Wang K, Sun J, Zhou S, Wan C, Qin S, Li C, et al. Prediction of drug-target interactions for drug repositioning only based on genomic expression similarity. PLoS Comput Biol. 2013;9(11):e1003315. pmid:24244130
  9. 9. Cristianini N, Shawe-Taylor J. An introduction to support vector machines and other kernel-based learning methods. Cambridge University Press; 2000.
  10. 10. Suykens JAK, Vandewalle J. Least squares support vector machine classifiers. Neural Processing Letters. 1999;9(3):293–300.
  11. 11. Nagamine N, Sakakibara Y. Statistical prediction of protein–chemical interactions based on chemical structure and mass spectrometry data. Bioinformatics. 2007;23(15):2004–2012. pmid:17510168
  12. 12. Bleakley K, Yamanishi Y. Supervised prediction of drug–target interactions using bipartite local models. Bioinformatics. 2009;25(18):2397–2403. pmid:19605421
  13. 13. Xia Z, Wu LY, Zhou X, Wong ST. Semi-supervised drug-protein interaction prediction from heterogeneous biological spaces. BMC Syst Biol. 2010;4(suppl 2):S6. pmid:20840733
  14. 14. van Laarhoven T, Nabuurs SB, Marchiori E. Gaussian interaction profile kernels for predicting drug–target interaction. Bioinformatics. 2011;27(21):3036–3043. pmid:21893517
  15. 15. van Laarhoven T, Marchiori E. Predicting drug-target interactions for new drug compounds using a weighted nearest neighbor profile. PloS One. 2013;8(6):e66952. pmid:23840562
  16. 16. Cheng F, Liu C, Jiang J, Lu W, Li W, Liu G, et al. Prediction of drug-target interactions and drug repositioning via network-based inference. PLoS Comput Biol. 2012;8(5):e1002503. pmid:22589709
  17. 17. Mei JP, Kwoh CK, Yang P, Li XL, Zheng J. Drug–target interaction prediction by learning from local information and neighbors. Bioinformatics. 2013;29(2):238–245. pmid:23162055
  18. 18. Yamanishi Y, Kotera M, Moriya Y, Sawada R, Kanehisa M, Goto S. DINIES: drug–target interaction network inference engine based on supervised analysis. Nucleic Acids Res. 2014;42(W1):W39–W45. pmid:24838565
  19. 19. Ding H, Takigawa I, Mamitsuka H, Zhu S. Similarity-based machine learning methods for predicting drug–target interactions: a brief review. Brief Bioinform. 2014;15(5):734–747. pmid:23933754
  20. 20. Li J, Zhu X, Chen JY. Building disease-specific drug-protein connectivity maps from molecular interaction networks and PubMed abstracts. PLoS Comput Biol. 2009;5(7):e1000450. pmid:19649302
  21. 21. Chen B, Ding Y, Wild DJ. Assessing drug target association using semantic linked data. PLoS Comput Biol. 2012;8(7):e1002574. pmid:22859915
  22. 22. Adomavicius G, Tuzhilin A. Toward the next generation of recommender systems: a survey of the state-of-the-art and possible extensions. IEEE Trans Knowl Data Eng. 2005;17(6):734–749.
  23. 23. Su X, Khoshgoftaar TM. A survey of collaborative filtering techniques. Advances in Artificial Intelligence. 2009;2009:4.
  24. 24. Gönen M. Predicting drug–target interactions from chemical and genomic kernels using Bayesian matrix factorization. Bioinformatics. 2012;28(18):2304–2310. pmid:22730431
  25. 25. Salakhutdinov R, Mnih A. Probabilistic matrix factorization. Adv Neural Inf Process Syst 21. 2008; p. 1257–1264.
  26. 26. Cobanoglu MC, Liu C, Hu F, Oltvai ZN, Bahar I. Predicting drug–target interactions using probabilistic matrix factorization. J Chem Inf Model. 2013;53(12):3399–3409. pmid:24289468
  27. 27. Zheng X, Ding H, Mamitsuka H, Zhu S. Collaborative matrix factorization with multiple similarities for predicting drug-target interactions. KDD’13: Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2013; p. 1025–1033.
  28. 28. Gonen M, Kaski S. Kernelized Bayesian matrix factorization. IEEE Trans Pattern Anal Mach Intell. 2014;36(10):2047–2060. pmid:26352634
  29. 29. Jacob L, Vert JP. Protein-ligand interaction prediction: an improved chemogenomics approach. Bioinformatics. 2008;24(19):2149–2156. pmid:18676415
  30. 30. Knox C, Law V, Jewison T, Liu P, Ly S, Frolkis A, et al. DrugBank 3.0: a comprehensive resource for ‘omics’ research on drugs. Nucleic Acids Res. 2011;39(suppl 1):D1035–D1041. pmid:21059682
  31. 31. Yamanishi Y, Araki M, Gutteridge A, Honda W, Kanehisa M. Prediction of drug–target interaction networks from the integration of chemical and genomic spaces. Bioinformatics. 2008;24(13):i232–i240. pmid:18586719
  32. 32. Kanehisa M, Goto S, Hattori M, Aoki-Kinoshita KF, Itoh M, Kawashima S, et al. From genomics to chemical genomics: new developments in KEGG. Nucleic Acids Res. 2006;34(suppl 1):D354–D357. pmid:16381885
  33. 33. Schomburg I, Chang A, Ebeling C, Gremse M, Heldt C, Huhn G, et al. BRENDA, the enzyme database: updates and major new developments. Nucleic Acids Res. 2004;32(suppl 1):D431–D433. pmid:14681450
  34. 34. Hattori M, Okuno Y, Goto S, Kanehisa M. Development of a chemical structure comparison method for integrated analysis of chemical and genomic information in the metabolic pathways. J Am Chem Soc. 2003;125(39):11853–11865. pmid:14505407
  35. 35. Johnson CC. Logistic matrix factorization for implicit feedback data. NIPS 2014 Workshop on Distributed Machine Learning and Matrix Computations. 2014;.
  36. 36. Hu Y, Koren Y, Volinsky C. Collaborative filtering for implicit feedback datasets. ICDM’08: Proceedings of the 8th IEEE International Conference on Data Mining. 2008; p. 263–272.
  37. 37. Liu Y, Wei W, Sun A, Miao C. Exploiting geographical neighborhood characteristics for location recommendation. CIKM’14: Proceedings of the 23rd ACM International Conference on Information and Knowledge Management. 2014; p. 739–748.
  38. 38. Hwang T, Atluri G, Xie M, Dey S, Hong C, Kumar V, et al. Co-clustering phenome–genome for phenotype classification and disease gene discovery. Nucleic Acids Res. 2012;40(19):e146. pmid:22735708
  39. 39. Gligorijević V, Janjić V, Pržulj N. Integration of molecular network data reconstructs gene ontology. Bioinformatics. 2014;30(17):i594–i600. pmid:25161252
  40. 40. Duchi J, Hazan E, Singer Y. Adaptive subgradient methods for online learning and stochastic optimization. J Mach Learn Res. 2011;12:2121–2159.
  41. 41. Pahikkala T, Airola A, Pietilä S, Shakyawar S, Szwajda A, Tang J, et al. Toward more realistic drug–target interaction predictions. Briefings Bioinf. 2015;16:325–337.
  42. 42. Bergstra J, Bengio Y. Random search for hyper-parameter optimization. J Mach Learn Res. 2012;13(1):281–305.
  43. 43. Davis J, Goadrich M; ACM. The relationship between precision-recall and roc curves. ICML’06: Proceedings of the 23rd International Conference on Machine Learning. 2006; p. 233–240.
  44. 44. Gönen M, Alpaydın E. Multiple kernel learning algorithms. J Mach Learn Res. 2011;12:2211–2268.
  45. 45. Liu Y, Zhao P, Sun A, Miao C. A boosting algorithm for item recommendation with implicit feedback. IJCAI’15: Proceedings of the 24th International Conference on Artificial Intelligence. 2015; p. 1792–1798.