Skip to main content
Advertisement
  • Loading metrics

Vicus: Exploiting local structures to improve network-based analysis of biological data

Abstract

Biological networks entail important topological features and patterns critical to understanding interactions within complicated biological systems. Despite a great progress in understanding their structure, much more can be done to improve our inference and network analysis. Spectral methods play a key role in many network-based applications. Fundamental to spectral methods is the Laplacian, a matrix that captures the global structure of the network. Unfortunately, the Laplacian does not take into account intricacies of the network’s local structure and is sensitive to noise in the network. These two properties are fundamental to biological networks and cannot be ignored. We propose an alternative matrix Vicus. The Vicus matrix captures the local neighborhood structure of the network and thus is more effective at modeling biological interactions. We demonstrate the advantages of Vicus in the context of spectral methods by extensive empirical benchmarking on tasks such as single cell dimensionality reduction, protein module discovery and ranking genes for cancer subtyping. Our experiments show that using Vicus, spectral methods result in more accurate and robust performance in all of these tasks.

Author summary

Networks are a representation of choice for many problems in biology and medicine including protein interactions, metabolic pathways, evolutionary biology, cancer subtyping and disease modeling to name a few. The key to much of network analysis lies in the spectrum decomposition represented by eigenvectors of the network Laplacian. While possessing many desirable algebraic properties, Laplacian lacks the power to capture fine-grained structure of the underlying network. Our novel matrix, Vicus, introduced in this work, takes advantage of the local structure of the network while preserving algebraic properties of the Laplacian. We show that using Vicus in spectral methods leads to superior performance across fundamental biological tasks such as dimensionality reduction in single cell analysis, identifying genes for cancer subtyping and identifying protein modules in a PPI network. We postulate, that in tasks where it is important to take into account local network information, spectral-based methods should be using Vicus matrix in place of Laplacian.

This is a PLOS Computational Biology Methods paper.

Introduction

Networks are a powerful paradigm for representing relations among objects from micro to macro level. It is no surprise that networks became a representation of choice for many problems in biology and medicine including gene-gene and protein-protein interaction networks [1], diseases [2] and their interrelations [3], cancer subtyping [4], genetic diversity [5], image retrieval [6], dimensionality reduction [7, 8] and many other applications. Computational biologists routinely use networks to represent data and analyze networks to obtain better understanding of patterns and local structures hidden in the complex data they encode. One of the most standard graph-based methods to analyze networks is to decompose it into eigenvectors and eigenvalues, i.e. apply spectral methods to the network to understand its structure. At the heart of spectral methods is the so-called Laplacian matrix. Spectral clustering relies on the fact that the principle eigenvectors of the Laplacian capture membership of nodes in implicit network clusters. This principle is essential to clustering and dimensionality reduction.

The traditional formulation of the Laplacian captures the global structure of the matrix, which is often insufficient in biology where local topologies are what needs to be sought and exploited. Moreover, recently algorithms designed to capture the local structure of the data have been shown to significantly outperform global methods [9, 10]. These approaches aim to reconstruct each data point using its local neighbours and have been shown to be robust and powerful for unweighted networks. Weighted networks are richer representations of underlying data than unweighted networks: in biological networks weights can represent the strength of interactions or the strength of the evidence underlying each interaction, in patient networks weights represent the degree of similarity between patients [4]. In this paper, we provide a local formulation of the Laplacian for weighted networks which we call the Vicus matrix (), from the Latin word ‘neighborhood’. Using Vicus in place of the Laplacian allows spectral methods to exploit local structures and makes them a lot more relevant to a variety of biological applications.

In this paper we introduce Vicus and compare its performance to the Laplacian across a wide range of tasks. Our experiments include single cell dimensionality reduction, protein module discovery, feature ranking and large scale network clustering. Since we consider such a diverse set of biological questions, in each case we also compare to appropriate state-of-the-art methods corresponding to each question. Spectral clustering using Vicus outperforms competing approaches in all of these tasks. Our experiments show that Vicus is a more robust alternative to traditional Laplacian matrix for network analysis.

Results

Simulations: Laplacian vs Vicus

In this section we consider predetermined 2D and 3D structures, represent them as a graph and analyze the performance of local Vicus as compared to traditional Laplacian in the task of graph-based dimensionality reduction.

First, let us consider a particular type of protein fold that has a complex structure in which four pairs of antiparallel beta sheets, only one of which is adjacent in sequence, are wrapped in three dimensions to form a barrel shape. This structure known as jelly roll or Swiss roll is particularly common in viral proteins and is schematically depicted in Fig 1A. Spectral methods assume that clusters of data points can be well described by the Euclidean distance. Though it looks relatively unambiguous to a human, this task is computationally challenging since the assumption that Euclidean proximity translates to similarity does not hold in the original data space for the Swiss roll structure. As expected, standard spectral decomposition fails to find a lower dimensional representation of the data due to the inability to capture the underlying manifolds in Fig 1A. Using Vicus in place of the Laplacian matrix helps spectral decomposition to transform the original data to the latent space with reduced complexity while preserving the contiguity and the cluster memberships of the original data.

thumbnail
Fig 1. Four examples of three dimensional manifolds (left column) and their embeddings by global Laplacian (middle column) and the proposed Vicus (right column).

A shows an example of the structure known as jelly roll or Swiss roll which is particularly common in viral proteins. B shows five random non-overlapping clusters in 3D space connected by sparsely measured channels. C shows a toroidal helix containing a circle as its basic geometric shape. D is an example of sampling in 3D space where we sample points from a solid bowl-shaped figure non-uniformly: the top of the bowl is more densely sampled, gradually reducing sampling towards the bottom of the bowl graph. On all the cases, Vicus is able to recover the underlying distributions of the input data more robustly.

https://doi.org/10.1371/journal.pcbi.1005621.g001

Another simulation that we considered is a typical example in bioinformatic imaging, structured 3D data. A schematic of clustered signal within brain regions and connecting channels between them is captured in Fig 1B. Given five random non-overlapping clusters in 3D space connected by sparsely measured channels, Vicus maps the clusters into dense points while preserving the lines connecting them. This embedding indicates that, by considering local structures, local spectrum can highlight the obvious cluster structures without disregarding the structure of the data between clusters. By comparison, Laplacian-based embedding highlights the dense clusters while making the connectivity between them more ambiguous (Fig 1B). This example sheds light on how Vicus can preserve local structure of the data.

A very common structure in protein folding is a helix. Among such foldings are toroidal helices, where the helix is wrapped around a toroid. These structures have a pore in the middle that allows unfolded DNA to pass through. The toroidal helix in Fig 1C has a circle as its basic geometric shape. Our local spectrum recovers the underlying 2D circle by considering the labels in local neighborhoods while the Laplacian finds a circle distorted by similarities of points in the 3D dimension. The distortions by the global spectrum result from a fundamental limitation in descriptive power of Euclidean distance in high dimensional spaces, while our local spectrum can avoid such limitation by focusing on the local rather than the global manifold structures.

Our final example is the task of sampling in 3D space, such as sampling an image of a cell shape in a cell morphology study. We sampled points from a solid bowl-shaped figure (Fig 1D) non-uniformly: the top of the bowl is more densely sampled, gradually reducing sampling towards the bottom of the bowl graph. The Laplacian based 2D embedding has considerable bias towards the densely sampled region while Vicus’ embedding recognizes that sampling was done on a solid shape, again by capturing the labels in the local neighbourhoods.

These examples show the benefits of capturing local structure in a network (graph) decomposition, which gives a better understanding of patterns and neighborhoods hidden in complex networks.

A case study in single-cell RNA-seq analysis

Single-cell RNA sequencing (scRNA-seq) technologies have recently emerged as a powerful means to measure gene expression levels of individual cells [11]. Quantifying the variation across gene expression profiles of individual cells is key to the dissection of the heterogeneity and the identification of new populations among cells. The unique challenges associated with single-cell RNA-seq data include large noise in quantification of transcriptomes and high dropout rates, therefore reducing the usability of traditional unsupervised clustering methods. Vicus, employing local structures hidden in high-dimensional data, is able to tackle these challenges and improve many types of single-cell analyses including visualization, clustering and gene selection.

We benchmark our method on four recently published single-cell RNA-seq datasets with validated cell populations:

  • Pollen data set [12] consists of 11 cell populations including neural cells and blood cells.
  • Usoskin data set [13] consists of neuronal cells with sensory subtypes. This data set contains 622 cells from the mouse dorsal root ganglion, with an average of 1.14 million reads per cell. The authors divided the cells into four neuronal types: peptidergic nociceptors, non-peptidergic nociceptors, containing neurofilament and containing tyrosine hydroxylase.
  • Buettner data set [14] consists of embryonic stem cells in different cell cycle stages. This dataset was obtained from a controlled study that quantified the effect of the cell cycle on gene expression level in individual mouse embryonic stem cells (mESCs).
  • Kolodziejczyk data set [15] consists of pluripotent cells under different environmental conditions. This data set was obtained from a stem cell study on how different culture conditions influence pluripotent states of mESCs. Preprocessed data was obtained directly from [11].

The main reason we chose these four single-cell datasets is that their ground-truth labels have been validated either experimentally or computationally in their original studies. We formulate the problem of clustering cells from RNA-seq data in terms of networks. First, cell-to-cell similarity networks (Materials and methods) are constructed from single-cell RNA-seq data. The advantage of using networks to represent this data are in network’s ability to capture a set of relationship between all pairs of cells. After the construction of cell-to-cell networks, we can apply our Vicus to obtain a low-dimensional representation that contains local structures in the networks and potential cluster memberships of cells.

To demonstrate the representative power of the low-dimensional representations by Vicus, we ran t-SNE [16], the most common visualization method in single-cell studies, on the obtained low-dimensional representations and compare the 2-D visualization of both Vicus and Laplacian across the four single-cell datasets in Fig 2. Note that we are only using t-SNE for the purpose of visualization of Laplacian and Vicus. The cells, color-coded by the ground-truth labels from original studies [1215], are clearly separated by Vicus (Fig 2), indicating greater power of Vicus to capture fine-grained structures in cell-to-cell similarity networks.

thumbnail
Fig 2. Visualization of low-dimensional representations for single cells learned by Vicus and global Laplacian.

Four columns represent the embedding results for Buettner data, Kolodziejczyk data, Pollen data, and Usoskin data respectively. In each dataset, cells are color-coded as their ground-truth labels. Larger separations between different clusters usually indicate better performances in low-dimensional embeddings.

https://doi.org/10.1371/journal.pcbi.1005621.g002

We compare spectral decomposition using Vicus with spectral methods using traditional global Laplacian along with 6 other popular dimensionality reduction methods. The six methods include linear methods such as Principle Component Analysis (PCA), Factor Analysis(FA), and Probabilistic PCA (PPCA) and nonlinear methods such as multidimensional scaling (MDS), Kernel PCA, Maximum Variance Unfolding (MVU), Locality Preserving Projection (LPP) and Sammon mapping. We use a widely-used toolbox [16] implementing all these popular dimensionality reduction methods. Further, we also compare Vicus with three widely used state-of-the-art network-based clustering algorithms: InfoMap [17], modularity-based Louvian [18], and Affinity Propagation (AP) [19]. To compare these 11 methods we adopted two metrics: Normalized Mutual Information(NMI) [20] and Adjusted Rand Index(ARI) [21] (Materials and methods), evaluating the concordance of obtained label and the ground-truth. Higher values of these evaluation metrics indicate better ability of correctly identifying cell populations.

Results in Table 1 illustrate Vicus’ superior performances compared to all ten other methods in most of the considered cases. It is noticeable that Vicus outputs much better module detection results than all other methods on Buettner data set [14]. This is due to the fact that Buettner data set [14] contains cells in three different continuous cell stages which are hard to detect due to large noise. In addition, PCA performs the best on Pollen data set [12] because the ground-truth is obtained by simple clustering with PCA on a set of pre-selected genes. Further, compared with the three network-based module detection methods (InfoMap, Louvian and AP), our Vicus is able to achieve much more accurate module discovery on each of the same networks.

thumbnail
Table 1. Clustering results comparison on the four single-cell datasets.

https://doi.org/10.1371/journal.pcbi.1005621.t001

Vicus captures rare cell populations

One of the major challenges in single-cell analysis is to detect rare populations of cells from noisy single-cell RNA-seq data. The signals of rare populations can be easily neglected due to the existence of various sources of noises. Our approach based on Vicus matrix is able to discover weak signals of rare populations by exploiting local structures while global Laplacian fails. We applied our method on a scRNA-seq data consisting of 2700 peripheral blood mononuclear cells (PBMC). It is generated by 10x Genomics GemCode platform, a droplet-based high-throughput technique and 2700 cells with UMI counts were identified by their customized computational pipeline [22]. This cell population includes five major immune cell types in a healthy human as well as a rare population of metakaryocytes (less than 0.5% abundance in PBMC). The processed data is available in [11] and was originally published in [22]. Vicus captures the rare population consisting of 11 cells (Fig 3A) while global Laplacian fails to find such rare population. Vicus is also able to detect differential genes that define each cluster (Fig 3B). Note that we used Vicus score to rank important genes (Materials and methods) and we only show top 5 genes for each cluster.

thumbnail
Fig 3. Vicus detects rare population in PBMC data.

A: a 3-D mapping of the learned low dimension by Vicus. Each cell is colored according to its ground-truth. The rare population of Megakaryocytes is shown in yellow. B: The top 5 differential genes for each cell types detected by Vicus.

https://doi.org/10.1371/journal.pcbi.1005621.g003

Stability in clustering E. coli PPI network

Identification of functional modules in Protein-protein interaction (PPI) networks is an important challenge in bioinformatics. Network module detection algorithms can be employed to extract functionally homogenous proteins. In this application, first submodules are detected and subsequently these submodules are investigated for enrichment of proteins with a particular biological function. Stability is one of the essential goals of the multi-scale module detection problem [23]. It measures how robust the employed algorithm is able to recover the most dense subnetworks enriched to certain biological functions or physical interactions. Inside the definition of stability (Materials and methods), the Laplacian is used in a Markov process on the network which allows to compare and rank partitions at each iteration.

To analyze the stability of our method we partition a Protein-Protein Interaction(PPI) network, which consists of 7,613 interactions between 2,283 Escherichia coli proteins [24]. This task is more challenging than traditional clustering problems due to the intrinsic complexity of the cell captured by the PPI network. Due to large noise in experimental measurements of protein interactions, proteins in the same pathway do not necessarily have higher density of interactions. This fact poses particular challenges to traditional network partition algorithms which usually fail to infer the true membership of proteins to their underlying pathways. Vicus-based spectrum exhibits higher stability along the Markovian timeline (Fig 4A) compared to the global Laplacian. Global spectrum and Vicus-based local spectrum exploit different modes of variation in the network (Fig 4B). Global spectrum tends to find large components in networks to reduce the variation and increase stability while local spectrum exploits deeper substructures of the large components and detects partitions in more fine-grained fashion.

thumbnail
Fig 4. Stability and variations along different Markov time spans.

Stability in Panel A indicates the robustness of the community detection algorithms (Vicus vs Laplacian) while Variations in Panel B show how the corresponding algorithm exploits the community membership information in the network.

https://doi.org/10.1371/journal.pcbi.1005621.g004

Ranking genes associated with cancer subtypes

One of the holy grails of computational medicine is identification of robust biomarkers associated with the phenotype of interest. Here we consider the question of identifying genes associated with cancer subtyping in 5 cancers from 6 microarray datasets. These are benchmark datasets for feature selection in computational biology from http://featureselection.asu.edu/datasets.php. Table 2 shows the statistics of these six datasets.

thumbnail
Table 2. Statistics summary of the six macro-array datasets for cancer subtypes.

https://doi.org/10.1371/journal.pcbi.1005621.t002

In the standard formulation of spectral clustering, the ranking of features (in this case, genes) is done using Laplacian score. Laplacian Score is a score derived based on the network spectrum that is commonly used to rank features in the order of their importance and relevance to the clusters. Given a feature f, the corresponding Laplacian Score (LS) is defined as follows: (1)

Unfortunately, LS has difficulty identifying features that are only relevant to one of the clusters (a certain local subnetwork) but not the whole network. Traditional LS will prefer features that are globally relevant to all the clusters, even if they are not as strongly indicative of any cluster in particular. We thus, propose to substitute the Laplacian matrix with our Vicus matrix . We define our Vicus Score (VS) analogously to Laplacian Score: (2)

For each data set presented in Table 2, we rank the features by Laplacian Score and Vicus Score. We take N highest ranked features and then apply simple k-means clustering. If the feature ranking algorithm correctly ranks the relevant features, the clustering accuracy should be higher compared to the accuracy of the method that uses the same number of chosen but less relevant features. We varied the number of chosen features and plotted the accuracy of the ranking algorithms in Fig 5. Again, we use NMI and ARI as the evaluation metrics for the clustering results. We observe that features ranked using the Vicus matrix result in better accuracy when the number of chosen features is small, confirming that the most discriminative features are ranked among the top by Vicus.

thumbnail
Fig 5. The results of feature ranking by Laplacian and Vicus.

Experiments are performed on 6 cancer datasets. On each dataset, we vary the number of selected features (genes) and use k-means to report the clustering accuracy. NMI and ARI are used to measure the goodness of selected features. It is consistently observed across six datasets that Vicus can select better features than Laplacian.

https://doi.org/10.1371/journal.pcbi.1005621.g005

Discussion

The proposed Vicus matrix for weighted networks exhibits greater power to represent the underlying cluster structures of the networks than the traditional global Laplacian. The key observation is the ability of our local spectrum to make the top eigenvectors more robust to noise and hyper parameters in the process of constructing such weighted networks. The proposed Vicus-based local spectrum can supplant the usage of Laplacian-based spectral methods for weighted networks in various tasks such as clustering, community detection, feature ranking and dimensionality reduction. Sharing similar algebraic properties with global Laplacian, our local spectrum helps to understand the underlying structures of the noisy weighted networks. As demonstrated, local spectrum is robust with respect to noise and outliers. Finally, we have parallelized Vicus to achieve scalability. While the discussed applications contained at most a few thousands nodes, we have performed experiments on networks with up to 500,000 nodes. On this very large network, Laplacian based spectral clustering took 7.5min while Vicus took 12.9min with better performance (higher NMI). Thus, Vicus is not only more accurate but it can scale to very large networks, a property which will become important as we start constructing, for example, DNA co-methylation probe-based networks with hundreds of thousands of probes.

Conclusion

The power of local network neighborhoods has become abundantly clear in many fields where the networks are used. Principled methods are needed to take advantage of the local network structure. In this work we have proposed the Vicus matrix, a new formulation that shares algebraic properties with the traditional Laplacian and yet improves the power of spectral methods across a wide range of tasks necessary to gain deeper understanding into biological data and behavior of the cell. Taking advantage of the local network structure, we showed improved performance in single cell RNA-seq clustering, feature ranking for identifying biomarkers associated with cancer subtyping and dimensionality reduction in single cell RNA-seq data. Further, we have shown that our method is amenable to parallelization which allows it to be performed in time comparable to the traditional methods.

Materials and methods

Laplacian matrix

Suppose we have a network with a set of V nodes and weighted edges. Let be the weighted ajacency matrix of this network, where |V| is the number of nodes. Here, Wij represents the weight of the edge between the ith and jth nodes. Let diagonal matrix D be W’s degree matrix, where . The classical formulation of the Laplacian of W is then matrix also known as the combinatorial Laplacian. A common variant of the Laplacian is which is called the normalized Laplacian.

Traditional state-of-the-art spectral clustering [25] aims to minimize RatioCut, an objective function that effectively combines MinCut and equipartitioning, by solving the following optimization problem: (3) where C is the number of clusters, n is the number of nodes and Q = [q1, q2, …, qC] is the set of eigenvectors, capturing the structure of the graph. Eigenvectors associated with the Laplacian matrix of the weighted network are used in many tasks (e.g., face clustering, dimensionality reduction, image retrieval, feature ranking, etc). These eigenvectors suffer from some limitations. For example, the top eigenvectors, in spite of their ability to map the data to a low-dimensional space, are sensitive to noisy measurements and outliers encoded by pairwise similarities (S1 Fig) [4]. Additionally, the Laplacian is very sensitive to the hyper-parameters used to construct the similarity matrices (Materials and methods, S2 Fig) [25].

Local spectrum matrix: Vicus

Our Vicus Matrix () is similar to the Laplacian () in functionality and in addition captures the local structure inherent in the data. The intuition behind Vicus is that we use local information from neighboring nodes, akin to label propagation [26] or random walks [27]. As we demonstrate, relying on local subnetworks makes the matrix more robust to noise, helping to alleviate the influence of outliers.

Let our data be a set of points {x1, x2, …, xn}. Then, each vertex vi, in the weighted network , represents a point xi and represents xi’s neighbours, not including xi. We constrain the neighbourhood size to be held constant across nodes (i.e., ).

Our main assumption is that the labels (such as cluster assignments 1 … C for C clusters) of neighbouring points in the network are similar. Specifically, we assume that the cluster indicator value of the ith datapoint (xi) can be inferred from the labels of its direct neighbors (). First, we extract a subnetwork such that and which represents the edges connecting all the nodes in Vi. The similarity matrix associated with the subgraph is , representing the weights for all the edges associated with all the nodes in Vi. Using the label diffusion algorithm [28], we can reconstruct a virtual label indicator vector such that (4) where α is a constant (0 < α < 1, empirically set to 0.9 in all our experiments, as suggested in [28]) and is the scaled cluster indicator vector of the subnetwork . Si represents the normalized transition matrix of Wi, i.e., . Note that we do not actually perform any diffusion, since our setting is completely unsupervised. Instead we use pk to estimate qik. is a vector of K + 1 elements, where is the estimate of how likely datapoint i belongs to cluster k based on its neighbours. As we want maximal concordance between and , we set , where is the row of the matrix (1 − α)(IαSi)−1, representing label propagation at its final state. Here, βi represents the convergence of the label propagation for the datapoint i (Note that the original matrix was constructed as the concatenation of the neighborhood of i and datapoint i as the last row). Hence (5) where βi[1: K] represents the first K elements of βi and βi[K + 1] is the K + 1st element in βi, corresponding to the ith datapoint.

We can construct a matrix B, that represents a linear relationship , (k = 1, …, C), such that (6)

Our objective is to minimize the difference between and qk: (7) Setting , we arrive at our novel local version of spectral clustering: (8)

Similarly to the original spectral clustering formulation (Eq 3), our clustering results can be obtained by performing eigen-decomposition of matrix [25] to solve Eq 8. The final grouping of datapoints into clusters is achieved by performing k-means clustering on Q as in [29].

Analysis of Vicus properties.

Our Vicus Matrix shares many properties with Laplacian [25]:

  1. 1. Both matrices are symmetric and positive semi-definite.
  2. 2. The smallest eigenvalue of both matrices is 0, the corresponding eigenvector is the constant vector l.
  3. 3. Both matrices have n non-negative, real-valued eigenvalues 0 = λ1λ2 ≤ … ≤ λn
  4. 4. The multiplicity of the eigenvalue 0 of both and equals the number of connected components in the network.

Here n is the number of nodes in the network. To prove the first property, we note that , thus is symmetric. Also, for any non-zero vector x, we have .

For the second property, we first prove that B l = l, i.e., matrix B has an eigenvalue of 1 corresponding to an all-one constant vector. To prove the above, the following statement must be true: According to Eq 6, . Note that βi is the last row of the transition kernel , hence we have , considering the sum of each row of Li is all one. Thus we prove and therefore Bl = l.

It is then easy to verify that

Hence we proved that matrix always has an eigenvalue of 0 corresponding to the eigenvector l. Property 3 follows directly from properties 1 and 2. Finally, the last property can be easily proven using an arguments similar to [25]. The above properties verify that our proposed Vicus matrix is a proper alternative to Laplacian including the desirable algebraic properties.

Similarity network constructions

Given a feature set that describes a collection of objects, denoted as X = {x1, x2, …, xn}, we want to construct a similarity network in which indicates the similarity between the i-th and j-th object. The most widely used method is to assume a Gaussian distributions across pairwise similarites: Here σ is a hyper-parameter that needs careful manual setting. More advanced methods of constructing similarity networks can be seen in [4].

Normalized mutual information

Throughout the paper, we used Normalized Mutual Information (NMI) [20] to evaluate the consistency between the obtained clustering and the groundthuth. Given two clustering results U and V on a set of data points, NMI is defined as: I(U, V)/ max{H(U), H(V)}, where I(U, V) is the mutual information between U and V, and H(U) represents the entropy of the clustering U. Specifically, assuming that U has P clusters, and V has Q clusters, the mutual information is computed as follows: where |Up| and |Vq| denote the cardinality of the p-th cluster in U and the q-th cluster in V respectively. The entropy of each cluster assignment is calculated by and

Details can be found in [20]. NMI is a value between 0 and 1, measuring the concordance of two clustering results. In the simulation, we calculate the obtained clustering with respect to the ground-truth. Therefore, a higher NMI refers to higher concordance with truth, i.e. a more accurate result.

Adjusted Rand Index

The Adjusted Rand Index (ARI) is another widely-used metric for measuring the concordance between two clustering results. Given two clustering U and V, we calculate the following four quantities:

  1. a: number of objects in a pair are placed in the same group in U and in the same group in V;
  2. b: number of objects in a pair are placed in the same group in U and in different groups in V;
  3. c: number of objects in a pair are placed in the same group in V and in different groups in U;
  4. d: number of objects in a pair are placed in different groups in U and in different groups in V.

The (normal) Rand Index (RI) is simply . It basically weights those objects that were classified together and apart in both U and V. There are some known problems with this simple version of RI such as the fact that the Rand statistic approaches its upper limit of unity as the number of clusters increases. With the intention to overcome these limitations, ARI has been proposed in [21] in the form of

Stability and variations in Markov clustering

Given a network on a set of N nodes with edge weights W, we first present a few related terms as follows

  1. L: the Laplacian matrix for the network. It can either be traditional Laplacian or our newly proposed Vicus;
  2. π: the stationary distribution vector with πL = 0
  3. di: the degree of node i as di = ∑j Wij
  4. Σ: the normalized degree matrix with nonzero values only on the diagonal .
  5. H: the partition indicator matrix with Hij = 1 if the node i is classified with cluster j and Hij = 0 otherwise.

Then the stability measure on time t is defined in terms of the clustered auto-covariance matrix as follows: and the stability curve of the network is obtained by maximizing this measure over all possible partitions: A good clustering over time t will have large stability, with a large trace of Rt over such a time span.

The variation is defined in terms of the asymptotic stability induced by going from the ‘finest’ to the ‘next finest’ partitions is: where u2 is the normalized Fiedler eigenvector with its corresponding eigenvalue λ2. We refer the mathematical details in deriving these two definitions to [23].

Hyper-parameters settings for Vicus

There are mainly three hyper-parameters in Vicus: first the number of neighbors K, the variance in network construction σ, and the diffusion parameter α. Details about the meaning of these hyper-parameters can be seen in [30]. In all our experiments, we use the same setting of hyper-parameters as follows:

The proposed Vicus is very robust to the choice of σ and α (S3 Fig). For the choice of K, we usually increase K as the number of nodes in the networks get larger (S3 Fig). We also provide a range of recommended choices for these hyper-parameters:

We also want to emphasize that, when performing clustering tasks, Vicus does not specify the number of clusters since Vicus is only providing a new form of Laplacian that captures local structures in the network. In our experiments of single-cell applications, we only feed the number of clusters to the clustering algorithms (i.e, K-means algorithm) as the true number of clusters.

Supporting information

S1 Fig. An illustrative example showing Vicus is more robust to noise and outliers compared to Laplacian.

Panel A shows the underlying ground-truth network heatmap consisting of 3 connected components. Given this perfect network, we manually add random noise. The random noise is generated from uniform distribution between [0, δ]. Larger δ indicates bigger magnitude of the noise therefore stronger corruption on the network. Panel B shows an example of the noisy network after corruption when δ = 1.2. Panel C is the clustering accuracy measured by NMI if we vary the number of noise strength δ.

https://doi.org/10.1371/journal.pcbi.1005621.s001

(EPS)

S2 Fig. An illustrative example of comparison between Laplacian and Vicus to illustrate their sensitivity to hyper-parameters used in the construction of similarity network.

The first column shows the groundtruth of the data distribution. Panel A is the 3D scattering of the data points used in the experiment. Panel E shows the corresponding 2D ground-truth distribution generating the data. This is also a desired output of low-dimensional embedding we want to recover. Panels B-D shows the results of low-dimensional embedding by Laplacian while Panels F-H are for Vicus using different values of hyper-parameters.

https://doi.org/10.1371/journal.pcbi.1005621.s002

(EPS)

S3 Fig. Sensitivity test for three hyper-parameters in Vicus.

We apply Vicus on the Buettner data set of single-cell RNA-seq. Panel A shows both NMI and ARI with different choices of number of neighbors K with fixed σ = 0.5 and α = 0.9. Panel B shows both NMI and ARI with different choices of σ with fixed K = 10 and α = 0.9. Panel C shows both NMI and ARI with different choices of α with fixed σ = 0.5 and K = 10.

https://doi.org/10.1371/journal.pcbi.1005621.s003

(TIFF)

Author Contributions

  1. Conceptualization: BW AG.
  2. Data curation: BW.
  3. Formal analysis: BW AG.
  4. Funding acquisition: AG.
  5. Investigation: BW AG SB AK.
  6. Methodology: BW LH YZ AG.
  7. Project administration: AG SB AK.
  8. Resources: BW.
  9. Software: BW LH YZ.
  10. Supervision: AG.
  11. Validation: BW LH YZ.
  12. Visualization: BW.
  13. Writing – original draft: BW LH AG.
  14. Writing – review & editing: BW LH YZ SB AK AG.

References

  1. 1. De Las Rivas J, Fontanillo C. Protein–protein interactions essentials: key concepts to building and analyzing interactome networks PLoS Comput Biol, 6(6):e1000807, 2010. pmid:20589078
  2. 2. Stam CJ. Modern network science of neurological disorders. Nature Reviews Neuroscience, 15(10):683–695, 2014. pmid:25186238
  3. 3. Barabási AL, Gulbahce N, Loscalzo J. Network medicine: a network-based approach to human disease. Nature Reviews Genetics, 12(1):56–68, 2011. pmid:21164525
  4. 4. Wang B, Mezlini AM, Demir F, Fiume M, Tu Z, Brudno M, et al. Similarity network fusion for aggregating data types on a genomic scale. Nature methods, 11(3):333–337, 2014. pmid:24464287
  5. 5. Halary S, Leigh JW, Cheaib B, Lopez P, Bapteste E. Network analyses structure genetic diversity in independent genetic worlds. Proceedings of the National Academy of Sciences, 107(1):127–132, 2010.
  6. 6. He X, Min W, Cai D, Zhou K. Laplacian optimal design for image retrieval. Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval, 119–126, 2007.
  7. 7. Wang B, Zhu J, Pourshafeie A, Ursu O, Batzoglou S, Kundaje A. Unsupervised Learning from Noisy Networks with Applications to Hi-C Data. Advances in Neural Information Processing Systems, pp. 3305-3313, 2016.
  8. 8. Belkin M, Niyogi P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural computation, 15(6):1373–1396, 2003.
  9. 9. Roweis ST, Saul LK. Nonlinear dimensionality reduction by locally linear embedding. Science, 290(5500):2323–2326, 2000. pmid:11125150
  10. 10. Wu M, Schölkopf B. A local learning approach for clustering. In Proceedings of Neural Information Processing Systems, pp. 1529–1536, 2006.
  11. 11. Wang B, Zhu J, Pierson E, Ramazzotti D, Batzoglou S. Visualization and analysis of single-cell RNA-seq data by kernel-based similarity learning. Nature Methods, 14(4):414–416, 2017. pmid:28263960
  12. 12. Pollen AA, Nowakowski TJ, Shuga J, Wang X, Leyrat AA, Lui JH, et al. Low-coverage single-cell mRNA sequencing reveals cellular heterogeneity and activated signaling pathways in developing cerebral cortex. Nature biotechnology, 32(10):1053–1058, 2014. pmid:25086649
  13. 13. Usoskin D, Furlan A, Islam S, Abdo H, Lönnerberg P, Lou D, et al. Unbiased classification of sensory neuron types by large-scale single-cell RNA sequencing. Nature neuroscience, 18(1):145–153, 2015. pmid:25420068
  14. 14. Buettner F, Natarajan KN, Casale FP, Proserpio V, Scialdone A, Theis FJ, et al. Computational analysis of cell-to-cell heterogeneity in single-cell RNA-sequencing data reveals hidden subpopulations of cells. Nature biotechnology, 33(2):155–160, 2015. pmid:25599176
  15. 15. Kolodziejczyk AA, Kim JK, Tsang JC, Ilicic T, Henriksson J, Natarajan KN, et al. Single cell RNA-sequencing of pluripotent states unlocks modular transcriptional variation. Cell stem cell, 17(4):471–485, 2015 pmid:26431182
  16. 16. Van Der Maaten L, Postma E, Van den Herik J. Dimensionality reduction: a comparative. J Mach Learn Res, 10:66–71, 2009.
  17. 17. Rosvall M, Bergstrom CT. Maps of random walks on complex networks reveal community structure. Proceedings of the National Academy of Sciences, 105(4):1118–1123, 2008.
  18. 18. Newman ME. Modularity and community structure in networks. Proceedings of the National Academy of Sciences, 103(23):8577–8582, 2006.
  19. 19. Frey BJ, Dueck D. Clustering by passing messages between data points. Science, 315:972–976, 2007. pmid:17218491
  20. 20. Vinh NX, Epps J, Bailey J. Information Theoretic Measures for Clusterings Comparison: Variants, Properties, Normalization and Correction for Chance. J Mach Learn Res, 11(18):2837–2854, 2010.
  21. 21. Rand WM Objective criteria for the evaluation of clustering methods Journal of the American Statistical association, 66(336), 846–850, 1971.
  22. 22. Zheng GX, Terry JM, Belgrader P, Ryvkin P, Bent ZW, Wilson R, et al. Massively parallel digital transcriptional profiling of single cells. Nature Communications, 8:14049, 2017. pmid:28091601
  23. 23. Delvenne JC, Yaliraki SN, Barahona M. Stability of graph communities across time scales. Proceedings of the National Academy of Sciences, 107(29):12755–12760, 2010.
  24. 24. Peregrín-Alvarez JM, Xiong X, Su C, Parkinson J. The modular organization of protein interactions in escherichia coli. PLoS computational biology, 5(10):e1000523, 2009. pmid:19798435
  25. 25. Luxburg UV. A tutorial on spectral clustering. Statistics and computing, 17(4):395–416, 2007.
  26. 26. Gregory S. Finding overlapping communities in networks by label propagation. New Journal of Physics, 12(10):103018, 2010.
  27. 27. Fallani FD, Nicosia V, Latora V, Chavez M. Nonparametric resampling of random walks for spectral network clustering. Physical Review E, 89(1):012802, 2014.
  28. 28. Zhou DY, Bousquet O, Lal TN, Weston J, Schölkopf B. Learning with local and global consistency. In Proceedings of Neural Information Processing Systems, volume 16, pages 321–328, 2003.
  29. 29. Cour T, Benezit F, Shi J. Spectral segmentation with multiscale graph decomposition. In Conference on Computer Vision and Pattern Recognition, 2:1124–1131, 2005.
  30. 30. Wang B, Jiang J, Wang W, Zhou ZH, Tu Z. Unsupervised metric fusion by cross diffusion. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2997–3004, 2012.