 Research
 Open Access
 Published:
Breaking the hierarchy  a new cluster selection mechanism for hierarchical clustering methods
Algorithms for Molecular Biology volume 4, Article number: 12 (2009)
Abstract
Background
Hierarchical clustering methods like Ward's method have been used since decades to understand biological and chemical data sets. In order to get a partition of the data set, it is necessary to choose an optimal level of the hierarchy by a socalled level selection algorithm. In 2005, a new kind of hierarchical clustering method was introduced by Palla et al. that differs in two ways from Ward's method: it can be used on data on which no full similarity matrix is defined and it can produce overlapping clusters, i.e., allow for multiple membership of items in clusters. These features are optimal for biological and chemical data sets but until now no level selection algorithm has been published for this method.
Results
In this article we provide a general selection scheme, the level independent clustering selection method, called LInCS. With it, clusters can be selected from any level in quadratic time with respect to the number of clusters. Since hierarchically clustered data is not necessarily associated with a similarity measure, the selection is based on a graph theoretic notion of cohesive clusters. We present results of our method on two data sets, a set of drug like molecules and set of proteinprotein interaction (PPI) data. In both cases the method provides a clustering with very good sensitivity and specificity values according to a given reference clustering. Moreover, we can show for the PPI data set that our graph theoretic cohesiveness measure indeed chooses biologically homogeneous clusters and disregards inhomogeneous ones in most cases. We finally discuss how the method can be generalized to other hierarchical clustering methods to allow for a level independent cluster selection.
Conclusion
Using our new cluster selection method together with the method by Palla et al. provides a new interesting clustering mechanism that allows to compute overlapping clusters, which is especially valuable for biological and chemical data sets.
Background
Clustering techniques have been used for decades to find entities that share common properties. Regarding the huge data sets available today, which contain thousands of chemical and biochemical molecules, clustering methods can help to categorize and classify these tremendous amounts of data [1–3]. In the special case of drug design their importance is reflected in their widerange application from drug discovery to lead molecule optimization [4]. Since structural information of molecules is easier to obtain than their biological activity, the main idea behind using clustering algorithms is to find groups of structurally similar molecules in the hope that they also exhibit the same biological activity. Therefore, clustering of druglike molecules is a great help to reduce the search space of unknown biologically active compounds.
Several methods that intend to locate clusters have been developed so far. The methods that are used most in chemistry and biochemistry related research are Ward's hierarchical clustering [5], single linkage, complete linkage and group average methods [6]. All of them build hierarchies of clusters, i.e., on the first level of the hierarchy all molecules are seen as similar to each other, but further down the hierarchy, the clusters get more and more specific. To find one single partition of the data set into clusters, it is necessary to determine a level that then determines the number and size of the resultant clusters, e.g., by using the Kelleyindex [7]. Note that a too high level will most often lead to a small number of large, unspecific clusters, and that a too low level will on the other hand lead to more specific but maybe very small and too many clusters. A cluster that contains pairwise very similar entities can be said to be cohesive. Thus, a level selection algorithm tries to find a level with not too many clusters that are already sufficiently specific or cohesive.
Other commonly used clustering methods in chemistry and biology are not based on hierarchies, like the Kmeans [8] and the JarvisPatrick method [9]. Note however, that all of the methods mentioned so far rely on a total similarity matrix, i.e., on total information about the data set which might not always be obtainable.
A group of clustering techniques which is not yet so much applied in the field of bio and chemoinformatics is based on graph theory. Here, molecules are represented by nodes and any kind of similarity relation is represented by an edge between two nodes. The big advantage of graph based clustering lies in those cases where no quantifiable similarity relation is given between the elements of the data set but only a binary relation. This is the case, e.g., for proteinproteininteraction data where the interaction itself is easy to detect but its strength is difficult to quantify; another example are metabolic networks that display whether or not a substrate is transformed into another one by means of an enzyme. The most wellknown examples of graph based clustering methods were proposed by Girvan and Newman [10] and Palla et al [11].
The latter method, the socalled kclique community clustering (CCC), which was also independently described in [12, 13], is especially interesting since it cannot only work with incomplete data on biological networks but is also able to produce overlapping clusters. This means that any of the entities in the network can be a member of more than one cluster in the end. This is often a natural assumption in biological and chemical data sets:

1.
proteins often have many domains, i.e., many different functions. If a set of proteins is clustered by their function, it is natural to require that some of them should be members of more than one group;

2.
similarly, drugs may have more than one target in the body. Clustering in this dimension should thus also allow for multiple membership;

3.
molecules can carry more than one active group, i.e., pharmacophore, or one characteristic structural feature like heteroaromatic ring systems. Clustering them by their functional substructures should again allow for overlapping clusters.
This newly proposed method by Palla et al. has already been proven useful in the clustering of Saccharomyces cerevisiae[11, 14] and human proteinproteininteractions networks [15]. To get a valid clustering of the nodes, it is again necessary to select some level k, as for other hierarchical clustering methods. For the CCC the problem of selecting the best level is even worse than in the classic hierarchical clustering methods cited above: while Ward's and other hierarchical clustering methods will only join two clusters per level and thus monotonically decrease the number of clusters from level to level, the number of clusters in the CCC may vary wildly over the levels without any monotonicity as we will show in 'Palla et al.'s clustering method'.
This work proposes a new way to cut a hierarchy to find the best suitable cluster for each element of the data set. Moreover, our method, the levelindependent cluster selection or LInCS for short does not choose a certain level which is optimal but picks the best clusters from all levels, thus allowing for more choices. To introduce LInCS and prove its performance, section 'Methods: the LInCS algorithm' provides the necessary definitions and a description of the new algorithmic approach. Section 'Data sets and experimental results' describes the data and section 'Results and discussion' the experimental results that reveal the potential of the new method. Finally, we generalize the approach in section 'Generalization of the approach' and conclude with a summary and some future research problems in section 'Conclusions'.
Methods: the LInCS algorithm
In this section we first present a set of necessary definitions from graph theory in 'Graph theoretical definitions' and give a general definition of hierarchical clustering with special emphasis on the CCC method by Palla et al. in 'Hierarchical clustering and the level selection problem'. Then we introduce the new hierarchy cutting algorithm called LInCS in 'Finding cohesive kclique communities: LInCS'.
Graph theoretical definitions
Before we start with sketching the underlying CCC algorithm by Palla et al. and our improvement, the LInCS method, we describe the necessary graphbased definitions.
An undirected graph G = (V, E) consists of a set V of nodes, and a set of edges E ⊆ V × V that describes a relation between the nodes. If {v_{ i }, v_{ j }} ∈ E then v_{ i }and v_{ j }are said to be connected with each other. Note that (v_{ i }, v_{ j }) will be used to denote an undirected edge between v and w. The degree deg(v) of a node v is given by the number of edges it is contained in. A path P(v, w) is an ordered set of nodes v = v_{0}, v_{1},..., v_{ k }= w such that for any two subsequent nodes in that order (v_{ i }, v_{i+1}) is an edge in E. The length of a path in an unweighted graph is given by the number of edges in it. The distance d(v, w) between two nodes v, w is defined as the minimal length of any path between them. If there is no such path, it is defined to be ∞. A graph is said to be connected if all pairs of nodes have a finite distance to each other, i.e., if there exists a path between any two nodes.
A graph G' = (V', E') is a subgraph of G = (V, E) if V' ⊆ V, E' ⊆ E and E' ⊆ V' × V'. In this case we write G' ≤ G. If moreover V' ≠ V then G' is a proper subgraph, denoted by G' < G. Any subgraph of G that is connected and is not a proper subgraph of a larger, connected subgraph, is called a connected component of G.
A kclique is any (sub)graph consisting of k nodes where each node is connected to every other node. A kclique is denoted by K_{ k }. If a subgraph G' constitutes a kclique and G' is no proper subgraph of a larger clique, it is called a maximal clique. Fig. 1 shows examples of a K_{3}, a K_{4}, and a K_{5}.
We need the following two definitions given by Palla et al. [11]. See Fig. 2 for examples:
Definition 1 A kclique A is kadjacent with kclique B if they have at least k  1 nodes in common.
Definition 2 Two kcliques C_{1}and C_{ s }are kcliqueconnected to each other if there is a sequence of kcliques C_{1}, C_{2},..., C_{s1}, C_{ s }such that C_{ i }and C_{i+1}are kadjacent for each i = 1,..., s  1.
This relation is reflexive, i.e., clique A is always kcliqueconnected to itself by definition. It is also symmetric, i.e., if clique B is kcliqueconnected to clique A then A is also kcliqueconnected to B. In addition, the relation is transitive since if clique A is kcliqueconnected to clique B and clique B is kcliqueconnected to C then A is kcliqueconnected to C. Because the relation is reflexive, symmetric and transitive it belongs to the class of equivalence relations. Thus this relation defines equivalence classes on the set of kcliques, i.e., there are unique maximal subsets of kcliques that are all kcliqueconnected to each other. A kclique community is defined as the set of all kcliques in an equivalence class [11]. Fig. 2(a), (b) and 2(c) give examples of kclique communities. A knode cluster is defined as the union of all nodes in the cliques of a kclique community. Note that a node can be member of more than one kclique and thus it can be a member of more than knode cluster, as shown in Fig. 3. This explains how the method produces overlapping clusters.
We will make use of the following observations that were already established by Palla et al. [11]:
Observation 1 Let A and B be two cliques of at least size k that share at least k  1 nodes. It is clear that A containscliques of size k and B containscliques of size k. Note that all of these cliques in A and B are kcliqueconnected. Thus, we can generalize the notion of kadjacency and kcliqueconnectedness to cliques of size at least k and not only to those of strictly size k.
We want to illustrate this observation by an example. Let C_{1} be a clique of size 6 and C_{2} a clique of size 8. C_{1} and C_{2} share 4 nodes, denoted by v_{1}, v_{2}, v_{3}, v_{4}. Note that within C_{1} all possible subsets of 5 nodes build a 5clique. It is easy to see that all of them are 5cliqueconnected by definition 1 and 2. The same is true for all possible 5cliques in C_{2}. Furthermore, there is at least one 5clique in C_{1} and one in C_{2} that share the nodes v_{1}, v_{2}, v_{3}, v_{4}. Thus, by the transitivity of the relation as given in definition 2, all 5cliques in C_{1} are kcliqueconnected to all 5cliques in C_{2}.
Observation 2 Let C ⊂ C' be a kclique that is a subset of another clique then C is obviously kcliqueconnected to C'. Let C' be kcliqueconnected to some clique then due to the transitivity of the relation, C is also kcliqueconnected to B. Thus, it suffices to restrict the set of cliques of at least size k to all maximal cliques of at least size k.
As an illustrative example, let C_{1} denote a 4clique within a 6clique C_{2}. C_{1} is 4cliqueconnected to C_{2} because they share any possible subset of 3 nodes out of C_{1}. If now C_{2} shares another 3 nodes with a different clique C_{3}, by the transitivity of the kcliqueconnectedness relation, C_{1} and C_{3} are also 3cliqueconnected. With these graph theoretic notions we will now describe the idea of hierarchical clustering.
Hierarchical clustering and the level selection problem
A hierarchical clustering method is a special case of a clustering method. A general clustering method produces nonoverlapping clusters that build a partition of the given set of entities, i.e., a set of subsets such that each entity is contained in exactly one subset. An ideal clustering partitions the set of entities into a small number of subsets such that each subset contains only very similar entities. Measuring the quality of a clustering is done by a large set of clustering measures, for an overview see, e.g., [16]. If a good clustering can be found, each of the subsets can be meaningfully represented by some member of the set leading to a considerable data reduction or new insights into the structure of the data. With this sketch of general clustering methods, we will now introduce the notion of a hierarchical clustering.
Hierarchical clusterings
The elements of a partition P = {S_{1}, S_{2},..., S_{ k }} are called clusters (s. Fig. 4(a)). A hierarchical clustering method produces a set of partitions on different levels 1,..., k with the following properties: Let the partition of level 1 be just the given set of entities. A refinement of a partition P = {S_{1}, S_{2},..., S_{ j }} is a partition such that each element of P' is contained in exactly one of the elements of P. This containment relation can be depicted as a tree or dendogramm (s. Fig. 4(b)).
The most common hierarchical clustering methods start at the bottom of the hierarchy with each entity in its own cluster, building the socalled singletons. These methods require the provision of a pairwise distance measure, often called similarity measure, of all entities. From this a distance between any two clusters is computed, e.g., the minimum or maximum distance between any two members of the clusters, resulting in singlelinkage and completelinkage clustering [6]. In every step, the two clusters S_{ i }, S_{ j }with minimal distance are merged into a new cluster. Thus, the partition of the next higher level consists of nearly the same clusters minus S_{ i }, S_{ j }and plus the newly merged cluster S_{ i }∪ S_{ j }.
Since a hierarchical clustering computes a set of partitions but a clustering consists of only one partition, it is necessary to determine a level that defines the final partition. This is sometimes called the klevel selection problem. Of course, the optimization goals for the optimal clustering are somewhat contradicting: on the one hand, a small number of clusters is wanted. This favors a clustering with only a few large clusters within which not all entities might be very similar to each other. But if, on the other hand, only subsets of entities with high pairwise similarity are allowed, this might result in too many different maximal clusters which does not allow for a high data reduction. Several level selection methods have been proposed to solve this problem so far; the best method for most purposes seems to be the Kelleyindex [7], as evaluated by [3]. To find clusters with high inward similarity Kelley et al. measure the average pairwise distance of all entities in one set. Then they create a penalty score out of this value and the number of clusters on every level. They suggest to select the level at which this penalty score is lowest.
We will now shortly sketch Palla et al.'s clustering method, show why it can be considered a hierarchic clustering method although it produces overlapping clusters and work out why Kelley's index cannot be used here to decide the level selection problem.
Palla et al.'s clustering method
Recently, Palla et al. proposed a graph based clustering method that is capable of computing overlapping clusters [11, 17, 18]. This method has already been proven to be useful, especially in biological networks like proteinproteininteraction networks [14, 15]. It needs an input parameter k between 1 and the number of nodes n with which the algorithm computes the clustering as follows: for any k between 1 and n compute all maximal cliques of size at least k. From this a metagraph can be built: Represent the maximal cliques as nodes and connect any two of them if they share at least k 1 nodes (s. Fig. 5). These cliques are obviously kcliqueconnected by observations 1 and 2. Any path in the metagraph connects by definition cliques that are kcliqueconnected. Thus, a simple connected component analysis in the metagraph is enough to find all kclique communities. From this, the clusters on the level of the original entities can be easily constructed by merging the entities of all cliques within a kclique community. Note that on the level of the maximal cliques the algorithm constructs a partition, i.e., each maximal clique can only be in one kclique community. Since a node can be in different maximal cliques (as illustrated in Fig. 5 for nodes 4 and 5) it can end up in as many different clusters on the knode cluster level.
Note that for k = 2 the 2clique communities are just the connected components of the graph without isolated nodes. Note also that the kclique communities for some level k do not necessarily cover all nodes but only those that take part in at least one kclique. To guarantee that all nodes are in at least one cluster, those that are not contained in at least one knode cluster are added as singletons.
We will now show that the kclique communities on different klevels can be considered to build a hierarchy with respect to the containment relation. We will first show a more general theorem and then relate it to the buildup of a hierarchy.
Theorem 3 If k > k' ≥ 3 and two nodes v, u are in the same knode cluster, then there is a k'node cluster containing both u and v.
This theorem states that if two nodes u, v are contained in cliques that belong to some kclique community, then, for every smaller k' until 3, there will also be a k'clique community that contains cliques containing u and v. As an example: if C_{1} and C_{2} are 6cliqueconnected, then they are also 5, 4, and 3cliqueconnected.
Proof: By definition 2 u and v are in the same kclique community if there is a sequence of kcliques C_{1}, C_{2},..., C_{s1}, C_{ s }such that C_{ i }and C_{i+1}are kadjacent for each i = 1,..., s 1, and such that u ∈ C_{1}, v ∈ C_{ s }. In other words, there is a sequence of nodes u = v_{1}, v_{2},..., v_{s+k1}= v, such that v_{ i }, v_{i+1},..., v_{i+k1}is a kclique for each 1 ≤ i ≤ s.
It is easy to see that in this case the subset of nodes v_{ i }, v_{i+1},..., v_{i+k'1}constitutes a k'clique for each 1 ≤ i ≤ s + k  k'. Thus by definition there is a k'clique community that contains both u and v. ■
The proof is illustrated in Fig. 6. Moreover the theorem shows that if two cliques are kclique connected, they are also k'clique connected for each k > k' ≥ 3. This general theorem is of course also true for the special case of k' = k  1, i.e., if two cliques are in a kclique community, they are also in at least one k  1clique community. We will now show that they are only contained in at most one k  1clique community:
Theorem 4 Let the different kclique communities be represented by nodes and connect node A and node B by a directed edge from A to B if the corresponding kclique community C_{ A }of A is on level k and B's corresponding community C_{ B }is on level k  1 and C_{ A }is a subset of or equal to C_{ B }. The resulting graph will consist of one or more trees, i.e., the kclique communities are hierarchic with respect to the containment relation.
Proof: By Theorem 3 each kclique community with k > 3 is contained in at least one k 1clique community. Due to the transitivity of the kconnectedness relation, there can be only one k  1clique community that contains any given kclique community. Thus, every kclique community is contained in exactly one k  1clique community.
There are two important observations to make: ■
Observation 3 Given the set of all knode clusters (instead of the kclique communities) for all k, these could also be connected by the containment relationship. Note however that this will not necessarily lead to a hierarchy, i.e., one knode cluster can be contained in more than one k  1node cluster (s. Fig. 7).
Observation 4 Note also that the number of knode clusters might neither be monotonically increasing nor decreasing with k (s. Fig. 7).
It is thus established that on the level of kclique communities, the CCC builds a hierarchical clustering. Of course, since maximal cliques have to be found in order to build the kclique communities, this method can be computationally problematic [19], although in practice it performs very well. In general, CCC is advantageous in the following cases:

1.
if the given data set does not allow for a meaningful, realvalued similarity or dissimilarity relationship, defined for all pairs of entities;

2.
if it is more natural to assume that clusters of entities might overlap.
It is clear that this clustering method bears the same klevel selection problem as other hierarchical clustering methods. Moreover, the number and size of clusters can change strongly from level to level. Obviously, since quantifiable similarity measures might not be given, Kelley's index cannot be used easily. Moreover, it might be more beneficial to select not a whole level, but rather to find for each maximal clique the one kclique community that is at the same time cohesive and maximal. The next section introduces a new approach to finding such a kclique community for each maximal clique, the level independent cluster selection mechanism (LInCS).
Finding cohesive kclique communities: LInCS
Typically, at lower values of k, e.g., k = 3, 4, large clusters are discovered, which tend to contain the majority of entities. This suggests a low level of similarity between some of them. Conversely, small clusters at larger kvalues are more likely to show higher level of similarity between all pairs of entities. A cluster in which all pairs of entities are similar to one another will be called a cohesive cluster. Note that a high value of k might also leave many entities as singletons since they do not take part in any clique of size k.
Since the CCC is often used on data sets where no meaningful pairwise distance function can be given, the question remains of cohesion within a cluster can be meaningfully defined. It does not seem to be possible on the level of the knode clusters. Instead, we use the level of the kclique communities and define a given kclique community to be cohesive if all of its constituting kcliques share at least one node (s. Fig. 8):
Definition 5 A kclique community satisfies the strict clique overlap criterion if any two kcliques in the kclique community overlap (i.e., they have a common node). The kclique community itself is said to be cohesive.
A kclique community is defined to be maximally cohesive if the following definition applies:
Definition 6 A kclique community is maximally cohesive if it is cohesive and there is no other cohesive kclique community of which it is a proper subset.
The CCC was implemented by Palla et al., resulting in a software called the CFinder[20]. The output of CFinder contains the set of all maximal cliques, the overlapmatrix of cliques, i.e., the number of shared nodes for all pairs of maximal cliques, and the kcliquecommunities. Given this output of CFinder, we will now show how to compute all maximally cohesive kclique communities.
Theorem 7 A kclique community is cohesive if and only if it fulfills one of the following properties:
1. A kclique community is cohesive if and only if either it contains only one clique and this contains less than 2k nodes, or
2. if the union of any two cliques K_{ x }and K_{ y }in the community has less than 2k nodes. Note that this implies that the number of shared nodes z has to be larger than x + y  2k.
This theorem states that we can also check the cohesiveness of a kclique community if we do not know all constituting kcliques but only the constituting maximal cliques. I.e., the latter can contain more than k nodes. Since our definition of cohesiveness is given on the level of kcliques, this new theorem helps to understand its significance on the level of maximal cliques. The proof is illustrated in Fig. 9.
Proof: (1): If the kclique community consists of one clique of size ≥ 2k then one can find two disjoint cliques of size k, contradicting the strict clique overlap criterion. If the clique consists of less than 2k nodes it is not possible to find two disjoint cliques of size k.
(2) Note first that since the kclique community is the union of cliques with at least size k, it follows that x, y ≥ k. Assume that there are two cliques K_{ x }and K_{ y }and let K_{x∩y}:= K_{ x }∩ K_{ y }denote the set of shared nodes. Let furthermore K_{x∩y} = z, K_{ x∪y }:= K_{ x }∪ K_{ y }and let their union have at least 2k nodes: K_{x∪y} = x + y  z ≥ 2k. It follows that z > x + y  2k. If now x  z ≥ k choose any k nodes from K_{ x }\K_{ y }and any k nodes from K_{ y }. These two sets constitute kcliques that are naturally disjoint. If x z < k add any k (x z) nodes from K_{x∩y}, building kclique C_{1}. Naturally, K_{ y }\C_{1} will contain at least y  (k  x + z)) = y  k + x  z > k nodes. Pick any k nodes from this to build the second kclique C_{2}. C_{1} and C_{2} are again disjoint. It thus follows that if the union of two cliques contains at least 2k nodes, one can find two disjoint cliques of size k in them. If the union of the two cliques contains less than 2k distinct nodes it is not possible to find two sets of size k that do not share a common node which completes the proof. ■
With this, a simple algorithm to find all cohesive kclique communities is given by checking for each kclique community on each level k first whether it is cohesive:

1.
Check whether any of its constituting maximal cliques has a size larger than 2k  then it is not cohesive. This can be done in O(1) in an appropriate data structure of the kclique communities, e.g., if stored as a list of cliques. Let γ denote the number of maximal cliques in the graph. Since every maximal clique is contained in at most one kclique community on each level, this amounts to O(k _{ max } γ).

2.
Check for every pair of cliques K _{ x }, K _{ y }in it whether their overlap is larger than x + y  2k  then it is not cohesive. Again, since every clique can be contained in at most one kclique community on each level, this amounts to O(k _{ max } γ ^{2}).
The more challenging task is to prove maximality. In a naive approach, every of the kclique communities has to be checked against all other kclique communities whether it is a subset of any of these. Since there are at most k_{ max }γ many kclique communities with each at most γ many cliques contained in them, this approach results in a runtime of . Luckily this can be improved to the following runtime:
Theorem 8 To find all maximally cohesive kclique communities given the cliqueclique overlap matrix M takes O(k_{ max }· γ^{2}).
The proof can be found in the Appendix.
Of course, γ can in the worst case be an exponential number [19]. However, CFinder has proven itself to be very useful in the analysis of very large data sets with up to 10, 000 nodes [21]. Realworld networks neither tend to have a large k_{ max }nor a large number of different maximal cliques. Thus, although the runtime seems to be quite prohibitive it turns out that for the data sets that show up in biological and chemical fields the algorithm behaves nicely. Of course, there are several other algorithms for computing the set of all maximal cliques, especially on special graph classes, like sparse graphs or graphs with a limited number of cliques. A good survey on these algorithm can be found in [22]. The algorithm in [23] runs in O(nmγ), with n the number of nodes and m the number of edges in the original graph. Determining the cliqueclique overlap matrix takes O(nγ^{2}) time, and with this we come to an overall runtime of O(n(mγ + γ^{2})). Computing the kclique communities for a given k, starting from 3 and increasing it, can be done by first setting all entries smaller than k  1 to 0. Under the reasonable assumption that the number of different maximal cliques in realworld networks can be bound by a polynomial the whole runtime is polynomial.
Algorithm LInCS
Input: CliqueCliqueOverlap Matrix M
for k = 3 to k_{ max }do
Build graph G(k) in which two cliques C_{ i }, C_{ j }are connected by an
edge if M [i] [j] ≥ k  1
(k) ← compute components in G(k)
for all components C in (k) do
if isCohesive(C, M) then
Insert C into the list of recognized maximally cohesive kclique communities
Remove all maximal cliques in C from M
end if
end for
end for
Bool function isCohesive
for i = 1 to number of cliques in kcliquecommunity C do
if clique1 has more than 2k nodes then
return FALSE
end if
end for
for all pairs of cliques C_{ i }and C_{ j }do
if C_{ i }is a K_{ x }clique and C_{ j }is a K_{ y }clique and M [i] [j] < x + y  2k then
return FALSE
end if
end for
return TRUE
In the following section we will describe some results on the performance of the LInCSalgorithm on different data sets.
Data sets and experimental results
In subsection 'Data sets' we introduce the data sets that were used to evaluate the quality of the new clustering algorithm. Subsection 'Performance measurement of clustering molecules' describes how to quantify the quality of a clustering of some algorithm with a given reference clustering.
Data sets
We have applied LInCS to two different data sets: the first data set consists of druglike molecules and provides a natural clustering into six distinct clusters. Thus, the result of the clustering can be compared with the natural clustering in the data set. Furthermore, since this data set allows for a pairwise similarity measure, it can be compared with the result of a classic Ward clustering with level selection by Kelley. This data set is introduced in 'Druglike molecules'. The next data set on proteinprotein interactions shows why it is necessary to allow for graph based clustering methods that can moreover compute overlapping clusters.
Druglike molecules
157 druglike molecules were chosen as a reference data set to evaluate the performance of LInCS with respect to the most used combination of Ward's clustering plus level selection by Kelley et al.'s method. The molecules were downloaded from the ZINC database which contains commercially available druglike molecules [24]. The chosen molecules belong to six groups that all have the same scaffold, i.e., the same basic ring systems, enhanced by different combinations of side chains. The data was provided by the former ComGenex Inc., now Albany Molecular Research Inc. [25]. Thus, within each group, the molecules have a basic structural similarity; the six groups are set as reference clustering. Fig. 10 shows the general structural scheme of each of the six subsets and gives the number of compounds in each library. Table 1, 2 gives the IDs of all 157 molecules with which they can be downloaded from ZINC.
As already indicated, this data does not come in the form of a graph or network. But it is easy to define a similarity function for every pair of molecules, as sketched in the following.
Similarity metric of molecules
The easiest way to determine the similarity of two molecules is to use a socalled 2Dfingerprint that encodes the twodimensional structure of each molecule. 2D molecular fingerprints are broadly used to encode 2D structural properties of molecules [26]. Despite their relatively simple, graphtheory based information content, they are also known to be useful for clustering molecules [4].
Although different fingerprinting methods exist, hashed binary fingerprint methods attracted our attention due to their simplicity, computational costefficiency and good performance in the case of combinatorial libraries [2, 3]. One of the most commonly used hashed fingerprint algorithms is the Daylightalgorithm [27]. A similar algorithm was implemented by the ChemAxon Inc. [28] to produce ChemAxon hashed binary fingerprints.
The binary hashed fingerprint generating procedure first explores all the substructures in the molecule up to a predefined number of bonds (typically 6 bonds). The following step converts each discovered fragment into an integer based on a scoring table. The typical length, i.e., the number of bits of a hashed fingerprint are 1024, 2048, or higher, and initially the value of bits are set to 0. The presence of a discovered fragment in the fingerprint is represented by turning the value of the bit from 0 to 1 at the bit position computed by the score of the fragment. In summary, the fingerprints try to map substructures in a molecule into numbers with the idea that molecules with similar overall structure will have many substructures in common and thus also their fingerprints will be alike.
In this study we used the freely available ChemAxon fingerprint method [28]. Fingerprints for molecules were produced with a length of 4096 bits by exploring up to 6 bonds.
In order to quantify the level of similarity between the fingerprints of two molecules we applied the wellknown Tanimotosimilarity coefficient [29]. The Tanimotosimilarity coefficient of molecules A and B (T_{ AB }) is computed according to Eq. 1. In the formula, c denotes the number of common bits in the fingerprint of molecule A and B, and a and b stand for the number of bits contained by molecule A and B, respectively. The value of the Tanimotosimilarity coefficient ranges from 0 (least similar) to 1 (most similar).
The Tanimotosimilarity coefficient of each pair of molecules can be computed and then stored in a matrix, denoted as the similarity matrix. As stated above, the CCC method and LInCS expect as input a graph or network that is not fully connected, otherwise the outcome is just the whole set of entities. In the following we will describe how a reasonable similaritythreshold t can be found to turn the similarity matrix into a meaningful adjacency matrix. With this, we will then define a graph by representing all molecules by nodes and connect two nodes if their similarity is higher than that threshold.
Generating a similarity network of molecules
It is clear that the choice of a threshold value t might strongly influence the resulting clustering; thus, a reasonable selection of t is very important. To our knowledge, no general solution exists for selecting the optimal t. It is clear that if t is too high, the resulting graph will have only a few number of edges and will consist of several components. But it can be expected that these components will show a high density of edges and will be almost cliquelike. This can actually be measured by the socalled clustering coefficient that was introduced by Watts and Strogatz [30]. The clustering coefficient of a single node v computes the ratio between the number e(v) of its neighbors that are directly connected to each other and the possible number of neighbors that could be connected to each other, given by deg(v)·(deg(v)  1)/2:
The clustering coefficient of a graph is defined as the average clustering coefficient of its nodes. Starting at t = 1, the resulting graph is empty. By decreasing t, more and more edges will be introduced, connecting the isolated components with each other and thus increasing the average clustering coefficient.
Below a certain t, it is likely that the new edge will be connecting more or less random pairs of nodes and thus decrease the clustering coefficient of the newly connected nodes. Of course, if the threshold value is set to 0, the resulting graph would be one big clique and the clustering coefficient would again rise to 1. Based on this general behavior, the question is whether there is a natural choice for t where the clustering coefficient is still high and t is not yet too low. To understand the dependency between the average clustering coefficient and t, we computed all graphs with thresholds t ranging from 0.34which is the average similarity between all molecules in the setto 1. Surprisingly, there is a local maximum at t = 0.46 with an average clustering coefficient of 0.9834 (s. Fig. 11). This value was then chosen as the optimal threshold value.
On this data set we have applied Ward's hierarchical clustering with a level selection according to Kelley's index and compared it with the reference clustering and the one computed by our LInCS algorithm. The results are discussed in 'Performance of LInCS on the druglike molecule data set'. We will now introduce the second data set that does not allow for a simple pairwise similarity measure but is rather described naturally by a graph.
Proteinprotein interaction network of yeast core
Proteinprotein interaction data present a very good example of biological data that does not come with a quantifiable pairwise similarity but with a binary relationship: either two proteins are either observed to interact or not. Thus, classic hierarchical clustering methods cannot be applied here. Furthermore, it is assumed that many proteins may have more than one function in the cell; e.g., in the yeast protein database one or more of 42 cellular roles can be assigned to a given protein. According to [31], 39% of the proteins were assigned multiple cellular roles. Here, a clustering method is clearly advantageous if it can produce overlapping clusters.
The data was used as provided by Palla et al. in their CFindersoftware package which is available on the Internet [20] under the name ScereCR20050417. In the whole it contains 2, 640 proteins. The data is based on experimental data which are curated manually as well as automatically. Some of the most used proteinprotein interaction detection methods, as the yeast twohybrid system are wellknown for the poor quality of their results; some estimate that the number of falsepositive results is up to 50% [32]. In this case a clustering method can help to find out those edges that are more likely to be correct. To create a graph, each protein is presented by a node and two nodes are connected by an edge if the corresponding proteins are reported to interact. The resulting graph is unweighted and undirected.
Performance measurement of clustering molecules
After having introduced the data sets, we will now show how a given clustering's quality can be quantified if some reference clustering is given.
Sensitivity and specificity
We will first discuss the case of a nonoverlapping reference clustering and then that of an overlapping reference clustering.
Given any nonoverlapping reference clustering, as the set of 157 molecules that naturally come in six clusters, the clustering of some algorithm X has to be compared with it in a reasonable way. The correctness of the clustering process can be quantified by the socalled sensitivity and specificity measures [33]. Both rely on the following idea, exemplified on the molecule data set: Let A_{ CL }denote a matrix of size 157 × 157 where there is a 1 in A_{ CL }[i][j] if any combinatorial library contained both molecule m_{ i }and m_{ j }(and a 1 otherwise). Let similarly A_{ X }denote the corresponding matrix that results from the clustering by algorithm X, i.e., A_{ X }[i][j] contains a 1 if algorithm X detects a cluster in which m_{ i }and m_{ j }are both contained and a 1, otherwise. Since A_{ CL }is the standard, an algorithm performs well if it makes the same decisions of whether a pair of molecules is in the same cluster or not. We distinguish the following four cases:

1.
A _{ CL }[i][j] = 1 & A _{ X }[i][j] = 1: true positive;

2.
A _{ CL }[i][j] = 1 & A _{ X }[i][j] = 1: true negative;

3.
A _{ CL }[i][j] = 1 & A _{ X }[i][j] = 1: false positive;

4.
A _{ CL }[i][j] = 1 & A _{ X }[i][j] = 1: false negative;
With this, the sensitivity is defined as the number of pairs of molecules that are correctly clustered together by algorithm X divided by all the pairs that were clustered together in the reference clustering. This latter number can be expressed by the sum of all true positive and false negative pairs as can be easily seen from the enumeration above:
If the sensitivity is 1, this means that algorithm X has clustered together exactly the same pairs of molecules as the reference clustering. However, this would also happen, e.g., if X simply produces one big cluster. To rule this out, the specificity is defined as the ratio of the number of pairs correctly separated into different clusters, divided by the number of pairs that are in different clusters in the reference clustering. This latter number can be analogously represented by the sum of the number of true negatives plus the number of false positives:
Thus, the clustering computed by some algorithm X has a specificity of 1 if and only if it puts all pairs of molecules into different clusters that are in different clusters in the reference clustering. If the clustering computed by some algorithm X shows a sensitivity and specificity of 1 it has found exactly the same clusters as the reference clustering. Note that for a hierarchical clustering the sensitivity of a clustering in level k is at least as large as the sensitivity of clusterings in higher levels while the specificity is at most as large as that of clusterings in higher levels.
The case is a bit more difficult for overlapping reference clusterings as in the case of the proteinproteininteraction network where no simple reference clustering exists. However, for many proteins annotation data is available that assigns different kind of properties to them, e.g.:

1.
the biological processes in which the protein is involved;

2.
the cellular component in which the protein is located;

3.
the molecular function of the protein.
Using the BINGO 2.0 plugin [34] of Cytoscape [35] it is possible to obtain for a given clustering of proteins all of the annotated functions from the three categories as described above. Actually, these are downloaded from different databases, as described by [34]. One of the possible applications of a clustering algorithm on this kind of data is that proteins that are not yet assigned to, e.g., a biological process might be assigned to a cluster in which almost all proteins are annotated to the same biological process. Since a cluster represents a set of strongly interacting proteins it is then very likely that also the not yet annotated protein takes part in this biological process. This can then be checked directly in the laboratory, possibly saving money and time for other tests. To understand whether the clusterings obtained by the CCC and LInCS method are pure enough for most annotation categories, we analyzed them with respect to all annotation categories. Each category is thought of as a cluster in the reference clustering. Since more than one category can be assigned to each protein, this can be thought of as an overlapping reference clustering. To make the quality measurements meaningful in this case we will have to alter them slightly as described in the following.
Let denote the set of all proteins and let P(bp) denote the set of all proteins that are assigned to at least one biological process. This amounts to 2, 198 proteins. Similarly, P(cc) denotes the subset of all proteins that are assigned to at least one cellular component with P(cc) = 2, 184, and P(mf) the subset of all proteins of which at least one molecular function is known with P(mf) = 1, 822.
Each single biological process, cellular component, or molecular function will in the following be called a category. C_{ X }, X ∈ {bp, mf, cc}, denotes the type of the category: C_{ bp }denotes a specific biological process and analogously C_{ mf }and C_{ cc }a cellular component. Let now (C) denote the subset of proteins that is assigned to category C. Of course, a clustering is sensitive if all proteins in (C) are in one cluster.Regarding a single category C_{ X }, e.g., a biological process, we will call a clustering specific if no protein from (C_{ X }) is in a cluster with any protein that is at least assigned to one biological process but not to C_{ X }, i.e., any protein in P (X)  P(C_{ X }). However, we do not care whether any pair of two proteins from (P(X)  P(C_{ X })) × (P(X)  P(C_{ X })) are together in a cluster or not. This is justified because these proteins in P(X)  P(C_{ X }) could be assigned to other categories of the same type that require them to be in the same cluster. This implies that we should not require them to be in separate clusters. We thus restrict the measurement of the sensitivity to all pairs in (C_{ X }) × (C_{ X }) and the specificity to all pairs in (C_{ X }) × (P(X)  P(C_{ X })).
After describing the data sets and the methods to quantify the result of a given clustering, we will now describe the results.
Results and discussion
Performance of LInCS on the druglike molecule data set
The molecule data set described in 'Druglike molecules' is the data set to which all methods can be applied: Ward's hierarchical clustering together with the levelselection mechanism defined by Kelleyindex and the two overlapping clustering algorithms CCC and LInCS. We first discuss the clusters resulting from Ward's hierarchical clustering together with the Kelleyindex, then discuss the clusters on different klevels as computed by the CCC and finally show the results for LInCS.
Clustering the 157 molecules by Ward's method and using the Kelleyindex for selecting the best level resulted in nine clusters, i.e., k = 9, instead of the expected six. This clustering results in a specificity value of 1.0 and a sensitivity value of 0.6389. Actually, in this case the Kelley index did not choose the optimal level with respect to the reference clustering: In order to exclude the potential bias of the Kelleyindex, the sensitivity and specificity values were computed for all possible levels of the clustering hierarchy (Fig. 12). Surprisingly, the fifth level with sensitivity 1 and specificity 0.9842 shows the optimal combination of both values, and not the nineth level as predicted by the Kelleyindexs. Note that none of the levels shows the correct clustering.
The CCC performs better in that it at least produces some level in which the six wanted clusters are found: As in Ward's algorithm, we tested the sensitivity and specificity of the resulting clusters at each level of k. Here, the situation is a bit better since for k = 5 the computed clustering is the same as the reference clustering (Fig. 13). However, without the reference clustering set it would of course not be possible to find the best k.
Only LInCS manages to choose from all of the 138 kcliquecommunities provided by the CCC those six clusters that are the same as the reference clusters in one step by determining the set of maximally cohesive kclique communities (s. 'Finding cohesive kclique communities: LInCS').
Note that these clusters came from different k levels although they could also have come from the same level. In any case, this confirms the ability of our LInCS algorithm to resolve the klevel selection problem by choosing the maximally cohesive clusters from different levels. It should be noted that the similarity network itself consists of four components, i.e., the six clusters are not directly mapped to six different connected components which would then be easy to find. This argues for the fact that the six resultant clusters were not produced by applying the similarity threshold to the full graph but were produced by LInCS. Both, sensitivity and specificity reach their maximal value of 1.0 (Fig. 14) at the previously determined optimal similarity threshold at t = 0.46. Note also that the clusters are nonoverlapping for the chosen similarity threshold of t = 0.46. However, generating the similarity network by applying lower threshold values t, for instance t = 0.40, some of the resultant LInCS clusters are overlapping. This is reasonable considering that the criterion of two molecules to be similar is weakened by decreasing the threshold value. Another consequence is that the sensitivity and specificity values are slightly weaker but still close to their maximal value of 1 as shown in Fig. 14. Note that the correct clustering can be found for t in the interval of 0.450.48.
LInCS was valuable in finding all six groups of molecules that had the highest similarity as defined by the six combinatorial libraries. We will now show its performance on a data set where no such reference clustering exists but where the homogeneity of the clusters can still be determined.
Analysis of yeast core proteinprotein interaction network
While designing the LInCS algorithm the most important objective was to provide a general method for finding cohesive groups in networks. In order to demonstrate the specific feature of LInCS we applied it to the proteinprotein interaction network of Saccharomyces cerevisiae. CCC and the consecutive LInCS algorithm can be applied to this network without any transformation and the use of any parameters. The performance of the clustering was measured by computing the sensitivity and specificity values for the biological properties assigned to the proteins as described in 'Performance measurement of clustering molecules'.
Note that most of the proteins in the data set will not be contained in any of the kclique communities on each possible level k (s. Table 3). Thus, all proteins not contained in any community have to be added as singletons to guarantee that each protein is contained in the clustering. We call the percentage of proteins contained in at least one community the coverage of a clustering.
It is clear that from the coverage, the k = 3 level would be most reasonable because here 1, 284 out of the 2, 640 proteins are contained in at least one kclique community. On the other hand, the clusters are not all very specific: the minimal specificity of all clusters regarding categories from biological processes, molecular functions or the cellular component is 0.8 in all three cases. Vice versa, the clusterings on higher levels are much more specific but they cover only a few proteins out of all: the minimal specificity of any cluster regarding categories from biological processes is 0.98, for molecular function categories it is 0.99, and for cellular component categories it is 0.99. In essence, the only reasonable levels seem to be located at k = 3 or 4. In their paper, Palla et al. picked the k = 4 level as the most biologically reasonable by manually looking at the resultant clusterings [11]. According to the following measures the clustering computed by LInCS lies between the two levels: it covers 880 proteins instead of only 609 in level 4 but it is already almost perfectly specific: for all clusters the minimal specificity concerning biological process categories is 0.99, for all molecular function categories it is 1.0 and for all cellular component categories it is 0.99.
But is the LInCS clustering really preferable over the CCC clustering at level 3 if it covers less proteins? As stated above, the main motivation for the cluster selection algorithm was to find at least one cluster for each protein that is cohesive. Of course, the idea is that a cohesive cluster at the graph theoretic level also represents a set of data that is similar to each other. If no such cluster can be found for a given protein it might be better not to include it in an incohesive cluster but rather add it as a singleton. To test whether our notion of cohesiveness really manages to identify the clusters that contain biologically similar proteins we measured the homogeneity of categories in the following way: We define the homogeneity H(C, k, i) of cluster i and category C at level k as the percentage of proteins assigned to category C. For all categories to which at least one protein of the cluster is assigned we computed its homogeneity. Fig. 15, 16 and 17 shows histograms that represent for each (nonsingleton) cluster how many categories had a homogeneity of at most 20%, between 20  40%, 40  60%, 60  80% and 80  100%. It can be clearly seen that LInCS almost exclusively only picks those clusters in which the homogeneity of the categories is at least between 20% and 40%. Furthermore, those that were not included contain many proteins assigned to categories with a very low homegeneity value. This implies that those clusters that were included to the LInCS clustering by the graph theoretic notion of cohesiveness are much more homogeneous on the biological level than those that were not included. This result strengthens our choice of a cohesiveness measure, but of course, for other data sets other measures of cohesiveness might be appropriate as we will discuss in section 'Generalization of the approach'.
In summary it turns out that LInCS, by picking clusters from different levels, interpolates between the advantages and disadvantages of level 3 and 4. A closer look at those clusters that were not included in LInCS' clustering revealed that they are biologically not very meaningful because the contained proteins do not agree in their biological role. This indicates that the graph theoretic based notion of cohesiveness captures a notion that is also biologically meaningful. Note that LInCS, like the Kelley index for data sets with a similarity measure, does not need any further parameters: Following the paper by Palla et al., different proteinprotein interaction networks have been analyzed based on the CCC [11, 14, 15]. Interestingly, none of the latter publications offered a systematic and general method for extracting one single (overlapping) clustering from the different levels. As shown above, selecting the kclusters by LInCS did not require any parameter or prior knowledge from the user. The experiments confirm that our approach is not only able to produce a cohesive clustering but that the resulting clustering is reasonable for the two data sets we analyzed.
As already seen above, LInCS can produce clusterings that put many nodes in singletons. In extreme cases it is possible that CCC produces only noncohesive kclique communities such that LInCS will produce a clustering in which all nodes are contained as singletons (s. Fig. 18). It is certainly necessary to analyze a data set more closely, as we have done it for the PPI data set above, to see whether this is really a feature of the data set or an artifact of the cohesiveness measure. We will thus now discuss how the idea of finding cohesive clusters can be generalized.
Generalization of the approach
The idea of choosing maximally cohesive clusters can actually be extended to all hierarchical clusterings to avoid choosing one optimal level instead of choosing from all the clusters on all levels at the same time. Of course, for other algorithms and data sets the notion of cohesiveness has to be replaced reasonably. To generalize our idea, the chosen cohesiveness measure should have the following property: If a cluster on level k is cohesive this implies that all of its subsets are cohesive. Thus, coming from the levels with the lowest k the maximally cohesive clusters are found first, i.e., in a socalled divisive clustering algorithm that starts with the full set and refines this iteratively the algorithm would stop the refinement at a given cluster the moment it is cohesive.
Conclusion
In this article we present a new kind of cluster selection method for hierarchical clusterings, introduced on the example of the CCC method by Palla et al. [11]. This method is especially suitable for biological and chemical data sets where no full similarity measure can be defined and in which multiple membership of items to clusters is expected. For these data sets, which can be represented by a graph, we propose to use the CCC method and our levelindependent cluster selection mechanism LInCS. We showed on two examples that this method yields biologically or chemically meaningful clusterings. The method is deterministic, the runtime is quadratic in the number of maximal cliques in the data set and linear in the size of the maximum clique. Under the reasonable assumption that both parameters are small in realworld data sets the runtime is feasible in practice. LInCS uses a graphtheory based and deterministic procedure to find socalled cohesive clusters which does not require prior knowledge and expectation on the number and size of the clusters. Although the proposed cohesiveness measure was very successful for the analyzed data sets, other data sets might require other kinds of measures. We thus sketched the generally necessary properties of any kind of cohesiveness measure to adapt our method to other kind of hierarchical clustering methods and data sets. Further research on this topic will have to show whether the level independent method of cluster selection is in general more successful than the classic level selection methods.
Appendix
Proof of Theorem 8
PROOF: Let C be any cohesive kclique community with k ≥ 3. As stated by Theorem 4, it might be the superset of one or more k + 1clique communities. It is easy to see that the k + 1clique communities contained in it will also be cohesive since any subset of a cohesive set of cliques is also cohesive. Thus, a cohesive superset ensures that all its contained jclique communities with j > k are also cohesive. Since it is as least as large as those, we do not have to check these jclique communities since they cannot be maximally cohesive. On the other hand, the kclique community C is itself a subset of exactly one k  1clique community on level k  1. Thus, if this community is cohesive, C is not maximally cohesive. It is now easy to see that by trying for smallest k first, and checking all kclique communities on that level for cohesiveness, any cohesive kclique community is at the same time maximally cohesive. Thus, all its maximal cliques can be removed from the cliquecliqueoverlap matrix, thereby reducing the amount of search time on higher levels.
To achieve the runtime, the cliqueclique overlap matrix is viewed as an adjacency matrix that builds a graph with at most γ nodes and at most γ^{2} edges. Set k to 3 and let every entry greater than or equal to k  1 = 2 in the matrix constitute one edge. Computing this graph G(k) = G(3) costs O(γ^{2}). A simple component analysis, e.g., by a breadth first search analysis, yields all 3clique communities. These can then be checked for cohesiveness and due to the arguments above, any cohesive 3clique community will at the same time be maximally cohesive. We can now equate a cohesive kclique community with the component in the graph and define a component to be cohesive if the corresponding kclique community is cohesive. All cohesive components can now be deleted from the graph because of the above given argument that they are maximally cohesive. After all components have been checked and while the graph is still nonempty, k is increased by one. Then a new graph G(k) is built where again an edge is drawn between two cliques if their overlap is at least k 1. Then, the components in G(k) are computed and checked for cohesiveness. This cycle of increasing k, building G(k), the component and cohesiveness analysis and deleting the cohesive components from the graph is then repeated until the graph G(k) is empty. Note that the graph must eventually be empty since the highest entry in the cliqueclique overlap matrix is ≤ k_{ max }.
Since a component analysis can be done in linear time in the number of edges in a graph, i.e., in O(γ^{2}), and there are at most k_{ max }rounds, this amounts to O(k_{ max }γ^{2}). Checking for cohesiveness costs the same as argued above, and thus the total runtime is in O(k_{ max }γ^{2}). See Algorithm for a pseudocode description of LInCS. ■
References
 1.
Downs GM, Willett P: Similarity searching and clustering of chemicalstructure databases using molecular property data. J Chem Inf Comput Sci. 1994, 34: 10941102.
 2.
Willett P: Chemical similarity searching. J Chem Inf Comput Sci. 1998, 38: 983996.
 3.
Wild DJ, Blankley CJ: Comparison of 2D fingerprint types and hierarchy level selection methods fo structural grouping using Ward's clustering. J Chem Inf Comput Sci. 2000, 40: 155162.
 4.
Brown RD, Martin YC: Use of structureactivity data to compare structurebased clustering methods and descriptors for use in compound selection. J Chem Inf Comput Sci. 1996, 36: 572584.
 5.
Ward JH: Hierarchical grouping to optimize an objective function. J Amer Statist Assoc. 1963, 58: 236244. 10.2307/2282967.
 6.
Leach AR: Molecular modeling, principles and applications. 1997, AddisonWesley Publishing Company
 7.
Kelley LA, Gardner SP, Sutcliffe MG: An automated approach for clustering an ensemble for NMRderived protein structures into conformationally related subfamilies. Protein Eng. 1996, 9: 10631065.
 8.
Hartigan JA, Wong MA: A Kmeans clustering algorithm. Applied Statistics. 1979, 28: 100108. 10.2307/2346830.
 9.
Jarvis RA, Patrick EA: Clustering using a similarity measure based on shared near neighbors. IEEE Trans Comput. 1973, C22: 10251034. 10.1109/TC.1973.223640.
 10.
Girvan M, Newman MEJ: Community structure in social and biological networks. Proceedings of the National Academy of Sciences. 2002, 99: 78217826. 10.1073/pnas.122653799.
 11.
Palla G, Derényi I, Farkas I, Vicsek T: Uncovering the overlapping community structure of complex networks in nature and society. Nature. 2005, 435: 814818.
 12.
Jardine N, Sibson R: The construction of hierarchic and nonhierarchic classifications. Comp J. 1968, 11: 177
 13.
Cole AJ, Wishar D: An improved algorithm for the JardineSibson method of generating overlapping clusters. Comp J. 1970, 13: 156163. 10.1093/comjnl/13.2.156.
 14.
Zhang S, Ning X, Zhang XS: Identification of functional modules in a PPI network by clique percolation clustering. Computational Biology and Chemistry. 2006, 30 (6): 445451.
 15.
Futschik ME, Chaurasia G, Tschaut A, Russ J, Babu MM, Herzel H: Functional and transcriptional coherency of modules in the human protein interaction network. Journal of Integrative Bioinformatics. 2007, 4 (3): doi:10.2390/biecolljib200776
 16.
Gaertler M: Network analysis: Methodological foundations. 178215. SpringerVerlag 2005 chap. Clustering
 17.
Adamcsek B, Palla G, Farkas IJ, Derényi I, Vicsek T: CFinder: Locating cliques and overlapping modules in biological networks. Bioinformatics. 2006, 22: 10211023.
 18.
Derényi I, Palla G, Vicsek T: Clique percolation in random networks. Phys Rev Lett. 2005, 94: 160202
 19.
Garey MR, Johnson DS: Computers and intractability  a guide to the theory of NPcompleteness. 1979, W. H. Freeman and Company, New York
 20.
CFinder. http://cfinder.org
 21.
Personal communication with Gergely Palla.
 22.
Bonze I, Budinich M, Pardalos P, Pelillo M: Handbook of combinatorial optimization. 1999, 4: 1921. Kulwer Academic Publishers, chap The maximum clique problem
 23.
Tsukiyama S, Ide H, Ariyoshi H, Shirakawa I: A new algorithm for generating all the maximal independent sets. SIAM J Comput. 1977, 6 (3): 505517. 10.1137/0206036.
 24.
Irwin JJ, Shoiche BK: ZINC  a free database of commercially available compounds for virtual screening. J Chem Inf Model. 2005, 45: 177182.
 25.
Albany Molecular Research Inc. http://www.amriglobal.com/
 26.
Maldonado AG, Doucet JP, Petitjean M, Fan BT: Molecular similarity and diversity in chemoinformatics: from theory to applications. Molecular Diversity. 2006, 10 (1): 3979.
 27.
Daylight Chemical Information Systems Inc. http://www.daylight.com/dayhtml/doc/theory/theory.finger.html
 28.
ChemAxon Ltd., Chemical hashed fingerprints. http://www.chemaxon.com/jchem/doc/user/fingerprint.html
 29.
Tanimoto TT: Tech. rep., IBM Internal Report. 1957
 30.
Watts DJ, Strogatz SH: Collective dynamics of 'smallworld' networks. Nature. 1998, 393: 440442.
 31.
Schwikowski B, Uetz P, Fields S: A network of proteinprotein interactions in yeast. Nature Biotechnology. 2000, 18: 12571261.
 32.
Deane C, Salwiñski L, Xenarios I, Eisenberg D: Protein interactions: two methods for assessment of the reliability of high throughput observations. Mol Cell Proteomics. 2002, 1 (5): 349356.
 33.
Altman DG, Bland JM: Diagnostic tests 1: Sensitivity and specificity. BMJ. 1994, 308: 1552
 34.
Maere S, Heymans K, Kuiper M: BiNGO: a Cytoscape plugin to assess overrepresentation of Gene Ontology categories in Biological Networks. Bioinformatics. 2005, 21 (16): 34483449.
 35.
Shannon P, Markiel A, Ozier O, Baliga NS, Wang JT, Ramage D, Amin N, Schwikowski B, Ideker T: Cytoscape: a software environment for integrated models of biomolecular interaction networks. Genome Res. 2003, 13 (11): 24982504.
Acknowledgements
The project supported by the Hungarian National Research Fund and by the National Office for Research and Technology (Grant Number OTKA 67651), by European Community's Seventh Framework Programme (FP7/20072013)/ERC grant agreement no. 208319 and National Technology Programme, grant agreement no. TECH_08_A1/22008106, by the DELTA Informatika Inc., by a grant from the National Research and Technological Office (NKTH Textrend). KAZ was supported by a grant by the Deutsche Akademie der Naturforscher Leopoldina (BMBF LPD 9901/8182). The authors thank Imre Derényi and Gergely Palla for fruitful discussions.
Author information
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors' contributions
LAZ, GYK and GZK designed the algorithm, proved its correctness, and LAZ implemented it. KAZ and GZK designed the experiments, GZK performed them, KAZ generalized the idea and wrote most of the text. All authors contributed to the text, read and approved the final manuscript.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
About this article
Cite this article
Zahoránszky, L.A., Katona, G.Y., Hári, P. et al. Breaking the hierarchy  a new cluster selection mechanism for hierarchical clustering methods. Algorithms Mol Biol 4, 12 (2009). https://doi.org/10.1186/17487188412
Received:
Accepted:
Published:
Keywords
 Cluster Coefficient
 Maximal Clique
 Hierarchical Cluster Method
 Average Cluster Coefficient
 Reference Cluster