- Research
- Open access
- Published:

# A polynomial time algorithm for computing the area under a GDT curve

*Algorithms for Molecular Biology*
**volume 10**, Article number: 27 (2015)

## Abstract

### Background

Progress in the field of protein three-dimensional structure prediction depends on the development of new and improved algorithms for measuring the quality of protein models. Perhaps the best descriptor of the quality of a protein model is the *GDT* function that maps each distance cutoff *θ* to the number of atoms in the protein model that can be fit under the distance *θ* from the corresponding atoms in the experimentally determined structure. It has long been known that the area under the graph of this function (*GDT*_*A*) can serve as a reliable, single numerical measure of the model quality. Unfortunately, while the well-known *GDT*_*TS* metric provides a crude approximation of *GDT*_*A*, no algorithm currently exists that is capable of computing accurate estimates of *GDT*_*A*.

### Methods

We prove that *GDT_A* is well defined and that it can be approximated by the Riemann sums, using available methods for computing accurate (near-optimal) *GDT* function values.

### Results

In contrast to the *GDT_TS* metric, *GDT*_*A* is neither insensitive to large nor oversensitive to small changes in model’s coordinates. Moreover, the problem of computing *GDT*_*A* is tractable. More specifically, *GDT*_*A* can be computed in cubic asymptotic time in the size of the protein model.

### Conclusions

This paper presents the first algorithm capable of computing the near-optimal estimates of the area under the *GDT* function for a protein model. We believe that the techniques implemented in our algorithm will pave ways for the development of more practical and reliable procedures for estimating 3D model quality.

## Background

Advances in the area of protein three-dimensional structure prediction depend on the ability to accurately measure the quality of a protein model. One of the most popular and most reliable measure of the protein model quality is *GDT*_*TS*. It is defined as the average value of *GDT_P*
_{
θ
} computed for four distance cutoffs \(\theta = 2^{i}\), \(i = \overline{0,3}\), where *GDT_P*
_{
θ
} is the percentage of model residues (represented by their *C*
_{
α
} atoms) that can be placed under *θ* ångströms from the corresponding residues in the experimental structure [1, 2]. In a “high-accuracy” version of *GDT*_*TS*, denoted by *GDT*_*HA*, the distance cutoffs are cut in half (\(\theta = 2^{i}\), \(i = \overline{ - 1,2}\)) [3]. In both approaches, the underlying assumption is that the experimental (crystallographic or NMR) structure is close to the real (native) structure (which is sometimes not true due to experimental errors).

Several methods exist for computing *GDT*_*TS*. The LGA algorithm [4] can estimate *GDT*_*TS* quickly, but those estimates deviate from the true *GDT*_*TS* values in about 10 % of the cases [5]. Rigorous algorithms for computing *GDT*_*TS* have also been developed [6–9], but they are computationally much more expensive.

The *GDT*_*TS* is commonly interpreted as an approximation of the area under the *GDT* curve, denoted by *GDT*_*A* [10–12]. Unfortunately, since the measure is approximated using the *GDT* function values at only several distance cutoffs, the errors in the area approximation are large. As we demonstrate later, *GDT*_*TS* is not only overly sensitive to small but also insensitive to large changes in the protein model’s coordinates.

In this paper, we present a polynomial time algorithm for computing *GDT*_*A*. Our method runs on the order \(\widetilde{O}(n^{3} )\), where *n* represents the length of the protein model (and \(\widetilde{O}\) hides the log factor). The algorithm returns “near-optimal” *GDT*_*A* scores, meaning that the errors in our estimates can be made arbitrary small i.e., smaller than any upfront specified vale. Although our method is theoretical, we believe that its parallel implementations, coupled with carefully designed speed up techniques, can result in a practical and widely used software tool.

The rest of this paper is structured as follows. First, we present three examples that illustrate drawbacks of *GDT*_*TS* and advantages of *GDT*_*A*. Then, we place our theory on a firm mathematical ground, which enables us to formally define the *GDT*_*A* computation problem. Finally, we describe the actual algorithm for *GDT*_*A* and provide its running time analysis.

## Methods

### Definition of the GDT function

The *GDT* function is a mapping that relates each distance cutoff *θ* to the percentage of model residues that can be placed at distance ≤*θ* from the corresponding residues in the experimentally determined structure. The graph of a *GDT* function provides a valuable insight into the quality of a protein model (Fig. 1). More specifically, the closer the graph runs to the horizontal axis (in other words, the smaller the area under the graph), the better the model.

As a single numerical measure of the model quality, *GDT*_*TS* is extensively used at CASP to rank different models for the same target [13, 14]. Since it represents the average of *GDT*_*P*
_{
θ
} at several distance cutoffs, *GDT*_*TS* is often viewed as an approximation of the area under the *GDT* curve (*GDT*_*A*) [10–12]:

However, as we demonstrate below, such a sparse sampling of the values of *GDT* function compromises the reliability of *GDT*_*TS*.

In our first example, we analyze the protein model for the target T0482, submitted by the group TS208 at CASP8 (Fig. 1). The *GDT*_*TS* score of this particular model was not even among the best dozen at CASP8, despite the fact that it fits the largest number of residues at distance \(\le\sim 4\) from the corresponding residues in the experimental structure. In fact, the blue model (Fig. 1a) can be superimposed onto the experimental structure so that all of its residues are at distance \(\le 8\) from the residues in the experimental structure (Fig. 1b), while no such superposition exists for any other model, even for the distance cutoff of 10Å. Interestingly, according to the MAMMOTH algorithm [15], the blue model is the best model for this particular target, while the DALI [16] algorithm ranks it as the second best.

Although it is impossible to tell whether #13 *GDT*_*TS* rank is more or less fair than #1 and #2 rank assigned by MAMMOTH and DALI, respectively, it is also not difficult to see that the ranking by the area under the *GDT* plot (*GDT*_*A*) would serve as a good compromise between these extremes.

The next two examples illustrate further disadvantages of *GDT*_*TS*. As seen in Fig. 2, better *GDT*_*TS* scores can be assigned to obviously worse models. Moreover, as demonstrated in Fig. 3, very similar models can have significantly different *GDT*_*TS* scores.

### Mathematical formalism

Strictly speaking, the *GDT* function is not well-defined. Zooming into the plot of the model highlighted in Fig. 1a, we see a set of many small vertical segments, meaning that each point on the horizontal axis is mapped to zero or more points on the vertical axis (Fig. 4). On the other hand, the inverse function (mapping each distance cutoffs *θ* to the percentage of residues in the model structure that can be fit under the distance *θ* from the corresponding residues in the experimental structure) is obviously well defined. This allows us to define the area under the *GDT* plot as the complement of the area under the inverse function:

where *Total*_*Area* represents the area of the rectangular region under consideration (100 × 10). We start our mathematical formalism by first defining a protein structure.

###
**Definition 1**

A *protein structure*
*a* is a sequence of points in the three dimensional Euclidean space \({\mathbb{R}}^{3}\)

The sequence elements *a*
_{
i
} can represent individual atoms, but it is more typical (in particular in protein structure prediction experiments) to assume that each point *a*
_{
i
} corresponds to the alpha-carbon atom of the protein’s *ith* amino acid.

In what follows, we formally define the *GDT* function [17]. For simplicity of presentation, we will modify the codomain of *GDT* to represent the “fraction of residues” (ranging from 0 to 1) instead of “percentages of residues” (ranging from 0 to 100). We note that this simple rescaling of the ordinate values will have no effects on the results obtained in our study.

###
**Definition 2**

Let \(a = (a_{1} , \ldots , a_{n} )\) be a protein structure consisting of *n* amino acids, let \(b = (b_{1} , \ldots , b_{n} )\) be a 3D model of *a*, and let \(\overline{\theta }\) be a positive constant. The *Hubbard function (or GDT function)* is the function \(H_{b} :[0,\overline{\theta } ] \to (0,1]\), defined by \(H_{b} \left( \theta \right) = \mathop {\hbox{max} }\nolimits_{\tau } | \{ i \text{ }|\text{ } \| a_{i} - \tau \left( {b_{i} } \right) \|\le \theta \} |/n\), where denotes the Euclidean norm on \({\mathbb{R}}^{3}\) and *τ* is a rigid transformation (a composition of a rotation and a translation).

###
**Theorem 1**

*H*
_{
b
}
*is a stepwise function with finitely many steps*
\(\theta_{1} , \ldots , \theta_{k}\), 1 ≤ *k* ≤ *n* − 1.

###
*Proof*

Since *H*
_{
b
} is monotony non-decreasing and since the range of *H*
_{
b
} is a finite subset of (0,1], it follows that *H*
_{
b
} must be a stepwise function. To complete the proof, we note that the number of steps in *H*
_{
b
} matches the size of its range, which does not exceed *n* − 1, where *n* is the length of *b*.

For simplicity of presentation, from now on (and whenever the model *b* is implied), we will omit the subscript in *H*
_{
b
} and denote the Hubbard function only by *H*.

### Algorithm for GDT_A

The area under *H* is the sum of the areas of the rectangular regions (*θ*
_{
i
})(*θ*
_{
i
} − *θ*
_{
i−1}):

where *θ*
_{0} = 0 and \(\theta_{k + 1} = \overline{\theta }\) (Fig. 5). It would be trivial to compute *Area* had we known all *θ*
_{
i
} and all function values *H*(*θ*
_{
i
}). Unfortunately, even if we knew the step points *θ*
_{
i
}, it would be computationally very difficult to compute the function values at them, since the best to date algorithm for computing *H*(*θ*
_{
i
}) runs on the order of *O*(*n*
^{7}) [7]. Hence, we resort to using the Riemann sums to approximate (instead of to compute exactly) the area under the graph of *H*.

The following definition and an accompanying theorem can be found in virtually any mathematical analysis textbook.

###
**Definition 3**

If \(f:\left[ {a,b} \right] \to {\mathbb{R}}\) is a function then \( R = \sum\limits_{i = 1}^{n} {v_{i} \left( {x_{i} - x_{i - 1} } \right)}, \) where *a* = *x*
_{0} < *x*
_{1} < ··· < *x*
_{
n
} = *b* is the partition of the interval [*a*, *b*] and *v*
_{
i
} denotes the supremum of *f* over [*x*
_{
i−1}, *x*
_{
i
}], is called the *upper Riemann*
*sum* of *f* on [*a*, *b*].

###
**Theorem 2**

*Let f be a real, non-decreasing, Riemann integrable function on an interval* [*a*, *b*]. *Then*

*where*

*is the upper Riemann sum of f the and*
\(\Delta x = \max_{i} (x_{i} - x_{i - 1} ).\)

Observe that, since *H*
_{
b
} is piecewise continuous, it must be integrable on \([0,\overline{\theta } ]\). Thus, the area under the graph of *H* is

To approximate *Area* with a Riemann sum, one can define the partition points \(\epsilon ,2\epsilon , \ldots ,m\epsilon\), where \(m = \lceil {\overline{\theta}} / \epsilon \rceil\) (Fig. 6) and then compute an estimate *Area*(\(\epsilon\)) of *Area* as

The error |*Area* – *Area* (\(\epsilon\))| in the estimate (8) is below 2\(\epsilon\). Up to a half of this error is due to the error in the Riemann sum with the remaining error being due to the possible placement of the last partition point *m*
\(\epsilon\) outside the interval \([0,\overline{\theta } ]\).

Unfortunately, computing the area estimates according to (8) is still a challenging problem, because (as we mentioned above), there is no computationally effective procedure for finding the function values *H*(*i*
\(\epsilon\)). To circumvent the problem, we utilize an efficient algorithm capable of computing the lower bound estimates *H*
_{
i
} of *H*(*i*
\(\epsilon\)), satisfying \(H\left( {(i - 1)\epsilon} \right) \le H_{i} \le H\left( {i\epsilon} \right)\), \(i = \overline{1,m}\). We then compute an estimate \(\widetilde{Area}(\epsilon)\) of *Area* as

Since \(\left| {\widetilde{Area}\left(\epsilon\right) - Area(\epsilon)} \right| < 2\epsilon\), it follows that \(\widetilde{Area}(\epsilon)\) is a \(4 \epsilon\)-approximation of *Area*. Below we show how to compute all *H*
_{
i
}’*s*, and, in turn, \(\widetilde{Area}(\epsilon)\) in time \(O\left({n^{3} logn/\epsilon^{6} } \right)\), where *n* is the length of *b*. Our algorithm takes advantage of an efficient procedure for computing near optimal *GDT*_*TS* values [5].

Let *T*(*b*) denotes the image of the model structure *b* under the transformation *T*. Denote by \(MAX(T, \theta )\) the largest fraction of residues from *T*(*b*) that are at distance ≤*θ* from the corresponding residues in the experimental structure *a*. To find each *H*
_{
i
}, it is enough to compute a rigid body transformation *T*
_{
i
} satisfying \(H\left({(i - 1)\epsilon} \right) \le MAX\left({T_{i} ,i}\epsilon \right) \le H\left({i} \epsilon\right)\).

Denote by *T*
_{
θ
} a transformation that places a largest subset *b*
_{
θ
} of residues from *b* at distance ≤*θ* from the corresponding residues in the experimental structure. Given *T*
_{
θ
}, one can easily compute *b*
_{
θ
} by calculating all *n* distances between the residues *a*
_{
i
} and *T*
_{
θ
}(*b*
_{
i
}). Note that *P*(*T*
_{
θ
}, *θ*) = *H*(*θ*). We approximate the transformation *T*
_{
θ
} by a so-called “near-optimal” transformation i.e., a transformation that places at least as many residues from the model structure under distance \(\theta + \epsilon\) as the optimal transformation *T*
_{
θ
} places under the distance *θ*. From now on, we will use \(T_{\theta}^{\epsilon}\) to denote a “near-optimal” transformation and the corresponding set of residues will be denoted by \(b_{\theta}^{\epsilon}\). Observe that \(P\left({T_{\theta}^{\epsilon}, \theta + \epsilon} \right) \ge P\left({T_{\theta}, \theta} \right) = H(\theta)\).

Building upon any procedure for computing \(T_{\theta}^{\epsilon}\), one can develop an algorithm for \(\widetilde{Area}(\epsilon)\) by substituting \(P\left({T_{{\theta_{i}}}^{\epsilon}, \theta_{i} + }\epsilon \right)\) for *H*
_{
i
} in (10), where \(\theta_{i} = (i - 1)\epsilon\). Several existing methods can be modified and made suitable for finding \(T_{\theta}^{\epsilon}\). The most efficient such method relies on the concept of “radial pair” [5].

###
**Definition 4**

Let \(S = \left\{ {s_{1} , \ldots , s_{n} } \right\}\) be a set of points in the three-dimensional Euclidean space. An ordered pair of points \((s_{i} , s_{j} )\) is called a *radial pair* of *S* if *s*
_{
j
} is the furthest point from *s*
_{
i
} among all points in *S*.

###
**Theorem 3**

*Let T*
_{1}
* and*
*T*
_{2}
*be two transformations and let*
\((s_{k} , s_{l} )\)
* be a radial pair of S*. *If*
\(\|T_{1} \left({s_{k}} \right) - T_{2} (s_{k})\| <\epsilon /3\)
*and*
\(\|T_{1} \left({s_{l}} \right) - T_{2} (s_{l})\| < \epsilon/3\)
* then there exists a rotation R*
*around the line through T*
_{1}(*s*
_{
k
}) *and*
\(T_{1} (s_{l} )\)
*such that*
\(\|R\left({T_{1} \left({s_{p} } \right)} \right) - T_{2} (s_{p}) \|<\epsilon\), *for any s*
_{
p
}
* in*
*S*. *The rotation R can be found in time O* (*nlogn*), *where n is the size of S*.

A proof of the above theorem can be found in [5]. The algorithm for finding *R* is fairly straightforward and it relies on the so-called *plane*-*sweep* approach [18].

The Theorem 3 implies that one choice for the near-optimal transformation \(T_{\theta}^{\epsilon}\) is the transformation *R* ∘ *T*, where *T* is any transformation that maps the points *b*
_{
k
} and *b*
_{
l
} from the radial pair (*b*
_{
k
}, *b*
_{
l
}) of *b*
_{
θ
} to the \(\epsilon/3\) neighborhoods of *T*
_{
θ
}(*b*
_{
k
}) and *T*
_{
θ
}(*b*
_{
l
}), respectively, and *R* is the rotation around the radial axis \(\overline{{T\left( {b_{k} } \right)T\left( {b_{l} } \right)}}\) that maps the remaining points from *T*(*b*
_{
θ
}) to the \(\epsilon\)-neighborhoods of the corresponding points from *T*
_{
θ
}(*b*
_{
θ
}).

In search for a radial pair of *b*
_{
θ
}, the algorithm in [5] explores all *n*
^{2} possible pairs of residues in *b*. For each candidate radial pair \((b_{k} , b_{l} )\), the algorithm generates a finite, representative set of transformations that map *b*
_{
k
} and *b*
_{
l
} into \(\theta + \epsilon/3\) neighborhoods of *a*
_{
k
} and *a*
_{
l
}, respectively (see the paragraph below for more details). For every such transformation *T*, a plane-sweep algorithm [18] is used to find a rotation *R* around the axis \(\overline{{T\left( {b_{k} } \right)T\left( {b_{l} } \right)}}\) that maximizes the number of residues from *R*(*T*(*b*)) that can be placed at distance \(<\theta + \epsilon\) from the corresponding residues in *a*.

A finite set of transformations that map the residues *b*
_{
k
} and *b*
_{
l
} into the \(\theta + \epsilon/3\) neighborhoods of *a*
_{
k
} and *a*
_{
l
}, respectively, is constructed in such a way to ensure that for at least one of those transformation *T*, \(\| T\left({b_{k}} \right) - T_{\theta} (b_{k} )\| < \epsilon/3\) and \(\|T\left({b_{l}} \right) - T_{\theta} (b_{l}) \|< \epsilon/3\). This can be achieved by partitioning \({\mathbb{R}}^{3}\) into small cubes of side length slightly smaller than \(\sqrt 3\epsilon /9\) and then collecting the vertices of the cubes that are inside the open ball of radius \(\theta + \epsilon/6\) around *a*
_{
k
} (Fig. 7). The elements of this set, denoted by *A*
_{
k
}, are the candidate points *T*(*b*
_{
k
}). The number of points in *A*
_{
k
} is \(O(1/\epsilon^{3} )\) and at least one of them must be at distance \(<\epsilon/6\) from *T*
_{
θ
}(*b*
_{
k
}) (Fig. 7). For each point *a*
^{k} ∊ *A*
_{
k
}, the set *A*
_{
l
}(*a*
^{k}) of possible images of *b*
_{
l
} under *T* is computed by discretizing the spherical cap \(S(a^{k}, \| b_{k} - b_{l}\| ) \cap B(a_{l} ,\theta + \epsilon/3)\), where *S*(*a*, *r*) and *B*(*a*, *r*) denote the sphere and the open ball in \({\mathbb{R}}^{3}\) with center *a* and radius *r*, respectively, in such a way that at least one point from *A*
_{
l
}(*a*
^{k}) is found at distance \(< \epsilon/3\) from *T*
_{
θ
}(*b*
_{
l
}) (Fig. 7). We note that size of *A*
_{
l
}(*a*
^{k}) is \(O(1/\epsilon^{2} )\). Hence, the total number of candidate pairs of points (*T*(*b*
_{
k
}), *T*(*b*
_{
l
})) is \(O(1/\epsilon^{5})\).

An obvious to compute \(T_{{\theta_{1} }}^{\epsilon} , \ldots ,T_{{\theta_{m} }}^{\epsilon}\) is to run the just described algorithm *m* times in succession, for \(\theta = \theta_{1} , \ldots , \theta = \theta_{m}\). However, such an approach results in many unnecessary repeated calculations as the area around *a*
_{
k
} and the corresponding spherical cap in the neighborhoods of *a*
_{
l
} are discretized over and over again. Moreover, all transformations *T* and *R*, generated and inspected during the procedure for finding \(T_{{\theta_{i}}}^{\epsilon}\), are inspected again during the procedure for finding \(T_{{\theta_{j}}}^{\epsilon}\), for each *j* > *i*.

We show that all transformations \(T_{{\theta_{1}}}^{\epsilon} , \ldots ,T_{{\theta_{m} }}^{\epsilon}\) and the corresponding values *H*
_{1}, …, *H*
_{
m
} can be computed, at once, during the procedure of finding the last transformation, namely \(T_{{\theta_{m} }}^{\epsilon}\). As demonstrated in the pseudocode above, the transformation *T* is generated only once for each pair of points \(\left( {a^{k} , a^{l} } \right) \in A_{k} \times A_{l} \left( {a^{k} } \right)\) and a sweep-plane algorithm for finding *R* is called only once for each *i* satisfying \(\|a_{k} - a^{k}\| < \theta_{i} + \epsilon/6\) and \(\|a_{l} - a^{l} \|< \theta_{i} + \epsilon/3\). The values of *H*
_{
i
} are updated on the fly.

### Running time analysis

To analyze the algorithm’s running time, we note that the number of iterations of the first* for* loop is equal to the number of candidate radial pairs (*b*
_{
k
}, *b*
_{
l
}), which is \(O\left( {n^{2} } \right).\) The number of iterations of the second *for* loop matches the number of pairs of grid points around *a*
^{k} and *a*
^{l}, which is \(O\left( {1/\epsilon^{3} } \right) \times O\left( {1/\epsilon^{2} } \right) = O\left( {1/\epsilon^{5} } \right)\). Each one of \(O\left( m \right) = O(\lceil\bar{\theta }/\epsilon\rceil) = O(1/\epsilon)\) iterations of the third *for* loop calls a *O*(*nlogn*) plane-sweep procedure to compute an optimal rotation and (if needed) to update the value *H*
_{
i
}. Hence, the asymptotic time complexity of the three nested *for* loops is \(O\left( {n^{3} logn/\epsilon^{6} } \right).\)

## Conclusions

Estimating the quality of a protein 3D model is a challenging task. Automatically generated *GDT*_*TS* score is helpful as the first raw approximation but this measure is neither sensitive nor selective enough to be exclusively relied upon in ranking different models for the same target. In this paper, we show that using a more accurate approximation of the area under the *GDT* curve as the criterion of model quality addresses many of the drawbacks of *GDT*_*TS*. We also present a rigorous \(\widetilde{O}\left( {n^{3} } \right)\) algorithm for computing the area under the *GDT* curve for a given model, where *n* is the model’s length. The area estimate returned by our method is “near-optimal”, meaning that the error in the estimate can be made smaller than any upfront specified value.

Despite the cubic asymptotic running time with a relatively large hidden constant, we believe that the techniques presented in this paper can guide a future development of a computationally efficient computer program, in particular since our methodology is amenable to parallel implementations. A heuristic version of the algorithm for estimating the area under the *GDT* plot can be found at http://bioinfo.cs.uni.edu/GDT_A.html.

## References

Zemla A, Venclovas C, Moult J, Fidelis K. Processing and analysis of CASP3 protein structure predictions. Proteins. 1999;37(S3):22–9.

Zemla A, Venclovas C, Moult J, Fidelis K. Processing and evaluation of predictions in CASP4. Proteins. 2001;45(S5):13–21.

Read RJ, Chavali G. Assessment of CASP7 predictions in the high accuracy template-based modeling category. Proteins. 2007;69(S8):27–37.

Zemla A. LGA—a method for finding 3D similarities in protein structures. Nucleic Acids Res. 2003;31:3370–4.

Li SC, Bu D, Xu J, Li M. Finding nearly optimal GDT scores. J Comput Biol. 2011;18(5):693–704.

Li SC, Ng YK. On protein structure alignment under distance constraint. Theor Comput Sci. 2011;412:4187–99.

Choi V, Goyal N. A combinatorial shape matching algorithm for rigid protein docking. CPM Lecture Notes Comput Sci. 2004;3109:285–96.

Akutsu T. Protein structure alignment using dynamic programming and iterative improvement. IEICE Trans Ins Syst. 1995; E79-D(12):1629–36.

Poleksic A. Improved algorithms for matching r-separated sets with applications to protein structure alignment. IEEE/ACM Trans Comput Biol Bioinform. 2013;10(1):226–9.

Kryshtafovych A, Fidelis K, Moult J. CASP8 results in context of previous experiments. Proteins. 2009;77(9):217–28.

Tramontano A, Cozzetto D, Giorgetti A, Raimondo D. The assessment of methods for protein structure prediction. Methods Mol Biol. 2008;413:43–57.

Kryshtafovych A, Milostan M, Szajkowski L, Daniluk P, Fidelis K. CASP6 data processing and automatic evaluation at the protein structure prediction center. Proteins. 2005;61(S7):19–23.

Huang YJ, Mao B, Aramini JM, Montelione GT. Assessment of template-based protein structure predictions in CASP10. Proteins. 2014;82(S2):43–56.

Tai CH, Bai H, Taylor TJ, Lee B. Assessment of template-free modeling in CASP10 and ROLL. Proteins. 2014;82(S2):57–83.

Ortiz AR, Strauss CE, Olmea O. MAMMOTH (matching molecular models obtained from theory): an automated method for model comparison. Protein Sci. 2002;11:2606–21.

Holm L, Sander C. Protein structure comparison by alignment of distance matrices. J Mol Biol. 1993;233:123–38.

Pevsner J. Bioinformatics and functional genomics. 2nd edn. Wiley-Blackwell; 2009.

Alt H, Mehlhorn K, Wagener H, Welzl E. Congruence, similarity, and symmetries of geometric objects. Dicrete Comput Geom. 1988;3:237–56.

## Acknowledgements

This project was supported by the University of Northern Iowa Professional Development Award. The structure alignment figures were prepared in Jmol (http://www.jmol.org).

### Competing interests

The author declares that there is no competing interests regarding the publication of this article.

## Author information

### Authors and Affiliations

### Corresponding author

## Rights and permissions

**Open Access** This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

## About this article

### Cite this article

Poleksic, A. A polynomial time algorithm for computing the area under a GDT curve.
*Algorithms Mol Biol* **10**, 27 (2015). https://doi.org/10.1186/s13015-015-0058-0

Received:

Accepted:

Published:

DOI: https://doi.org/10.1186/s13015-015-0058-0