- Open Access
Characteristics of predictor sets found using differential prioritization
© Ooi et al; licensee BioMed Central Ltd. 2007
- Received: 27 September 2006
- Accepted: 04 June 2007
- Published: 04 June 2007
Feature selection plays an undeniably important role in classification problems involving high dimensional datasets such as microarray datasets. For filter-based feature selection, two well-known criteria used in forming predictor sets are relevance and redundancy. However, there is a third criterion which is at least as important as the other two in affecting the efficacy of the resulting predictor sets. This criterion is the degree of differential prioritization (DDP), which varies the emphases on relevance and redundancy depending on the value of the DDP. Previous empirical works on publicly available microarray datasets have confirmed the effectiveness of the DDP in molecular classification. We now propose to establish the fundamental strengths and merits of the DDP-based feature selection technique. This is to be done through a simulation study which involves vigorous analyses of the characteristics of predictor sets found using different values of the DDP from toy datasets designed to mimic real-life microarray datasets.
A simulation study employing analytical measures such as the distance between classes before and after transformation using principal component analysis is implemented on toy datasets. From these analyses, the necessity of adjusting the differential prioritization based on the dataset of interest is established. This conclusion is supported by comparisons against both simplistic rank-based selection and state-of-the-art equal-priorities scoring methods, which demonstrates the superiority of the DDP-based feature selection technique. Reapplying similar analyses to real-life multiclass microarray datasets provides further confirmation of our findings and of the significance of the DDP for practical applications.
The findings have been achieved based on analytical evaluations, not empirical evaluation involving classifiers, thus providing further basis for the usefulness of the DDP and validating the need for unequal priorities on relevance and redundancy during feature selection for microarray datasets, especially highly multiclass datasets.
- Feature Selection
- Marker Gene
- Microarray Dataset
- Manhattan Distance
- Feature Selection Technique
The aim of feature selection is to form, from all available features in a dataset, a relatively small subset of features capable of producing the optimal classification accuracy. This subset is called the predictor set. A feature selection technique is made of two components: the predictor set scoring method (which evaluates the goodness of a candidate predictor set); and the search method (which searches the gene subset space for the predictor set based on the scoring method). The technique becomes wrapper-based when classifiers are invoked in the predictor set scoring method. Otherwise, the technique is filter-based, which is the focus of this study.
An important principle behind most filter-based feature selection studies can be summarized by the following statement: A good predictor set should contain features highly correlated to the target class concept, and yet uncorrelated with each other . The predictor set attribute referred to in the first part of this statement, 'relevance', is the backbone of rank-based feature selection techniques. The aspect alluded to in the second part, 'redundancy', refers to pairwise relationships between all pairs of features in the predictor set. The relevance of a predictor set tells us how well the predictor set is able to distinguish among different classes. The redundancy in a predictor set indicates the amount of similarity among the members of the predictor set, or rather, the amount of repetitions in terms of the information conveyed by the members of the predictor set.
Previous studies [1, 2] have based their feature selection techniques on the concept of relevance and redundancy having equal importance in the formation of a good predictor set. We call the predictor set scoring methods used in such correlation-based feature selection techniques equal-priorities scoring methods. On the other hand, it is demonstrated in  using a 2-class problem that seemingly redundant features may improve the discriminant power of the predictor set instead, although it remains to be seen how this scales up to multiclass domains with thousands of features. A study was implemented on the effect of varying the importance of minimizing redundancy in predictor set evaluation in . However, due to its use of a relevance score that is inapplicable to multiclass problems, the study was limited to only binary classification.
Currently, when it comes to the use of filter-based feature selection for multiclass molecular classification, three popular recommendations are: 1) no selection [5, 6]; 2) select based on relevance alone [5, 7]; and finally, 3) select based on relevance and redundancy [2, 8]. Thus, so far, relevance and redundancy are the two existing criteria which have ever been used in predictor set scoring methods for multiclass molecular classification.
To these two criteria we introduce one modification and a new criterion in our previous study :
• Antiredundancy, which is a parameter opposite to redundancy in terms of quality and thus is to be maximized along with relevance. Accordingly, instead of maximizing relevance and minimizing redundancy, we now maximize both relevance and antiredundancy.
• Aside from relevance and antiredundancy/redundancy, there is a third criterion in feature selection which is necessary for the formation of the predictor set. The third criterion is the degree of differential prioritization (DDP), which represents the relative importance placed between relevance and antiredundancy.
DDP compels the search method to prioritize the optimization of one of the two criteria (of relevance or antiredundancy) at the cost of the optimization of the other. In other words, DDP controls the balance between the two requirements in feature selection (maximizing relevance and maximizing antiredundancy). Therefore, unlike other existing correlation-based techniques, the novelty of the DDP-based feature selection technique is that it does not take for granted that the optimizations of both elements of relevance and antiredundancy are to have equal priorities in the search for the predictor set [10, 11].
DDP is represented by a variable α which can take any value from 0 to 1. Decreasing the value of α forces the search method to put more priority on maximizing antiredundancy at the cost of maximizing relevance. Raising the value of α increases the emphasis on maximizing relevance (and at the same time decreases the emphasis on maximizing antiredundancy) during the search for the predictor set [10, 11].
A predictor set found using a larger value of α contains more features with strong relevance to the target class concept, but also more redundancy among these features. Conversely, a predictor set obtained using a smaller value of α contains less redundancy among its member features, but at the same time also has fewer features with strong relevance to the target class concept. At α = 0.5, we get an equal-priorities scoring method. At α = 1, the feature selection technique becomes rank-based. Thus, the beauty of the DDP concept is that it subsumes the two existing concepts in feature selection which are represented by equal-priorities scoring methods and rank-based techniques.
A large body of our work has provided empirical support regarding the efficacy of the DDP concept in feature selection [9–12], including comparisons to other feature selection techniques on highly multiclass microarray datasets in . However, we have yet to establish the fundamental strengths and merits of the DDP-based feature selection technique. This is precisely the aim of this paper, which is to be realized through a simulation study involving vigorous analyses of predictor sets found using the DDP-based feature selection technique and simple but illustrative examples using toy datasets.
To generate toy datasets for this purpose, we employ two models which are well-known and recognized not only in the domains of molecular classification and microarray analysis but also conventional data mining . Later in this paper, we also show how close conditions in real-life multiclass microarray datasets resemble those of our toy datasets. Additional advantages of toy datasets include the unlimited number of datasets we can generate (vs. the limited number of available real-life microarray datasets ); the control we are able to exercise over dataset characteristics such as the number of classes and features; and prior knowledge of the members of the ideal predictor set, which provides the ultimate means for measuring the efficacy of the feature selection technique without involving the inductions of actual classifiers.
The organization of the paper is as follows: Beginning with descriptions of the models used to produce the toy datasets: the OVA (one-vs.-all) and PW (pairwise) models, we proceed to analyze the characteristics of the predictor sets obtained from each of the toy datasets and then summarize the properties of the predictor sets which are dependent on the associated DDP values. After reapplying the same set of analyses to eight real-life multiclass microarray datasets, we demonstrate how the DDP works for datasets with different number of classes. We then follow with further discussion of the results and present the conclusions of the study. Finally, in the Methods section, we describe the DDP-based feature selection technique and the real-life datasets used in this study.
The aim of toy datasets is to provide simple but clear and demonstrative examples on the importance of the correct choice of the value of the DDP in forming the best predictor set. Furthermore, another advantage of toy datasets is the fact that we know exactly just how large a predictor set should be for each case, facilitating the task of determining the value of the maximum size of the predictor set, P.
It is widely accepted that over-expression or under-expression (suppression) of genes causes the difference in phenotype among samples of different classes. The categorization of gene expression is given as follows.
• A gene is over-expressed: if its expression value is above baseline.
• A gene is under-expressed: if its expression value is below baseline.
• Baseline interval: the normal range of expression value.
As one of the data processing steps recommended in , logarithmic transformation are applied on microarray datasets: base 10 log for data derived from oligonucleotide (Affymetrix) platform and base 2 log for data derived from cDNA (two-color) platform. Later, another of the data processing steps, normalization, is conducted. Normalization involves the standardization of the gene expression data by mean-centering so that the samples have mean 0 across genes . The purpose of normalization is to prevent the expression levels in one particular sample from dominating the average expression levels across samples . (This normalization is not to be confused with dye normalization, which is performed in an earlier stage of data processing.)
Since the result of normalization is that the mean expression across all genes in a sample is 0, the 'average' genes in a sample have expression values of or close to 0. As the 'average' genes are associated with the baseline or the normal range of expression, the value 0 denotes the center of the baseline interval. Over-expression is represented by positive values and under-expression by negative values. With this categorization, we next employ two well-known paradigms leading to the OVA and PW models, which are then used to generate two different sets of toy datasets.
One-vs.-all (OVA) model
The crux of the OVA concept has gained wide, albeit tacit, acceptance among researchers involved in gene expression analysis. The fact that particular genes are only over-expressed in tissues of a certain type of cancer, and not any other types of cancer or normal tissues , is part of the domain knowledge. Hence the term 'marker' – for genes that mark the particular cancer associated with them. In the OVA model, certain groups of genes, also called the 'marker genes' are only over-expressed (or under-expressed) in samples belonging to a particular class and never in samples of other classes. This model emphasizes that a group of marker genes is specific to one class. Therefore for a K-class dataset, there are K different groups of marker genes.
Let us denote as G the number of genes in each group of marker genes, Xmax and Xmin the maximum and minimum limits, respectively, to the absolute value of the class means for the whole dataset. Thus, for the g-th gene in a group of marker genes, the maximum limit to the absolute value of the class means is defined as:
xmax,g= Xmax - (ΔX)(g - 1)
The purpose of equations (1), (2), and (3) is to produce the following effect: We would like to vary the class means such that there is an imbalance or inequality in terms of class means among the K classes. The reasons are firstly to mimic a condition prevalent in multiclass microarray datasets (imbalance among classes in terms of class means even after normalization), especially in datasets with large number of classes; and secondly, to present a challenge to the feature selection technique in choosing sufficiently relevant but non-redundant genes. We will provide further elucidation on the second reason later in this section.
Another purpose of the equations is to generate genes with varying relevance in each group of marker genes. Based on equations (1), (2), and (3), the first gene in a group of marker genes (the gene associated with g = 1) has the strongest relevance among the members of that group of marker genes. Accordingly, the gene with the weakest relevance is the last gene in a group of marker genes (the gene associated with g = G). The reason for doing this is also to present a challenge to the feature selection technique in choosing sufficiently relevant but non-redundant genes.
Next, initialize a matrix M: = (μi, k)N × Kof zeros where N is the total number of genes in the dataset, and, in this case, is the product of G and K. This is the matrix of class means, whose element, μi,k, represents the mean of gene i across samples belonging to class k (k = 1, 2, ..., K):
μ(g - 1)K + k, k= (-1 g )[xmax, g- (Δx g )(k - 1)]
The [(g - 1)K + k]-th gene is the g-th member of the k-th group of marker genes and therefore has non-zero class mean for class k and zero class means for all other classes – the archetypal OVA trait. The term (-1 g ) serves to change the sign of the class mean at different values of g so as to produce both over- and under-expressed marker genes.
Standard deviation among samples of the same class, or class standard deviation, is set to 1 for all instances, σi,k= 1 for all k and i. For all k, a total of m samples are generated for class k using Gaussian distribution of mean μi,kand standard deviation σi,kfor gene i.
A 4-class example from the OVA model. μi, krepresents the mean of gene i across samples belonging to class k.
0.5(ΔX - Xmax)
ΔX - Xmax
μ (G - 1)K+1,k
(-1 G )Xmin
μ (G - 1)K+2,k
0.5(-1 G ) Xmin
μ (G - 1)K+3,k
-0.5(-1 G ) Xmin
μ (G - 1)K+4,k
-(-1 G ) Xmin
The ostensible answer would be equal weights, which is the foundation of existing equal-priorities scoring methods. But as mentioned previously in the Background section, it has been implied that antiredundancy is not as important as relevance for the 2-class problem  – this is obvious in case of our OVA toy dataset; any subset of sufficiently relevant genes is capable of differentiating between the two classes. Hence we ask the questions which motivate the concept of the DDP: If at K = 2, antiredundancy is not as important as relevance, will this change as the number of classes increases (an important theme in multiclass classification studies)? As K increases, might not the importance of antiredundancy (w.r.t. relevance) increase as well? If yes, is there a point where antiredundancy eventually overcomes relevance in terms of importance as a criterion in feature selection? These questions are to be answered from the analyses in this study.
Pairwise (PW) model
In the PW model, for a given pair of classes, a group of marker genes only distinguishes samples from one class of the pair of classes against samples from the other class of the pair of classes. As implied by its name, this model represents the 1-vs.-1 paradigm as opposed to the 1-vs.-others of the OVA model. For a K-class dataset, there are different groups of marker genes in the PW model. is the number of unique pairs of classes in a K-class dataset; it is also known as K C2.
As is the case in the OVA model, we denote as G the number of genes in each group of marker genes, Xmax and Xmin the maximum and minimum limits, respectively, to the absolute value of the class means for the whole dataset. The definitions of xmax,g, ΔX, and Δx g are the same as for the OVA model.
for b = 1 and b = 2. For the PW model, the -th gene is the g-th member of the q-th group of marker genes and therefore has non-zero class means for classes c1,qand c2,q, and zero class means for all other classes – which is the typical PW characteristic.
The procedure for the generation of datasets is similar to that of the OVA model.
In this study, for both models, Xmax and Xmin are set to 100 and 1 respectively, while the number of samples per class, or class size, m, is set to 100 uniformly for all classes.
Ten values of α are tested from 0.1 to 1 with equal intervals of 0.1, α denoting the value of the DDP. For both models, the number of genes in each group of marker genes, G, is set to 3, 5, 10, 20, and 30. We test for K = 2 to K = 30, K denoting the number of classes in a dataset. Since no inductions of classifiers are to be implemented in this study, whole datasets are used as training sets during feature selection.
For toy datasets generated from the OVA model, the minimum predictor set size necessary to differentiate among the K classes is K - 1. The optimal predictor set is actually any subset of K - 1 genes from the first K of the marker genes (i.e., at g = 1) generated using the class means defined in equation (4).
where |S| = K - 2, , , and C qi represents the q i -th pair of classes as defined previously in the subsection on the PW model. In other words, the optimal predictor set contains representatives from enough groups of marker genes such that all K classes are represented in pairs of classes associated to those groups of marker genes.
Therefore for datasets generated from the OVA model, P is set to K - 1 and for those from the PW model, P is set to K - 2.
Separation of classes
A natural way to measure separation of classes is the distance between pairs of class centers. We use two popular metrics, the Euclidean and the Manhattan (or taxicab) distances. At the end of the "One-vs.-all (OVA) model" subsection under the Results section, we discuss a preceding study on feature selection  which inspired the DDP concept. The authors of that study employ a form of separation of classes to demonstrate that a redundant feature may enhance the predictor set's ability to distinguish between two classes in a 2-class problem (thus implying that antiredundancy is not as important as relevance for the 2-class problem). This form of separation of classes corresponds to the Manhattan distance used in our study.
In a 2-class problem, the authors of  first present two features from a toy dataset which are both relevant but redundant w.r.t each other, contained in a predictor set distinguishing between the two classes. Then, after a 45° rotation of those two features, the authors of that study show that the Manhattan distance between the class centers along one axis is now greater by a factor of than the corresponding Manhattan distance in the original plane – thus increasing the separation of classes. For a predictor set with two members, the aforementioned 45° rotation is akin to the transformation by principal component analysis, which we will implement later in this study.
We observe that the Euclidean distance remains the same before and after the transformation in that study. Therefore we have included the Euclidean distance as another form of separation of classes to study, if any, the differences between the two distances in the context of the DDP. Moreover, the Euclidean distance is as popularly used as the Manhattan distance in the field of intelligent data analysis.
If there is more than one value of α satisfying equations (9) or (12), the mean among these values is taken as or . Since these values are generally observed to be adjacent to each other, taking the mean will still provide a good picture of how the DDP affects separation of classes.
Principal component analysis (PCA)
PCA linearly transforms the data such that the greatest amount of variance among samples comes to lie along the axis representing the first principal component (PC). Similarly, the second PC contains the second largest variance among samples, and so on. An important property of the PCs is that a PC is always orthogonal to the adjacent PC.
In addition to analyzing the predictor sets in the original projection, we investigate the characteristics of the predictor sets after transformation by PCA. In the original form, the data are characterized along axes representing members of the predictor set (original feature space). After transformation by PCA, data are characterized along axes representing the PCs derived from the members of the predictor set (PCA-transformed space or PC space).
The input data matrix is never mean-centered throughout the transformation procedures – this is to enable comparisons in terms of distance metrics between data in original feature space and data in PC space later in this study. (For instance, in this manner, the Euclidean distance remains constant in both original feature space and PC space.) The sole effect of not mean-centering the dataset is that the first PC will span the variance characterized by the overall distance of the dataset from the origin . In case of our models (OVA and PW), marker genes contain non-zero class mean for each of the classes (OVA model) or non-zero class means for each of the pairs of classes (PW model) that they mark, and zero class means for all other classes. Thus for both models, even without mean-centering, the variance contained by the first PC will still be variance among classes, because for both models, the distance of a data point (a sample) from the origin as measured by each gene is actually characterized by the class of that data point.
The main use of PCA in this case is to rotate the data from the original sets of axes (represented by the members of the predictor set) so that the data are now projected along new sets of axes (represented by the PCs) which are orthogonal and hence minimally correlated to each other. In this study, PCA is conducted only on the members of the predictor set, not on the whole dataset. The reason we apply PCA in this manner is to expand on the finding in  which we discuss in the beginning of the "Separation of classes" subsection. Therefore, each of the PCs in this study contains information only from the predictor set, and never from any gene which is not a member of the predictor set.
Antiredundancy of PCA-transformed predictor sets
Separation of classes in PCA-transformed predictor sets
The Euclidean distance remains the same whether the predictor sets have been transformed by PCA or not; hence we do not repeat the analysis described in the previous subsection (on separation of classes) for PCA-transformed predictor sets. The Manhattan distance, however, is affected by the transformation. The procedures involved in computing the DDP value leading to the best separation of classes are the same as in case of the untransformed predictor sets described in the previous subsection (on separation of classes). To distinguish between the DDP value associated with untransformed predictor sets and the DDP value associated with PCA-transformed predictor sets, we will denote the latter as . Similarly, separation of classes in PC space in terms of the Manhattan distance as measured by a predictor set found using the DDP value of α, S α , is denoted as .
Indeed, the observation regarding provides the link between the study in  and the DDP concept. In almost all cases, a predictor set which is obtained using the DDP value of shows enhanced separation of classes in PC space compared to separation of classes in the original feature space (measured using the Manhattan distance) – a finding which is reflected in that study described earlier in the beginning of the "Separation of classes" subsection. Therefore at the optimal value of the DDP, separation of classes as measured using the Manhattan distance in PC space is maximized because of this enhancement.
Summary of analyses
We have found that as K increases, three parameters decrease in an exponential-like manner:
• the value of the DDP producing the best separation of classes in terms of the Euclidean distance;
• the value of the DDP producing the highest antiredundancy in PC space; and
• the value of the DDP producing the best separation of classes in terms of the Manhattan distance in PC space.
We have shown that regardless of the model type (OVA or PW) or the value set to the model parameter, G (3, 5, 10, 20, or 30), each of these three characteristics can be optimized by choosing the right value of the DDP, and that this value, in turn, is determined by the number of classes in the dataset.
Investigating the imbalance of class means in real-life datasets
Range of class means in real-life datasets.
Reapplying the analyses on real-life datasets
For real-life datasets, the analyses are implemented separately upon the training set of each split, there being a total of 10 splits of training and test sets. The mean across all splits is taken for the characteristics measured in the analyses: , , and , and then used to find the corresponding value of the DDP which optimizes each of the aforementioned characteristics.
We will assume that the optimal P for each real-life dataset is also directly proportional to K (as is the case for toy datasets). However, allowing for remnant noise (left even after data preprocessing), we assign larger values to P for real-life datasets (30K) than for toy datasets with similar K. Furthermore, we conduct two versions of the analyses involving PCA-transformed predictor sets:
• in the first version, only the top three PCs are used
• in the second version, all PCs are used, as is the case for toy datasets.
The reason for this is given as follows: The large percentage of the total variance among samples which is represented by the top PCs is 'relevant' variance (i.e., variance which is due to the difference among classes and thus is relevant w.r.t. the target class concept). On the other hand, the last PCs contain the remainder (small) percentage of the total variance, which is most likely caused by noise or variance within class (i.e., 'non-relevant' variance as opposed to the first type of variance since variance within class is not relevant w.r.t. the target class concept). Variance within class in real-life datasets, unlike variance within class in toy datasets (which is fixed at 1 in this study), differs from class to class even within the same dataset, and is likely to be larger than 1. This is the reason for the first version of the analysis.
In this section, we demonstrate how the DDP works for datasets with different number of classes. Then we discuss the reasons for the difference between the behaviors of and as K increases and the causes of the discrepancies between the plots for toy datasets and real-life datasets.
A look at how the DDP concept works
For the 4-class toy dataset, the values of the DDP which maximize separation of classes in terms of the Euclidean distance are 0.7, 0.8, 0.9, and 1; we can observe from Figure 7 that the predictor set produced from these DDP values do not contain non-relevant genes (genes which are generated at large values of g and hence barely differentiate among the K classes). On the other hand, predictor sets obtained using the DDP values of 0.5 and smaller do comprise such non-relevant genes.
Figure 8 shows that for the 6-class toy dataset, only predictor sets obtained at the DDP values of (0.5, 0.6, or 0.7) are able to differentiate samples among the maximum number of classes (3 classes: class 1, class 5, and class 6). Predictor sets obtained at α = 1 and DDP values less than 0.5 (e.g., α = 0.2) merely manage to distinguish samples between 2 classes at most (class 1 and class 6). The predictor set obtained at α = 1 contains redundant marker genes from both the first and sixth groups of marker genes. The predictor set obtained at α = 0.2 contains no such redundancy but has non-relevant genes among its members.
Note that at smaller values of K, tends to have multiple values; this is because more than one value of α satisfies the requirement for datasets where K < 9.
For the 10-class toy dataset, the predictor set obtained at = 0.4 contains representatives from the largest number of groups of marker genes, 6 (Figure 9). This means that this predictor set is able to distinguish samples among those 6 classes (class 1, class 2, class 3, class 8, class 9, and class 10). The predictor set found from the equal-priorities scoring method (α = 0.5) is only capable of telling apart samples from 5 classes and contains more redundancy than the optimal predictor set. For instance, S0.5 has two redundant marker genes from the first group of marker genes, whereas S0.4 has only one redundant marker gene from that same group of marker genes. The rank-based predictor set (α = 1) comprises redundant marker genes from only 2 groups of marker genes and therefore can only differentiate samples of the associated 2 classes from any other classes.
Figure 10 shows that for the 14-class toy dataset, the predictor set obtained at = 0.3, S0.3 contains genes from 8 groups of marker genes and thus is able to distinguish samples among more classes than predictor sets obtained using any other values of the DDP. The equal-priorities scoring method produces the predictor set S0.5 which is not able to differentiate samples of class 4 and class 11 from all other classes because the members of S0.5 come from only 6 groups of marker genes. S0.5 also contains more redundancy than S0.3. S0.5 has 2 and 3 redundant genes from the first and 14-th groups of marker genes, respectively, while S0.3 has only 1 and 2 redundant genes from the first and 14-th groups of marker genes, respectively. The rank-based predictor set S1 fares worse since it can only tell apart samples from 4 classes due to the high redundancy among its members (i.e., redundant genes from 3 out of the 4 groups of marker genes).
In summary, a predictor set obtained at contains representative genes from more groups of marker genes and thus has lower redundancy compared to predictor sets found using any other values of the DDP. As mentioned previously, the value of changes depending on the value of K. Therefore the equal-priorities scoring method is only capable of producing the optimal predictor set for a certain range of K. Below that range (small number of classes), rank-based selection may work better than equal-priorities scoring method (supporting the implication from the 2-class example of ), while DDP values smaller than or equal to 0.5 will select non-relevant genes. Above that range (large number of classes), the value of the DDP producing the optimal predictor set is always less than 0.5; S0.5 will contain more redundancy and, for a given P, is able to tell apart samples from smaller number of classes than the predictor set found using .
At P equal to K - 1 for OVA-based toy datasets we observe that none of the DDP values (whether , 0.5, or 1) are able to produce predictor sets which contain representatives from at least K - 1 different groups of marker genes. This is definitely achievable with greater P; we also do not expect the findings in this study to change significantly if a greater value of P is used. The predictor set found at will always contain representatives from more different groups of marker genes than predictor sets obtained using any other values of the DDP, and the value of is not necessarily 0.5 (equal-priorities scoring method) or 1 (rank-based selection), but varies according to K. Similar findings are also observed for toy datasets generated from the PW model, but for the sake of brevity, are not included in this paper.
The difference between the behaviors of and as K increases
The difference that we observe between the - K plots (Figure 1) and the - K plots (Figure 2) is rooted in the fundamental difference between the Euclidean and the Manhattan distances. The Euclidean distance is a square root of the sum of the squared differences between class means, as indicated in equation (7). In contrast, the Manhattan distance is simply the sum of the absolute differences between class means, as indicated in equation (10). Because of this, if there are h features that are redundant w.r.t. each other in the predictor set, all h features contribute
• the square root of their corresponding differences between class means to the Euclidean distance
• simply the absolute value of their corresponding differences between class means to the Manhattan distance.
If for the sake of simplicity we assume that all h features contain the same corresponding differences between class means, Δd, then the Euclidean distance is Δd· and the Manhattan distance is Δd·h. Recall that the higher the value of h, the higher the redundancy in the predictor set (and the lower the antiredundancy). Note that redundant features are not necessarily non-relevant. Furthermore, we assume that the other (P - h) members of the predictor set are equally relevant (i.e., contain the same corresponding differences between class means, Δd).
Since the Manhattan distance in the original feature space gives equal weight to the contributions from the bulk of redundant features (represented by h) and from the individual relevance (represented by Δd), it is maximized when relevance is maximized (Δd is maximized) and/or antiredundancy is minimized (h is maximized), regardless of K. In the context of the DDP, there is greater emphasis on maximizing relevance than on maximizing antiredundancy in the range [0.8,1], where majority of the points in the - K plots are located (Figure 2). We can then deduce that the Manhattan distance is a pure embodiment of relevance because it does not matter whether the total relevance results from redundancy or otherwise – any relevant feature, whether redundant w.r.t. the other members of the predictor set or not, will increase separation of classes as measured using the Manhattan distance.
Given that the Manhattan distance, d m , represents relevance and h represents redundancy, then in maximizing separation of classes as measured using the Euclidean distance, d e , we are maximizing relevance while at the same time maximizing antiredundancy. This is the reason that, unlike (Figure 2), does not occur exclusively in the range [0.8,1]. The maximizations of relevance and antiredundancy which happen in the maximization of the Euclidean distance are not always given equal weights either ('equal' as in the sense denoted by α = 0.5), since decreases as K increases and is not constantly located in the range [0.4,0.6] (Figure 1). Also, factors which have been oversimplified by the previous two assumptions (which state that Δd is constant for all members of the predictor set) might contribute to the observation in Figure 1. Indeed, this aspect of the analysis provides us with the scope for future study on the differences between the two distances in the context of the DDP.
Causes of the discrepancies between the plots for toy datasets and real-life datasets
The causes of the discrepancies between the plots for toy datasets and real-life datasets are based on the differences in dataset characteristics between toy datasets and real-life datasets.
• Small class sizes: compared to the class sizes in toy datasets (100 samples per class), some classes in the real-life datasets are comparatively small (e.g., 5 samples in the central nervous system tumor type in the NCI60 dataset).
• Varying class sizes: in toy datasets, the class size is kept fixed for all classes and all datasets (m = 100), whereas class size varies in real-life datasets. For example, the lung cancer class consists of 47 samples, whereas there are only 6 bladder tumor samples in the BRN dataset.
• Heterogeneity of some of the classes and the residual noise (variance within class) which remain even after discarding the fourth and the subsequent PCs: as mentioned previously, variance within class in real-life datasets, unlike the variance within class in toy datasets (which is fixed at 1), differs from class to class even within the same dataset.
Despite the aforementioned discrepancies and their probable causes, the general trends in the - K, - K, and - K plots for real-life datasets in Figure 5 can be said to correspond to the general trends in the corresponding plots for toy datasets (Figures 1, 3, and 4). The overall picture provided by Figure 5 indicates that the effect of K on the values of the DDP which optimize , , and in real-life datasets is the same as the effect in toy datasets.
For each dataset from both collections of toy datasets and real-life microarray datasets, we have shown that there exists a value of the DDP which optimizes several characteristics representing the goodness of the predictor set. In turn, this optimal value of the DDP is influenced by the number of classes in the dataset.
We have also demonstrated, through selected examples from toy datasets and comparisons against both simplistic rank-based selection and state-of-the-art equal-priorities scoring methods, how the DDP concept works for datasets with different number of classes. A predictor set obtained at the optimal value of the DDP contains representative genes from more groups of marker genes than predictor sets found using any other values of the DDP. Thus the predictor set obtained using the optimal value of the DDP contains lower redundancy, and is capable of telling apart samples from more classes than predictor sets found using other, sub-optimal values of the DDP.
These findings have been achieved without turning to empirical experiments involving inductions of classifiers (which have previously proved the usefulness of the DDP for both artificial and real-life datasets), thus establishing the fundamental underpinnings for the DDP concept.
The DDP-based feature selection technique
For gene expression datasets, the terms gene and feature may be used interchangeably. From the total of N genes, the objective is to form the subset of genes, called the predictor set S, which gives the optimal classification accuracy.
where I(.) is an indicator function returning 1 if the condition inside the parentheses is true, otherwise it returns 0. is the average of the expression of gene i across all training samples in T. is the average of the expression of gene i across training samples belonging to class k. xi,jis the expression of gene i in sample j. F(i) is the BSS/WSS (between-groups sum of squares/within-groups sum of squares) ratio first used in  for multiclass molecular classification. It indicates the gene's ability in discriminating among samples belonging to K different classes.
The absolute value of the Pearson product moment correlation coefficient between genes i and j, |R(i,j)|, is used to measure the redundancy of gene i w.r.t. gene j (and vice-versa).
The power factor α ∈ (0, 1] in equation (17) denotes the DDP between maximizing relevance and maximizing antiredundancy. We posit that different datasets will require different values of the DDP between maximizing relevance and maximizing antiredundancy in order to come up with the most efficacious predictor set. Therefore the optimal range of α (leading to the predictor set giving the best accuracy) is dataset-specific.
The linear incremental search [2, 8] is conducted as follows: The first member of S is chosen by selecting the gene with the highest F(i) score. To find the second and the subsequent members of the predictor set, the remaining genes are screened one by one for the gene that gives the maximum WA,S. Since the combination of our predictor set scoring method and this search method does not specify an output as to the final size of the predictor set, P, the value of P will have to be predetermined by the user.
Descriptions of real-life datasets. N is the number of features after preprocessing. K is the number of classes in the dataset.
Training set size:Test set size
The PDL dataset  consists of 6 classes, each class representing a diagnostic group of childhood leukemia. The SRBC dataset  consists of 4 subtypes of small, round, blue cell tumors (SRBCTs). In the 5-class lung dataset , 4 classes are subtypes of lung cancer; the fifth class consists of normal samples. The MLL dataset  contains 3 subtypes of leukemia: ALL, MLL, and AML. The AML/ALL dataset  also contains 3 subtypes of leukemia: AML, B-cell, and T-cell ALL.
Except for the BRN and SRBC datasets (which are only available as preprocessed in their originating studies), datasets are preprocessed and normalized based on the recommended procedures  for Affymetrix and cDNA microarray data. Except for the GCM dataset, for which the ratio of training set size to test set size used in the originating study  is maintained to enable comparison with previous studies, for all datasets we employ the standard 2:1 split ratio.
A glossary of terms used in this manuscript is shown in Table 4
This work is supported by Monash Graduate Scholarship (MGS) and International Postgraduate Research Scholarship (IPRS) awards. The authors thank the reviewers for their valuable comments on the manuscript.
- Hall MA, Smith LA: Practical feature subset selection for machine learning. 1998, 181-191.Google Scholar
- Ding C, Long F, Peng H: Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2005, 27(8): 1226-1238.Google Scholar
- Guyon I, Elisseeff A: An introduction to variable and feature selection. J Machine Learning Research. 2003, 3: 1157-1182. 10.1162/153244303322753616.Google Scholar
- Knijnenburg TA, Reinders MJT, Wessels LFA: The selection of relevant and non-redundant features to improve classification performance of microarray gene expression data: Heijen, NL.2005Google Scholar
- Li T, Zhang C, Ogihara M: A comparative study of feature selection and multiclass classification methods for tissue classification based on gene expression. Bioinformatics. 2004, 20: 2429-2437.PubMedView ArticleGoogle Scholar
- Ramaswamy S, Tamayo P, Rifkin R, Mukherjee S, Yeang CH, Angelo M, Ladd C, Reich M, Latulippe E, Mesirov JP, Poggio T, Gerald W, Loda M, Lander ES, Golub TR: Multi-class cancer diagnosis using tumor gene expression signatures. Proc Natl Acad Sci USA. 2001, 98: 15149-15154.PubMedPubMed CentralView ArticleGoogle Scholar
- Chai H, Domeniconi C: An evaluation of gene selection methods for multi-class microarray data classification. 2004, 3-10.Google Scholar
- Yu L, Liu H: Redundancy based feature selection for microarray data. 2004, 737-742.Google Scholar
- Ooi CH, Chetty M, Gondal I: The role of feature redundancy in tumor classification. 2004, 197-208.Google Scholar
- Ooi CH, Chetty M, Teng SW: Relevance, redundancy and differential prioritization in feature selection for multiclass gene expression data. 2005, 367-378.Google Scholar
- Ooi CH, Chetty M, Teng SW: Differential prioritization between relevance and redundancy in correlation-based feature selection techniques for multiclass gene expression data. BMC Bioinformatics. 2006, 7: 320-PubMedPubMed CentralView ArticleGoogle Scholar
- Ooi CH, Chetty M, Teng SW: Modeling microarray datasets for efficient feature selection. 2005, 115-129.Google Scholar
- Dudoit S, Fridlyand J, Speed T: Comparison of discrimination methods for the classification of tumors using gene expression data. J Am Stat Assoc. 2002, 97: 77-87. 10.1198/016214502753479248.View ArticleGoogle Scholar
- Yang YH, Dudoit S, Luu P, Speed TP: Normalization for cDNA microarray data. Microarrays: Optical Technologies and Informatics. Edited by: Bittner ML, Chen Y, Dorsel AN, Dougherty ER. 2001Google Scholar
- Beavis RC, Colby SM, Goodacre R, Harrington PB, Reilly JP, Sokolow S, Wilkerson CW: Artificial Intelligence and Expert Systems in Mass Spectrometry. Encyclopedia of Analytical Chemistry. 2000, 11558-11597.Google Scholar
- Munagala K, Tibshirani R, Brown P: Cancer characterization and feature set extraction by discriminative margin clustering. BMC Bioinformatics. 2004, 5: 21-PubMedPubMed CentralView ArticleGoogle Scholar
- Park M, Hastie T: Hierarchical classification using shrunken centroids. Department of Statistics, Stanford University Technical Report [http://www-statstanfordedu/~hastie/Papers/hpampdf]. 2005Google Scholar
- Ross DT, Scherf U, Eisen MB, Perou CM, Rees C, Spellman P, Iyer V, Jeffrey SS, Van de Rijn M, Waltham M, Pergamenschikov A, Lee JCF, Lashkari D, Shalon D, Myers TG, Weinstein JN, Botstein D, Brown PO: Systematic variation in gene expression patterns in human cancer cell lines. Nat Genet. 2000, 24: 227-235.PubMedView ArticleGoogle Scholar
- Yeoh EJ, Ross ME, Shurtleff SA, Williams WK, Patel D, Mahfouz R, Behm FG, Raimondi SC, Relling MV, Patel A, Cheng C, Campana D, Wilkins D, Zhou X, Li J, Liu H, Pui CH, Evans WE, Naeve C, Wong L, Downing JR: Classification, subtype discovery, and prediction of outcome in pediatric lymphoblastic leukemia by gene expression profiling. Cancer Cell. 2002, 1(2): 133-143. 10.1016/S1535-6108(02)00032-6.View ArticleGoogle Scholar
- Khan J, Wei JS, Ringner M, Saal LH, Ladanyi M, Westermann F, Berthold F, Schwab M, Antonescu CR, Peterson C, Meltzer PS: Classification and diagnostic prediction of cancers using expression profiling and artificial neural networks. Nat Med. 2001, 7: 673-679.PubMedPubMed CentralView ArticleGoogle Scholar
- Bhattacharjee A, Richards WG, Staunton JE, Li C, Monti S, Vasa P, Ladd C, Beheshti J, Bueno R, Gillette M, Loda M, Weber G, Mark EJ, Lander ES, Wong W, Johnson BE, Golub TR, Sugarbaker DJ, Meyerson M: Classification of human lung carcinomas by mRNA expression profiling reveals distinct adenocarcinoma subclasses. Proc Natl Acad Sci USA. 2001, 98: 13790-13795.PubMedPubMed CentralView ArticleGoogle Scholar
- Armstrong SA, Staunton JE, Silverman LB, Pieters R, den Boer ML, Minden MD, Sallan SE, Lander ES, Golub TR, Korsmeyer SJ: MLL translocations specify a distinct gene expression profile that distinguishes a unique leukemia. Nat Genet. 2002, 30: 41-47.PubMedView ArticleGoogle Scholar
- Golub TR, Slonim DK, Tamayo P, Huard C, Gaasenbeek M, Mesirov JP, Coller H, Loh ML, Downing JR, Caligiuri MA, Bloomfield CD, Lander ES: Molecular classification of cancer: Class discovery and class prediction by gene expression monitoring. Science. 1999, 286: 531-537.PubMedView ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.