Skip to main content
  • Methodology article
  • Open access
  • Published:

Discovering biclusters in gene expression data based on high-dimensional linear geometries

Abstract

Background

In DNA microarray experiments, discovering groups of genes that share similar transcriptional characteristics is instrumental in functional annotation, tissue classification and motif identification. However, in many situations a subset of genes only exhibits consistent pattern over a subset of conditions. Conventional clustering algorithms that deal with the entire row or column in an expression matrix would therefore fail to detect these useful patterns in the data. Recently, biclustering has been proposed to detect a subset of genes exhibiting consistent pattern over a subset of conditions. However, most existing biclustering algorithms are based on searching for sub-matrices within a data matrix by optimizing certain heuristically defined merit functions. Moreover, most of these algorithms can only detect a restricted set of bicluster patterns.

Results

In this paper, we present a novel geometric perspective for the biclustering problem. The biclustering process is interpreted as the detection of linear geometries in a high dimensional data space. Such a new perspective views biclusters with different patterns as hyperplanes in a high dimensional space, and allows us to handle different types of linear patterns simultaneously by matching a specific set of linear geometries. This geometric viewpoint also inspires us to propose a generic bicluster pattern, i.e. the linear coherent model that unifies the seemingly incompatible additive and multiplicative bicluster models. As a particular realization of our framework, we have implemented a Hough transform-based hyperplane detection algorithm. The experimental results on human lymphoma gene expression dataset show that our algorithm can find biologically significant subsets of genes.

Conclusion

We have proposed a novel geometric interpretation of the biclustering problem. We have shown that many common types of bicluster are just different spatial arrangements of hyperplanes in a high dimensional data space. An implementation of the geometric framework using the Fast Hough transform for hyperplane detection can be used to discover biologically significant subsets of genes under subsets of conditions for microarray data analysis.

Background

In DNA microarray experiments, discovering groups of genes that share similar transcriptional characteristics is instrumental in functional annotation, tissue classification and motif identification [1, 2]. In many situations, an interesting cellular process is active only under a subset of conditions, or a single gene may participate in multiple pathways that may or may not be co-active under all conditions [3, 4]. In addition, the data to be analyzed often include many heterogeneous conditions from many experiments. In these instances, it is often unrealistic to require that related genes behave similarly across all measured conditions and conventional clustering algorithms, such as the k-means and hierarchical clustering algorithms [5, 6] and the self-organizing map [7], often cannot produce a satisfactory solution.

When a subset of genes shares similar transcriptional characteristics only across a subset of measures, the conventional algorithm may fail to uncover useful information between them. In Fig. 1a, we see a data matrix clustered using the hierarchical clustering algorithm, where no coherent pattern can be observed by naked eyes. However, Fig. 1b indicates that an interesting pattern actually exists within the data if we rearrange the data appropriately.

Figure 1
figure 1

An illustrative example where conventional clustering fails but biclustering works: (a) A data matrix, which appears random visually even after hierarchical clustering. (b) A hidden pattern embedded in the data would be uncovered if we permute the rows or columns appropriately.

The hidden pattern in Fig. 1b is called a bicluster. One of the criteria to evaluate a biclustering algorithm is what kind of bicluster patterns an algorithm is able to find. In this paper, we address six major classes of numerical biclusters. Fig. 2 shows different patterns that are of interest to us: (a) constant values, (b) constant rows, (c) constant columns, (d) additive coherent values, where each row or column is obtained by adding a constant to another row or column, (e) multiplicative coherent values, where each row or column is obtained by multiplying another row or column by a constant value, and (f) linear coherent values, where each column is obtained by multiplying another column by a constant value and then adding a constant. Among these patterns, the first 5 patterns have been introduced by Madeira and Oliveira [8]. Patterns (a-c) are compatible to (d) or (e), in the sense that an algorithm which can detect additive patterns can also detect constant rows/column since the latter are two special cases of the former, while (d) and (e) are mutually independent. Most existing algorithms are based on either the additive model (d) or the multiplicative model (e). The linear coherent pattern of (f) is a generalization proposed by us and subsumes all patterns in Fig. 2. These patterns can be more easily understood based on our geometric perspective introduced below.

Figure 2
figure 2

Examples of different bicluster patterns: (a) constant values, (b) constant rows, (c) constant columns, (d) additive coherent values, (e) multiplicative coherent values, and (f) linear coherent values.

In this work, we deal with numerical biclusters only. There are also works [9, 10] that focus on biclusters containing symbolic data or the so-called coherent evolution biclusters, where the evolution (i.e., up, down, or no change) of the elements in a numerical data matrix is considered instead of the numerical values themselves. We choose to focus on the numerical data based on the following considerations. First, a numerical biclustering algorithm can be used to analyze symbolic data by assigning appropriate numerical values to the symbols. Second, many gene expression data analysis tasks, such as gene regulation network analysis, require numerical biclustering results.

Previous work on biclustering

Throughout the paper, we use F ∈ ℜN × Mto denote a gene expression data matrix with N genes and M arrays or experiment conditions. In the matrix F, a row F i ∈ ℜ1 × Mrepresents the expression of the gene i in M arrays. For simplicity, we only introduce biclustering algorithm for constant/coherent rows below, the corresponding algorithm for constant/coherent columns is similar and can be easily deduced.

Bicluster of constant values is obviously the simplest type. A bicluster of constant values can be modeled as

F(i, j) = u IJ + ε(i, j),

where u IJ is the typical value of the bicluster and ε(i,j) is a small perturbation. Hartigan [11] split the original matrix into a predetermined set of submatrices, and use the variance to evaluate each submatrix V A R ( I , J ) = ∑ i ∈ I , j ∈ j F ( i , j ) − u I J MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOvayLaemyqaeKaemOuaiLaeiikaGIaemysaKKaeiilaWIaemOsaOKaeiykaKIaeyypa0ZaaabuaeaacqWGgbGrcqGGOaakcqWGPbqAcqGGSaalcqWGQbGAcqGGPaqkcqGHsislcqWG1bqDdaWgaaWcbaGaemysaKKaemOsaOeabeaaaeaacqWGPbqAcqGHiiIZcqWGjbqscqGGSaalcqWGQbGAcqGHiiIZcqWGQbGAaeqaniabggHiLdaaaa@4B65@ to determine whether a bicluster should be accepted.

If the noise is additive, a bicluster of constant rows can be modeled as

>F(i,j) = u IJ + f i + ε(i,j).

where f i is the i-th row offset. The straightforward method to detect a bicluster of constant row is to normalize the rows of the bicluster using the row mean. By doing so, a bicluster of constant row can be transformed into a bicluster of constant values and hence becomes detectable using algorithms for biclusters of constant values. Getz et al. [12] have developed a method based on this consideration and even extended it to detect biclusters of coherent values. However, methods based on data normalization have a dilemma: for a good normalization, we need to estimate the parameter f i for each row of a bicluster. However, for an accurate estimate of f i , we need to know the location of a bicluster, which is exactly the problem we need to solve. The noise ε(i,j) in the data further complicates the estimation of f i . Instead of relying on data normalization, Califano et al. [13] have developed a method to find some small biclusters first with each row satisfying

>max(F(i,j)) - min(F(i,j)) <δ, ∀j ∈ J

and then add additional rows or columns into it to produce a bicluster that is as large as possible. Sheng et al. [14] have assumed that the multinomial distributions for different columns in a bicluster are mutually independent and used the Gibbs sampling for parameter estimation.

A bicluster of additive coherent values with additive noise can be modeled as

F(i, j) = u IJ + f i + g j + ε(i, j).

Cheng and Church [15] are the first who applied biclustering to microarray data analysis. In their method, the mean squared residue 1 | I | | J | ∑ i ∈ I , j ∈ J ε ( i , j ) 2 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqaIXaqmaeaacqGG8baFcqWGjbqscqGG8baFcqGG8baFcqWGkbGscqGG8baFaaGcdaaeqbqaaiabew7aLjabcIcaOiabdMgaPjabcYcaSiabdQgaQjabcMcaPmaaCaaaleqabaGaeGOmaidaaaqaaiabdMgaPjabgIGiolabdMeajjabcYcaSiabdQgaQjabgIGiolabdQeakbqab0GaeyyeIuoaaaa@48A5@ in (2) is minimized. Cho et al. [16] have improved this mean-squared-residue based method by using the variance as the second measure. Lazzeronic and Owen [17] have introduced a plaid model and proposed the general additive model to identify biclusters of constant rows, constant columns and additive coherent values. Prelic et al. [18] have compared many biclustering algorithms using the additive model.

A bicluster of multiplicative coherent values with additive noise can be modeled as

F(i, j) = u IJ × f i × g j + ε(i, j)

Kluger et al. [19] have studied the checkerboard structure of this type of biclusters using a normalization scheme based on the above equation. Tang et al. [20] have developed a method to compute the cosine value of the angle between each normalized row vector and a predefined stable pattern and then measure the similarity between two rows or two columns. Getz et al. [12] have introduced the Couple Two-Way Clustering by repeatedly performing one-way clustering on the rows and columns of the data matrix.

Madeira and Oliveira [8] are the first to classify many existing numerical biclustering algorithms systematically based on the additive and multiplicative bicluster models. It should be pointed out that some symbolic, coherent evolution or numerical biclusters, such as those produced by cMonkey [9], SAMBA [10] and some statistical criteria, cannot be classified as additive or multiplicative patterns directly. For example, in cMonkey, additional information besides the usual gene expression value, such as motif co-occurrence and association network relationships, are taken into account. Moreover, cMonkey attempts to ensure that a greater percentage of genes that are observed in the data set are included in at least one cluster, while reducing redundancy between overlapping biclusters and maximizing the number of conditions that are included in each bicluster. These features cannot be modeled directly using the additive and multiplicative coherent patterns.

Although the classification into additive or multiplicative patterns is not perfect, it is nevertheless applicable to many existing biclustering algorithms, which can all be formulated using the general linear model proposed in this paper. In fact, in most biclustering algorithms that deal with expression values only, the underlying theme is the coherency in expression values within the biclusters. Our general linear model of Fig. 2(f) therefore conveniently captures the zero and first order coherent relationships within a bicluster.

A high-dimensional geometric method for biclustering

As pointed out in [8], existing approaches are often based on searching for sub-matrices within a data matrix by optimizing certain heuristically defined merit functions. Obviously, the form of the merit function depends greatly on the bicluster pattern to be uncovered. In these methods, when the data contain different types of biclusters, multiple merit functions or different data normalizations or transformations are needed. This often results in a high computational complexity and the optimization procedure is NP-hard in general can be easily trapped at a local optimal point.

In this paper, we extend our previous work [21] and present a novel perspective for biclustering problem through a geometric interpretation. Such a new perspective allows us to regard biclusters with different coherent patterns as hyperplanes in a high dimensional space, and facilitates the use of any generic plane finding algorithm for detecting them. The geometric viewpoint of our approach provides a unified framework to handle different types of linear patterns simultaneously by matching a specific set of linear geometries. It also reveals the existence of the general linear model, which can unify the additive and the multiplicative models. As a particular realization of our framework, we implemented a Hough Transform-based hyperplane detection algorithm. The experimental results on human lymphoma gene expression dataset show that our algorithm is highly effective for gene expression data biclustering.

Results

We tested our algorithm using synthetic dataset and human lymphoma dataset. For synthetic dataset, we use a test model proposed in [14], but deal with both additive and multiplicative biclusters. In the Gibbs sampling method [14], only additive biclusters are used. For human lymphoma dataset, we detect biclusters based on additive, multiplicative and general linear models, and investigate whether the detected biclusters are biological meaningful. Our experiments show that the proposed linear coherent model can produce biologically significant groups enriched by the genes in biclusters.

Synthetic dataset

We generated a synthetic dataset containing four overlapping biclusters of constant columns, constant rows, and multiplicative coherent values, and tested the ability of our approach to detect these patterns simultaneously. To test noise resistance of our method, we embedded the biclusters into a noisy background generated by a uniform distribution U(-5, 5). Gaussian noise with variance of 0.3 was used to degrade the biclusters. The dataset has 200 rows by 40 columns, and the embedded biclusters have the following sizes (rows × columns): 40 × 7 for Bicluster 1 of constant row, 25 × 10 for Bicluster 2 of constant column, and 35 × 8 for Bicluster 3 of constant column, 40 × 8 for Bicluster 4 of multiplicative coherent values with the multiplicative coefficients for each row given in Fig. 3g. As shown in the main plot of Fig. 3a, Bicluster 1 overlaps with Bicluster 2 in two columns, and Bicluster 3 overlaps with Bicluster 2 in five rows and three columns. Random row and column permutations are then performed in Fig. 3a to obtain the final test dataset.

Figure 3
figure 3

A synthetic dataset with multiple overlapping biclusters of different patterns and the biclusters extracted using the proposed method. (a) The data matrix before random row and column permutation, (b) bicluster 1 of constant rows, (c) bicluster 2 of constant columns, (d) bicluster 3 of constant columns, (e) bicluster 4 of multiplicative coherent values, (f) the extra bicluster extracted by the proposed method, and (g) the multiplicative coefficients of each row in bicluster 4.

In this experiment, the three biclusters contain additive coherent values, and both the Gibbs sampling method [14] and our algorithm can identify all of them, but with different accuracies. The Gibbs sampling method misses 2 genes in bicluster 2 and 4 genes in bicluster 3, whereas our algorithm detects all genes perfectly (Fig. 3). Interestingly, a new bicluster with 3 conditions and 60 rows was also reported by our method (Fig. 3f). This bicluster is located in the overlap region of biclusters 2 and 3 and comprises of last three columns of bicluster 2 and first three columns of bicluster 3 and all rows of the two biclusters. Although unexpected, this is a reasonable result since the extra bicluster detected is a valid bicluster by itself. In contrast, the Gibbs sampling method fails to detect this extra, but valid bicluster. The detection of this new bicluster further shows the efficacy of our algorithm in handling overlapping biclusters.

Biological Data: Human Lymphoma Dataset

We apply our algorithm to the lymphoma dataset [22]. This dataset is characterized by well defined expression patterns differentiating three types of lymphoma: diffuse large B-cell lymphoma, chronic lymphocytic leukaemia and follicular lymphoma. The dataset consists of expression data from 128 Lymphochip microarrays for 4026 genes in 96 normal and malignant lymphocyte samples. Missing values in the dataset are imputed using POCSimpute [23].

We compare our algorithm with six existing algorithms, i.e., OPSM [24], Bimax [18], Iterative Signature Algorithm, ISA [25], SAMBA [10], Cheng and Church's algorithm, CC [15] and xMotif [26], using the procedure proposed by Prelic et al. [18]. Since most existing numerical biclustering algorithms do not detect biclusters with general linear coherent values, we only compare the performance for the additive model. Similar to the validation method proposed by Tanay et al. [10], we investigate whether the gene groups produced by different algorithms show significant enrichment with respect to a specific Gene Ontology (GO) annotation. We know that biclustering algorithms aim to classify the genes involved in the same Molecular Function or Biological Process into a group, so a better biclustering algorithm can find more or larger groups that show significant enrichment. Specifically, in our experiment, biclusters are evaluated by computing the hyper-geometric functional enrichment score [27] based on the GO Biological Process annotations, and the resulting scores are adjusted for multiple testing using the Westfall and Young procedure [27, 28].

The histogram in Fig. 4 presents the proportion of biclusters produced by each method for which one or more GO categories are overrepresented at different levels of significance. Best results are obtained from OPSM and the proposed algorithm. Our algorithm is very competitive even when we only consider additive biclusters.

Figure 4
figure 4

Proportion of biclusters significantly enriched by a GO Biological Process category for the six selected biclustering methods. The columns are grouped method-wise, and different bars within a group represent the results obtained for five different significance levels α.

Our method is also capable of detecting biclusters with general linear coherent values. Fig. 5a shows one of these biclusters detected in the lymphoma dataset. The linearity amongst the columns in this bicluster is verified using the scatter plots in Fig. 5b and a good fit can be observed. By defining the column of this bicluster as F0, F1, ..., F10, the pattern of this bicluster can be expressed as F0 = 0.57F1 - 0.08 = 0.38F2 - 0.24 = 0.27F3 - 0.15 = 0.36F4 - 0.26 = 0.36F5 - 0.27 = 0.30F6 - 0.25 = 0.37F7 - 0.22 = 0.28F8 - 0.27 = 0.27F9 - 0.28 = 0.22F10 - 0.29. The detailed results from the GOTermFinder at significance level of 5% are provided in Fig. 6. The result from the GO analysis shows that these linear coherent biclusters are indeed biological meaningful.

Figure 5
figure 5

Biclusters detected in the lymphoma dataset. (a) A bicluster of linear coherent values detected by our algorithm (for the full size image, please see Additional File 5), (b) scatter plots showing the linearity amongst the columns in this bicluster.

Figure 6
figure 6

The GO-based evaluation for the bicluster of Fig. 5a using the GOTermFinder. The upper table is from the biological process ontology; the middle table is from the molecular function ontology; and the lower table is from the cellular component ontology.

In Additional File 1, we provide the algorithmic parameters used in the experiment for the lymphoma dataset. In our experiment, more than 600 biclusters are detected. In Additional File 2, we provide a list of all biclusters with 1 showing corresponding genes/arrays covered by the bicluster while 0 is the contrary. In Additional File 3, we selected 6 biclusters as an example for GO annotation. All the biclusters with full data are given in Additional File 4. The full-sized image of Fig. 5a is shown in Additional File 5.

Conclusion

We analyzed the different type of numerical biclusters and proposed a general linear coherent bicluster model that effectively captures the zero and first order coherent relationships within a bicluster. Then, we presented a novel interpretation of the biclustering problem in terms of the geometric distributions of data points in a high dimensional data space. In this perspective, the biclustering problem becomes that of detecting structures of known linear geometries, i.e., hyperplanes, in the high dimensional data space. We have shown that many common types of bicluster are just different spatial arrangements of the hyperplanes in the high dimensional data space. This novel perspective allows us to perform biclustering geometrically using a hyperplane detection algorithm. The experiment results on both synthetic and real gene expression datasets have demonstrated that our algorithm is very effective.

Method

Although the six patterns in Fig. 2 appear to be substantially different from each other, if we treat each measurement (column) as a variable in the 4D space [x, y, z, w] and each object (row) as a point in the 4D space, the six pattern in Figs. 2(a) to 2(f) would correspond to the following six geometric structures respectively: (a) a cluster at a single point with coordinate [x, y, z, w] = [1.2, 1.2, 1.2, 1.2], (b) a cluster defined by the lines x = y = z = w, (c) a cluster at a single point with coordinate [x, y, z, w] = [1.2, 2.0, 1.5, 3.0], (d) a cluster defined by the lines x = y - 1 = z + 1 = w - 2, (e) a cluster defined by the lines x = 0.5y = 2z = 2w/3, and (f) a cluster defined by the lines x = 0.5(y - 0.1) = 2(z - 0.1) = 2(w - 0.2)/3. Each object (row) in a cluster is a point lying on one of these points or lines.

When a pattern is embedded in a larger data matrix with extra measurements, i.e., a bicluster that covers only part of the measurements in the data, the points or lines defined by the bicluster would sweep out a hyperplane in a high dimensional data space. Assume that we have a three-measurement experiment with the measurements denoted by x, y, and z. If a bicluster covers measurements x and z, then there exists a plane where all data points in the bicluster would lie on. The plane is defined by:

β0 + β1x + β3z = 0

where β i , (i = 0, 1, 3) are constants and β2y is omitted since β2 = 0. The coordinates that appeared in Eq. (4) denote the measurements the bicluster covers, and the points on the plane denote the objects or genes in that bicluster. In Fig. 7, an example of such a plane is shown. We select 3 columns from the data matrix of Fig. 1a and form a new data matrix with a 2-column bicluster embedded inside. The new data matrix is then plotted in a 3D space. We can see that there exists an obvious plane, which provides clues about the hidden bicluster in the data. The linear model has been used in the clustering method OSCAR developed by Bondell and Reich [29]. A major difference between OSCAR and our algorithm is that OSCAR carries out clustering or classification in one direction only, while our algorithm performs biclustering or simultaneous clustering in both row and column directions of the data matrix.

Figure 7
figure 7

If we visualize the data in Fig. 1a in a high-dimensional space, the hidden pattern stands out. Due to the difficulties in visualizing data beyond 3D, we only select columns 32, 41 and 45 in Fig. 1a to form a new data matrix with a 2-column bicluster embedded inside. In this figure, there exists an obvious plane, which provides clues about the hidden bicluster in the data.

In general, different bicluster patterns discussed above can be uniquely defined by specific geometric structures (lines, planes or hyperplanes) in a high dimensional data space. In a 3D space, if we denote the three measurements as x, y and z respectively, and assume a bicluster covers x and z only, we can generate 3D geometric views for different patterns as shown in Fig. 8. When the dimension of the data space is more than three, it becomes difficult to visualize the data points, but the geometric structures are still similar. In addition, this geometric perspective provides valuable insight to the property of biclusters. For example, current algorithms often deal with additive and multiplicative models separately. When we analyze the 3D geometries of these two types of biclusters (Figs. 8d and 8e), it is obvious that there are geometries (Fig. 8f) covering both of them. These new kind of geometries denote a new type of biclusters – the linear coherent model, which our method can deal with easily.

Figure 8
figure 8

Different geometries (lines or planes) in the 3D data space for corresponding bicluster patterns. In each table, the shaded columns are covered by a bicluster. (a) A bicluster with constant values: represented by one of the lines that are parallel to the y-axis and lie in the plane x = z (the T-plane), (b) a bicluster with constant rows: represented by the T-plane, (c) a bicluster with constant columns: represented by one of the lines parallel to the y-axis, (d) a bicluster with additive coherent values: represented by one of the planes parallel to the T-plane, (e) a bicluster with multiplicative coherent values: represented by one of the planes that include the y-axis, and (f) a bicluster with linear coherent values: represented by one of the planes that are parallel to the y-axis.

Based on the geometric perspective discussed above, we propose a geometric gene expression biclustering framework that involves the following two steps. First, we detect the hyperplanes that exist in the gene expression data. Then we analyze whether a required pattern exists for the genes that lie in these hyperplanes.

A powerful technique for line detection in noisy 2-D images and for plane detection in noisy 3-D data called the Hough transform (HT) [30] is widely used in pattern recognition. The HT has been extensively studied in image processing and is well known to be robust against noise for line detection in poor quality images. This robustness is especially useful in microarray data analysis since the data are often heavily corrupted by noise. The method has recently been applied successfully to two and three-color microarray data analysis [31, 32]. Interested readers are referred to the survey paper [33] on the properties and general applications of the HT.

However, it may be difficult to use the standard HT for more than 3 dimensions because of the large computational complexity and storage requirement. In this work, we use the Fast Hough transform (FHT) [34] as our plane detection algorithm since it gives considerable speedup and requires less storage than the conventional HT. The FHT has a very simple and efficient high-dimensional extension. Furthermore, the FHT uses a coarse-to-fine mechanism and provides good noise resistibility. In the following, we briefly discuss the basic principles of the FHT.

Plane detection using the fast Hough transform

We use {F0, F1... FM-1} to denote the coordinates of M arrays. For each gene j {j = 1, 2... N}, the expression vector is given as [F0(j), F1(j), ..., FM-1(j)].

In a 2-D space, a line can be described by

y = mx + c,

where (m, c) are two parameters: the slope and the intercept of the line with y axis. However, a problem with the (m, c) parameterization of lines is its inability to describe vertical lines, i.e., m → ∞. Therefore, Eq. (5) is only used for lines with |m| ≤ 1. The second equation that swaps the roles of the x and y axes,

x = m'y + c'

is used for lines with |m| > 1. With |m| ≤ 1 and |m'| < 1, Eq. (2) and (3) describe all lines in a 2-D space without overlap. A similar method can be used to describe hyperplanes in a high dimensional space. In this paper, this parameterization method is used for our hyperplane detection algorithm.

Suppose that among all the observed data [F0(j), F1(j), ..., FM-1(j)], {j = 1, 2, ..., N}, there exists a target hyperplane described by the following equation

F 0 = ∑ i = 1 M − 1 β i F i + β M , MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOray0aaSbaaSqaaiabicdaWaqabaGccqGH9aqpdaaeWbqaaiabek7aInaaBaaaleaacqWGPbqAaeqaaOGaemOray0aaSbaaSqaaiabdMgaPbqabaGccqGHRaWkcqaHYoGydaWgaaWcbaGaemyta0eabeaaaeaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGnbqtcqGHsislcqaIXaqma0GaeyyeIuoakiabcYcaSaaa@427E@
(7)

where {F0, F1, ..., FM-1} are coordinates of points in observed data space and {β1, β2, ..., β M } are M parameters. Define a set Ω with all the indices of the genes that lies on this hyperplane. For each j ∈ Ω, we have F 0 ( j ) = ∑ i = 1 M − 1 β i F i ( j ) + β M MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOray0aaSbaaSqaaiabicdaWaqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGH9aqpdaaeWbqaaiabek7aInaaBaaaleaacqWGPbqAaeqaaOGaemOray0aaSbaaSqaaiabdMgaPbqabaGccqGGOaakcqWGQbGAcqGGPaqkaSqaaiabdMgaPjabg2da9iabigdaXaqaaiabd2eanjabgkHiTiabigdaXaqdcqGHris5aOGaey4kaSIaeqOSdi2aaSbaaSqaaiabd2eanbqabaaaaa@4779@ . The inversion of (7) indicates that all these points on the target surface satisfy

∑ i = 1 M − 1 F i ( j ) β i + β M − F 0 ( j ) = 0 for all  j ∈ Ω . MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeqabeGaaaqaamaaqahabaGaemOray0aaSbaaSqaaiabdMgaPbqabaGccqGGOaakcqWGQbGAcqGGPaqkcqaHYoGydaWgaaWcbaGaemyAaKgabeaakiabgUcaRiabek7aInaaBaaaleaacqWGnbqtaeqaaOGaeyOeI0IaemOray0aaSbaaSqaaiabicdaWaqabaGccqGGOaakcqWGQbGAcqGGPaqkaSqaaiabdMgaPjabg2da9iabigdaXaqaaiabd2eanjabgkHiTiabigdaXaqdcqGHris5aOGaeyypa0JaeGimaadabaGaeeOzayMaee4Ba8MaeeOCaiNaeeiiaaIaeeyyaeMaeeiBaWMaeeiBaWMaeeiiaaIaemOAaOMaeyicI4SaeuyQdCLaeiOla4caaaaa@58C4@
(8)

We find that the parameters {β1, β2, ..., β M } are given by the intersection of many hyperplanes given by Eq. (8).

Suppose that we know the initial ranges of value {β1, β2, ..., β M } are centered at {P1, P2, ..., P M } and with half-length {L1, L2, ..., L M }. We can divide these ranges into very small "array accumulators" so that each array accumulator can determine a unique array of values {β1, β2, ..., β M } within the acceptable tolerance. According to Eq. (8), one feature point in the observed signal space is mapped into many points (e.g., hyperplanes) in the parameter space. An accumulator in the parameter space containing many mapped points (e.g., the intersection of many hyperplanes) reveals the potential feature of interest.

According to above analysis, the FHT-based plane detection method includes three parts. First, we need a hyperplane formulation as in Eq. (8). Second, we divide the parameter space into accumulators that is small enough so that the desired resolution is satisfied. Third, for the accumulators, let every point in the observed data vote for them. If the votes that an accumulator receives is more than a selected threshold, we detect a hyperplane in the observed data space as given by Eq. (7), where the values of {β1, β2, ..., β M } are given by the accumulator. Now we introduce each part of the algorithm in details.

Hyperplane formulation

The FHT does not use Eq. (8) directly. Suppose that we know the initial ranges of values {β1, β2, ..., β M } are centered at {P1, P2, ..., P M } and with half-length {L1, L2, ..., L M }. According to Eq. (8), we have

∑ i = 1 M − 1 F i ( j ) L i W ( j ) L M β i L i + β M W ( j ) L M − F 0 ( j ) W ( j ) L M = 0 for all  j ∈ Ω , MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeqabeGaaaqaamaaqahabaqcfa4aaSaaaeaacqWGgbGrdaWgaaqaaiabdMgaPbqabaGaeiikaGIaemOAaOMaeiykaKIaemitaW0aaSbaaeaacqWGPbqAaeqaaaqaaiabdEfaxjabcIcaOiabdQgaQjabcMcaPiabdYeamnaaBaaabaGaemyta0eabeaaaaWaaSaaaeaacqaHYoGydaWgaaqaaiabdMgaPbqabaaabaGaemitaW0aaSbaaeaacqWGPbqAaeqaaaaakiabgUcaRKqbaoaalaaabaGaeqOSdi2aaSbaaeaacqWGnbqtaeqaaaqaaiabdEfaxjabcIcaOiabdQgaQjabcMcaPiabdYeamnaaBaaabaGaemyta0eabeaaaaGccqGHsisljuaGdaWcaaqaaiabdAeagnaaBaaabaGaeGimaadabeaacqGGOaakcqWGQbGAcqGGPaqkaeaacqWGxbWvcqGGOaakcqWGQbGAcqGGPaqkcqWGmbatdaWgaaqaaiabd2eanbqabaaaaaWcbaGaemyAaKMaeyypa0JaeGymaedabaGaemyta0KaeyOeI0IaeGymaedaniabggHiLdGccqGH9aqpcqaIWaamaeaacqqGMbGzcqqGVbWBcqqGYbGCcqqGGaaicqqGHbqycqqGSbaBcqqGSbaBcqqGGaaicqWGQbGAcqGHiiIZcqqHPoWvcqGGSaalaaaaaa@73A5@
(9)

where W(j) is a weighting scale used to ensure that ∑ i = 1 M a i 2 ( j ) = 1 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaabmaeaacqWGHbqydaqhaaWcbaGaemyAaKgabaGaeGOmaidaaOGaeiikaGIaemOAaOMaeiykaKcaleaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGnbqta0GaeyyeIuoakiabg2da9iabigdaXaaa@3B29@ . Let X i = β i L i MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemiwaG1aaSbaaSqaaiabdMgaPbqabaGccqGH9aqpjuaGdaWcaaqaaiabek7aInaaBaaabaGaemyAaKgabeaaaeaacqWGmbatdaWgaaqaaiabdMgaPbqabaaaaaaa@35FD@ (i = 1, ..., M), F i ( j ) L i W ( j ) L M = a i ( j ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqWGgbGrdaWgaaqaaiabdMgaPbqabaGaeiikaGIaemOAaOMaeiykaKIaemitaW0aaSbaaeaacqWGPbqAaeqaaaqaaiabdEfaxjabcIcaOiabdQgaQjabcMcaPiabdYeamnaaBaaabaGaemyta0eabeaaaaGccqGH9aqpcqWGHbqydaWgaaWcbaGaemyAaKgabeaakiabcIcaOiabdQgaQjabcMcaPaaa@4256@ (i = 1, ..., M-1), a 0 ( j ) = − F 0 ( j ) W ( j ) L M MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemyyae2aaSbaaSqaaiabicdaWaqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGH9aqpcqGHsisljuaGdaWcaaqaaiabdAeagnaaBaaabaGaeGimaadabeaacqGGOaakcqWGQbGAcqGGPaqkaeaacqWGxbWvcqGGOaakcqWGQbGAcqGGPaqkcqWGmbatdaWgaaqaaiabd2eanbqabaaaaaaa@3FC2@ and a M ( j ) = 1 W ( j ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemyyae2aaSbaaSqaaiabd2eanbqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGH9aqpjuaGdaWcaaqaaiabigdaXaqaaiabdEfaxjabcIcaOiabdQgaQjabcMcaPaaaaaa@3862@ , Eq. (9) can be rewritten as

∑ i = 1 M a i ( j ) X i + a 0 ( j ) = 0 for all  j ∈ Ω . MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeqabeGaaaqaamaaqahabaGaemyyae2aaSbaaSqaaiabdMgaPbqabaGccqGGOaakcqWGQbGAcqGGPaqkcqWGybawdaWgaaWcbaGaemyAaKgabeaakiabgUcaRiabdggaHnaaBaaaleaacqaIWaamaeqaaOGaeiikaGIaemOAaOMaeiykaKIaeyypa0JaeGimaadaleaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGnbqta0GaeyyeIuoaaOqaaiabbAgaMjabb+gaVjabbkhaYjabbccaGiabbggaHjabbYgaSjabbYgaSjabbccaGiabdQgaQjabgIGiolabfM6axjabc6caUaaaaaa@5304@
(10)

In fact, it is not necessary for the dimension of the parameter space X to be equal to the dimension of observed signal, M. We use k to replace M for a more general expression

∑ i = 1 k a i ( j ) X i + a 0 ( j ) = 0 for all  j ∈ Ω , MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeqabeGaaaqaamaaqahabaGaemyyae2aaSbaaSqaaiabdMgaPbqabaGccqGGOaakcqWGQbGAcqGGPaqkcqWGybawdaWgaaWcbaGaemyAaKgabeaakiabgUcaRiabdggaHnaaBaaaleaacqaIWaamaeqaaOGaeiikaGIaemOAaOMaeiykaKIaeyypa0JaeGimaadaleaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGRbWAa0GaeyyeIuoaaOqaaiabbAgaMjabb+gaVjabbkhaYjabbccaGiabbggaHjabbYgaSjabbYgaSjabbccaGiabdQgaQjabgIGiolabfM6axjabcYcaSaaaaaa@533C@
(11)

where X i is the i-th dimension of the parameter space. Each a i (j) is a function of observed feature points and is normalized such that ∑ i = 1 k a i 2 ( j ) = 1 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaabmaeaacqWGHbqydaqhaaWcbaGaemyAaKgabaGaeGOmaidaaOGaeiikaGIaemOAaOMaeiykaKIaeyypa0JaeGymaedaleaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGRbWAa0GaeyyeIuoaaaa@3B5B@ . The initial range for each X i is an interval of length 2, with center at P i /L i . All these ranges comprise a hypercube in the parameter space (X1, ...., X k ).

Vote counting scheme

As mentioned before, every point in the observed data votes for supporting accumulators. We know that each accumulator corresponds to a group of range values of (X1, X2, ..., X M ). For each point j in the observed data, if ∑ i = 1 k a i ( j ) X i + a 0 ( j ) = 0 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaabCaeaacqWGHbqydaWgaaWcbaGaemyAaKgabeaakiabcIcaOiabdQgaQjabcMcaPiabdIfaynaaBaaaleaacqWGPbqAaeqaaOGaey4kaSIaemyyae2aaSbaaSqaaiabicdaWaqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGH9aqpcqaIWaamaSqaaiabdMgaPjabg2da9iabigdaXaqaaiabdUgaRbqdcqGHris5aaaa@43D0@ can be satisfied when the values of (X1, X2, ..., X M ) lie in this accumulator, and it will give a vote to this accumulator. An accumulator receiving votes more than a threshold reveals a corresponding hyperplane in the observed data space.

So, to determine whether an accumulator received a vote from a point j in observed signals, we only need to determine whether a hypercube (accumulator) intersect with a particular hyperplane ∑ i = 1 k a i ( j ) X i + a 0 ( j ) = 0 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaabCaeaacqWGHbqydaWgaaWcbaGaemyAaKgabeaakiabcIcaOiabdQgaQjabcMcaPiabdIfaynaaBaaaleaacqWGPbqAaeqaaOGaey4kaSIaemyyae2aaSbaaSqaaiabicdaWaqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGH9aqpcqaIWaamaSqaaiabdMgaPjabg2da9iabigdaXaqaaiabdUgaRbqdcqGHris5aaaa@43D0@ . We can use a simpler conservative test to check whether the hyperplane intersects the hypercube's circumscribing hypersphere. Assume the center of the accumulator is at [C1, ..., C k ] and r is the radius of the hypersphere. We check whether

a 0 ( j ) + ∑ i = 1 k a i ( j ) C i ≤ r MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemyyae2aaSbaaSqaaiabicdaWaqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGHRaWkdaaeWbqaaiabdggaHnaaBaaaleaacqWGPbqAaeqaaOGaeiikaGIaemOAaOMaeiykaKIaem4qam0aaSbaaSqaaiabdMgaPbqabaGccqGHKjYOcqWGYbGCaSqaaiabdMgaPjabg2da9iabigdaXaqaaiabdUgaRbqdcqGHris5aaaa@4522@
(12)

If Eq. (12) is satisfied, gene j will give a vote to the corresponding accumulator.

K-tree representation

For simplicity, we have assumed above that the parameter space was directly divided into very small accumulators. Actually, this is not necessary. The FHT algorithm recursively divides the parameter space into hypercubes from low to high resolutions. It performs the subdivision and the subsequent "vote counting" is done only in hypercubes with votes exceeding a selected threshold. This hierarchical approach leads to a significant reduction in both computational time and storage space compared to the conventional HT.

For the FHT, we represent the parameter space as a nested hierarchy hypercube. We can associate a k-tree with the representation. The root node of the tree corresponds to a hypercube centered at vector C0 with side-length S0. Each node of the tree has 2kchildren arising when that node's hypercube is halved along each of its k dimensions. Each child has a child index, a vector b = [b1, ..., b k ], where each b i is - 1 or 1. The child index is interpreted as follows: if a node at level l of the tree has center C l , then the center of its child node with index [b1, ..., b k ] is

C l + S l + 1 2 + b , MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaacbeGae83qam0aaSbaaSqaaiabdYgaSbqabaGccqGHRaWkjuaGdaWcaaqaaiabdofatnaaBaaabaGaemiBaWMaey4kaSIaeGymaedabeaaaeaacqaIYaGmaaGccqGHRaWkcqWFIbGycqGGSaalaaa@38D9@
(13)

where Sl+1is the side length of the child at level l+1 and Sl+1= S l /2.

Since we use a coarse-to-fine mechanism, for each accumulator at different levels we need to make a test using Eq. (12). For an accumulator of level l, the radius of its circumscribing hypersphere r is equal to k S l / 2 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaOaaaeaacqWGRbWAaSqabaGccqWGtbWudaWgaaWcbaGaemiBaWgabeaakiabc+caViabikdaYaaa@31F7@ . Based on the K-tree structure, an incremental formula can be used to calculate the left part of Eq. (12). If we divide the left part of Eq. (12) by S l , the normalized distance can be computed incrementally for a child node at level l with child index [b1, ..., b k ] as follows,

R 0 ( j ) = a 0 ( j ) S 0 + ∑ i = 1 k a i ( j ) ⋅ C 0 ( i ) S 0 ; MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOuai1aaSbaaSqaaiabicdaWaqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGH9aqpjuaGdaWcaaqaaiabdggaHnaaBaaabaGaeGimaadabeaacqGGOaakcqWGQbGAcqGGPaqkaeaacqWGtbWudaWgaaqaaiabicdaWaqabaaaaOGaey4kaSYaaabCaeaacqWGHbqydaWgaaWcbaGaemyAaKgabeaakiabcIcaOiabdQgaQjabcMcaPiabgwSixNqbaoaalaaabaGaem4qam0aaSbaaeaacqaIWaamaeqaaiabcIcaOiabdMgaPjabcMcaPaqaaiabdofatnaaBaaabaGaeGimaadabeaaaaaaleaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGRbWAa0GaeyyeIuoakiabcUda7aaa@53F0@
(14)
R l ( j ) = 2 R l − 1 ( j ) + 1 2 ∑ i = 1 k a i ( j ) b i . MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOuai1aaSbaaSqaaiabdYgaSbqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGH9aqpcqaIYaGmcqWGsbGudaWgaaWcbaGaemiBaWMaeyOeI0IaeGymaedabeaakiabcIcaOiabdQgaQjabcMcaPiabgUcaRKqbaoaalaaabaGaeGymaedabaGaeGOmaidaaOWaaabCaeaacqWGHbqydaWgaaWcbaGaemyAaKgabeaakiabcIcaOiabdQgaQjabcMcaPiabdkgaInaaBaaaleaacqWGPbqAaeqaaaqaaiabdMgaPjabg2da9iabigdaXaqaaiabdUgaRbqdcqGHris5aOGaeiOla4caaa@4F9E@
(15)

Test of Eq. (12) can now be expressed as: for the gene j and a child node with child index [b1, ..., b k ] at level l, if

| R l ( j ) | ≤ k / 2 , MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaeiiFaWNaemOuai1aaSbaaSqaaiabdYgaSbqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGG8baFcqGHKjYOdaGcaaqaaiabdUgaRbWcbeaakiabc+caViabikdaYiabcYcaSaaa@3AE7@
(16)

gene j will generate a vote for this child node.

According to the above analysis, the FHT is a mapping from an observed data space into a parameter space. Each feature point in the data space generates "votes" for a set of sub-areas (hypercubes) in the parameter space. A sub-area in the parameter space that receives many votes reveals the feature of interest. The FHT algorithm recursively divides the parameter space into hypercubes from low to high resolutions. It performs the subdivision and the subsequent "vote counting" is done only in hypercubes with the number of votes exceeding a selected threshold. A hypercube with acceptable resolution and with votes exceeding a selected threshold indicate a detected hyperplane in the observed data.

The proposed geometric biclustering algorithm and parameter selection

To summarize, when given a set of genes expression data [F0(j), F1(j), ..., FM-1(j)], j = 1, 2,..., N under diverse experimental conditions, our geometric biclustering algorithm can be summarized as follows:

Parameters that need to be predetermined:

  1. (1)

    The minimum votes count "T" as threshold and the desired finest resolution "q".

  2. (2)

    A transformation that maps gene expression data [F0(j), F1(j), ..., FM-1(j)] into a hyperplane in the parameter space represented by ∑ i = 1 k a i ( j ) X i + a 0 ( j ) = 0 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaabCaeaacqWGHbqydaWgaaWcbaGaemyAaKgabeaakiabcIcaOiabdQgaQjabcMcaPiabdIfaynaaBaaaleaacqWGPbqAaeqaaOGaey4kaSIaemyyae2aaSbaaSqaaiabicdaWaqabaGccqGGOaakcqWGQbGAcqGGPaqkcqGH9aqpcqaIWaamaSqaaiabdMgaPjabg2da9iabigdaXaqaaiabdUgaRbqdcqGHris5aaaa@43D0@ for j = 1,2,⋯,N. Based on the transformation, determine the initial bound of each X i and the root hypercube.

Biclustering procedure:

  1. (1)

    Map gene expression data onto the parameter space.

  2. (2)

    Compute the initial normalized distance from the hyperplane to the root node and perform the voting procedure for the root node. For each gene, if Eq. (16) is satisfied, add one to the vote count of the root node. If the vote count for root node is larger than the threshold T and the resolution is coarser than q, subdivide the root node into the K-tree child nodes.

  3. (3)

    Vote for each child node and subdivide them if needed. A similar vote-and-subdivide mechanism is performed for each new node until no new node appears.

  4. (4)

    When there is no node with resolution equal to q and the vote count larger than T, record the node with the finest resolution. This is the most probable solution. When there are several nodes with resolution equal to q and vote counts larger than T, collect the planes associated with these nodes that have the same genes into a bundle.

  5. (5)

    For each bundle of hyperplanes, check the common conditions (variables) and compare the hyperplanes with the models corresponding to different types of biclusters. A bundle of hyperplanes that are not consistent with any patterns in Fig. 2 or the corresponding bicluster covers too few samples will be discarded. If the bundle survives this process, it will be output as a bicluster. Repeat this step until all bundles are processed.

In the procedure above, there are two parameters: minimum vote count "T" and the desired finest resolution "q". The minimum vote count "T" denotes the minimum number of genes in a bicluster. T depends on the experiment objective and may be selected by the user. For example, the minimum may be 4, that is, a bicluster must contain at least 4 genes. The desired finest resolution "q" depends on the variance of noise in the data. For a perfect bicluster (for example, a perfect constant bicluster where all values are equal), "q" can be arbitrarily large, that is, one can use an arbitrarily fine resolution. However, in practical applications, perfect biclusters are rarely found and "q" reflects how much noise (or inconsistency) is permitted in the detected biclusters. If we wish to detect strongly coherent biclusters (i.e., near perfect bicluster with very little noise), q should be set to a large number. Smaller q can be used to detect biclusters that exhibit more inconsistency due to noise. In general, larger q results in biclusters of smaller size.

In many situations, one has no knowledge about the noise in the data. An appropriate range of q can be determined experimentally to return meaningful biclusters. Recall that the FHT uses a coarse-to-fine mechanism. At coarse resolution, there are fewer accumulator cells and the number of hyperplanes detected is small. At finer resolution, there are more accumulator cells. However, in this case the accumulator cells are also smaller and it is more difficult for a feature point to generate a hit. Many accumulators therefore cannot gain enough votes (exceeding the threshold) to ensure the existence of the corresponding hyperplane. So, if q is set too large or too small, fewer hyperplanes will be detected. Hence, the range of q can be chosen to be one that returns a reasonably large number of hyperplanes.

Computational complexity

For FHT, the following theorem from [34] limits the computational complexity. The "thin tree" property resulting from the theorem guarantees that the complexity of the FHT does not go beyond the bound due to the chosen q.

Theorem [34]:

Assume that all M hyperplanes in the parameter space intersect at a single point C and that they are uniformly distributed in orientation. Given a minimum vote threshold T, the number of hypercubes of size q that can receive T or more votes is less than some number K that does not depend on q", where

K = 4 K 1 2 with K 1 = ceil ( 2 2 / tan ( π T 2 M ) ) + 1 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeqabeWaaaqaaiabdUealjabg2da9iabisda0iabdUealnaaDaaaleaacqaIXaqmaeaacqaIYaGmaaaakeaacqqG3bWDcqqGPbqAcqqG0baDcqqGObaAaeaacqWGlbWsdaWgaaWcbaGaeGymaedabeaakiabg2da9iabbogaJjabbwgaLjabbMgaPjabbYgaSnaabmaabaWaaSGbaKqbagaadaWcaaqaamaakaaabaGaeGOmaidabeaaaeaacqaIYaGmaaaakeaacyGG0baDcqGGHbqycqGGUbGBdaqadaqcfayaamaalaaabaGaeqiWdaNaemivaqfabaGaeGOmaiJaemyta0eaaaGccaGLOaGaayzkaaaaaaGaayjkaiaawMcaaiabgUcaRiabigdaXaaaaaa@521B@

The FHT algorithm is highly parallel. As shown, the processing for the hypercubes or accumulators is independent of each other. Furthermore, the intersection test for a hyperplane does not depend on that of other hyperplanes. Actually, in our implementation, some simple multi-processing optimization, such as OpenMP or OpenMPI library, can achieve a high level of speedup.

In the above discussion, we assume that all the possible linear hyperlanes are to be detected using the FHT. In practice, detecting a small portion of hyperplanes is already enough for our biclustering algorithm. For example, in a dataset [F0(j), F1(j), ..., FM-1(j)], {j = 1, 2, ..., N, using Equation F 0 = ∑ i = 1 M − 1 β i F i + β M MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOray0aaSbaaSqaaiabicdaWaqabaGccqGH9aqpdaaeWbqaaiabek7aInaaBaaaleaacqWGPbqAaeqaaOGaemOray0aaSbaaSqaaiabdMgaPbqabaGccqGHRaWkcqaHYoGydaWgaaWcbaGaemyta0eabeaaaeaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGnbqtcqGHsislcqaIXaqma0GaeyyeIuoaaaa@4146@ we can find all the biclusters covering F0. However, using Equation F 0 = ∑ i = 1 M − 1 β i F i + β M MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOray0aaSbaaSqaaiabicdaWaqabaGccqGH9aqpdaaeWbqaaiabek7aInaaBaaaleaacqWGPbqAaeqaaOGaemOray0aaSbaaSqaaiabdMgaPbqabaGccqGHRaWkcqaHYoGydaWgaaWcbaGaemyta0eabeaaaeaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGnbqtcqGHsislcqaIXaqma0GaeyyeIuoaaaa@4146@ with β i = 0 or t and t is a scale, i.e., a constraint that requires all the non-zero gradients to be equal, we can also find all the bicusters covering F0. The second equation can significantly lower the comptational burden1. Another optimization direction is to take advantage of the property of the gene expression data. Since the gene expression data values are distributed in the range of [-5 5], the hyperplanes ∑ i = 1 M 0.3 X i + 20 = 0 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaabCaeaacqaIWaamcqGGUaGlcqaIZaWmcqWGybawdaWgaaWcbaGaemyAaKgabeaakiabgUcaRiabikdaYiabicdaWiabg2da9iabicdaWaWcbaGaemyAaKMaeyypa0JaeGymaedabaGaemyta0eaniabggHiLdaaaa@3CD1@ or 60 X 1 + ∑ i = 2 M 0.3 X i + 0.2 = 0 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaeGOnayJaeGimaaJaemiwaG1aaSbaaSqaaiabigdaXaqabaGccqGHRaWkdaaeWbqaaiabicdaWiabc6caUiabiodaZiabdIfaynaaBaaaleaacqWGPbqAaeqaaOGaey4kaSIaeGimaaJaeiOla4IaeGOmaiJaeyypa0JaeGimaadaleaacqWGPbqAcqGH9aqpcqaIYaGmaeaacqWGnbqta0GaeyyeIuoaaaa@42E0@ do not have any practical significance and can be disregarded. So scanning the dataset to determine the range of hyperplane parameters before biclustering can significantly lower the computational burden.

In certain special cases, we can simplify the problem according to the bicluster model. For example, if we extract biclusters of constant row, we only need to detect all the hyperplanes with ∑ i = 1 M a i X i = 0 MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaabCaeaacqWGHbqydaWgaaWcbaGaemyAaKgabeaakiabdIfaynaaBaaaleaacqWGPbqAaeqaaOGaeyypa0JaeGimaadaleaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGnbqta0GaeyyeIuoaaaa@3A25@ , a i = 0, 1 or -1, and if we extract multiplicative biclusters, we only need to detect those hyperplanes without intercept.

In term of CPU time, our algorithm is computationally intensive in its un-optimized general form. Based on the complexity of the FHT, the computational demands of the proposed biclustering algorithm depends on how many biclusters exist in the dataset. To give an indication of the computational cost, we run the un-optimized algorithm on a small test dataset on a personal computer (Linux OS with 2.0 G Intel Core 2 Duo processor and 1 GB memory) and record the CPU time.

We randomly select 16 conditions in Human Lymphoma Dataset to produce a 4026 × 16 matrix. The CPU time for over 800 biclusters is 1953 seconds (32.55 minutes). We can adjust the parameters to exclude small and noisy biclusters and reduce the computing time. For example, the CPU time reduces to 397 seconds (6.62 minutes) if we discard biclusters with less than 8 conditions.

For larger dataset, we need to run our algorithm on a computer cluster. For the entire 4026 × 96 Human Lymphoma Dataset, we run our algorithm on a computer cluster of 8 nodes with 2 processors each and it takes about 22 hours. Hence, the proposed algorithm is very time-consuming for large datasets if we search through the entire high-dimensional Hough space to obtain the optimal solution and detect all possible additive and multiplicative coherent patterns in the data.

The computing time can be substantially reduced if we allow the solution to the sub-optimal. For example, we can divide 96 conditions into 6 sets with 16 conditions in each set. Then, only 39.7 (6 × 6.62) minutes are needed on Linux computer described above for the biclustering process. The biclusters from the 6 sets can then be combined. Such a strategy has already been used in [18]. We can also consider two conditions at a time and then combine sub-biclusters gradually to form large ones [32].

Note

1 This method is easy to implement by only testing the hyperplane/accumulator with equal non-zero gradients. Assume there are t subranges for each β i . If we do not consider the coarse-to-fine optimization of FHT, the first equation need to process t Maccumulator while the second equation only need to process about t2*2M-1. In the case of t = 5, M = 40, the computational burden of the second scheme is 1.5*10-15 times that of the first scheme.

Abbreviations

GO:

Gene Ontology, 2D: Two dimensional, 3D: Three dimensional, NP: Non-deterministic polynomial time, HT: Hough transform.

References

  1. Rew DA: DNA microarray technology in cancer research. European Journal of Surgical Oncology. 2001, 27 (5): 504-508.

    Article  CAS  PubMed  Google Scholar 

  2. Golub TR, Slonim DK, Tamayo P, Huard C, Gaasenbeek M, Mesirov JP, Coller H, Loh ML, Downing JR, Caligiuri MA: Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science. 1999, 286 (5439): 531-537.

    Article  CAS  PubMed  Google Scholar 

  3. Laub MT, McAdams HH, Feldblyum T, Fraser CM, Shapiro L: Global analysis of the genetic network controlling a bacterial cell cycle. Science. 2000, 290 (5499): 2144-2148.

    Article  CAS  PubMed  Google Scholar 

  4. DeRisi JL, Iyer VR, Brown PO: Exploring the metabolic and genetic control of gene expression on a genomic scale. Science. 1997, 278 (5338): 680-686.

    Article  CAS  PubMed  Google Scholar 

  5. Tavazoie S, Hughes JD, Campbell MJ, Cho RJ, Church GM: Systematic determination of genetic network architecture. Nature genetics. 1999, 22 (3): 281-285.

    Article  CAS  PubMed  Google Scholar 

  6. Eisen MB, Spellman PT, Brown PO, Botstein D: Cluster analysis and display of genome-wide expression patterns. Proceedings of the National Academy of Sciences of the United States of America. 1998, 95 (25): 14863-14868.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  7. Tamayo P, Slonim D, Mesirov J, Zhu Q, Kitareewan S, Dmitrovsky E, Lander ES, Golub TR: Interpreting patterns of gene expression with self-organizing maps: methods and application to hematopoietic differentiation. Proceedings of the National Academy of Sciences of the United States of America. 1999, 96 (6): 2907-2912.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  8. Madeira SC, Oliveira AL: Biclustering algorithms for biological data analysis: a survey. IEEE/ACM Trans Comput Biol Bioinform. 2004, 1 (1): 24-45.

    Article  CAS  PubMed  Google Scholar 

  9. Reiss DJ, Baliga NS, Bonneau R: Integrated biclustering of heterogeneous genome-wide datasets for the inference of global regulatory networks. BMC bioinformatics. 2006, 7: 280-

    Article  PubMed Central  PubMed  Google Scholar 

  10. Tanay A, Sharan R, Shamir R: Discovering statistically significant biclusters in gene expression data. Bioinformatics. 2002, 18 (Suppl 1): S136-144.

    Article  PubMed  Google Scholar 

  11. Hartigan JA: Direct Clustering of a Data Matrix. Journal of the American Statistical Association. 1972, 67 (337): 123-129.

    Article  Google Scholar 

  12. Getz G, Levine E, Domany E: Coupled two-way clustering analysis of gene microarray data. Proceedings of the National Academy of Sciences of the United States of America. 2000, 97 (22): 12079-12084.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  13. Califano A, Stolovitzky G, Tu Y: Analysis of gene expression microarrays for phenotype classification. Proceedings of the International Conference on Intelligent Systems for Molecular Biology. 2000, 75-85.

    Google Scholar 

  14. Sheng Q, Moreau Y, De Moor B: Biclustering microarray data by Gibbs sampling. Bioinformatics. 2003, 19 (suppl_2): ii196-205.

    PubMed  Google Scholar 

  15. Cheng Y, Church GM: Biclustering of Expression Data. Proceedings of the Eighth International Conference on Intelligent Systems for Molecular Biology. 2000, AAAI Press

    Google Scholar 

  16. Cho H, Dhillon I, Guan Y, Sra S: Minimum sum squared residue co-clustering of gene expression data. Proceedings of the Fourth SIAM International Conference on Data Mining. 2004, 114-125.

    Google Scholar 

  17. Lazzeroni L, Owen AB: Plaid models for gene expression data. Statistica Sinica. 2002, 12 (1): 61-86.

    Google Scholar 

  18. Prelic A, Bleuler S, Zimmermann P, Wille A, Buhlmann P, Gruissem W, Hennig L, Thiele L, Zitzler E: A systematic comparison and evaluation of biclustering methods for gene expression data. Bioinformatics. 2006, 22 (9): 1122-1129.

    Article  CAS  PubMed  Google Scholar 

  19. Kluger Y, Basri R, Chang JT, Gerstein M: Spectral biclustering of microarray data: coclustering genes and conditions. Genome Res. 2003, 13 (4): 703-716.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  20. Tang C, Zhang L, Zhang A, Ramanathan M: Interrelated two-way clustering: an unsupervised approach for gene expression data analysis. Proceedings of the IEEE 2nd International Symposium on Bioinformatics and Bioengineering Conference. 2001, 41-48.

    Chapter  Google Scholar 

  21. Gan X, Liew AWC, Yan H: Biclustering gene expression data based on a high dimensional geometric method. Proceedings of the International Conference on Machine Learning and Cybernetics. 2005, 3388-3393.

    Google Scholar 

  22. Alizadeh AA, Eisen MB, Davis RE, Ma C, Lossos IS, Rosenwald A, Boldrick JC, Sabet H, Tran T, Yu X: Distinct types of diffuse large B-cell lymphoma identified by gene expression profiling. Nature. 2000, 403 (6769): 503-511.

    Article  CAS  PubMed  Google Scholar 

  23. Gan X, Liew AWC, Yan H: Microarray missing data imputation based on a set theoretic framework and biological knowledge. Nucleic Acids Res. 2006, 34 (5): 1608-1619.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  24. Ben-Dor A, Chor B, Karp R, Yakhini Z: Discovering local structure in gene expression data: the order-preserving submatrix problem. Proceedings of the Sixth Annual International Conference on Computational Biology. 2002, 49-57.

    Google Scholar 

  25. Ihmels J, Friedlander G, Bergmann S, Sarig O, Ziv Y, Barkai N: Revealing modular organization in the yeast transcriptional network. Nature genetics. 2002, 31 (4): 370-377.

    CAS  PubMed  Google Scholar 

  26. Murali TM, Kasif S: Extracting conserved gene expression motifs from gene expression data. Proceedings of the Pacific Symposium on Biocomputing. 2003, 77-88.

    Google Scholar 

  27. Berriz GF, King OD, Bryant B, Sander C, Roth FP: Characterizing gene sets with FuncAssociate. Bioinformatics. 2003, 19 (18): 2502-2504.

    Article  CAS  PubMed  Google Scholar 

  28. Westfall PH, Young SS: Resampling-based multiple testing: examples and methods for P-value adjustment. 1993, New York, Chichester, Wiley

    Google Scholar 

  29. Bondell HD, Reich BJ: Simultaneous Regression Shrinkage, Variable Selection, and Supervised Clustering of Predictors with OSCAR. Biometrics. 2008, 64 (1): 115-123.

    Article  PubMed Central  PubMed  Google Scholar 

  30. Ballard DH, Brown CM: Computer vision. 1982, Englewood Cliffs, N.J., Prentice-Hall

    Google Scholar 

  31. Zhao H, Yan H: HoughFeature, a novel method for assessing drug effects in three-color cDNA microarray experiments. BMC Bioinformatics. 2007, 8: 256-

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  32. Zhao H, Liew AW, Xie X, Yan H: A new geometric biclustering algorithm based on the Hough transform for analysis of large-scale microarray data. J Theor Biol. 2008, 251 (2): 264-274.

    Article  CAS  PubMed  Google Scholar 

  33. Illingworth J, Kittler J: A survey of the Hough transform. Comput Vision Graph Image Process. 1988, 44 (1): 87-116.

    Article  Google Scholar 

  34. Li H, Lavin MA, Master RJL: Fast Hough transform: A hierarchical approach. Comput Vision Graph Image Process. 1986, 36 (2–3): 139-161.

    Article  Google Scholar 

Download references

Acknowledgements

This work is supported by a grant from the Hong Kong Research Grant Council (project CityU122506). X.Gan is now supported by EPSRC grant EP/D062012/1.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Alan Wee-Chung Liew.

Additional information

Authors' contributions

XG worked on the hyperplane modeling, implementation and experimental analysis when he was a Ph.D. student at City University of Hong Kong. AWCL proposed the geometric perspective for biclustering, problem formulation, and algorithm design. Both XG and AWCL contributed equally to this work and should be considered as joint first author. HY initiated the project and worked on the Hough transform. All authors read and approved the final manuscript.

Electronic supplementary material

12859_2007_2194_MOESM1_ESM.doc

Additional file 1: Information for additive biclusters detection on the Human Lymphoma Dataset. The parameters used in the proposed biclustering algorithm for the Human Lymphoma Dataset are given. (DOC 26 KB)

12859_2007_2194_MOESM2_ESM.txt

Additional file 2: All detected biclusters. A list of all biclusters with 1 showing corresponding genes/arrays covered by the bicluster while 0 is the contrary. (TXT 2 MB)

12859_2007_2194_MOESM3_ESM.doc

Additional file 3: GO annotation of six selected biclusters. The expression heat map and GO annotation table of six biclusters are given here. (DOC 1 MB)

12859_2007_2194_MOESM4_ESM.zip

Additional file 4: All detected biclusters with full data. All the detected biclusters with full data are given here. (ZIP 7 MB)

12859_2007_2194_MOESM5_ESM.pdf

Additional file 5: A bicluster of linear coherent values in the lymphoma dataset. A full size image showing the linear coherent bicluster detected. (PDF 684 KB)

Authors’ original submitted files for images

Rights and permissions

Open Access This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License ( https://creativecommons.org/licenses/by/2.0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Gan, X., Liew, A.WC. & Yan, H. Discovering biclusters in gene expression data based on high-dimensional linear geometries. BMC Bioinformatics 9, 209 (2008). https://doi.org/10.1186/1471-2105-9-209

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1471-2105-9-209

Keywords