Skip to main content
  • Research article
  • Open access
  • Published:

Context based mixture model for cell phase identification in automated fluorescence microscopy

Abstract

Background

Automated identification of cell cycle phases of individual live cells in a large population captured via automated fluorescence microscopy technique is important for cancer drug discovery and cell cycle studies. Time-lapse fluorescence microscopy images provide an important method to study the cell cycle process under different conditions of perturbation. Existing methods are limited in dealing with such time-lapse data sets while manual analysis is not feasible. This paper presents statistical data analysis and statistical pattern recognition to perform this task.

Results

The data is generated from Hela H2B GFP cells imaged during a 2-day period with images acquired 15 minutes apart using an automated time-lapse fluorescence microscopy. The patterns are described with four kinds of features, including twelve general features, Haralick texture features, Zernike moment features, and wavelet features. To generate a new set of features with more discriminate power, the commonly used feature reduction techniques are used, which include Principle Component Analysis (PCA), Linear Discriminant Analysis (LDA), Maximum Margin Criterion (MMC), Stepwise Discriminate Analysis based Feature Selection (SDAFS), and Genetic Algorithm based Feature Selection (GAFS). Then, we propose a Context Based Mixture Model (CBMM) for dealing with the time-series cell sequence information and compare it to other traditional classifiers: Support Vector Machine (SVM), Neural Network (NN), and K-Nearest Neighbor (KNN). Being a standard practice in machine learning, we systematically compare the performance of a number of common feature reduction techniques and classifiers to select an optimal combination of a feature reduction technique and a classifier. A cellular database containing 100 manually labelled subsequence is built for evaluating the performance of the classifiers. The generalization error is estimated using the cross validation technique. The experimental results show that CBMM outperforms all other classifies in identifying prophase and has the best overall performance.

Conclusion

The application of feature reduction techniques can improve the prediction accuracy significantly. CBMM can effectively utilize the contextual information and has the best overall performance when combined with any of the previously mentioned feature reduction techniques.

Background

Quantitating the changes in cell cycle timing before and after drug treatment is useful for effective drug discovery research. Knowledge of the cell cycle progression, e.g., interphase, prophase, metaphase, and anaphase, is important to improving our understanding of the effects of various drugs on cancer cells [14]. Cell cycle progress can be identified by measuring changes in the nucleus as a function of time. Automated time-lapse fluorescence microscopy imaging provides an effective method to observe and study nuclei dynamically and is an important quantitative technique in the fields of cell biology and systems biology [24]. Nevertheless, the vast amount and complexity of image data acquired from automated microscopy renders manual analysis unreasonably time-consuming. Accurate automatic classification of cell nuclei into interphase, prophase, metaphase, or anaphase, is an unresolved issue in cell biology studies using fluorescence microscopy. Murphy et al. [59] have proposed different feature extraction, feature reduction, and classification algorithms for a similar problem of classification of subcellular location patterns in fluorescence microscope images. Methods have also been proposed to identify the cell cycle phase recognition. Gallardo et al. [10] used Hidden Markov Models (HMMs) to classify the feature vector sequences that are extracted from the segmented, potential mitotic cells. Chen et al. [11] proposed an automated system to segment, classify, and track individuals in live cell population, in which the KNN classifier with a set of seven features was used. A novel hybrid fragments merging method based on watershed segmentation and HMMs is also proposed for cell phase identification [1, 2].

In this work, an automated analytical system [1, 2] is used to acquire images, track cell nuclei and generate features of each cell nucleus in a population of thousands of cells. In these specific time-lapse fluorescence microscopy images, nuclei are bright objects protruding out from a relatively uniform dark background; see an example in Figure 1. The cell nuclei are segmented from the acquired images and represented by a group of features for phase identification. To extract features from these time-lapse fluorescence images, four operating steps are conducted [1, 2]: image preprocessing, cell nuclei segmentation, fragment merging, and cell nuclei tracking. After that, 145 features are extracted from each cell nucleus. But there are many noisy and functionally redundant features. Thus it is necessary to remove the noisy, irrelevant, and redundant features with feature reduction techniques. Many feature reduction methods have been proposed to improve the efficiency and effectiveness of cell phase identification [1, 6, 7, 12, 13].

Figure 1
figure 1

A gray level image of a population of cells showing only nuclei channel. The image shows the nuclei after image enhancements.

Feature reduction techniques can be generally classified into feature extraction and feature selection approaches [13, 1519]. Commonly used feature extraction algorithms, such as PCA [16, 17], Linear Discriminant Analysis (LDA) [17, 18], and Maximum Margin Criterion (MMC) [19] are investigated in this work. In PCA, the linear projections of the greatest variance from the top eigenvectors of the covariance matrix are computed, which works well when the data lies close to a flat manifold. LDA is one of the most commonly used supervised feature extraction algorithms. It is used to locate a lower dimensional space that best discriminates the samples from different classes. LDA explicitly utilizes the label information of the samples and thus is suitable for classification problems. However, it often suffers from small sample size when dealing with the high dimensional image data. Moreover, while LDA is guaranteed to find the best directions when each class has a Gaussian density with a common covariance matrix, it may fail if the class densities are more generalized. To solve the limitations of LDA, MCC, a supervised approach, has recently been proposed. The computation complexity of MMC is lower than LDA. In addition, the SDAFS is proposed as the best approach in feature selection for cell phase identification [7]. Other approaches, such as MIFS [15, 20], GAFS [21], and T-test based Feature Selection (TFS) [22], are also evaluated. It is a NP-hard problem to determine the optimal feature subset for MIFS by global search. So we adopt a greedy searching algorithm, in which the features with the highest average mutual information are selected. Genetic algorithm [17, 21] is a classical random optimization method, which mimics the evolutionary process of survival of the fittest. A T-test is used to generate the initial individual feature subset, which belongs to the "Population" in the GA algorithm. To classify cell nuclei, some researchers have used KNN [11, 17], BPNN [8], and SVM [23, 24] to classify cell nuclei. Though acceptable results have been reported, both of them ignored the contextual information of time-lapse microscopy. To utilize the contextual information, we propose a Context Based Mixture Model. Finally, as a standard practice in machine learning, a systematic comparison is conducted to select an optimal combination of a feature reduction technique and a classifier.

Feature reduction techniques can effectively improve the prediction accuracy, and more features do not necessarily guarantee better performance. Our finding indicates that CBMM outperforms SVM, BPNN, and KNN in identifying prophase and achieves the best overall performance.

Results

Key Steps

The experiments consist of the following six steps.

Step 1. Image processing: Include image pre-processing, thresholding, fragment merging, and tracking. The cell nuclei are segmented from the background and tracked as cell sequences, refer to [1, 2, 2527] for a detailed description.

Step 2. Feature generation: Generate 145 features for each cell nucleus in all tracked sequences.

Step 3. Data labeling and splitting: Label 100 cell sequences manually as interphase, prophase, metaphase, and anaphase.

Step 4. Feature reduction: Use the six different approaches introduced in this paper to reduce the dimension of vector space.

Step 5. Classifier training: Train the classifiers using reduced training data, the four classifiers are trained using the reduced training data obtained from step 4.

Step 6. Phase identification: Identify cell cycle phases of the cells from the reduced testing data.

Materials

The data is generated from HeLa H2B GFP cells imaged during a 2-day period with images acquired 15 minutes apart using an automated time-lapse fluorescence microscopy. H2B GFP is a recombinant protein that localizes to DNA and is fluorescent. Each image has a resolution of 672*512 pixels. To get more reliable training data from each tracked cell sequence, we select the frames where the cell is in mitosis, thus including interphase, prophase, metaphase, and anaphase. In addition, the subsequence is supposed to start from a cell in interphase and end with a cell in anaphase. There are totally 100 manually labeled subsequences. A typical 200-frames sample of digital microscope images contains at least 18,000 interphase cells and the other types of cells sum up to less than 1,000. Obviously, the data sets are critically imbalanced. To handle this problem, we down-sample the interphase cells greatly while keeping other three classes of cells. But there are still serious problems with unequal distribution of the training examples, e.g., there are only 100 prophase cells and 306 anaphase cells.

Image Processing and Feature Generation

During cell phase identification, it is critical to separate nuclei from the background. Nuclei are bright objects protruding out from a relatively uniform dark background. Digital images usually require pre-processing to remove noise, discard undesirable features, and correct illumination artifacts. In a sequence of cell images, our pre-processing procedure includes four steps: image enhancement, adaptive shareholding, morphological filtering, and distance transformation [14, 11]. Although the adaptive threshold can segment all the cells from the background effectively, it cannot separate touching nuclei clusters. To solve this problem, our system utilizes a watershed algorithm. Traditional watershed segmentation, however, will lead to over-segmentation. Thus, a hybrid fragment merging approach that combines the roughness score and Probability Distribution Function (PDF) score of each cell is used [14, 2527]. This algorithm can effectively segment separated nuclei and most of the touching ones. The dynamic behaviors of cell nuclei are tracked by distance and size. After tracking, the performance of fragment merging is improved by the contextual information. The revised segmentation results are then used to reinforce the tracking performance.

After obtaining the segmented nuclei, feature vectors that each contains 145 features are generated to represent the cells. They compose of twelve general image features about shape, size, and intensity (max intensity, min intensity, deviation of gray level, average intensity, length of long axis, length of short axis, long axis/short axis, area, perimeter) [11]; 14 Haralick co-occurrence textural features [6, 28]; 49 Zernike moment features [6, 29], and 70 features generated by Gabor transformation [2, 30].

Parameters

To determine the dimensionality of the lower space for feature extraction algorithms, we vary the ratio of the energy preserved by feature extraction algorithms from 80% to 95% and compare the performance of different classifiers. Our results show that when the reduced dimensionality is 15, almost all the classifiers reach their best performance. Therefore, the reduced dimensionality for feature extraction algorithms is estimated as 15, with which 90% of the energy is preserved by PCA. With the FS algorithms, 10, 20, 30, 40, and 50 features are selected to compare the performance. We reduce the dimensions to 20 for all FS approaches. Both linear kernel and RBF kernel are used for the SVM classifier. The genetic algorithm is used in feature selection, and the parameters are as follows: populations size of 200, maximum generation size of 200, the portion of crossover is set to 0.5-th of the feature length, and the mutation rate is 0.3. One of the 200 populations is initialized with t-test feature selection method. The best performance of KNN is achieved by selecting K = 7 for cell phase identification. A BPNN [8] with a single hidden layer of 20 nodes is used to classify the four classes of cell phases, and is trained with back propagation algorithm. The best performance of SVM is always achieved by linear kernel except for the case without feature reduction. Only linear FE approaches are used in this paper due to their efficiency in contrast to nonlinear ones [31]. Thus, as indicated in Table 1 and 2, the best performance of SVM using linear kernel and RBF kernel is reported.

Table 1 The precision of the combinations of various classifiers and feature reduction algorithms.
Table 2 The sensitivity of the combinations of various classifiers and feature reduction algorithms.

Measurements

We use precision and sensitivity as the measurements for our experimental results. Suppose TP, FP, and FN stand for the number of true positive, false positive and false negative samples respectively after the completion of cell phase identification. Precision is defined as precision = TP/(TP+FP), and sensitivity is defined as sensitivity = TP/(TP+FN). In other words, precision is the portion of cells identified positively that are really positive. Sensitivity refers to the ability to identify positive cells correctly. We can calculate the precision and sensitivity for each class if we treat one class as positive and other classes as negative. The average precision and sensitivity of the four classes is used to indicate the overall performance. The ten-fold cross validation is used for testing the trained classifiers. To address the statistical significance of differences in classification result, the confidence intervals of classification accuracies are estimated to be at the 90% confidence level.

Testing Results

We first reduce the data to 15 dimensional vector spaces with different FE approaches and 20 for all FS approaches. Table 1 and 2 shows the results of all four classes (class1-interphase, class2-prophase, class3-metaphase, and class4-anaphase). We describe the details of all four classes instead of only their average, since prophase and metaphase identifications are very important for drug discovery application [14, 11].

According to Table 1 and 2, it can be observed that CBMM has the best overall performance regardless of which feature reduction algorithm is coupled with. PCA combined with the CBMM classifier outperforms other combinations. Although LDA achieves similar performance, its use is not recommended due to the singularity problem [18].

Since prophase is more important than the other three classes for drug screening, we show the specific measurement (precision and sensitivity) of prophase besides the average measurements of the four classes in Figure 2 and 3. CBMM outperforms SVM, BPNN, and KNN (Figure 2 and 3, (B), (D)) in classifying prophase. On the other hand, SDAFS has been reported to be the best among all the FS approaches [7], its performance when combined with CBMM is shown in Table 3 and 4. From Table 3 and 4, it can be observed that the optimal dimension for CBMM is 20 and its performance cannot be enhanced significantly by the increasing of the subspace dimension.

Figure 4
figure 4

Changes in the appearance of a nucleus during cell mitosis. From (a) to (h) consecutive image subframes form a sequence showing nuclei size and shape changes during cell mitosis.

Figure 5
figure 5

An example of Continuous Gaussian Mixture Hidden Markov Model. M = 4, R = 2, the prior probability of phases are π i , i = 1,2,3,4 which are ignored in this picture.

Table 3 The precision of SDAFS combined with CBMM when compared over a range of subspace dimensions.
Table 4 The sensitivity of SDAFS combined with CBMM when compared over a range of subspace dimensions.

These conclusions are drawn based on the preliminary analysis, which can serve as a guideline for future research. With the accumulation of new data, a more detailed and conclusive analysis will be presented in our future work.

Implementation

The functions for PCA, LDA, MMC, SDAFS, and t-test are developed in Matlab 6.5. MIFS is adapted from previously reported source code [1]. The feature generation tools are adapted from [2, 11] and [30], whereas the twelve general features are generated by Matlab. We use Libsvm [23] as the SVM classifier and implement the CBMM, BPNN, and KNN classifier in Matlab 6.5.

Discussion

Our method can successively classify over 80% of the cell phases. Although such precision is acceptable for most biological applications, additional heuristic rules and online-training will improve the precision further. Since the cell cycle is confined by biological constraints, knowledge-driven heuristic rules can be applied to compensate for certain phase identification errors. For example, we are going to implement the following three biological rules to enhance the system performance:

  • Phase progression rule: Once a cell enters a defined cell-cycle phase, it cannot go back to its previous phase.

  • Phase timing rule: The time period that a cell stays in a phase also obey certain biological rules. Cells will usually stay in prophase no more than 45 minutes, metaphase for about 1 hour in untreated cell sequences, and anaphase under 1 hour. On the other hand, a cell can stay in interphase for more than 20 hours. In time-lapse sequences in a drug-treated cell population, certain cells can stay in metaphase for an even longer period of time.

  • Phase continuation rule: Cells cannot skip the one cell cycle phase and enter next phase following the one it skipped, e.g., cells cannot jump from metaphase to interphase or from anaphase to metaphase.

It is worth noting that the problems with unequal distribution of training examples can be solved in a supervised framework while the unsupervised approach heavily depends on the distribution of training examples. For example, we may first down-sample the training sets with appropriate sampling method, then train the SVMs by assigning the training samples with different cost weights according to class size [23, 24, 32]. In the "weighted" SVM [23], the prediction accuracy of prophase can be increased by 10%~20% at the expense of slightly decrease of the classes with large samples using the reduced features. The weights for interphase, prophase, metaphase, and anaphase are 1, 10, 10, and 10 respectively.

Conclusion

This paper proposes a new Context-Based Mixture Model for dealing with the time-series cell cycle sequence information, which outperforms other traditional classifiers in identifying prophase. The application of feature reduction techniques can effectively improve the prediction accuracy, whereas more features do not necessarily guarantee better performance.

Methods

Feature Reduction

From the Feature Reduction (FR) perspective, the traditional and the state-of-the-art dimensionality reduction methods can be generally classified into Feature Extraction (FE) and Feature Selection (FS) [1521]. FE aims to project high dimensional data to a lower dimensional space by algebraic transformation according to certain criteria while FS identifies a subset of the most representative features according to pre-defined criteria, thus the features are ranked according to their individual predictive power. We compare certain commonly used feature reduction approaches below.

Series of cell divisions can be represented in lineage trees. For this study, we only use one of the daughter cells after each division. This results in only one cell being tracked in any given frame, thus a sequence of cell nuclei are tracked for all of the frames in the current experiment. In other words, a sequence of these cell nuclei is mathematically represented by a n × d matrix X Rn × d, where d is the number of features and n is the length of a cell sequence. Each cell nuclei is represented with a feature vector. XTis used to denote the transpose of matrix X. M sequences of cell nuclei (or M cell sequences) are denoted by a mn × d matrix X ˜ MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaaieqacuWFybawgaacaaaa@2DFA@ Rmn × d. Each cell is denoted by a row vector x i , i = 1, 2,, mn. Assume that these feature vectors belong to c different classes and the sample number of the jthclass is n j , we use c j to represent class j, j = 1, 2,, c. The mean vector of c j is m j = 1 n j x i c j x i MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaacqWGTbqBdaWgaaWcbaGaemOAaOgabeaakiabg2da9maaliaabaGaeGymaedabaGaemOBa42aaSbaaSqaaiabdQgaQbqabaaaaOWaaabuaeaaieqacqWF4baEdaWgaaWcbaGaemyAaKgabeaaaeaacqWG4baEdaWgaaadbaGaemyAaKgabeaaliabgIGiolabdogaJnaaBaaameaacqWGQbGAaeqaaaWcbeqdcqGHris5aaaa@4134@ . The mean vector of all the cell nuclei is m a l l = 1 m n i = 1 m n x i MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaacqWGTbqBdaWgaaWcbaGaemyyaeMaemiBaWMaemiBaWgabeaakiabg2da9maaliaabaGaeGymaedabaGaemyBa0MaemOBa4gaamaaqadabaacbeGae8hEaG3aaSbaaSqaaiabdMgaPbqabaaabaGaemyAaKMaeyypa0JaeGymaedabaGaemyBa0MaemOBa4ganiabggHiLdaaaa@4238@ . The feature reduction problem can be framed as the problem of finding a function f : RdRpaccording to an objective function J, where p is the dimension of data after the dimensionality reduction, so that an object x i Rdis transformed into y i = f(x i ) Rp.

Classifiers

We select the established classifiers KNN [17], BPNN [8, 17], and SVM [23, 24] for comparison. In addition, the CBMM is also proposed to incorporate the contextual information.

Context based Mixture Model Classifier

Figure 4 provides an example of cell mitosis process. The occurrence of phases in a sequence can be regarded as a stochastic process; hence, the cell sequence can be represented as a Markov chain where phases are in hidden states. The occurrence of the first phase in the sequence is characterized by the initial probability of the Markov chain. The occurrence of the other phases, which is given by the occurrence of its previous phase, is characterized by the transition probability. We calculate initial and transition probabilities for Markov chains with a set of training nuclei sequences. In addition, we assume each hidden state can generate a group of continuous visible states described by R Gaussian Mixtures. We optimize these Gaussian mixtures with the Expectation-Maximization (EM) algorithm. Those initial probabilities and the optimized Gaussian mixtures are regarded as a continuous Hidden Markov Model (HMM) [1, 10] for the training sequences.

Figure 2
figure 2

Charts of average precision and sensitivity by CBMM and SVM. Charts (A), (C) show the average precision and sensitivity of different classifiers; while (B), (D) show the precision and sensitivity of the prophase cell identified by different classifiers.

Mathematically, suppose a set of N training sequences (χ1, χ2,χ N ) is given and the phases of all the cell nuclei in these sequences are known. Each sequence is a cell nucleus in T l different frames χ l = { x 1 l , x 2 l , ... x T l l } MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaaiiGacqWFhpWydaWgaaWcbaGaemiBaWgabeaakiabg2da9iabcUha7Hqabiab+Hha4naaDaaaleaacqaIXaqmaeaacqWGSbaBaaGccqGGSaalcqGF4baEdaqhaaWcbaGaeGOmaidabaGaemiBaWgaaOGaeiilaWIaeiOla4IaeiOla4IaeiOla4Iae4hEaG3aa0baaSqaaiabdsfaunaaBaaameaacqWGSbaBaeqaaaWcbaGaemiBaWgaaOGaeiyFa0haaa@464D@ , where each cell nucleus is denoted by a p dimensional vector x t l = { x t 1 l , x t 2 l , ... x t p l } MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaaieqacqWF4baEdaqhaaWcbaGaemiDaqhabaGaemiBaWgaaOGaeyypa0Jaei4EaSNaemiEaG3aa0baaSqaaiabdsha0jabigdaXaqaaiabdYgaSbaakiabcYcaSiabdIha4naaDaaaleaacqWG0baDcqaIYaGmaeaacqWGSbaBaaGccqGGSaalcqGGUaGlcqGGUaGlcqGGUaGlcqWG4baEdaqhaaWcbaGaemiDaqNaemiCaahabaGaemiBaWgaaOGaeiyFa0haaa@4A77@ , t = 1,2,..., T. On the other hand, let S = {s1, s2,, s M } be the set of M hidden states in Markov chain, we also consider the N training sequences as a group of T l T l length random variables, Θ l = { θ 1 l , θ 2 l , , θ T l l } MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaacqqHyoqudaWgaaWcbaGaemiBaWgabeaakiabg2da9iabcUha7HGaciab=H7aXnaaDaaaleaacqaIXaqmaeaacqWGSbaBaaGccqGGSaalcqWF4oqCdaqhaaWcbaGaeGOmaidabaGaemiBaWgaaOGaeiilaWIaeS47IWKaeiilaWIae8hUde3aa0baaSqaaiabdsfaunaaBaaameaacqWGSbaBaeqaaaWcbaGaemiBaWgaaOGaeiyFa0haaa@46E1@ . The Sample Space of these variables is S. By applying the Markov assumption (the state of an object at time t only depends on the state of it at time t-the conditional probabilities is: Pr ( θ t l = s j | θ t 1 l = s i , θ t 2 l = s i 2 θ 1 l = s i 1 ) = Pr ( θ t l = s j | θ t 1 l = s i ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaacyGGqbaucqGGYbGCcqGGOaakiiGacqWF4oqCdaqhaaWcbaGaemiDaqhabaGaemiBaWgaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdQgaQbqabaGccqGG8baFcqWF4oqCdaqhaaWcbaGaemiDaqNaeyOeI0IaeGymaedabaGaemiBaWgaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdMgaPbqabaGccqGGSaalcqWF4oqCdaqhaaWcbaGaemiDaqNaeyOeI0IaeGOmaidabaGaemiBaWgaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdMgaPnaaBaaameaacqaIYaGmaeqaaaWcbeaakiabl+Uimjab=H7aXnaaDaaaleaacqaIXaqmaeaacqWGSbaBaaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemyAaK2aaSbaaWqaaiabigdaXaqabaaaleqaaOGaeiykaKIaeyypa0JagiiuaaLaeiOCaiNaeiikaGIae8hUde3aa0baaSqaaiabdsha0bqaaiabdYgaSbaakiabg2da9iabdohaZnaaBaaaleaacqWGQbGAaeqaaOGaeiiFaWNae8hUde3aa0baaSqaaiabdsha0jabgkHiTiabigdaXaqaaiabdYgaSbaakiabg2da9iabdohaZnaaBaaaleaacqWGPbqAaeqaaOGaeiykaKcaaa@780A@ , i, j = 1,2,M. The trained model is represented by a group of parameters Λ = {∏, A, μ kr , Σ kr , c kr , k = 1,2,M, r = 1,2,R}, where Π stands for the initial probability of each phase and A = {a ij } stands for the transition probability of Markov model.

For the Continuous Hidden Markov Model (CHMM), we assume each hidden state can generate R visible Gaussian mixtures ϕ(x, μ kr , Σ kr ), k = 1, 2,, M, r = 1,2,R, where u kr and Σ kr are means and covariance matrices of Gaussian mixtures respectively. In addition, we have a group of coefficients, c kr , to weight the Gaussian mixtures of each hidden state. Figure 5 provides an example of CHMM. The Gaussian mixtures of each phase can be initialized by Fuzzy K-means [17, 19]. Eventually u kr and Σ kr are initialized based on the results of Fuzzy K-means. After initialization, the parameters u kr and Σ kr , c kr , k = 1,2,M, r = 1,2,R are optimized by EM algorithm iteratively.

Figure 3
figure 3

Charts of average precision and sensitivity by KNN and BPNN. Charts (A), (C) show the average precision and sensitivity of different classifiers; while (B), (D) show the precision and sensitivity of the prophase cell identified by different classifiers.

The next issue is how to use this model to predict cell phases. According to traditional Continuous Gaussian Mixture HMM, the probability of a cell x t belonging to phase s m , i.e. θ t = s m , should only depend basically on x t and xt-1. Thus the probabilities we need for the categorization of cells could be denoted as p(θ t = s m |x t , xt-1). Based on the Bayesian formula, we rewrite them as:

p ( θ t = s m | x t , x t 1 ) = i = 1 M p ( θ t = s m , θ t 1 = s i | x t , x t 1 ) = i = 1 M p ( x t , x t 1 | θ t = s m , θ t 1 = s i ) p ( θ t = s m , θ t 1 = s i ) j = 1 M k = 1 M p ( x t , x t 1 | θ t = s j , θ t 1 = s k ) p ( θ t = s j , θ t 1 = s k ) , ( 1 ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaafaqaaeWabaaabaGaemiCaaNaeiikaGccciGae8hUde3aaSbaaSqaaiabdsha0bqabaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemyBa0gabeaakiabcYha8Hqabiab+Hha4naaBaaaleaacqWG0baDaeqaaOGaeiilaWIae4hEaG3aaSbaaSqaaiabdsha0jabgkHiTiabigdaXaqabaGccqGGPaqkaeaacqGH9aqpdaaeWbqaaiabdchaWjabcIcaOiab=H7aXnaaBaaaleaacqWG0baDaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabd2gaTbqabaGccqGGSaalcqWF4oqCdaWgaaWcbaGaemiDaqNaeyOeI0IaeGymaedabeaakiabg2da9iabdohaZnaaBaaaleaacqWGPbqAaeqaaOGaeiiFaWNae4hEaG3aaSbaaSqaaiabdsha0bqabaGccqGGSaalcqGF4baEdaWgaaWcbaGaemiDaqNaeyOeI0IaeGymaedabeaakiabcMcaPaWcbaGaemyAaKMaeyypa0JaeGymaedabaGaemyta0eaniabggHiLdaakeaacqGH9aqpdaaeWbqaamaalaaabaGaemiCaaNaeiikaGIae4hEaG3aaSbaaSqaaiabdsha0bqabaGccqGGSaalcqGF4baEdaWgaaWcbaGaemiDaqNaeyOeI0IaeGymaedabeaakiabcYha8jab=H7aXnaaBaaaleaacqWG0baDaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabd2gaTbqabaGccqGGSaalcqWF4oqCdaWgaaWcbaGaemiDaqNaeyOeI0IaeGymaedabeaakiabg2da9iabdohaZnaaBaaaleaacqWGPbqAaeqaaOGaeiykaKIaemiCaaNaeiikaGIae8hUde3aaSbaaSqaaiabdsha0bqabaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemyBa0gabeaakiabcYcaSiab=H7aXnaaBaaaleaacqWG0baDcqGHsislcqaIXaqmaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdMgaPbqabaGccqGGPaqkaeaadaaeWbqaamaaqahabaGaemiCaaNaeiikaGIae4hEaG3aaSbaaSqaaiabdsha0bqabaGccqGGSaalcqGF4baEdaWgaaWcbaGaemiDaqNaeyOeI0IaeGymaedabeaakiabcYha8jab=H7aXnaaBaaaleaacqWG0baDaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdQgaQbqabaGccqGGSaalcqWF4oqCdaWgaaWcbaGaemiDaqNaeyOeI0IaeGymaedabeaakiabg2da9iabdohaZnaaBaaaleaacqWGRbWAaeqaaOGaeiykaKIaemiCaaNaeiikaGIae8hUde3aaSbaaSqaaiabdsha0bqabaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemOAaOgabeaakiabcYcaSiab=H7aXnaaBaaaleaacqWG0baDcqGHsislcqaIXaqmaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdUgaRbqabaGccqGGPaqkaSqaaiabdUgaRjabg2da9iabigdaXaqaaiabd2eanbqdcqGHris5aaWcbaGaemOAaOMaeyypa0JaeGymaedabaGaemyta0eaniabggHiLdaaaaWcbaGaemyAaKMaeyypa0JaeGymaedabaGaemyta0eaniabggHiLdaaaOGaeiilaWIaaCzcaiaaxMaadaqadaqaaiabigdaXaGaayjkaiaawMcaaaaa@E9D0@

where m = 1,2,...M, p(θ t = s j , θt-1= s i ) = p(θ t = s j |θt-1= s i )p(θt-1= s i ) = a ij π i . The p(x t , xt-1|θ t = s j , θt-1= s i ) means given θ t = s j and θt-1= s i , the probability of Gaussian mixtures ϕ(x, μ jr , Σ jr ) and ϕ(x, μ kr , Σ kr ), r = 1, 2,..., R can generate vectors x t and xt-1.

Traditional CHMM has utilized the information of the previous time point to predict the state of current time point. In our application, we know both the information of "left point" and "right point", since we have obtained the cell trace and the features of all cells. In contrast to the traditional CHMM introduced above, we propose to utilize the contextual information for cell phase identification, i.e., we propose to use Gaussian mixture models based on two-pixels. This model is called the Context Based Mixture Model Classifier.

p ( θ t = s m | x t , x t 1 , x t + 1 ) = i = 1 M j = 1 M p ( θ t = s m , θ t 1 = s i , θ t + 1 = s j | x t , x t 1 , x t + 1 ) = i = 1 M j = 1 M p ( x t , x t 1 , x t + 1 | θ t = s m , θ t 1 = s i , θ t + 1 = s j ) P ( m , i , j ) m = 1 M i = 1 M j = 1 M p ( x t , x t 1 , x t + 1 | θ t = s m , θ t 1 = s i , θ t + 1 = s j ) P ( m , i , j ) ( 2 ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaafaqaaeGabaaabaGaemiCaaNaeiikaGccciGae8hUde3aaSbaaSqaaiabdsha0bqabaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemyBa0gabeaakiabcYha8Hqabiab+Hha4naaBaaaleaacqWG0baDaeqaaOGaeiilaWIae4hEaG3aaSbaaSqaaiabdsha0jabgkHiTiabigdaXaqabaGccqGGSaalcqGF4baEdaWgaaWcbaGaemiDaqNaey4kaSIaeGymaedabeaakiabcMcaPiabg2da9maaqahabaWaaabCaeaacqWGWbaCcqGGOaakcqWF4oqCdaWgaaWcbaGaemiDaqhabeaakiabg2da9iabdohaZnaaBaaaleaacqWGTbqBaeqaaOGaeiilaWIae8hUde3aaSbaaSqaaiabdsha0jabgkHiTiabigdaXaqabaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemyAaKgabeaakiabcYcaSiab=H7aXnaaBaaaleaacqWG0baDcqGHRaWkcqaIXaqmaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdQgaQbqabaGccqGG8baFcqGF4baEdaWgaaWcbaGaemiDaqhabeaakiabcYcaSiab+Hha4naaBaaaleaacqWG0baDcqGHsislcqaIXaqmaeqaaOGaeiilaWIae4hEaG3aaSbaaSqaaiabdsha0jabgUcaRiabigdaXaqabaGccqGGPaqkaSqaaiabdQgaQjabg2da9iabigdaXaqaaiabd2eanbqdcqGHris5aaWcbaGaemyAaKMaeyypa0JaeGymaedabaGaemyta0eaniabggHiLdaakeaacqGH9aqpdaWcaaqaamaaqahabaWaaabCaeaacqWGWbaCcqGGOaakcqGF4baEdaWgaaWcbaGaemiDaqhabeaakiabcYcaSiab+Hha4naaBaaaleaacqWG0baDcqGHsislcqaIXaqmaeqaaOGaeiilaWIae4hEaG3aaSbaaSqaaiabdsha0jabgUcaRiabigdaXaqabaGccqGG8baFcqWF4oqCdaWgaaWcbaGaemiDaqhabeaakiabg2da9iabdohaZnaaBaaaleaacqWGTbqBaeqaaOGaeiilaWIae8hUde3aaSbaaSqaaiabdsha0jabgkHiTiabigdaXaqabaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemyAaKgabeaakiabcYcaSiab=H7aXnaaBaaaleaacqWG0baDcqGHRaWkcqaIXaqmaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdQgaQbqabaGccqGGPaqkcqWGqbaucqGGOaakcqWGTbqBcqGGSaalcqWGPbqAcqGGSaalcqWGQbGAcqGGPaqkaSqaaiabdQgaQjabg2da9iabigdaXaqaaiabd2eanbqdcqGHris5aaWcbaGaemyAaKMaeyypa0JaeGymaedabaGaemyta0eaniabggHiLdaakeaadaaeWbqaamaaqahabaWaaabCaeaacqWGWbaCcqGGOaakcqGF4baEdaWgaaWcbaGaemiDaqhabeaakiabcYcaSiab+Hha4naaBaaaleaacqWG0baDcqGHsislcqaIXaqmaeqaaOGaeiilaWIae4hEaG3aaSbaaSqaaiabdsha0jabgUcaRiabigdaXaqabaGccqGG8baFcqWF4oqCdaWgaaWcbaGaemiDaqhabeaakiabg2da9iabdohaZnaaBaaaleaacqWGTbqBaeqaaOGaeiilaWIae8hUde3aaSbaaSqaaiabdsha0jabgkHiTiabigdaXaqabaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemyAaKgabeaakiabcYcaSiab=H7aXnaaBaaaleaacqWG0baDcqGHRaWkcqaIXaqmaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdQgaQbqabaGccqGGPaqkcqWGqbaucqGGOaakcqWGTbqBcqGGSaalcqWGPbqAcqGGSaalcqWGQbGAcqGGPaqkaSqaaiabdQgaQjabg2da9iabigdaXaqaaiabd2eanbqdcqGHris5aaWcbaGaemyAaKMaeyypa0JaeGymaedabaGaemyta0eaniabggHiLdaaleaacqWGTbqBcqGH9aqpcqaIXaqmaeaacqWGnbqta0GaeyyeIuoaaaaaaOGaaCzcaiaaxMaadaqadaqaaiabikdaYaGaayjkaiaawMcaaaaa@1AC6@

where p(m, i, j) = p(θ t = s m , θt-1= s i , θt+1= s j ) is the prior probability defined as:

P ( θ t = s m , θ t 1 = s i , θ t + 1 = s j ) = # s t a t u s o f ( i , j , m ) i = 1 M j = 1 M m = 1 M # s t a t u s o f ( i , j , m ) ( 3 ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaacqWGqbaucqGGOaakiiGacqWF4oqCdaWgaaWcbaGaemiDaqhabeaakiabg2da9iabdohaZnaaBaaaleaacqWGTbqBaeqaaOGaeiilaWIae8hUde3aaSbaaSqaaiabdsha0jabgkHiTiabigdaXaqabaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemyAaKgabeaakiabcYcaSiab=H7aXnaaBaaaleaacqWG0baDcqGHRaWkcqaIXaqmaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabdQgaQbqabaGccqGGPaqkcqGH9aqpdaWcaaqaaiabcocaJiabdohaZjabdsha0jabdggaHjabdsha0jabdwha1jabdohaZjabbccaGiabd+gaVjabdAgaMjabbccaGiabcIcaOiabdMgaPjabcYcaSiabdQgaQjabcYcaSiabd2gaTjabcMcaPaqaamaaqahabaWaaabCaeaadaaeWbqaaiabcocaJiabdohaZjabdsha0jabdggaHjabdsha0jabdwha1jabdohaZjabbccaGiabd+gaVjabdAgaMjabbccaGiabcIcaOiabdMgaPjabcYcaSiabdQgaQjabcYcaSiabd2gaTjabcMcaPaWcbaGaemyBa0Maeyypa0JaeGymaedabaGaemyta0eaniabggHiLdaaleaacqWGQbGAcqGH9aqpcqaIXaqmaeaacqWGnbqta0GaeyyeIuoaaSqaaiabdMgaPjabg2da9iabigdaXaqaaiabd2eanbqdcqGHris5aaaakiaaxMaacaWLjaWaaeWaaeaacqaIZaWmaiaawIcacaGLPaaaaaa@8E79@

Since the denominator in (2) is the same for each class, in implementation we can neglect this term, i.e., p ( θ t = s m | x t , x t 1 , x t + 1 ) i = 1 M j = 1 M p ( x t , x t 1 , x t + 1 | θ t = s m , θ t 1 = s i , θ t + 1 = s j ) P ( m , i , j ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaacqWGWbaCcqGGOaakiiGacqWF4oqCdaWgaaWcbaGaemiDaqhabeaakiabg2da9iabdohaZnaaBaaaleaacqWGTbqBaeqaaOGaeiiFaWNaeCiEaG3aaSbaaSqaaiabdsha0bqabaGccqGGSaalcqWH4baEdaWgaaWcbaGaemiDaqNaeyOeI0IaeGymaedabeaakiabcYcaSiabhIha4naaBaaaleaacqWG0baDcqGHRaWkcqaIXaqmaeqaaOGaeiykaKIaeyyhIu7aaabCaeaadaaeWbqaaiabdchaWjabcIcaOiabhIha4naaBaaaleaacqWG0baDaeqaaOGaeiilaWIaeCiEaG3aaSbaaSqaaiabdsha0jabgkHiTiabigdaXaqabaGccqGGSaalcqWH4baEdaWgaaWcbaGaemiDaqNaey4kaSIaeGymaedabeaakiabcYha8jab=H7aXnaaBaaaleaacqWG0baDaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabd2gaTbqabaGccqGGSaalcqWF4oqCdaWgaaWcbaGaemiDaqNaeyOeI0IaeGymaedabeaakiabg2da9iabdohaZnaaBaaaleaacqWGPbqAaeqaaOGaeiilaWIae8hUde3aaSbaaSqaaiabdsha0jabgUcaRiabigdaXaqabaGccqGH9aqpcqWGZbWCdaWgaaWcbaGaemOAaOgabeaakiabcMcaPaWcbaGaemOAaOMaeyypa0JaeGymaedabaGaemyta0eaniabggHiLdaaleaacqWGPbqAcqGH9aqpcqaIXaqmaeaacqWGnbqta0GaeyyeIuoakiabdcfaqjabcIcaOiabd2gaTjabcYcaSiabdMgaPjabcYcaSiabdQgaQjabcMcaPaaa@8E4B@ .

To estimate p(x t , xt-1, xt+1|θ t = s m , θt-1= s i , θt+1= s j ), the simplest way is to assume that the phase of three states are independent, then it can be simply approximated by the product of the three items p(x t |θ t = s m ), p(xt-1|θt-1= s i ) and p(xt+1|θt+1= s j ). This assumption is not always true in realistic applications. In this case, we assume that they are dependent. More complex models are trained than the standard continuous models described above. We collect the cells in status satisfying (θ t = s m , θt-1= s i , θt+1= s j ). For example, if (θ t = 2, θt-1= 1, θt+1= 3), the interphase cells are collected if they lie between the interphase and anaphase to estimate p(x i , xt-1, xt+1|θ t = 2, θt-1= 1, θt+1= 3) using the EM algorithm. Then we have 4 × 4 × 4 = 64 hidden states altogether. The hidden states with less than five training samples are assigned zero prior probabilities. Each hidden state can generate one visible Gaussian mixture ϕ(x, μ k , Σ k ), k = 1,2,...,64, where μ k and Σ k sare means and covariance matrices of Gaussian mixtures respectively. The parameters μ k and Σ k are optimized by the EM algorithm iteratively. Then we get p(θ i = s m |x i , xt-1, xt+1), m = 1,..., M, and classify cell X t to phase sm*, such that

θ t = s m *  iff  s m * = arg max m = 1 , 2 , M { p ( θ t = s m | x t , x t 1 , x t + 1 ) } MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaaiiGacqWF4oqCdaWgaaWcbaGaemiDaqhabeaakiabg2da9iabdohaZnaaBaaaleaacqWGTbqBcqGGQaGkaeqaaOGaeeiiaaIaeeyAaKMaeeOzayMaeeOzayMaeeiiaaIaem4Cam3aaSbaaSqaaiabd2gaTjabcQcaQaqabaGccqGH9aqpdaWfqaqaaiGbcggaHjabckhaYjabcEgaNjGbc2gaTjabcggaHjabcIha4bWcbaGaemyBa0Maeyypa0JaeGymaeJaeiilaWIaeGOmaiJaeiilaWIaeS47IWKaemyta0eabeaakiabcUha7jabdchaWjabcIcaOiab=H7aXnaaBaaaleaacqWG0baDaeqaaOGaeyypa0Jaem4Cam3aaSbaaSqaaiabd2gaTbqabaGccqGG8baFcqWH4baEdaWgaaWcbaGaemiDaqhabeaakiabcYcaSiabhIha4naaBaaaleaacqWG0baDcqGHsislcqaIXaqmaeqaaOGaeiilaWIaeCiEaG3aaSbaaSqaaiabdsha0jabgUcaRiabigdaXaqabaGccqGGPaqkcqGG9bqFaaa@6EE0@

Abbreviations

CHMM :

Continuous Hidden Markov Model

NN :

Neural Networks

BPNN :

Back-Propagation Neural Networks

References

  1. Yan J, Zhou X, Yang Q, Liu N, Cheng Q, Wong STC: An efficient system for optical microscopy cell image segmentation, tracking and cell phase identification. IEEE International symposium on Image Processing: 2006; Atlanta 2006, 1536–1537.

    Google Scholar 

  2. Zhou X, Wong STC: High content cellular imaging for drug development. IEEE Signal Processing Magazine 2006, 23: 170–174.

    Article  Google Scholar 

  3. Zhou X, Cao XH, Perlman Z, Wong STC: A computerized cellular imaging system for high content analysis in Monastrol suppressor screens. Journal of Biomedical Informatics 2006, 39: 115–125.

    Article  PubMed  Google Scholar 

  4. Zhou X, Wong STC: Informatics challenges of high-throughput cellular and molecular microscopy. IEEE Signal Processing Magazine 2006, 23: 63–72.

    Article  Google Scholar 

  5. Chen SC, Murphy RF: A graphical model approach to automated classification of protein subcellular location patterns in multi-cell images. BMC Bioinformatics 2006, 7: 90.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  6. Murphy RF, Velliste M, Porreca G: Robust numerical features for description and classification of subcellular location patterns in fluorescence microscope images. J VLSI Sig Proc 2003, 35: 311–321.

    Article  Google Scholar 

  7. Huang K, Velliste M, Murphy RF: Feature reduction for improved recognition of subcellular location patterns in fluorescence microscope image. Proc SPIE 2003, 4962: 307–318.

    Article  CAS  Google Scholar 

  8. Boland MV, Murphy RF: A neural network classifier capable of recognizing the patterns of all major subcellular structures in fluorescence microscope image of HeLa cells. Bioinformatcs 2001, 17: 1213–1223.

    Article  CAS  Google Scholar 

  9. Boland MV, Markey MK, Murphy RF: Automated recognition of pattern characteristic of subcellular structures in fluorescence microscopy images. Cytometry 1998, 33: 366–375.

    Article  CAS  PubMed  Google Scholar 

  10. Gallardo G, Ianzini F, Mackey MA, Sonka M, Yang F: Mitotic Cell Recognition with hidden Markov Models. Medical Imaging 2004, 5367: 661–668.

    Google Scholar 

  11. Chen X, Zhou X, Wong STC: Automated segmentation, classification, and tracking of cancer cell nuclei in time-lapse microscopy. IEEE Transactions on Biomedical Engineering 2006, 53: 762–766.

    Article  PubMed  Google Scholar 

  12. Li T, Zhang C, Ogihara M: A comparative study of feature selection and multiclass classification methods for tissue classification based on gene expression. Bioinformatics 2004, 20(15):2429–2437.

    Article  CAS  PubMed  Google Scholar 

  13. Huang K, Veliste M, Murphy RF: Feature reduction for improved recognition of subcellular location pattern in fluorescence microscope images. Proceedings of SPIE 2003, 4962: 307–318.

    Article  CAS  Google Scholar 

  14. Pham TD, Tran DT, Zhou X, Wong STC: Cell Phase Identification by Vector Quantization and Markov Models. In Progress in Cell Research. Edited by: Pham T. New York City: Nova Science; 2006.

    Google Scholar 

  15. Yan J, Zhang BY, Liu N, Yan S, Cheng Q, Fan W, Yang Q, Xi W, Chen Z: Effective and efficient dimensionality reduction for large scale and streaming data preprocessing. IEEE Transactions on Knowledge and Data Engineering 2006, 18(2):320–333.

    Article  Google Scholar 

  16. Jolliffe IT: Principal Component Analysis. Springer-Verlag; 1986.

    Chapter  Google Scholar 

  17. Duda O, Hart PE, Stork DG: Pattern classification. 2nd edition. John Wiley; 2001.

    Google Scholar 

  18. Oja E: Subspace methods of pattern recognition. Pattern recognition and image processing series 1983., 6:

    Google Scholar 

  19. Li H, Jiang T, Zhang K: Efficient and robust feature extraction by maximum margin criterion. In Proceeding of the Advances in Neural Information Processing Systems 16: 2004. Vancouver, Canada; 2004:97–104.

    Google Scholar 

  20. Avrim L, Blum , Langley P: Selection of relevant features and examples in machine learning. Artificial Intelligence 1997, 97(1–2):245–271.

    Google Scholar 

  21. Yang J, Honavar V: Feature subset selection using a Genetic Algorithm. IEEE Intelligence Systems 1997, 13: 44–49.

    Article  Google Scholar 

  22. Webb AR: Statistical Pattern Recognition. 2nd edition. John Wiley; 2002.

    Chapter  Google Scholar 

  23. Chang C, Lin C: LIBSVM: a library for support vector machines. Technical report Computer Science and Information Engineering, National Taiwan University; 2001. [http://www.csie.ntu.edu.tw/~cjlin/libsvm]

    Google Scholar 

  24. Vapnik VN: Statistical Learning Theory. New York: John Wiley; 1998.

    Google Scholar 

  25. Lindblad J, Wahlby C, Bengtsson E, Zaltsman A: Image analysis for automatic segmentation of cytoplasms and classification of Rac1 activation. Cytometry A 2004, 57: 22–33.

    Article  PubMed  Google Scholar 

  26. Lin G, Adiga U, Olson K, Guzowski J, Barnes C, Roysam B: A hybrid 3-D watershed algorithm incorporating gradient cues and object models for automatic segmentation of nuclei in confocal image stacks. Cytometry A 2003, 56: 23–36.

    Article  PubMed  Google Scholar 

  27. Beucher S: The watershed transformation applied to image segmentation. Scanning Microscopy International 1992, 6: 299–314.

    Google Scholar 

  28. Haralick RM, Shanmugam K, Dinstein I: Textural features for image classification. IEEE Transactions on Systems, Man, and Cybertinetics 1973, 3: 610–621.

    Article  Google Scholar 

  29. Teague MR: Image analysis via the general theory of moments. J Opt Soc Am 1980, 70(8):920–930.

    Article  Google Scholar 

  30. Manjunath BS, MA WY: Texture features for browsing and retrieval of image data. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI-Special issuer on Digital Libraries) 1996, 18: 837–842.

    Article  Google Scholar 

  31. Tenenbaum JB, Silva VD, Langford JC: A global geometric framework for nonlinear dimensionality reduction. Science 2000, 290: 2319–2323.

    Article  CAS  PubMed  Google Scholar 

  32. Wang M, Yang J, Liu GP, Xu ZJ, Chou KC: Weighted-support vector machines for predicting membrane protein types based on pseudo-amino acid composition. PEDS 2005, 17: 509–516.

    Google Scholar 

Download references

Acknowledgements

The authors would like to thank Jeremy Huckins of the Department of Cell Biology, Harvard Medical School for his review of this paper. The authors would also like to acknowledge the earlier work of Mr Jun Yan, now at Microsoft Research China, who contributes the feature extraction algorithms, as well as Baillie Yip and Ning Liu for their assistance in the evaluation and validation of the algorithms. This research is funded by the HCNR Center for Bioinformatics Research Grant, Harvard Medical School and a NIH R01 LM008696 Grant (STCW).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Stephen TC Wong.

Additional information

Authors' contributions

MW andXZ played equal roles in investigating the proposed approach and conducting the experiments. RWK generated the images of Hela cell line. STCW directed the project and guided the research discussion. All authors have read and approved the final manuscript.

Authors’ original submitted files for images

Rights and permissions

Open Access This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License ( https://creativecommons.org/licenses/by/2.0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Wang, M., Zhou, X., King, R.W. et al. Context based mixture model for cell phase identification in automated fluorescence microscopy. BMC Bioinformatics 8, 32 (2007). https://doi.org/10.1186/1471-2105-8-32

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1471-2105-8-32

Keywords