Automatic recognition of different tissue types in histological images is an essential part in the digital pathology toolbox. Texture analysis is commonly used to address this problem; mainly in the context of estimating the tumour/stroma ratio on histological samples. However, although histological images typically contain more than two tissue types, only few studies have addressed the multi-class problem. For colorectal cancer, one of the most prevalent tumour types, there are in fact no published results on multiclass texture separation. In this paper we present a new dataset of 5,000 histological images of human colorectal cancer including eight different types of tissue. We used this set to assess the classification performance of a wide range of texture descriptors and classifiers. As a result, we found an optimal classification strategy that markedly outperformed traditional methods, improving the state of the art for tumour-stroma separation from 96.9% to 98.6% accuracy and setting a new standard for multiclass tissue separation (87.4% accuracy for eight classes). We make our dataset of histological images publicly available under a Creative Commons license and encourage other researchers to use it as a benchmark for their studies.
Human solid tumours are complex structures that typically contain several distinct tissue types. Apart from clonal tumour cells, they consist of tumour stroma, immune cell infiltration, necrotic areas and islets of remaining non-malignant tissue. These different tissue types can be distinguished by histopathological evaluation of Hematoxylin and Eosin (H&E) stained tissue sections. In colorectal cancer (CRC), one of the most prevalent cancer types, tumour architecture changes during tumour progression1 and is related to patient prognosis2. Quantifying the tissue composition in CRC is therefore a relevant task in histopathology.
While manual evaluation of histological slides is still indispensable in clinical routine, automated image processing can provide quantitative and high-throughput analysis of the tumour tissue. In principle, automatic separation of tissue types in histological images can be achieved by different supervised machine learning approaches: in cell morphology based methods, individual cells are segmented and then classified into different categories such as tumour cells, stroma cells and immune cells. This approach has been successfully used by various groups (see Xu et al.3 for an overview) and has led to new candidate biomarkers4,5,6. A different type of tissue classification methods is based on texture. The term texture refers to specific properties of the internal structure of image regions, for example rough versus smooth or oriented versus randomly dispersed (among others)7,8,9. In medical image analysis, texture based methods are very useful to classify tissue types10,11. Typically, these methods extract texture features first8,12,13,14, then feed the features into a classifier to predict the tissue type9,15,16.
However, when it comes to classifying tissue types in CRC histological images, all published methods invariably show two common limitations: first, they consider only two categories of tissue (tumour and stroma), which makes these approaches unsuitable for more heterogeneous parts of the tumour8,12; second, all studies used their own image data set which prohibits quantitative comparison of classification performance. Whereas publicly available benchmarking datasets exist for image classification problems such as face recognition17, handwriting recognition18, universal computer vision problems19 and texture classification20,21, no such data are available for histopathological tissue classification.
The aim of this study is to fill this gap. To this end we assembled, tested and publicly released a comprehensive image set of all relevant types of tissue within colorectal cancer samples. We used the dataset to compare several state of the art texture features and classifiers and to determine which combination is best suited for a multiclass tissue classification problem.
Material and Methods
All experiments were approved by the institutional ethics board (medical ethics board II, University Medical Center Mannheim, Heidelberg University, Germany; approval 2015-868R-MA). The institutional ethics board waived the need for informed consent for this retrospective analysis of anonymized samples. All experiments were carried out in accordance with the approved guidelines and with the Declaration of Helsinki.
Ten anonymized H&E stained CRC tissue slides were obtained from the pathology archive at the University Medical Center Mannheim (Heidelberg University, Mannheim, Germany). Low-grade and high-grade tumours were included in this set; no further selection was applied. The slides were first digitized as described before22. Then, contiguous tissue areas were manually annotated and tessellated, creating 625 non-overlapping tissue tiles of dimension 150 px × 150 px (74 μm × 74 μm). Thus, texture features of different scales were included, ranging from individual cells (approx. 10 μm, e.g. Fig. 1d) to larger structures such as mucosal glands (>50 μm, e.g. Fig. 1f). The following eight types of tissue were selected for analysis:
Simple stroma (homogeneous composition, includes tumour stroma, extra-tumoural stroma and smooth muscle);
Complex stroma (containing single tumour cells and/or few immune cells);
Immune cells (including immune-cell conglomerates and sub-mucosal lymphoid follicles);
Debris (including necrosis, hemorrhage and mucus);
Normal mucosal glands;
Background (no tissue).
Together, the resulting 625 × 8 = 5000 images represented the training and testing set of the classification problem described in the following sections. The first 10 images of each class are shown in Fig. 1. Average staining intensity considerably varied between the tissue samples, reflecting the usual variability in routine histopathological slides. We took care that each of the classes listed above contained both bright and dark samples so that no bias in terms of average greyscale intensity was introduced (Fig. 1). In addition to these images, we also extracted ten larger images of dimension 5000 px × 5000 px from tissue regions different from those used for the smaller images. These ten images constituted an application set and were used to test the different combinations of texture features/classifiers in a realistic setting.
Data usage statement
We release all raw data under a Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/). The data can be accessed via the following DOI: 10.5281/zenodo.53169. All source codes used for this study are available under the MIT license (http://opensource.org/licenses/MIT) and can be accessed via the following DOI: 10.5281/zenodo.53735.
To describe the texture of histological images we considered six distinct sets of descriptors that are detailed in the following sections. All images were preliminarily converted to greyscale before computing the texture features. Yet, in the dataset we provide, images are native red/green/blue (RGB) images so that it can also be used to benchmark colour-based texture classifiers.
Lower-order and higher-order histogram features
Lower-order statistics can be used to describe texture23,24. We used the gray level histogram of a given image to construct two simple feature sets: 1) one set containing the mean, variance, skewness, kurtosis and the 5th central moment of the histogram (five features); 2) another set composed of the central moments from 2nd to 11th (ten features). In the remainder we refer to the two sets of features as ‘histogram-lower’ and ‘histogram-higher’, respectively. Note that the latter does not contain the mean therefore it is invariant to changes in the average greyscale intensity of the input image (and is therefore less sensitive to staining differences).
Local binary patterns (LBP)
The third feature set was based on local binary patterns (LBP)25. The Local Binary Patterns (LBP) operator considers the probability of occurrence of all the possible binary patterns that can arise from a neighbourhood of predefined shape and size. In this work we considered a neighbourhood of eight equally-spaced points arranged along a circle of radius 1px. This is usually referred to as the ‘8, 1’ configuration26. For each position of the neighbourhood a corresponding binary pattern is obtained by thresholding the intensity values of the eight points on the circle at the value of the central point. In our study, the resulting histogram was reduced to the 38 rotationally-invariant Fourier features proposed by Ahonen et al.27. Other LBP variants have been used for histological texture analysis in other studies12,13.
Gray-level co-occurrence matrix (GLCM)
The fourth feature set was based on GLCM features9,24. In particular, we used four directions (0°, 45°, 90° and 135°) and five displacement vectors (from 1px to 5px). To make this texture descriptor invariant with respect to rotation, we averaged the GLCMs obtained from all four directions for each displacement vector. From each of the resulting co-occurrence matrices we extracted the following four global statistics: contrast, correlation, energy and homogeneity9, thereby obtaining 5 × 4 = 20 features for each input image.
The fifth set of features was based on Gabor filtering28. We applied a bank of Gabor filters to the greyscale image and computed the mean intensity of the resulting Gabor-transformed magnitude images. In particular, we used six directions (0°, 30°, 60°, 90°, 120° and 150°) and six wavelengths (2, 4, 6, 8, 10 and 12 px/cycle). We chose these particular wavelengths because we subjectively observed that the texture structures of interest in histological images (cells, cell nuclei or collagen fibres) typically ranged between 2 px and 12 px. To make this texture descriptor invariant with respect to rotation, we averaged the results obtained from all Gabor filters with identical wavelength over all orientations, thereby obtaining 6 features for each input image.
The sixth set included features based on image perception. These are intrinsically different from most texture features, such as LBP or GLCM based features, which are, by contrast, not easily understandable. Tamura et al.7 showed that the human visual system discriminates texture through several specific attributes that were later on refined and tested by Bianconi et al.8. The features used in this study were the following five: coarseness, contrast, directionality, line-likeness and roughness. A detailed description of these features is given by Bianconi et al.8.
Combined feature sets
Lastly, we investigated whether discriminatory power of the feature sets could be improved by merging features into a concatenated feature vector. As opposed to the pure feature sets described before, we subsequently refer to those as combined feature sets. First, we ranked the feature sets based on their classification accuracy as described below (histogram-lower > LBP > histogram-higher > GLCM > Perceptual > Gabor). The procedure for accuracy estimation was based on 10-fold cross validation with full sampling. The subdivision into train and test set was repeated 10 times; in each subdivision 90% of the images of the whole dataset was used to train the classifier and the remaining 10% to test it. Accuracy for each classification round was computed as the ratio between the number of images of the test set correctly classified and the total number of images of the test set. The overall accuracy was estimated as the average over the 10 classification rounds.
Then, we successively added pure feature sets to the combined feature sets: best2 (histogram-lower and LBP), best3 (best2 and histogram-higher, removing the duplicate features that belonged both to histogram-low and histogram-high), best4 (best3 and GLCM), best5 (best4 and Perceptual) and all6 (best5 and Gabor). The different range of the feature vectors was accounted for by standardizing mean and variance of each column of the feature matrix before SVM classification.
We used four classification strategies: 1) 1-nearest neighbour, 2) linear SVM, 3) radial-basis function SVM and 4) decision trees. We recall the basics of each classifier in the following sections.
The Euclidean-distance 1-nearest neighbour (1-NN) is a very simple classifier that is independent of tuning parameters, is easy to implement and has a low risk of overfitting29. Before training the classifier, the feature vectors were standardized to have equal mean and variance.
Linear and radial basis function support vector machine
We employed support vector machines (SVM) with one-versus-one class decisions in an error-correcting output code multiclass model (ECOC)30,31. We compared linear SVM and radial basis function (rbf, Gaussian) SVM. Before training, the feature vectors were automatically normalized to have equal mean and variance.
Ensemble of decision trees
Finally, we considered an ensemble of decision trees using the RUSboost method. This method is especially suited for data with unequal group sizes. Although this is not the case in our study (the groups are perfectly balanced), we chose RUSboost because it is considered a fast and robust technique32.
Construction of training and testing set
To train the classifiers we used 10-fold cross validation. The 5000-item dataset was randomly subdivided in 10 parts and 10 rounds of training and testing were performed. For each subdivision a different 10% subset of the dataset was used for testing while the other 90% was used for training. Because the overall number of images was large (5000 images in total) and the group sizes balanced (625 images per set), randomly distributing the images into training and testing set yielded consistent group proportions, even without an explicit stratification approach.
Two types of classification problems were analysed: a multi-class problem (comprising all 5000 images in 8 classes, i.e. the full dataset) and a two-class problem (comprising only 1250 images, i.e. 625 images of “tumour epithelium” and 625 images of “simple stroma”). The two-class problem was addressed because tumour-stroma separation has been addressed by other studies8,12 and therefore these results could be quantitatively compared to the results of the present study.
After training and testing the classifiers we used them to segment an independent set of 10 images (application set, as described above). Each 5000-pixel square image contained regions of different tissue types: identifying these regions is a common problem in digital histopathology. Each input image was subdivided into 10,000 overlapping 150-pixel square tiles and for each tile the texture features were computed and submitted to the classifier.
The approaches described in the preceding sections have been implemented in Matlab® (R2015b, Mathworks, Natick, MA, USA) and the experiments were carried out on a standard computer workstation (2.2 GHz Intel Core i7, 16 GB 156 RAM). In addition to custom routines developed by the authors and Matlab’s built-in functions, we also used publicly available source code from Bianconi et al.8 and Ahonen et al.27. The entire code required to reproduce the experiments is freely available to the public (see “Data usage” section).
Performance of pure feature sets in a two-class and multiclass problem
We performed 4 × 6 × 2 = 48 supervised image classification experiments to estimate the accuracy of each combination of one of six feature sets, one of four classifiers for either two (tumour-stroma) or eight target categories (multiclass problem).
First, we tested all pure methods, i.e. sets of features obtained with a single texture description method. We found that in a conventional two-class problem, lower order histogram features outperformed the other feature sets (Fig. 2a). Comparing performance of different classifiers with identical feature sets, we found that radial basis function (rbf) support vector machine (SVM) yielded the lowest classification error rate in all but one experiment (Fig. 2a,b).
Specifically, using an rbf SVM in a two-class problem, classification error rate was 4.3% for histogram-lower, followed by 5.1% for LBP. Similarly, in a multiclass problem, histogram-lower and LBP yielded the best results with 19.2% and 23.8% error rates (Fig. 2b).
Combining feature sets markedly improves classification performance
Because the different pure feature sets are conceptually different and measure different aspects of texture, we investigated whether performance could be improved by merging these sets. We ranked the feature sets based on their performance in a multiclass problem and tested five combined sets. This approach markedly improved performance in a two-class setting: In a conventional two-class (tumour-stroma) classification problem, the best2 set already reached an accuracy of 98.3%, which was only slightly increased by considering more features (best2, best3: 98.3%; best4, best5, all6: 98.6%). To our knowledge, this accuracy is higher than previously reported accuracies for similar problems – see for instance refs 8,12. In a multiclass setting, the optimal performance was achieved by the best5 feature set (Fig. 2c) and was 87.4%. Also, computational performance was still acceptable in the best5 feature set (Fig. 2d) as compared to the all6 feature set. The confusion matrices and receiver operating characteristic (ROC) curves in Fig. 3 show that classification errors are approximately equally distributed among all classes.
Assessing classification performance in complex images
To subjectively assess classification performance, we used the best performing classification method (best5 feature set and rbf SVM) with our application set. This set consists of 10 images that were independent of the training/testing data and contained difficult, intermixed textures. Qualitatively, the resulting segmentation (Fig. 4) shows good separation among the eight tissue classes. Furthermore, the probability maps (Fig. 5) confirm that the class distribution correlates well with the subjective evaluation of the original image (Fig. 5i). To better visualize the classification of mixed tissue types, we also provide a false-colour representation of tumour-stroma separation in Fig. 6. As can be seen, simple stroma and complex stroma gradually fade into each other and complex stroma tends to cluster in the proximity of tumour epithelium.
Correlation analysis confirms usefulness of feature combination
In our study, the best classification performance was achieved by the best5 set, a combined feature set comprising histogram-based features as well as GLCM, LBP and perceptual features. We performed a correlation analysis of the concatenated 74 dimensional feature vectors and found that there was little correlation between the feature subsets (Fig. 7a). This indicates that the feature sets measure different aspects of texture and shows that combining pure feature sets may indeed be useful. A correlation analysis of all 5,000 feature vectors (one vector for every image) showed that images of a given class form distinct clusters (Fig. 7b).
In this paper we investigated the use of texture analysis for discriminating between eight different tissue types in colorectal cancer. We found that global lower-order texture measures (“histogram lower”) and the local texture measures GLCM and LBP were able to differentiate multiple tissue types in histological images of colorectal cancer (Fig. 2b) and that a combined approach was particularly effective (Figs 2c and 3c). Another analysis showed relatively little mutual correlation of the individual feature sets (Fig. 7a), thus supporting our approach to combine these different feature sets.
Texture measures and texture perception
Conceptually, there are many different approaches to measure texture (see Xie et al.33 and Beyerer et al.24 for an overview). Histogram-based features are first-order statistics describing the distribution of intensity values in an image. They measure the degree of dispersion of the grey values, the presence/absence of outliers and other properties which reflect the overall structure of the texture. Local texture descriptors such as GLCM9 or LBP26,27 are second-order statistics which consider the joint variability of the grey levels of pairs or groups of pixels. They are among the most used texture descriptors and proved effective in a wide range of applications34,35. Texture features mimicking the human perception at an abstract level have also been proposed in the literature7,8 and we also included these methods in our quantitative comparison experiments. Finally, we also tested Gabor filters, another common texture measure based on the response of a set of orientation- and frequency-selective filters28,36.
Multiparametric texture visualization
In our study, we became aware of a visualization problem of multiclass texture analysis that, as far as we know, has not been systematically addressed before: Multiclass texture analysis returns multidimensional parametric maps (one probability map for each tissue category). As previous studies mostly addressed two texture types, visualization of these textures was possible by a one-dimensional colour scale12. In a previous study, we investigated the use of two-dimensional colour scales to visualize histological imaging data37. However, in the present study, we generated an eight-dimensional dataset that cannot be visualized by a low dimensional colour scale. Thus, we implemented and applied three different visualization methods in this study (Figs 4, 5, 6). An alternative would be visualization as an interactive stack of channels.
Comparison to previous methods
Multi-class texture analysis has not been investigated in CRC histology yet, therefore direct comparison with other studies of the same type is not possible. Similar problems have however been addressed in closely related areas, such as prostate cancer histology. In the case of prostate cancer the neoplastic tissue has a different appearance which is usually classified through Gleason’s grades38. There have been structured efforts to automatically classify these grades with reported overall classification accuracies ranging from 74% to 97% (between 4 and 7 tissue categories)39,40,41. A similar approach applied to ovarian cancer histology has been reported to achieve 71.5% accuracy in distinguishing tumour epithelium from different stromal compartments42. Lastly, another method applied to breast cancer samples has been reported to achieve 89% accuracy for three tissue categories43.
If we compare our results with those just mentioned, we see that the accuracy achieved by our multi-class texture analysis approach is in the same range as was obtained in the other studies. Quite unfortunately, however, the results available in the literature are hard to reproduce and difficult to compare to each other owing to the fact that a) each study uses its own dataset and b) the datasets are usually not available to the public for further evaluations and comparisons. For this reason, it is not possible to quantitatively compare classification performance of these methods39,40,41,43 to our method.
Herein we presented an annotated dataset of 5,000 histological image patches along with a new state of the art classifier for two-class tissue separation and a new method for multiclass tissue separation. By publicly releasing our class-balanced database of histological CRC images we aimed at filling this gap in order to allow histological texture classifiers to be benchmarked on a standard and open-access dataset of colorectal cancer samples.
Whenever a pathologist evaluates a histological image, he or she mentally classifies tissue regions into categories such as “tumour epithelium”, “stroma”, “necrosis”, etc. The method we present in this paper can automate this task. Thus, histological images of colorectal carcinoma can be assessed in a reproducible and high-throughput manner. Automatic analysis is particularly useful when it comes to quantifying the extent of tissue regions. For example, the “tumour-stroma-ratio” (area covered by tumour epithelium divided by area covered by stroma) proved to be an important prognostic factor in a number of neoplastic disorders2,44,45,46,47. Likewise, the invasion depth of CRC carries profound consequences for the affected patients, but may be difficult to assess in some cases (e.g. when single tumour glands invade much deeper than the tumour bulk). Invasion depth could be automatically quantified by multiclass texture analysis. Another application could be automatic tumour grading, i.e. classification of tumour architecture into G1, G2 or G3. Today, this task is typically done manually (and therefore not always reproducible). Also, multiclass texture analysis could be applied to immunostained images in order to classify distribution patterns of a specific antigen. Another possible application of multiclass texture analysis would be to characterize the morphology of the invasive tumour margin that has been shown to be a powerful prognostic factor for patient survival48.
In addition to these technical advances, our texture analysis approach could be used to investigate biological hypotheses based on tissue morphology. For example, stroma tissue has a very heterogeneous morphology (as can be seen in Fig. 1b,c). There is no clear morphological definition of different stroma subtypes (e.g. normal stroma vs. tumour stroma). Multiclass texture analysis could be used to identify morphologically consistent stroma subtypes and investigate biological implications of these subtypes (potentially leading to new morphological biomarkers).
How to cite this article: Kather, J. N. et al. Multi-class texture analysis in colorectal cancer histology. Sci. Rep. 6, 27988; doi: 10.1038/srep27988 (2016).
Egeblad, M., Nakasone, E. S. & Werb, Z. Tumors as organs: complex tissues that interface with the entire organism. Dev Cell 18, 884–901 (2010).
Huijbers, A. et al. The proportion of tumor-stroma as a strong prognosticator for stage II and III colon cancer patients: validation in the VICTOR trial. Ann Oncol 24, 179–85 (2013).
Xu, J. et al. Stacked Sparse Autoencoder (SSAE) for Nuclei Detection on Breast Cancer Histopathology images. IEEE T Med Imaging 35, 119–30 (2015).
Lan, C. et al. Quantitative histology analysis of the ovarian tumour microenvironment. Sci Rep 5, 16317 (2015).
Ginsburg, S. B., Lee, G., Ali, S. & Madabhushi, A. Feature Importance in Nonlinear Embeddings (FINE): Applications in Digital Pathology. IEEE T Med Imaging 35, 76–88 (2015).
Yuan, Y. et al. Quantitative image analysis of cellular heterogeneity in breast tumors complements genomic profiling. Sci Transl Med 4, 157ra143 (2012).
Tamura, H., Mori, S. & Yamawaki, T. Textural Features Corresponding to Visual Perception. IEEE T Syst Man Cyb 8, 460–473 (1978).
Bianconi, F., Álvarez-Larrán, A. & Fernández, A. Discrimination between tumour epithelium and stroma via perception-based features. Neurocomputing 154, 119–126 (2015).
Haralick, R., Shanmugan, K. & Dinstein, I. Textural features for image classification. IEEE T Syst Man Cyb 3, 610–621 (1973).
Schad, L. R., Härle, W., Zuna, I. & Lorenz, W. J. Magnetic Resonance Imaging of Intracranial Tumors: Tissue Characterization by Means of Texture Analysis. Z Med Phys 2, 12–17 (1992).
Lerski, R. & Schad, L. The use of reticulated foam in texture test objects for magnetic resonance imaging. Magn Reson Imaging 16, 1139–1144 (1998).
Linder, N. et al. Identification of tumor epithelium and stroma in tissue microarrays using texture analysis. Diagn Pathol 7, 22 (2012).
Turkki, R. et al. Assessment of tumour viability in human lung cancer xenografts with texture-based image analysis. J Clin Pathol 68, 614–21 (2015).
Barker, J., Hoogi, A., Depeursinge, A. & Rubin, D. L. Automated Classification of Brain Tumor Type in Whole-Slide Digital Pathology Images Using Local Representative Tiles. Med Imag Anal 30, 60–71 (2015).
Zöllner, F. G., Emblem, K. E. & Schad, L. R. Support vector machines in DSC-based glioma imaging: suggestions for optimal characterization. Magn Reson Med 64, 1230–6 (2010).
Zöllner, F. G., Emblem, K. E. & Schad, L. R. SVM-based glioma grading: Optimization by feature reduction analysis. Z Med Phys 22, 205–14 (2012).
Phillips, P. J., Wechsler, H., Huang, J. & Rauss, P. J. The FERET database and evaluation procedure for face-recognition algorithms. Image Vis Comput 16, 295–306 (1998).
Marti, U.-V. & Bunke, H. The IAM-database: an English sentence database for offline handwriting recognition. Int J Doc Anal Recognit 5, 39–46 (2002).
Xiao, J., Hays, J., Ehinger, K. A., Oliva, A. & Torralba, A. SUN database: Large-scale scene recognition from abbey to zoo. in 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition 3485–3492 (IEEE, 2010). doi:10.1109/CVPR.2010.5539970.
Hossain, S. & Serikawa, S. Texture databases-A comprehensive survey. Pattern Recognit Lett 34, 2007–2022 (2013).
Bianconi, F. & Fernández, A. An appendix to ‘texture databases-A comprehensive survey’. Pattern Recognit Lett 45, 33–38 (2014).
Kather, J. et al. Continuous representation of tumor microvessel density and detection of angiogenic hotspots in histological whole-slide images. Oncotarget 6, 19163–19176 (2015).
Julesz, B. Textons, the elements of texture perception and their interactions. Nature 290, 91–97 (1981).
Beyerer, J., Puente León, F. & Frese, C. Texture Analysis in Machine Vision: Automated Visual Inspection: Theory, Practice and Applications 649–683 (Springer Berlin Heidelberg, 2015).
Pietikäinen, M., Hadid, A., Zhao, G. & Ahonen, T. Computer Vision Using Local Binary Patterns. doi: 10.1007/978-0-85729-748-8 (Springer London, 2011).
Ojala, T., Pietikäinen, M. & Mäenpää, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE T Pattern Anal 24, 971–987 (2002).
Ahonen, T., Matas, J., He, C. & Pietikäinen, M. Rotation Invariant Image Description with Local Binary Pattern Histogram Fourier Features in Lecture Notes in Computer Science 5575, 61–70 (Springer Berlin Heidelberg, 2009).
Tsia, D. M., Wu, S. K. & Chen, M. C. Optimal Gabor filters for texture segmentation using stochastic optimization. Image Vis Comput 19, 299–316 (2001).
Garcia, S., Derrac, J., Cano, J. R. & Herrera, F. Prototype selection for nearest neighbor classification: Taxonomy and empirical study. IEEE T Pattern Anal 34, 417–435 (2012).
Allwein, E. L., Schapire, R. E. & Singer, Y. Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers. J Mach Learn Res 1, 113–141 (2000).
Escalera, S., Pujol, O. & Radeva, P. On the decoding process in ternary error-correcting output codes. IEEE T Pattern Anal 32, 120–134 (2010).
Seiffert, C., Khoshgoftaar, T., Hulse, J. & Napolitano, A. RUSBoost: Improving clasification performance when training data is skewed. in 19th International Conference on Pattern Recognition 1–4 (2008).
Xie, X. & Mirmehdi, M. A Galaxy of Texture Features in Handbook of Texture Analysis (eds. Mirmehdi, M., Xie, X. & Suri, J. ) 375–407 (Imperial College Press, 2008).
Bianconi, F. & Fernandez, A. Rotation invariant co-occurrence features based on digital circles and discrete Fourier transform. Pattern Recogn Lett 48, 34–41 (2014).
Brahnam, S., Jain, L. C., Nanni, L. & Lumini, A. Local Binary Patterns: New Variants and Applications, doi: 10.1007/978-3-642-39289-4 (Springer Berlin Heidelberg, 2014).
Turner, M. R. Texture Discrimination by Gabor Functions. Biol Cybern 55, 71–82 (1986).
Kather, J. N. et al. New Colors for Histology: Optimized Bivariate Color Maps Increase Perceptual Contrast in Histological Images. PLoS One 10, e0145572 (2015).
Epstein, J. I. An update of the Gleason grading system. J Urol 183, 433–40 (2010).
Mattfeldt, T., Grahovac, P. & Luck, S. Multiclass Pattern Recognition of the Gleason Score of Prostatic Carcinomas Using Methods of Spatial Statistics. Image Anal Stereol 32, 155–165 (2013).
Huang, P.-W. & Lee, C.-H. Automatic classification for pathological prostate images based on fractal analysis. IEEE T Med Imaging 28, 1037–1050 (2009).
Doyle, S., Feldman, M. D., Shih, N., Tomaszewski, J. & Madabhushi, A. Cascaded discrimination of normal, abnormal and confounder classes in histopathology: Gleason grading of prostate cancer. BMC Bioinformatics 13, 282 (2012).
Signolle, N., Revenu, M., Plancoulaine, B. & Herlin, P. Wavelet-based multiscale texture segmentation: Application to stromal compartment characterization on virtual slides. Signal Process 90, 2412–2422 (2010).
Yang, L. et al. Virtual microscopy and grid-enabled decision support for large-scale analysis of imaged pathology specimens. IEEE T Med Imaging 13, 636–644 (2009).
Downey, C. L. et al. The prognostic significance of tumour-stroma ratio in oestrogen receptor-positive breast cancer. Brit J Cancer 110, 1744–7 (2014).
Dekker, T. J. A. et al. Prognostic significance of the tumor-stroma ratio: validation study in node-negative premenopausal breast cancer patients from the EORTC perioperative chemotherapy (POP) trial (10854). Breast Cancer Res Tr 139, 371–9 (2013).
Liu, J. et al. Tumor-stroma ratio is an independent predictor for survival in early cervical carcinoma. Gynecol Oncol 132, 81–6 (2014).
Wang, K. et al. Tumor-stroma ratio is an independent predictor for survival in esophageal squamous cell carcinoma. J Thorac Oncol 7, 1457–61 (2012).
Caie, P. D., Turnbull, A. K., Farrington, S. M., Oniscu, A. & Harrison, D. J. Quantification of tumour budding, lymphatic vessel density and invasion through image analysis in colorectal cancer. J Transl Med 12, 156 (2014).
We are grateful to Prof. Schultz of the department of Neuroanatomy at Medical Faculty Mannheim for the cooperation. Also, we want to thank Ms. Menge and her team for expert technical assistance. Furthermore, we want to thank Prof. Matti Pietikäinen (University of Oulu, Finland) for his permission to use and redistribute his source code for Local Binary Pattern generation.
The authors declare no competing financial interests.
About this article
Cite this article
Kather, J., Weis, C., Bianconi, F. et al. Multi-class texture analysis in colorectal cancer histology. Sci Rep 6, 27988 (2016). https://doi.org/10.1038/srep27988
Multiclass histology image retrieval, classification using Riesz transform and local binary pattern features
Computer Methods in Biomechanics and Biomedical Engineering: Imaging & Visualization (2020)
Evolutionary Intelligence (2020)
International Journal of Swarm Intelligence Research (2020)
Label-free multiphoton imaging allows brain tumor recognition based on texture analysis—a study of 382 tumor patients
Neuro-Oncology Advances (2020)
IEEE Access (2020)