Abstract
Breast cancer is the most common cancer in women. Classification of cancer/non-cancer patients with clinical records requires high sensitivity and specificity for an acceptable diagnosis test. The state-of-the-art classification model—convolutional neural network (CNN), however, cannot be used with such kind of tabular clinical data that are represented in 1-D format. CNN has been designed to work on a set of 2-D matrices whose elements show some correlation with neighboring elements such as in image data. Conversely, the data examples represented as a set of 1-D vectors—apart from the time series data—cannot be used with CNN, but with other classification models such as Recurrent Neural Networks for tabular data or Random Forest. We have proposed three novel preprocessing methods of data wrangling that transform a 1-D data vector, to a 2-D graphical image with appropriate correlations among the fields to be processed on CNN. We tested our methods on Wisconsin Original Breast Cancer (WBC) and Wisconsin Diagnostic Breast Cancer (WDBC) datasets. To our knowledge, this work is novel on non-image tabular data to image data transformation for the non-time series data. The transformed data processed with CNN using VGGnet-16 shows competitive results for the WBC dataset and outperforms other known methods for the WDBC dataset.
Similar content being viewed by others
Introduction
In recent times, there are growing interest in the development of machine learning (ML) models for medical datasets due to the advancements in digital technology and improvements in data collection methods. Increasingly, several ML-based systems have been designed as an early warning or diagnostic tool for chronic illnesses, for example diagnosing depression, diabetes and cancer1. Breast cancer is arguably one of the deadliest forms of cancer amongst women with millions of reported cases around the world of which many cases become fatal2,3. Breast cancer is caused by abnormal growth of some of the breast cells in the lining of the milk glands or ducts of the breast (ductal epithelium)4,5. Compared to healthy cells, these cells divide more rapidly and accumulate, forming a lump or mass. At this stage, the cells become malignant and may spread through the breast to lymph nodes or other parts of the body.
The study of breast cancer has attracted considerable attention in the past decades. Improving data collection and storage technologies has resulted in various types and amounts of data collected on breast cancer from around the world. These include data on Ribonucleic Acid (RNA) signatures for cell mutations that cause breast cancer6,7, mammogram images8,9 and data on symptoms and diagnosis10. Many traditional Computer-Aided Diagnosis (CADx) systems require hand-crafted feature extraction which is a challenging task11,12. Even conventional ML techniques require the extraction of an optimal set of features manually prior to model training. An extensive review on various feature selection and extraction techniques can be found in13,14. Some commonly used approaches for ML models are Principal Component Analysis (PCA)15, information gain16, GA-based feature selection17, recursive feature elimination (RFE)18, meta-heuristic methods19 and rough sets20. Feature selection and extraction, therefore, is an important consideration in the pre-processing step before applying any ML algorithm such as decision trees, Bayesian models, Support Vector Machines (SVM) and Artificial Neural Networks (ANN). The behavior of ML algorithms and their prediction accuracy is influenced by the choice of features selected21,22. Many times manual feature extraction or knowledge of domain experts is needed to have a good understanding on the relevance of the attributes23.
To address these issues surrounding the use of conventional ML algorithms has propelled the need for new approaches and methods to automatically extract features from large datasets. As a result, Deep Learning (DL) algorithms such as convolutional neural network (CNN or ConvNet) and Recurrent Neural Networks (RNNs) have emerged in recent times that can accept raw data and are automatically able to discover patterns in them24,25.
CNN is one of the most popular algorithms for deep learning which is mostly used for image classification, natural language processing, and time series forecasting. Its ability to extract and recognize the fine features has led to the state-of-the-art performance in various application domains such as computer vision, image recognition, speech recognition, natural and language processing26,27,28. CNN is an enhancement of a canonical Neural Networks architecture that is specifically designed for image recognition in29. Since then many variations have been added to the architecture of CNN to enhance its ability to produce remarkable solutions for deep learning problems such as AlexNet26, VGG Net27 and GoogLeNet30. CNN eliminates the need for manual feature extraction because the features are learned directly by different convolutional layers26,31. It does not require a separate feature extraction strategy which requires domain expert and other preprocessing techniques where complete features may still not be extracted32. Despite its huge success with image data, CNN is not designed to handle tabular non-image data in non-time series form. Note that all future referencing of non-image data are in non-time series form unless otherwise specified. Arguably, any problem that can represent the correlation of features of a given data example in a single map, maybe attempted via CNN.
CNNs have proven to work best on data that are in 2-D form, such as images and audio spectrograms33. This is attributed to the fact that the convolution technique in CNN requires data examples to have at least two dimensions. Conversely, CNN has been explored on application-specific 1-D data as well. These include gene sequencing data such as DNA sequences being treated as text data (sequence of words)34, and signals and sequences in text mining, word detection and natural language processing (NLP)35,36. More specifically, CNN for Time-Series Classification (TSC) has been recently explored with some new methods such as Multi-Scale CNN (MCNN)25 and an ensemble of CNN models with AlexNet on Inception-v4 architecture37,38. These methods have made significant improvement in the accuracy of the classifiers with the state-of-the-art ensemble methods such as Flat-COTE and HIVE-COTE39,40. Moreover, raw time-series data has also been used into 1-D CNN by calculating the area of the signal for convolution with better time complexity and scalability41,42. Nonetheless, much data still exists in a 1-D format such as clinical data of medical records, and therefore, opens challenging research questions on whether they can be effectively trained for classification using CNN. This paper is aimed at filling this gap by proposing a novel non-time series 1-D numerical data to 2-D data transformation methods and processing them with CNN. This would certainly help machine learners to train their data without being bothered about issues with feature extraction. This can also reduce a large feature vector to just a single image.
This paper is organized as follows: Section “Motivation” demonstrates the theoretical motivation of the proposed method. Section “Proposed methods” describes our three proposed methods of data wrangling from non-image Breast Cancer tabular data10 to image data. Section “Experiments” describes the complete methodology of the classification of breast cancer data with CNN. Section “Results” shows the experimental results and Section “Discussion” discusses the outcome of the experiments. Lastly, Section “Conclusion” concludes the paper by summarizing the results and proposing some further extensions to the research.
Motivation
The main motivation for this paper is to realize the potential of CNN for non-image clinical data for breast cancer because it eliminates the need for manual feature extraction. The features are learned directly by CNN whereby it also produces state-of-the-art recognition results43. The key difference between traditional ML and DL is in how features are extracted. Traditional ML approaches use handcrafted engineering features by applying several feature extraction algorithms and then apply the learning algorithms. On the other hand, in the case of DL, the features are learned automatically and are represented hierarchically at multiple levels. This is the strong point of DL against traditional machine learning approaches43.
We have proposed three novel methods to transform non-image clinical tabular data of breast cancer to 2-D feature map images in \(\mathbb {R}^2\) so that a large set of these kinds of data are not deprived of the services of CNN. This would also encourage other variations and/or methods for text to image transformation to be developed in the future. The scope of this paper is to broaden the usage of CNN to those applications where d-dimensional raw data has set of N, 1-D data vectors in \(\mathbb {R}\) as shown in Fig. 1. Each row represents a 1-D data vector with d elements where d, N \(\ge 1\). It is a sample of a Wisconsin Original Breast Cancer dataset (WBC) used in the experiments. This dataset from UCI10 is a record of medical examination of patients to diagnose breast cancer, where each row is a 1-D vector representing a numerical data example. We demonstrate our method of non-image breast cancer data transformation to image data—processed in CNN—produces exceptional results for classification accuracy. Some research demonstrates the use of 1-D convolutions on 1D datasets such as data in the form of signals and time sequences44. Though this provides a possibility of using 1-D convolutions in this research, our experiments revealed their unsuitability on our experimental datasets. Having applied the data in its raw form into 1-D CNN gave highly unpredictable results.
Proposed methods
We have proposed three basic techniques of data wrangling to convert Breast Cancer numerical tabular data to image data. The converted image must reflect some patterns to depict a given class. We have used Wisconsin Original Breast Cancer (WBC) and Wisconsin Diagnostic Breast Cancer (WDBC) datasets from the UCI library10 for the classification of numerical data in this work.
Equidistant bar graphs
The bar graph represents the measurement of every feature of a given dataset. There are lots of possibilities of drawing a bar graph but we have used a simplistic approach. The dataset is first normalized to [0, 1] then every feature is drawn based on its measured value. The width of the image in pixels is \(\psi d+\gamma (d+1)\) where d is total features, \(\psi\) is the width of a bar and \(\gamma\) is gap between two consecutive bars. The height of the image is normalized to produce a square image. We used \(1-\)pixel length for \(\psi\) and \(2-\)pixels length for \(\gamma\) in our experiments. This produces the square image of size \([3d\times 3d]\) approximately. Few data examples of WDBC dataset converted to bar graphs are shown in Fig. 4a with class labels—Benign and Malignant. The algorithm for this approach is given in the Fig. 2 (Algorithm 1).
These pictures are only useful to CNN if they depict a pattern in a convolved image. The first convolutional layer produces 6 features which are shown in Fig. 4b where some sort of distinguishing features have been reflected.
Intuitively, the “correct” order of the bars ought to give better results. The datasets of numerical data were reorganized where the related fields were put close to each other according to the order of their similarity. Firstly, a covariance matrix on data fields was generated then each value of the matrix is converted to ‘rank’ that determines how closely one field is related to the other. This is a shortest-path problem where algorithms such as dynamic programming or any metaheuristic algorithm45 such as Genetic Algorithm (GA)46, Particle Swarm Optimization47 or Reincarnation Algorithm (RA)48 can be used to get the optimum order of bars based on their respective rank. Thereafter, a new set of images was created using this new order of bars. This process has been elaborated more in Section “Discussion”.
Normalized distance matrix
The next method is the formation of a distance matrix which is a squared matrix of size \([d\times d]\) where d represents total features of a given example. Matrix elements are the difference between two features i.e., \(x_{ij}=x_i-x_j\) where \(x_i\) and \(x_j\) represent the measurement of a given feature with \(i,j\in [1,d]\). We used Euclidean distance in our experiments. The matrix is then normalized between \([0-1]\). This produces the square image of size \([d \times d]\) which has a gain of 3 folds in length compared to bar graphs described in Section “Equidistant bar graphs”. Few data examples of WDBC dataset converted to normalized distance matrix are shown in Fig. 4c with class labels. The images can be easily scaled up to \([3d\times 3d]\). The first convolutional layer produces 6 features similar to bar graphs is shown in Fig. 4d. Its pseudocode and further description is given in Fig. 2 (Algorithm 2).
Combination of options (bar graph, distance matrix, normalized numeric data)
Apparently, the above two strategies can be combined to give a third option for generating an image from numerical data. We create a colored image of 3 layers of size \([3d\times 3d]\) where the first layer has a normalized distance matrix, the second layer has bar graphs, and the third layer has a copy of numerical data stored row-wise, i.e., \(x_{ij}=x_i\) where \(i,j\in [1,d]\) shows row and column of a matrix and \(x_i\) represents the measurement of a given feature. Few data examples of WDBC dataset converted to the combination of options are shown in Fig. 5a with the class labels.
The first convolutional layer in this case, is not able to produce any distinct feature but the scaled up image shows different colors with some bars in Fig. 5b. The 3rd convolved block (12th layer) produces some blobs scattered in the images in Fig. 5c.
Experiments
CNN completes the classification process in two steps. The first step is the auto-feature extraction of the images and the second step is the classification of the same images with backpropagation neural networks49,50. In the case of a numerical dataset that is not in the form of images, first goes through the data wrangling process described in Section “Proposed methods”, where either of the three options is used for non-image to image data conversion. The transformed images may not make logical sense to human eyes but CNN is capable to extract relevant features out of it. Figure 3 illustrates the complete flowchart of the training process of CNN with non-image data sets. The process contains four important parts: Firstly, numeric input data (A) undergoes pre-processing of data wrangling (B) where it is normalized and converted to 2D image format using one of the data wrangling techniques described in Section “Proposed methods” (the figure shows distance matrix method of Section “Normalized distance matrix”). The generated image is filtered through the CNN convolution layers for feature extraction (C). The features are trained in the fully connected layers to obtain classification outputs (D).
Results
The objective of the experiment is to provide an alternative classification method with CNN for the non-image dataset of Breast Cancer and other similar datasets without any need for manual feature selection. We have used WBC and WDBC datasets from the UCI library10 for the experiments. The properties of these datasets are given in Table 2. We have tested the efficacy of our method with other published state-of-the-art methods used for breast cancer diagnosis, namely, variations of Neural Networks (NN)51, Support Vector Machine (SVM)16,52,53, Decision Tree (DT)54 and Naïve Bayes (NB)55. These methods are generally supported by additional feature selection methods such as IG, Rough set or weight NB.
For CNN, we used VGG1627 architecture with 4 convolutional blocks. Each convolutional block has 2D convolutional layer with the filter size of \([3\times 3]\), \(0.5\times Layer\times \left| \root \of {\parallel image \parallel }\right|\) filters, ReLU layer and lastly max pooling layer with of pool size and stride of \([2\times 2]\). Additionally, Bayesian optimization was used for parameter tuning. All parameter settings are shown in Table 1. For regularization and initial learning rate we used log transformation.
Initially, both datasets are divided into \(80\%\) training and \(20\%\) testing then \(20\%\) of training data is kept aside for validation data. After 30 attempts on each dataset, we have collected best and average classification accuracies on validation and test data sets shown in Tables 3 and 4 respectively. Bold figures represent the overall best result. CNN types 1, 2 and 3 represent equidistant bar graph, normalized distant matrix, and combined options respectively. px1 shows that the image is formed with bars of 1-pixel width only. Similarly px2 and px4 show width of 2 and 4 pixel sizes respectively for bars in an image.
Additionally, it is highly desirable in medical diagnosis to have high sensitivity and specificity measures. Sensitivity is the ability of a test to correctly identify those with the disease, and specificity is correctly identifying those without the disease. Alternatively, the F1 score can be used as a derived metric that merges both sensitivity and precision measures. Tables 6 and 7 show the best and average of these additional metrics respectively, for WDBC and WBC datasets on classification. The confusion matrix for the best cases is shown in Table 5. We have also performed experiments using CNN with 1-D convolutions on raw data without any sophisticated data transformation. However, we have obtained poor results when compared to our method with the average classification accuracy of 76.11% and 89.64% for WDBC and WBC datasets respectively.
The comparison of our methods with other state-of-the-art methods is shown in Table 8. The table shows different methods from 2009–2019. The results show accuracy, sensitivity and specificity of WBC and/or WDBC datasets. Authors in11 have used mammogram images of breast cancer as CNN works on images. In some cases, authors got 100% accuracy with 10-fold cross-validation for WBC dataset. Lower fold of cross-validation generally gives lower accuracy16,51,52.
Discussion
The experimental results of data transformation from non-image tabular breast cancer datasets to image have been promising for the utilization of CNN for classification accuracy. Although the proposed methods are in the early stages, the obtained results are very significant in the development of new strategies with data wrangling for deep learning. This also provides an opportunity to derive even better alternatives for CNN in the future. It was observed that our proposed combined approach, i.e. Type-3 transformation and bar width of 1 pixel i.e. px1, has been the most significant method as it carries the most information about the data in three dimensions of an image. It has outperformed other methods for the WDBC dataset by clocking 100% accuracy (with 1.0 sensitivity, specificity and F1 score). It has also shown very competitive results for the WBC dataset with 99.27% accuracy and 1.0 sensitivity 0.99 specificity and 0.99 F1 score.
As discussed in Section “Proposed methods”, different order of bar graphs for Type-1 and Type-3 transformations produce different images. A bar represents its corresponding field value of a given sample. We have tried to bring the related bars closer to each other by using a covariance matrix that determines the “closeness” of two fields. For example Fig. 6a shows the Adjacency Matrix of co-variance of each field for WBC dataset. The data is arranged row-wise such that each value represents the rank of ith row with jth column of a given field. To get the “best” arrangement of fields, we minimize the total co-variance rank by using a meta-heuristic algorithm GA to solve this shortest path problem. The process of minimization for WDBC is shown in Fig. 6b where the minimum rank is obtained by the end of 10th generation. The dataset fields were reorganized where the related fields were put close to each other according to the order of their similarity. The final order of fields for WBC and WDBC produced through minimum ranks are shown in Table 9. The images of these datasets were generated accordingly for the experiment. Notably, this order of fields does not have significant improvement over the original arrangement as the CNN produces similar convolved images.
The only shortcoming of the CNN algorithm is its high processing cost than other methods, especially with bigger sized images. Generally, it takes 9–15 s for a MATLAB 2018 program to complete the training process on DELL XPS i7-9700 @ 3GHz machine with 8 CPUs and NVIDIA GEFORCE RTX 2060 GPU. Despite this, the experimental results demonstrate the size of data has no direct impact on the performance of CNN. Additionally, with the advent of quantum computing56 and parallel GPUs with enough memory can produce results in a reasonable time frame. The data wrangling process of converting non-image data to the image is not too expensive either. The every-case time complexity of the bar graph approach has the order of O(Nd) and the normalized distance matrix has the order of \(O(Nd^2)\). The Matlab code and data is available at https://github.com/anuraganands.
Conclusion
The objective of this paper was to process non-image data (in a non-time series form) of Breast Cancer datasets WDBC and WBC into CNN due to its state-of-the-art performance and elimination of manual feature extraction for image recognition applications. The utilization of CNN has been confined largely to image data only except for some domain-specific data conversion techniques such as NLP and voice recognition. We have proposed three novel approaches to convert numerical non-time series data to image data. This process of conversion is very straightforward with the efficiency of the order of not more than \(O(Nd^2)\). The experimental results on classification accuracy show the competitiveness of these methods. There is also a high potential for improving these approaches further to have more outstanding results. For example, bar graphs with different shapes, sizes, color and even arrangements can be tried. Similarly, distance matrix can be enhanced to have more information such as the mean/variance of the neighboring elements. It still needs to be seen how other applications with various types and orientations of numerical data would respond to CNN after non-image data conversion to image data. Intuitively, the more the information on data would produce the better the results as observed with the combined approach. Moreover, the imminent future work is to try our methods on time-series data to have competitive results with its counterpart of 1-D transformation. Finally, the classification accuracy of numerical data without any sophisticated data transformation on 1-D CNN did not produce acceptable results.
Data availability
The datasets analysed during the current study are available in the UCI repository, [https://archive.ics.uci.edu/ml/datasets/breast+cancer+wisconsin+(diagnostic) and https://archive.ics.uci.edu/ml/datasets/breast+cancer+wisconsin+(original)].
References
Sourla, E., Sioutas, S., Syrimpeis, V., Tsakalidis, A. & Tzimas, G. Cardiosmart365: Artificial intelligence in the service of cardiologic patients. Adv. Artif. Intell. 2012, 2 (2012).
Gao, F. et al. SD-CNN: A shallow-deep CNN for improved breast cancer diagnosis. Comput. Med. Imaging Gr. 70, 53–62 (2018).
Tsai, M. L. et al. Effects of germline pathogenic variants, cancer subtypes, tumor-related characteristics, and pregnancy-associated diagnosis on outcomes. Clin. Breast Cancer. 21, 47–56 (2020).
Breast cancer—Latest research and news | Nature.
Breast cancer | definition of breast cancer by Medical dictionary.
Kaur, P., Porras, T. B., Ring, A., Carpten, J. D. & Lang, J. E. Comparison of TCGA and GENIE genomic datasets for the detection of clinically actionable alterations in breast cancer. Sci. Rep. 9, 1–15 (2019).
Larsen, M. J., Thomassen, M., Tan, Q., Sørensen, K. P. & Kruse, T. A. Microarray-based RNA profiling of breast cancer: Batch effect removal improves cross-platform consistency. BioMed Res. Int. 2014, 1–11 (2014).
Dembrower, K., Lindholm, P. & Strand, F. A multi-million mammography image dataset and population-based screening cohort for the training and evaluation of deep neural networks-the cohort of screen-aged women (CSAW). J. Digit. Imaging. 33, 408–413 (2020).
Bowyer, K. et al. The digital database for screening mammography. In Third International Workshop on Digital Mammography Vol. 58 27 (1996).
Dheeru, D. & Karra Taniskidou, E. UCI Machine Learning Repository (University of California, Irvine, School of Information and Computer Sciences, 2019).
Sun, W., Tseng, T.-L.B., Zhang, J. & Qian, W. Enhancing deep convolutional neural network scheme for breast cancer diagnosis with unlabeled data. Comput. Med. Imaging Gr. 57, 4–9 (2017).
Firmino, M., Angelo, G., Morais, H., Dantas, M. R. & Valentim, R. Computer-aided detection (CADe) and diagnosis (CADx) system for lung cancer with likelihood of malignancy. Biomed. Eng. Online 15, 2. https://doi.org/10.1186/s12938-015-0120-7 (2016).
Guyon, I. & Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003).
Kumar, V. & Minz, S. Feature selection: A literature review. SmartCR 4, 211–229 (2014).
Fodor, I. K. A Survey of Dimension Reduction Techniques. Tech. Rep., Lawrence Livermore National Lab., CA (US) (2002).
Liu, N., Qi, E.-S., Xu, M., Gao, B. & Liu, G.-Q. A novel intelligent classification model for breast cancer diagnosis. Inf. Process. Manag. 56, 609–623. https://doi.org/10.1016/j.ipm.2018.10.014 (2019).
Babatunde, O. H., Armstrong, L., Leng, J. & Diepeveen, D. A genetic algorithm-based feature selection. Int. J. Electron. Commun. Comput. Eng. (IJECCE). 5, 899–905 (2014).
Darst, B. F., Malecki, K. C. & Engelman, C. D. Using recursive feature elimination in random forest to account for correlated variables in high dimensional data. BMC Genet. 19, 65 (2018).
Sharma, M. & Kaur, P. A comprehensive analysis of nature-inspired meta-heuristic techniques for feature selection problem. Arch. Comput. Methods Eng. 28, 1103–1127 (2021).
Pawlak, Z. Rough Sets: Theoretical Aspects of Reasoning about Data (Springer Science & Business Media, 2012). Google-Books-ID: yeOoCAAAQBAJ.
Guyon, I. & Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003).
Singh, R. K. & SivaBalakrishnan, M. Feature selection of gene expression data for cancer classification: A review. In 2nd International Symposium on Big Data and Cloud Computing 52–57 (2015).
Mohamad, M. S., Deris, S., Yatim, S. M. & Othman, M. R. Feature selection method using genetic algorithm for the classification of small and high dimension data. In First International Symposium on Information and Communication Technologies (2004).
Kumar, D. & Sharma, D. Deep Learning in Gene Expression Modeling. in Handbook of Deep Learning Applications Vol. 136 (eds. Balas, V.etal.) 363–383 (Smart Innovation, Systems and Technologies, Springer, 2019).
Cui, Z., Chen, W. & Chen, Y. Multi-Scale Convolutional Neural Networks for Time Series Classification. arXiv:1603.06995 [cs] (2016). ArXiv: 1603.06995.
Krizhevsky, A., Sutskever, I. & Hinton, G. E. ImageNet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems Vol. 25 (eds Pereira, F. et al.) 1097–1105 (Curran Associates Inc., 2012).
Simonyan, K. & Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv:1409.1556 [cs] (2014). arXiv: 1409.1556.
Volokitin, A., Roig, G. & Poggio, T. A. Do deep neural networks suffer from crowding? In Advances in Neural Information Processing Systems Vol. 30 (eds Guyon, I. et al.) 5628–5638 (Curran Associates Inc., 2017).
LeCun, Y. et al. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1, 541–551. https://doi.org/10.1162/neco.1989.1.4.541 (1989).
Szegedy, C. et al. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 1–9 (2015).
Guo, T., Dong, J., Li, H. & Gao, Y. Simple convolutional neural network on image classification. In 2017 IEEE 2nd International Conference on Big Data Analysis (ICBDA) 721–724, https://doi.org/10.1109/ICBDA.2017.8078730 (2017).
Indolia, S., Goswami, A. K., Mishra, S. P. & Asopa, P. Conceptual understanding of convolutional neural network: A deep learning approach. Procedia Comput. Sci. 132, 679–688. https://doi.org/10.1016/j.procs.2018.05.069 (2018).
Li, W., Victor, B., Xiao, L. & Chen, H. Deep Learning: An Overview—Lecture Notes. https://studylib.net/doc/15672646/deep-learning-an-overview-university-of-arizona-1 (2015). [Online; accessed 10-Jan-2020].
Nguyen, N. G. et al. DNA sequence classification by convolutional neural network. J. Biomed. Sci. Eng. 9, 280 (2016).
Delakis, M. & Garcia, C. Text detection with convolutional neural networks. In VISAPP Vol. 2 290–294 (2008).
Xu, H. & Su, F. Robust seed localization and growing with deep convolutional features for scene text detection. In Proceedings of the 5th ACM on International Conference on Multimedia Retrieval, 387–394 (ACM, 2015).
Szegedy, C., Ioffe, S., Vanhoucke, V. & Alemi, A. A. Inception-v4, inception-ResNet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI’17, 4278–4284 (AAAI Press, San Francisco, California, USA, 2017).
Fawaz, H. I. et al. InceptionTime: Finding AlexNet for Time Series Classification. arXiv:1909.04939 [cs, stat] (2019). arXiv: 1909.04939 version: 2.
Lines, J., Taylor, S. & Bagnall, A. HIVE-COTE: The Hierarchical Vote Collective of Transformation-Based Ensembles for Time Series Classification. In 2016 IEEE 16th International Conference on Data Mining (ICDM) 1041–1046, https://doi.org/10.1109/ICDM.2016.0133 (2016). ISSN: 2374-8486.
Bagnall, A., Lines, J., Hills, J. & Bostrom, A. Time-series classification with COTE: The collective of transformation-based ensembles. IEEE Trans. Knowl. Data Eng. 27, 2522–2535. https://doi.org/10.1109/TKDE.2015.2416723 (2015).
Brownlee, J. Deep Learning for Time Series Forecasting: Predict the Future with MLPs, CNNs and LSTMs in Python (Machine Learning Mastery, 2018). Google-Books-ID: o5qnDwAAQBAJ.
Janos, N. & Roach, J. 1D Convolutional Neural Networks for Time Series Modeling—Nathan Ja (2020). Library Catalog: SlideShare.
Alom, M. Z. et al. A state-of-the-art survey on deep learning theory and architectures. Electronics 8, 292. https://doi.org/10.3390/electronics8030292 (2019).
Xiong, Z., Stiles, M. K. & Zhao, J. Robust ECG signal classification for detection of atrial fibrillation using a novel neural network. In 2017 Computing in Cardiology (CinC) 1–4 (2017).
Almufti, S. M. Historical survey on metaheuristics algorithms. Int. J. Sci. World 7, 1–12. https://doi.org/10.14419/ijsw.v7i1.29497 (2019).
Goldberg, D. E. Genetic Algorithms in Search, Optimization, and Machine Learning 1st edn. (Addison-Wesley Longman Publishing Co., Inc., Boston, MA, USA, 1989).
Eberhart, R. & Kennedy, J. A new optimizer using particle swarm theory. In Micro Machine and Human Science, 1995. MHS ’95., Proceedings of the Sixth International Symposium on 39 –43, https://doi.org/10.1109/MHS.1995.494215 (1995).
Sharma, A. A new optimizing algorithm using reincarnation concept. In 11th IEEE International Symposium on Computational Intelligence and Informatics (CINTI) 281 –288, https://doi.org/10.1109/CINTI.2010.5672231 (2010).
Khan, S. et al. A Guide to Convolutional Neural Networks for Computer Vision (Morgan & Claypool, 2018).
Saha, S. A Comprehensive Guide to Convolutional Neural Networks—The ELI5 Way (2018).
Bhardwaj, A. & Tiwari, A. Breast cancer diagnosis using Genetically Optimized Neural Network model. Expert Syst. Appl. 42, 4611–4620. https://doi.org/10.1016/j.eswa.2015.01.065 (2015).
Chen, H.-L., Yang, B., Liu, J. & Liu, D.-Y. A support vector machine classifier with rough set-based feature selection for breast cancer diagnosis. Expert Syst. Appl. 38, 9014–9022. https://doi.org/10.1016/j.eswa.2011.01.120 (2011).
Zheng, B., Yoon, S. W. & Lam, S. S. Breast cancer diagnosis based on feature extraction using a hybrid of K-means and support vector machine algorithms. Expert Syst. Appl. 41, 1476–1482. https://doi.org/10.1016/j.eswa.2013.08.044 (2014).
Liu, Y.-Q., Wang, C. & Zhang, L. Decision tree based predictive models for breast cancer survivability on imbalanced data. In 2009 3rd International Conference on Bioinformatics and Biomedical Engineering 1–4, https://doi.org/10.1109/ICBBE.2009.5162571 (2009). ISSN: 2151-7622.
Karabatak, M. A new classifier for breast cancer detection based on Naïve Bayesian. Measurement 72, 32–36. https://doi.org/10.1016/j.measurement.2015.04.028 (2015).
Arute, F. et al. Quantum supremacy using a programmable superconducting processor. Nature 574, 505–510. https://doi.org/10.1038/s41586-019-1666-5 (2019).
Author information
Authors and Affiliations
Contributions
A.S. conceived and presented the method, conducted the experiments, analysed the results and wrote the manuscript. D.S. wrote the introduction and provided graphics support to describe the process of the proposed method. Both authors reviewed the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Sharma, A., Kumar, D. Classification with 2-D convolutional neural networks for breast cancer diagnosis. Sci Rep 12, 21857 (2022). https://doi.org/10.1038/s41598-022-26378-6
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-022-26378-6
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.