Abstract
Breast adenocarcinoma is the most common of all cancers that occur in women. According to the United States of America survey, more than 282,000 breast cancer patients are registered each year; most of them are women. Detection of cancer at its early stage saves many lives. Each cell contains the genetic code in the form of gene sequences. Changes in the gene sequences may lead to cancer. Replication and/or recombination in the gene base sometimes lead to a permanent change in the nucleotide sequence of the genome, called a mutation. Cancer driver mutations can lead to cancer. The proposed study develops a framework for the early detection of breast adenocarcinoma using machine learning techniques. Every gene has a specific sequence of nucleotides. A total of 99 genes are identified in various studies whose mutations can lead to breast adenocarcinoma. This study uses the dataset taken from 4127 human samples, including men and women from more than 12 cohorts. A total of 6170 mutations in gene sequences are used in this study. Decision Tree, Random Forest, and Gaussian Naïve Bayes are applied to these gene sequences using three evaluation methods: independent set testing, self-consistency testing, and tenfold cross-validation testing. Evaluation metrics such as accuracy, specificity, sensitivity, and Mathew’s correlation coefficient are calculated. The decision tree algorithm obtains the best accuracy of 99% for each evaluation method.
Similar content being viewed by others
Introduction
Machine Learning plays a phenomenal role in solving many crucial issues in various fields of life. Adenocarcinoma is a type of cancer that begins in secretary cells. Breast Adenocarcinoma is the abnormal and uncontrolled growth of cells in the breast gland. It is the second most severe cancer among all the cancers present in the human body.
It mostly occurs in women. An estimated 0.3 million women are diagnosed with breast cancer each year in the United States of America. In 2021 estimated 44,130 deaths (43,600 women and 530 men) occurred due to breast cancer in the United States1. There are several reasons for breast cancer in women. Some are getting older with age, having a family breast cancer history, having a child after 35, starting menopause after 55, having high bone density, etc.
A Biopsy is a primary technique used for the detection of breast adenocarcinoma. It is the technique in which a small tissue is examined under a microscope2. The Artificial intelligence (AI) approach has potential effects in the field of medical science. Several AI techniques are used in the medical science field to detect various diseases inside the human body. In this study machine learning approach is used to detect breast cancer at its early stage. There are sequences of nucleotides in the human gene. Any change in the sequence is called a mutation, which mostly leads to cancer3. Figure 1 illustrates the process of mutation.
DNA is a sequence of 4 basic nucleotides Adenine (A), Guanine (G), Thymine (T), and Cytosine (C)4. Any change in the base sequence in genes leads to mutation. This change may be caused by insertion, deletion, or replication of the gene base and may cause damage to DNA. Different factors affect DNA. These factors include metabolic activities or environmental factors such as radiation, resulting in tens of thousands of individual DNA damage per cell per day5. The DNA molecule’s damage alters or eliminates the cell’s ability to transcribe the gene. DNA repair is when a cell identifies and corrects damage to the DNA6. This process is constantly active as it responds to damage in the DNA structure. When normal repair processes fail, the cellular apoptosis is disrupted, and DNA damage may not be repairable. This irreparable damage leads to malignant tumors or cancer as per the two-hit hypothesis7,8.
This study uses a machine learning framework for the identification of breast adenocarcinoma. Three machine learning algorithms: Decision tree, Gaussian Naïve Bayes, and Random Forest, are applied to three evaluation methods: self-consistency test, independent set test, and tenfold cross-validation test. After using the machine learning algorithms on evaluation methods, accuracy, specificity, sensitivity, and Mathew’s correlation coefficient is calculated, explained in the results and discussion section.
Literature review
Breast cancer is the second major cause of death in women. In 2021 estimated 44,130 deaths occurred due to breast cancer in the United States9. Breast cancer was discovered in the early 400 s B.C.E10,11. Breast cancer develops from breast tissue12. Breast cancer is a genetic disorder, and the development of breast cancer has a genetic component13,14,15,16. Breast adenocarcinoma develops in cells from the lining of milk ducts and lobules. Lobules supply these ducts with milk12. Breast cancer is more common in developed countries13. Oncogenomics mutations lead to the uncontrollable growth of cancer cells. Although every mutation in the sequence doesn’t cause cancer, every cell growth is not cancerous. The interruption in the balance of creating cells and destruction of cells causes the beginning of cancer. This happens because of the change in the functional characteristics of genes. Cancer driver genes drive the development of cancer. Therefore, the mutation caused in driver genes commonly leads to cancerous mutation while passenger mutations do not cause cancer2,17. Driver and passenger mutations have been identified by various researchers based on clinical data. A well-known database of cancer driver genes called IntoGen reports that there are 99 driver genes that can cause breast adenocarcinoma which is a malignant tumor18. Bioinformatics plays a vital role in the field of medical sciences. There are many machine learning and computational technologies used in the area of medicine for the detection and prevention of various diseases. A study conducted by Botstein et al. used a semi-supervised approach in 2000 to identify the subtypes of breast cancer but initialing curating a database of the genes involved in breast cancer19. There are many machine learning algorithms developed for breast adenocarcinoma detection and identification. In research, the leveraging Machine Learning algorithm is applied to the dataset of 683 patients taken from the M. G Cancer Hospital & Research Institute, Visakhapatnam, India. The dataset is preprocessed using Gaussian filters, and then the features are extracted20. The accuracy in detecting breast cancer by the Deep Neural Network with Support Value (DNNS) model was 97.21%. Researchers have employed different machine learning algorithms, including artificial neural networks (ANNs), support vector machines (SVMs), decision trees (DTs), and k-nearest neighbors (k-NNs) applied to the Wisconsin breast cancer database (WBCD) dataset for the detection of breast cancer21. Data mining also plays an important role in the detection of breast cancer. Data mining techniques are applied to Decision tree, Naïve Bayes, and Sequential Minimal Optimization algorithms9. Subsequently, random forest (RF), k-NNs, and Naïve Bayes (NB) algorithms were applied to the WBDC dataset. The accuracy of k-NNs, Random forest, and NB was 95.90%, 94.74% 94.47%, respectively. Figure 2 explains the performance measure for this work22.
Another technique used is the Fast Correlation-Based Filter (FCBF) method to predict and classify breast cancer. Five machine learning algorithms are applied in RF, SVM, k-NNs, NB, and multilayer perceptron (MLP). The highest accuracy for this system is given by SVM. The Accuracy of SVM is 97.9%23. The extended form of the Naïve Bayes algorithm, the Weighted Naïve Bayes algorithm, is applied to the UCI machine learning breast cancer dataset for breast cancer prediction. The accuracy of this model was 92%24. Another similar study was conducted in 202125, which implemented RF, SVM, and ANN and achieved accuracy, sensitivity, and specificity in an independent set test such as 91.06%, 81.27, and 87.26, respectively.
Methodology
This study uses machine learning techniques for the detection of Breast Adenocarcinoma. Different machine learning algorithms are involved in the study to identify cancer. The systematic diagram of the proposed system is shown the Fig. 3.
Figure 3 explains the working of the complete process step by step. Decision tree, Random Forest, and Gaussian Naïve Bayes are used in each evaluation method to identify mutation for detecting breast adenocarcinoma. Researchers can use the proposed framework to develop an early warning diagnostic system based on genomic data. It will enable oncologists to detect and treat breast adenocarcinoma more personalized. The following sections explain the algorithms in detail with their testing methods and ROC curves.
Benchmark dataset collection
The dataset is the most critical factor for any bioinformatics related study. Typically, the dataset is used for training, testing, and validation. This study aims to use a high-quality benchmark dataset that is highly accurate and relevant to the study to obtain the best results. A meaningful dataset of the Breast Adenocarcinoma driver gene sequences is selected. The normal gene sequences are taken from https://asia.ensembl.org/26. Mutation pieces of information are taken from the most recent version available on http://intogen.org/18. IntOgen database does not contain mutated sequences. It has only mutation information. Therefore an application is developed in python to incorporate this information in normal gene sequences, taken from https://asia.ensembl.org/, to construct mutated sequences. The passenger mutations are not carcinogenic; therefore, these are considered normal sequences. Driver mutations are carcinogenic mutations. For the proposed study, 4127 human samples are used with a total of 6170 mutations in a total of 99 genes involved in breast adenocarcinoma. Genes involved in Breast Adenocarcinoma are shown in Table 1.
Word cloud is a visualization technique in python to represent text data. The size of each word indicates its frequency and importance27. The word cloud in Fig. 4 shows the frequency and importance of each nucleotide in all gene sequences related to breast adenocarcinoma.
Synthetic minority over-sampling technique (SMOTE)
The SMOTE technique balances the dataset. An unbalanced dataset is a dataset in which classification is not equally represented. There are two standard techniques used to balance datasets oversampling and undersampling. In the under-sampling technique, the number of classes reduces to balance the dataset. The overall data records are reduced. While in the over-sampling technique, the number of minority classes is increased. Smote is an oversampling technique for balancing the dataset. SMOTE randomly selects the instances from the minority class. It uses the interpolation method to generate instances between the selected point and its nearby instances.
The steps involved in SMOTE algorithm are as follow28:
-
1.
Insert dataset and mark minority and majority classes from it.
-
2.
Calculate the number of instances generated from the percentage of oversampling.
-
3.
From minority classes, identify random instance \(K\) and find its neighbors \(N\).
-
4.
From any neighbors, find the difference between \(N\) and \(K\).
-
5.
Multiply the difference with any number between 0 to 1 and add this difference to \(K\).
-
6.
Repeat the process until the required number of instances are generated.
Figure 5 explains creating synthetic data points in SMOTE29.
The dataset for the proposed study is represented by a \(B\) defined by Eq. (1).
Here \(B+\) are the mutated gene sequences that cause cancer while \(B-\) are the normal gene sequences, and U is the union for both sequences.
Feature extraction
Here H defines the gene sequence25.
The following Eqs. (2) and (3) calculate Hahne’s was polynomial.
Here \(P\) is an integer value from any Q, \(Q-1\) positive integers31. Hahn moment for two-dimension data is found by Eq. (3).
The raw moment is used for data imputation. Imputation replaces the missing data values in the dataset with most substitute values to preserve the information32. The raw moment for the 2D data with order \(a+b\) is expressed by Eq. (4)33.
Centroids \((r, s)\) are required to compute the central moments visualized as the center of data. By exploiting the centroids, central moments can be computed as.
Position Relative incidence matrix (PRIM) is used to determine each gene’s positioning in the gene sequence of breast adenocarcinoma. PRIM formed matrix with the dimension of 20 by 20 is shown in Eq. (4)34.
Feature scaling allows each data sample to participate in detecting breast cancer30. In machine learning, the algorithm is considered more efficient in which the most relevant data has been extracted. PRIM did not extract all the information from the data. Reverse Position Relative incidence matrix (RPRIM) also works the same as PRIM works but in the reverse sequence.
The frequency matrix provides information about the occurrence of genes in the gene sequence. The accumulative absolute position incidence vector (AAPIV) includes information about the composition of the gene sequence. The relative positioning of the cancer gene is found by using AAPIV. Equation (7) illustrates the relative positioning of the gene sequences35.
The reverse Accumulative absolute position incidence vector (RAAPIV) works the same as AAPIV works but in the reverse order. The Eq. (8) for RAAPIV is as follows
Prediction algorithms
This study uses a Decision Tree, Gaussian Naïve Bayes, and Random Forest Classifier for the prediction of Breast Adenocarcinoma.
The decision tree is a supervised machine learning technique. It is mostly used for classification and regression problems. In a decision tree root, nodes can be used as input. These nodes are filtered through decision nodes and leaf nodes used for getting desired output35,36,37. Entropy controls how data will be split in the decision tree, and information gain tells how much information a feature gives about the respective class. Equations (9) and (10) explain the formula for calculating Entropy and information gain in the decision tree38.
In the decision tree, the data flow in nodes. Figure 6 explains the working of the decision tree algorithm39.
The Naive Bayes algorithm is mostly used in data mining algorithms based on the Bayes theorem and uses simple probabilities. The Eq. (11) of Bayes theorem is as follows40.
Here P refers to probability, and Y is the attribute of a class. Figure 7 explains the Naïve Bayes classification41.
The algorithm for NB is
Here \({D}_{t}\) is the set of training examples, \(i\) is the instance, and \({X}_{i}\) is the random variable42. It is an easy algorithm used in many fields of medical science43.
Random Forest (RF) is the third algorithm applied to all the evaluation methods. It is the collection of the tree predictions which use different data for different techniques, and each technique leads to a different result. It is the ensemble learning method for regression and classification by constructing a multitude of decision trees44. The result is merged to represent the average result. Figure 8 illustrates the working of Random forest algorithms45.
MSE measures the average of the square of errors. It is the difference between the actual values and calculated values. The mean square error in RF is measured by Eq. (12).
In the equation \({({f}_{1}-{y}_{1})}^{2}\), is the square of errors.
Where \({y}_{1}\) is the predicted values and \({f}_{1}\) is the actual values.
Entropy is used to measure uncertainty and disorder. In Eq. (13), p1 is the prior probability of each class, c, and the number of unique classes46.
Results
Four types of evaluation methods are applied for the proposed research. The DT, GNB, and RF results are discussed in this section. For each technique, accuracy, specificity, sensitivity, Mathew's correlation coefficient, and accuracy is measured by the following equations47,48,49.
In the equations: \(TPV\) = All the true positive values from the dataset, \(TNV\) = All the True Negative values, \(FNV\) = False Negative values, \(FPV\) = False positive values.
For the proposed study, sensitivity refers to the ability of tests that truly identify Breast Adenocarcinoma cancer. Specificity refers to the tests’ ability to truly identify those who did not have Breast Adenocarcinoma in the dataset40. \(TPV + FNV\) represents the total number of subjects with the given conditions in the equations. In comparison, \(TNV+FPV\) is the total number of subjects without disorders. \(TPV+FPV\) is the total number of subjects with positive tests, and \(TNV+FNV\) is the total number of negative results50.
Self-consistency test
It is the iterative process that stops when the test results are satisfied. The same data is used in this technique for training and testing purposes. Table 2 shows the Decision tree results, Gaussian Naïve Bayes, Random Forest of Breast Adenocarcinoma cancer while the self-consistency test is applied to it.
ROC Curve of DT using self-consistency test is shown in Fig. 9.
The ROC curve defines the result as between 0.99 and 1.0, which should be considered excellent. ROC Curve of GNB using self-consistency test is shown in Fig. 10.
The rapid increase in the curve shows the accuracy value increases rapidly. ROC Curve of RF using the self-consistency test is shown in Fig. 11.
The combined ROC curve of the self-consistency test is shown in Fig. 12.
The ROC curve represents that all the results are on the upper side of the diagonal (50%), which shows the results under consideration are the best results. A Decision tree of 99% achieves the best outcome for the self-consistency test.
Independent set testing
The first evaluation method for the proposed work is independent set testing. The values extracted from the confusion matrix are used to determine the model's accuracy. This test is the basic performance measuring method for the proposed model. 80% of values are used to train the algorithm from the dataset, and 20% are used for testing purposes. Table 3 illustrates the independent test results on DT, GNB, and RF.
The Receiver Operating Characteristic (ROC) curve of DT, GNB, and FR Implemented after applying independent set testing is shown in Figs. 13, 14 and 15, respectively.
The ROC curve shows the specificity of 0.99 on the graph. It is the false-positive values get from the dataset. When sensitivity against the specificity together is a plot on the graph, a point in the ROC space is got. The position from the point in the ROC space shows the transaction between sensitivity and specificity. For most conditions, this point is between the points 0 and 1 on the graph. If this point falls on the area above the diagonal (more than 50%), it represents a good result; otherwise, it will be considered a bad result51. From the ROC curve of DT, it is determined that the point is falling above the diagonal at the point at 0.99, so it will be considered an excellent result.
The ROC curve of GNB shows the relation between sensitivity and specificity. The curve falls inside the ideal coordinate (0, 1). The result represents the value is between 0.8 and 0.85.
The ROC curve of RF shows a rapid increase in the results. The accuracy for this curve is 0.95.
The combined ROC Curve of independent set testing is shown in Fig. 16.
The green graph line represents the curve of DT; the Yellow line shows the RF and the Blue line results for GNB. The Decision tree algorithm obtains the best accuracy.
Tenfold cross-validation test
The data is equally subsampled into ten groups in the tenfold cross-validation technique. Divide the training set into ten partitions and then treat each in the validation set, train the model, and average generalization performance across the tenfold to make choices about hyperparameters and architecture. Figure 17 shows the working process of the tenfold cross-validation technique.
Table 4 represents the result of the tenfold cross-validation technique.
Figures 18, 19, and 20 show that medical study gathered after applying tenfold cross validations on DT, GNB, and RF, respectively, for the proposed model.
As shown in the figures, there are ten results for each testing technique. Tenfold cross-validation classifies the training set into ten groups52. And for each group, different results are gathered, and then the average is calculated. This technique is used to avoid model overfitting. This validation technique equally distributes the training set, and for every iteration, the results are different from the previous one, as shown in the ROC curves. The Combined ROC Curve of 10-FCV is shown in Fig. 21.
In Fig. 21, the combined ROC curve of maximum values is taken from each DT, GNB, and RF.
Discussion
Breast adenocarcinoma is the most common type of cancer in women. Different types have been proposed to detect and treat this breast cancer in the past. Some researchers also present computational studies to predict breast adenocarcinoma, as discussed in the literature review section. But in those computational studies, the researcher uses small with a smaller number of entries for their research. Most of them used only one machine learning technique. The proposed technique shows the best possible results for the early detection of carcinogenic mutations in Breast Adenocarcinoma using three machine learning algorithms. An extensive dataset that includes 99 gene sequences with 6170 mutations, from 4127 human samples makes an excellent consideration dataset for this study covering all the possible scenarios. The best test data techniques for such datasets are implemented for this study. Each evaluation method has its accuracy, specificity, sensitivity, and Mathew's correlation coefficient. The ROC curves of each evaluation method are discussed in the results section. According to the AUC classification, all the accuracy results fall in the excellent category. Most results of ROC curves are on the upper side of the diagonal (50%). The decision tree shows the accuracy of 99% for each evaluation method. Gaussian Naïve Bayes gives the accuracy of 81% for the self-consistency test and independent set test and 85% for the tenfold cross-validation test. Random Forest shows 97%, 95%, and 92% accuracy for the self-consistency test, independent set test, and tenfold cross-validation test, respectively.
Conclusion
Breast adenocarcinoma is the most common cancer in women. This study puts a positive effort into the identification of carcinogenic mutations which cause breast adenocarcinoma. Three machine learning algorithms, Decision Tree, Gaussian Naïve Bayes, and Random Forest, are applied to three types of evaluation methods independent set testing, Self-consistency testing, and tenfold cross-validation test. Accuracy, specificity, sensitivity, and Mathew’s correlation coefficient are calculated for these evaluation methods. Decision tree algorithms obtain the best accuracy of 99% for each evaluation method. All the evaluation methods’ accuracy gives results above the diagonal (50%) of the AUC value.
This study put a positive effort into identifying breast adenocarcinoma using different machine learning algorithms with a huge dataset. In the future, there may be a system that uses a more extensive dataset use for the current study and gives better results than the proposed system using different computational techniques.
Data availability
All data generated or analyzed during this study are included in this published article [and its supplementary information files].
References
Smith, T. J. Breast cancer surveillance guidelines. J. Oncol. Pract. 9, 65–67 (2013).
Biopsy. Cancer.Net (2020). https://www.cancer.net/navigating-cancer-care/diagnosing-cancer/tests-and-procedures/biopsy (Accessed 23 April 2022).
Fitzgerald, D. M. & Rosenberg, S. M. What is mutation? A chapter in the series: How microbes “jeopardize” the modern synthesis. PLoS Genet. 15, e1007995 (2019).
Tolosa, S., Sansón, J. A. & Hidalgo, A. Theoretical study of adenine to guanine transition assisted by water and formic acid using steered molecular dynamic simulations. Front. Chem. 7, 414 (2019).
Jackson, S. P. & Bartek, J. The DNA-damage response in human biology and disease. Nature 461, 1071–1078 (2009).
Pegg, A. E. Multifaceted roles of alkyltransferase and related proteins in DNA repair, DNA damage, resistance to chemotherapy, and research tools. Chem. Res. Toxicol. 24, 618–639 (2011).
Zhu, X., Lee, H., Perry, G. & Smith, M. A. Alzheimer disease, the two-hit hypothesis: An update. Biochim. et Biophys. Acta Mol. Basis Dis. 1772, 494–502 (2007).
Zhu, X., Raina, A. K., Perry, G. & Smith, M. A. Alzheimer’s disease: The two-hit hypothesis. Lancet Neurol. 3, 219–226 (2004).
Mohammed, S. A., Darrab, S., Noaman, S. A. & Saake, G. Analysis of breast cancer detection using different machine learning techniques. Data Mining Big Data. https://doi.org/10.1007/978-981-15-7205-0_10 (2020).
Garber, J. Implications of genetic information at breast cancer diagnosis. The Breast 12, S6 (2003).
Winchester, D. J. & Winchester, D. J. Breast Cancer (B.C. Decker, 2006).
Breast Cancer Treatment (Adult) (PDQ—ncbi.nlm.nih.gov). https://www.ncbi.nlm.nih.gov/books/NBK65969/. (Accessed 27 April 2022).
Holm, N. V., Hauge, M. & Harvald, B. Etiologic factors of breast cancer elucidated by a study of unselected twins2. J. Natl. Cancer Inst. https://doi.org/10.1093/jnci/65.2.285 (1980).
Williams, W. R., Anderson, D. E. & Rao, D. C. Genetic epidemiology of breast cancer: Segregation analysis of 200 Danish pedigrees. Genet. Epidemiol. 1, 7–20 (1984).
Newman, B., Austin, M. A., Lee, M. & King, M. C. Inheritance of human breast cancer: Evidence for autosomal dominant transmission in high-risk families. Proc. Natl. Acad. Sci. 85, 3044–3048 (1988).
Houlston, R. S., McCarter, E., Parbhoo, S., Scurr, J. H. & Slack, J. Family history and risk of breast cancer. J. Med. Genet. 29, 154–157 (1992).
Cancer driver mutations in breast adenocarcinoma. IntOGen. https://intogen.org/search?cancer=BRCA. (Accessed 24 April 2022).
Pon, J. R. & Marra, M. A. Driver and passenger mutations in cancer. Annu. Rev. Pathol. 10, 25–50 (2015).
Perou, C. M. et al. Molecular portraits of human breast tumours. Nature 406, 747–752 (2000).
Vaka, A. R., Soni, B. & Sudheer Reddy, K. Breast cancer detection by leveraging machine learning. ICT Express 6, 320–324 (2020).
Yue, W., Wang, Z., Chen, H., Payne, A. & Liu, X. Machine learning with applications in breast cancer diagnosis and prognosis. Designs 2, 13 (2018).
Bazazeh, D. & Shubair, R. Comparative study of machine learning algorithms for breast cancer detection and diagnosis. In 2016 5th International Conference on Electronic Devices, Systems and Applications (ICEDSA). https://doi.org/10.1109/icedsa.2016.7818560 (2016).
Khourdifi, Y. & Bahaj, M. Feature selection with fast correlation-based filter for breast cancer prediction and classification using machine learning algorithms. In 2018 International Symposium on Advanced Electrical and Communication Technologies (ISAECT). https://doi.org/10.1109/isaect.2018.8618688 (2018).
Kharya, S. & Soni, S. Weighted naive Bayes classifier: A predictive model for breast cancer detection. Int. J. Comput. Appl. 133, 32–37 (2016).
Malebary, S. J. & Khan, Y. D. Evaluating machine learning methodologies for identification of cancer driver genes. Sci. Rep. https://doi.org/10.1038/s41598-021-91656-8 (2021).
Ensembl Genome Browser 106. https://asia.ensembl.org/ (Accessed 24 April 2022).
Generating word cloud in python. GeeksforGeeks (2021). https://www.geeksforgeeks.org/generating-word-cloud-python/#:~:text=Word%20Cloud%20is%20a%20data,highlighted%20using%20a%20word%20cloud. (Accessed 24 April 2022).
Kaur, P. & Gosain, A. Comparing the behavior of oversampling and undersampling approach of class imbalance learning by combining class imbalance problem with noise. Adv. Intell. Syst. Comput. https://doi.org/10.1007/978-981-10-6602-3_3 (2017).
Chawla, N. V., Bowyer, K. W., Hall, L. O. & Kegelmeyer, W. P. Smote: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 16, 321–357 (2002).
Shah, A. A. & Khan, Y. D. Identification of 4-carboxyglutamate residue sites based on position Based Statistical Feature and multiple classification. Sci. Rep. https://doi.org/10.1038/s41598-020-73107-y (2020).
Zhu, H., Shu, H., Zhou, J., Luo, L. & Coatrieux, J. L. Image analysis by discrete orthogonal dual Hahn Moments. Pattern Recogn. Lett. 28, 1688–1704 (2007).
Sohail, M. U., Shabbir, J. & Sohil, F. Imputation of missing values by using raw moments. Stat. Trans. New Ser. 20, 21–40 (2019).
Butt, A. H. & Khan, Y. D. Canlect-pred: A cancer therapeutics tool for prediction of Target Cancerlectins using experiential annotated proteomic sequences. IEEE Access 8, 9520–9531 (2020).
Barukab, O., Khan, Y. D., Khan, S. A. & Chou, K.-C. iSulfoTyr-PseAAC: Identify tyrosine sulfation sites by incorporating statistical moments via Chou’s 5-steps rule and pseudo components. Curr. Genomics 20, 306–320 (2019).
Navada, A., Ansari, A. N., Patil, S. & Sonkamble, B. A. Overview of use of decision tree algorithms in machine learning. In 2011 IEEE Control and System Graduate Research Colloquium. https://doi.org/10.1109/icsgrc.2011.5991826 (2011).
Malik, H. A. M. Complex network formation and analysis of online social media systems. Cmes-Comr Model Engg & Sci 130(3), 1737–1750. https://doi.org/10.32604/cmes.2022.018015 (2022).
Malik, H. A. M. Analysis of social media complex system using community detection algorithms. Int. J. Comput. Digit. Syst. 11(1), 663–670. https://doi.org/10.12785/ijcds/110153 (2022).
Which Test is More Informative?—homes.cs.washington.edu. https://homes.cs.washington.edu/~shapiro/EE596/notes/InfoGain.pdf (Accessed 23 April 2022).
Decision tree algorithm, explained. KDnugget. https://www.kdnuggets.com/2020/01/decision-tree-algorithm-explained.html (Accessed 24 April 2022).
Salmi, N. & Rustam, Z. Naïve bayes classifier models for predicting the colon cancer. IOP Conf. Ser. Mater. Sci. Eng. 546, 052068 (2019).
Kaviani, P. & Dhotre, M. S. Short survey on naive Bayes algorithm. Int. J. Adv. Eng. Res. Dev. 4, 40826 (2017).
Gu, J. et al. Recent advances in convolutional neural networks. Pattern Recogn. 77, 354–377 (2018).
Maheswari, S. & Pitchai, R. Heart disease prediction system using decision tree and naive Bayes algorithm. Curr. Med. Imaging Form. Curr. Med. Imaging Rev. 15, 712–717 (2019).
Awais, M., Hussain, W., Rasool, N. & Khan, Y. D. iTSP-PseAAC: Identifying tumor suppressor proteins by using fully connected neural network and PseAAC. Curr. Bioinform. 16, 700–709 (2021).
Schott, M. Random Forest algorithm for machine learning. Medium (2020). https://medium.com/capital-one-tech/random-forest-algorithm-for-machine-learning-c4b2c8cc9feb (Accessed 24 April 2022).
Schonlau, M. & Zou, R. Y. The Random Forest algorithm for statistical learning. Stata J. Promot. Commun. Stat. Stata 20, 3–29 (2020).
Trevethan, R. Sensitivity, specificity, and predictive values: Foundations, pliabilities, and pitfalls in research and Practice. Front. Public Health 5, 307 (2017).
van Stralen, K. J. et al. Diagnostic methods I: Sensitivity, specificity, and other measures of accuracy. Kidney Int. 75, 1257–1263 (2009).
Lalkhen, A. G. & McCluskey, A. Clinical tests: Sensitivity and specificity. Contin. Educ. Anaesth. Crit. Care Pain 8, 221–223 (2008).
Kulkarni, A., Chong, D. & Batarseh, F. A. Foundations of data imbalance and solutions for a data democracy. Data Democracy. https://doi.org/10.1016/b978-0-12-818366-3.00005-8 (2020).
Hoo, Z. H., Candlish, J. & Teare, D. What is an ROC curve? Emerg. Med. J. 34, 357–359 (2017).
Sengar, P. P., Gaikwad, M. J. & Nagdive, A. S. Comparative study of machine learning algorithms for breast cancer prediction. In 2020 Third International Conference on Smart Systems and Inventive Technology (ICSSIT). https://doi.org/10.1109/icssit48917.2020.9214267 (2020).
Acknowledgements
The authors would like to thank the Deanship of Scientific Research at Majmaah University, Saudi Arabia, for supporting this work.
Author information
Authors and Affiliations
Contributions
A.A.S., H.A.M.M. and Y.D.K.F. envisioned the idea for research designed, wrote the results, and discussed. A.M., A.A. and H.A.M.M. worked on the literature and discussion section. All authors provided critical feedback and reviewed the paper and approved the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Shah, A.A., Malik, H.A.M., Mohammad, A. et al. Machine learning techniques for identification of carcinogenic mutations, which cause breast adenocarcinoma. Sci Rep 12, 11738 (2022). https://doi.org/10.1038/s41598-022-15533-8
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-022-15533-8
This article is cited by
-
m1A-Ensem: accurate identification of 1-methyladenosine sites through ensemble models
BioData Mining (2024)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.