Abstract
Chronic diseases are one of the most severe health issues in the world, due to their terrible clinical presentations such as long onset cycle, insidious symptoms, and various complications. Recently, machine learning has become a promising technique to assist the early diagnosis of chronic diseases. However, existing works ignore the problems of feature hiding and imbalanced class distribution in chronic disease datasets. In this paper, we present a universal and efficient diagnostic framework to alleviate the above two problems for diagnosing chronic diseases timely and accurately. Specifically, we first propose a network-limited polynomial neural network (NLPNN) algorithm to efficiently capture high-level features hidden in chronic disease datasets, which is data augmentation in terms of its feature space and can also avoid over-fitting. Then, to alleviate the class imbalance problem, we further propose an attention-empowered NLPNN algorithm to improve the diagnostic accuracy for sick cases, which is also data augmentation in terms of its sample space. We evaluate the proposed framework on nine public and two real chronic disease datasets (partly with class imbalance). Extensive experiment results demonstrate that the proposed diagnostic algorithms outperform state-of-the-art machine learning algorithms, and can achieve superior performances in terms of accuracy, recall, F1, and G_mean. The proposed framework can help to diagnose chronic diseases timely and accurately at an early stage.
Similar content being viewed by others
Introduction
Chronic diseases have been a severe health issue in the world. In 2019, the World Health Organization pointed out that chronic diseases account for about 7 of the top 10 causes of death in the world2. Deaths caused by chronic diseases account for more than 63% of the total global deaths. Common chronic diseases include heart disease, diabetes, hypertension, etc., which are mainly caused by individual unhealthy lifestyles3. Once people suffer from chronic diseases, several vital organs (e.g., eye, brain, heart, kidney, etc.) will be damaged, and it is easy to cause a series of serious complications affecting work and life4. Patients with chronic diseases are particularly vulnerable to infectious diseases, such as the coronavirus disease 2019 (COVID-19)5. More than 48% of COVID-19 patients have a history of chronic diseases and are more likely to develop severe symptoms6,7. Additionally, chronic diseases will lead to expensive medical expenses8. The Centers for Disease Control and Prevention reports chronic diseases are leading drivers of the nation’s 3.8 trillion in annual health care costs9. The main reason for the high fatality rate and expensive medical expenses is that chronic diseases have some terrible clinical presentations such as a long onset cycle, insidious symptoms, irreversible development, and various complications10. The above information reminds us that we need to quickly strengthen the prevention, diagnosis, and treatment of chronic diseases. Therefore, the early diagnosis of chronic diseases is urgent and essential, which can motivate high-risk patients to change their unhealthy lifestyles, thereby reducing the incidence of complications and further improving their health and quality of life.
Since the onset of chronic diseases is imperceptible and there are no obvious clinical symptoms in the early stage, it is difficult for doctors to determine the risk of patients with chronic diseases. Nowadays, machine learning has become the hottest promising technology for the assisted diagnosis of diseases with its advantages of autonomous learning and low error rate11,12,13. Several state-of-the-art machine learning algorithms have been widely used in the early diagnosis of different chronic diseases (e.g., chronic kidney disease, diabetes), such as support vector machines (SVM)14, logistic regression (LR)15, k-nearest neighbor (KNN)16, decision trees (DT)17, and the ensemble of some algorithms18,19,20. However, existing works are mainly dedicated to data preprocessing (e.g., data regularization and feature selection) to improve the early diagnostic performance of only a certain chronic disease21,22. Besides, they ignore the problems of feature hiding and imbalanced class distribution in chronic disease datasets. Hence, these methods are not conducive to improving the performance of the diagnostic model and are not suitable for a universal and efficient diagnosis of chronic diseases.
The problem of feature hiding represents that the feature in the dataset maybe not be directly related to decision-making. It needs to be further comprehensively analyzed together with other features to obtain the features directly related to decision-making23. For example, based on the heart rate and body mass index in the data, it is not possible to directly decide whether a patient has heart disease. If the visible original features of the data are directly used, neither the doctor nor the machine learning may be able to make a wise decision. Therefore, we need to expand the feature space of the data to capture its potential features related to chronic disease diagnosis. Additionally, the imbalanced class distribution of the dataset refers to a significant skew that exists between the number of samples for the different classes, which is also called the class imbalance problem24. The dominant class is called the majority class, and the remaining classes are called the minority class. Learning from the dataset with the class imbalance problem will make the learned model unreliable, which is more concerned with identifying the majority class correctly and ignoring the minority class25,26. Especially, in the chronic disease dataset, the number of sick cases (minority class) is generally lower than the number of healthy cases (majority class). However, the cost of misdiagnosing a sick case as a healthy case is significantly higher than the cost of misdiagnosing a healthy case as a sick case. The former may cause the patient to miss the best treatment period27. Therefore, how to accurately identify sick cases from the class imbalanced chronic disease dataset without affecting the overall diagnostic performance is of crucial importance and also a very challenging task.
Deep neural networks have great potential for solving various engineering problems in many fields, by extracting high-level features from data to achieve superior classification performance28,29. However, most deep neural network algorithms are not friendly to small-scale datasets and are prone to data overfitting30,31. Additionally, as the collected chronic disease data are not generally abundant (i.e., small-scale datasets), some existing deep neural network algorithms cannot train a well diagnostic model for chronic diseases. Recently, the deep polynomial neural network (PNN) has received the attention of some researchers32,33,34. We investigate the advantage of PNN and find that PNN is very friendly to classification tasks on small-scale datasets compared to other deep neural network algorithms. Surprisingly, the ideal PNN is parameter-free and can reduce the training error to zero iteratively35. Each network node of PNN is a polynomial function of its input. Thus, PNN can represent any polynomial value over the input data. Particularly, similar to other deep neural network algorithms, the network architecture of PNN is constructed layer by layer, which can represent higher and higher level (hidden) features of the input data. In other words, PNN can hierarchically expand the feature space of its input, and effectively capture features related to chronic disease diagnosis. Finally, the output layer of PNN can be constructed by solving a simple convex optimization problem.
In this paper, we are motivated to investigate the issue of the early diagnosis of chronic diseases. To the best of our knowledge, we are the first to study a universal and efficient diagnostic framework for chronic diseases, which can extract high-level features and solve the class imbalance problem to diagnose chronic diseases timely and accurately. Specifically, to efficiently capture high-level features hidden in chronic disease datasets, we propose a network-limited PNN (NLPNN) algorithm to avoid the problem of over-fitting. NLPNN can be seen as data augmentation in terms of its feature space. Additionally, as collected chronic disease datasets generally have a serious class imbalance problem, that is, the number of positive samples (sick cases) is significantly less than the number of negative samples (healthy cases), the PNN-based diagnostic model cannot fully learn the knowledge of sick cases, resulting in costly misdiagnosis (low recall). To alleviate this class imbalance problem, we further consider empowering samples with attention (i.e., weight) to change the importance of each sample and propose an improved NLPNN algorithm, named attention-empowered NLPNN (AEPNN). AEPNN pays more attention to these samples that are misclassified by NLPNN, regarded as data augmentation in terms of its sample space. Thus, the main contributions of this paper are summarized as follows.
-
We study a universal and efficient diagnostic framework to make timely and accurate early diagnosis of chronic diseases with small-scale datasets.
-
We propose an NLPNN algorithm to avoid the problem of over-fitting, which can efficiently capture high-level features hidden in chronic disease datasets and achieve high classification accuracy.
-
We further propose an AEPNN algorithm to solve the class imbalance problem, which greatly improves the recall of the diagnostic model, that is, it can accurately diagnose the sick case.
-
We evaluate and compare the proposed methods against other state-of-the-art methods using nine chronic diseases datasets (partly with class imbalance) and extensive experimental results demonstrate that the proposed two diagnostic models outperform state-of-the-art machine learning algorithms, and can achieve superior accuracy and recall.
The rest of the paper is organized as follows. We discuss related work in “Related work” section. “Diagnostic framework for chronic diseases” section presents the proposed algorithms, and experiment results are shown in “Experimental results” section. Finally, “Conclusion” section concludes this paper.
Related work
Early diagnosis of chronic diseases
Several existing machine learning algorithms have been proposed to diagnose a certain chronic disease36,37,38. Heydari et al.36 compared the performance of various machine learning classification algorithms in the early diagnosis of type 2 diabetes. The simulation results showed that the performance of classification techniques depends on the nature and complexity of the dataset. Khan et al.37 developed a chronic disease risk prediction framework. To reduce the impact of outliers, Alirezaei et al.38 incorporated K-means clustering, SVM, and meta-heuristic algorithm to diagnose diabetes disease. However, they ignored the influence of data distribution and structural changes on model generalization performance. Under the premise of not changing the structure and distribution of data, the authors in13 proposed a diagnostic model based on XGBoost for chronic kidney disease (CKD). Sekar et al.39 used a hierarchical neural network fusion method (FHNN) for the stratified diagnosis of cardiovascular disease (CVD). However, the impact of FHNN mainly depends on the optimal choice of the sub-neural network. Some tree-based ensemble learning techniques applied to early diagnosis methods of diabetes were comprehensively studied by Tama et al.20, and the differential performance of different classification methods was evaluated through statistical significance tests. At the same time, Altan et al.40 also compared various machine learning algorithms for the early diagnosis of chronic obstructive pulmonary disease and proposed a deep learning model to analyze multi-channel lung sounds using statistical features of Hilbert-Huang transform, which successfully achieved high classification performance of accuracy, sensitivity, and specificity of 93.67%, 91%, and 96.33%, respectively.
Class imbalance
In medical datasets, the problem of class imbalance seriously affects the accuracy of classifiers27,24. In most cases, it directly leads to a high rate of misdiagnosis of the disease. This is because the class imbalance of the training data brings difficulties to the algorithm learning, and the algorithm pays more attention to the majority class41. However, the minority class in medical datasets (sick vs. healthy) is often more important from a data mining perspective, and it usually carries critical and useful knowledge. At present, many scholars have studied the class imbalance problem, among which there are three main methods to alleviate the class imbalance42,43. (1) Data-level methods: in the data preprocessing stage, re-sampling is used to reduce the size of the majority class or increase the size of the minority class (or both) to balance the training set and eliminate difference. (2) Algorithm-level methods: in the training phase, the learning algorithm is modified to be suitable for mining data with imbalanced distributions. (3) Hybrid methods: the advantages of the first two methods are combined to alleviate the adverse effects of class imbalance on the results.
Diagnostic framework for chronic diseases
Statement: I confirm that all methods were performed in accordance with the relevant guidelines and regulations.
In this section, we propose a universal and efficient diagnostic framework for diagnosing chronic diseases timely and accurately. The proposed framework consists of the NLPNN algorithm and AEPNN algorithm to alleviate the problems of feature hiding and class imbalance, respectively.
Network-limited polynomial neural network
The PNN algorithm is dedicated to learning the high-level polynomial feature representation of the data through multi-layer network architecture, and finally, output features hierarchically32,33. Although the PNN algorithm has been proven to run in polynomial time, it still has a limitation, that is, the depth and width of the network cannot be controlled. Its network depth and width are both adaptive, and the criterion for depth stopping is until the training error is zero35. In the worst case, the network depth can be infinitely deepened or the network width can be as large as the number of training samples n. This will lead to severe overfitting. Hence, we present an NLPNN algorithm for the early diagnosis of chronic diseases to avoid this issue. The structure of NLPNN is shown in Fig. 1a, and the details of the NLPNN algorithm applied to chronic diseases diagnosis be described below.
For the early diagnosis of chronic diseases, we denote the labeled training dataset as \({\varvec{D}}= ({\varvec{X}},{\varvec{y}})\), where \({\varvec{X}} \in {\mathbb {R}} ^{n\times d}\) is the set of n samples with d features; \({\varvec{y}}=\left( y_{1},y_{2},\ldots ,y_{n}\right) ^{T}\) is a n-dimensional column vector and \(y_{i}\in \left\{ -1,1\right\}\), \(\forall i=1,2,\ldots ,n\). Here, \(y_{i}\) = 1 means that the i-th sample is labeled as a sick case, and \(y_{i}\) = − 1 otherwise. The M-order multivariate polynomial on the sample \({\varvec{x}}_{i}=\left( x_{i1},\ldots ,x_{id}\right) \in {\varvec{X}}\) is written as
where \(\varvec{\alpha }^{(j)}\) is a d-dimensional vector composed of non-negative integers and \(\sum _{s=1}^{d} \alpha _{s}^{(j)}\!=\!j\); \(w_{\varvec{\alpha }^{(j)}}\) is a coefficient of monomial \(\prod _{s=1}^{d} x_{is}^{\alpha _{s}^{(j)}}\) of degree j. Represent the value of each polynomial \({\varvec{p}}\) on n samples by linear projection
According to linear algebra, there are n polynomials \({\varvec{p}}_{1},\ldots ,{\varvec{p}}_{n}\), and \(\left\{ \left( {\varvec{p}}_{i}\left( {\varvec{x}}_{1}\right) , \ldots , {\varvec{p}}_{i}\left( {\varvec{x}}_{n}\right) \right) ^T\right\} _{i=1}^{n}\) form a basis of \({\mathbb {R}} ^{n}\) space. Therefore, there is a coefficient vector \(\varvec{\nu } = (\nu _1,\ldots ,\nu _n)\), so that \(\sum _{i=1}^{n} \nu _{i} {\varvec{p}}_{i}\left( {\varvec{x}}_{j}\right) =y_{j}, \forall y_{j} \in (y_1,\ldots ,y_n)^T\in {\mathbb {R}} ^{n}\).
The network layer of PNN is constructed by solving the basis of polynomial hierarchically, and each node calculates a linear function or weighted product over its input. We denote the j-th node of the i-th layer as \(\eta _j^i(\cdot )\), which actually represents a feature (original or high-level) of the input data. For the first layer, the j-th node is the degree-1 polynomial (or linear) function \(\eta _{j}^{1}({\varvec{x}})=[1 \ {\varvec{x}}]{\varvec{w}}_{j}\), and the \(\left\{ \left( \eta _{j}^{1}\left( {\varvec{x}}_{1}\right) , \ldots , \eta _{j}^{1}\left( {\varvec{x}}_{n}\right) \right) ^{T}\right\} _{j=1}^{d+1}\) is the basis of all values obtained by a polynomial of degree 1 on the training dataset. They form the columns of matrix \(F^1\in {{\mathbb {R}}^{n\times (d+1)}}\) and \(F_{i, j}^{1}=\eta _{j}^{1}\left( {\varvec{x}}_{i}\right)\). So far, a single-layer network has been constructed, and its output spans all the values obtained by the linear function on the training sample.
Generally speaking, the basis of the degree-2,3,...M polynomial is also obtained in the same trick. However, we find that the basis of the degree-M multiple polynomials is composed of \((d+1)^M\) vector elements. The scale of the basis of the polynomial increases exponentially with its degree, which will run into a computational problem.
The work in35 indicates that any degree-m polynomial can be regraded as
where \({\varvec{g}}_{i}({\varvec{x}})\) and \({\varvec{h}}_{i}( {\varvec{x}})\) are degree-1 and degree-\((m-1)\) polynomials respectively; \({\varvec{k}}( {\varvec{x}})\) is a polynomial of degree not greater than \(m-1\). Since all degree-1 polynomials are spanned by the nodes at the first layer of PNN, any degree-2 polynomial can be written as
where \(\alpha _{j}^{\left( {\varvec{g}}_{i}\right) }\), \(\alpha _{r}^{\left( {\varvec{h}}_{i}\right) }\), \(\alpha _{j}^{({\varvec{k}})}\) are scalar multipliers. (4) implies that the construction of the second layer of the network is based on the first layer. The matrix \([F^{1} {\tilde{F}}^{2}]\) is formed by concatenating the columns of \(F^{1}\), \({\tilde{F}}^{2}\), which spans all values attainable by degree-2 polynomials, and
where the symbol \(\circ\) indicates the Hadamard product; \(F_{1}\) refers to the first column of F; \(\left| F\right|\) refers to the number of columns of F. Similar to degree-1 polynomial, the column subset \(F^{2}\) of \({\tilde{F}}^{2}\) should be found, so that the column of \([F^{1} F^{2}]\) are the basis of column of \([F^{1} {\tilde{F}}^{2}]\). The second layer of the PNN is constructed by the column of \(F^{2}\), which is the product of two nodes \(\eta _{i}^{1}(\cdot )\) and \(\eta _{j}^{1}(\cdot )\) in the first layer.
The next step is to repeat the above process. Successively, the \(m = 3, 4, \ldots , M\) layers of the network are constructed. We represent the matrix, written as
Thus, we find a linearly independent column subset \(F^{m}\) of \({\tilde{F}}^{m}\), which lets the columns of matrix \([F \ F^{m}]\) are a basis of the columns of the augmented matrix \([F \ {\tilde{F}}^{m}]\), where the columns of \(F=\left[ F^{1}\ F^{2}\ \ldots \ F^{m-1}\right]\) can span the values attained by all polynomials for degree at most \(m-1\) over the training dataset. In addition, it needs to be explained that the conversion of \({\tilde{F}}^{m}\) to \(F^{m}\) is achieved by
where the projection matrix \(W\in {\mathbb {R}}^{{|F^{m-1}|\times |F^{1}|}}\) and \(W_{i(s), j(s)}=\sqrt{n} /\left\| {\tilde{F}}_{s}^{m}\right\|\). Therefore, when the M-layer network of the PNN is constructed, all the values obtained by the polynomial of degree at most M over the training dataset can be spanned by the columns of the matrix F. In fact, F stores the high-level features of the input data, the deeper layer, the higher feature.
However, for the implementation of NLPNN, we use a parameter \(\Omega =(d+1,\cdots ,d+1) \in {\mathbb {Z}}^M\) to pre-limit the depth and width of the network, which represents that the network consists of M (\(|\Omega |\)) non-output layers and each layer has \(d+1\) nodes at most. In the first non-output layer, we use singular value decomposition on the augmented data matrix \([{\varvec{1}}\ {\varvec{X}}]\) to obtain its partial orthogonal basis, which forms the \(d+1\) nodes (select the first \(d+1\) main singular vectors). In the next non-output layer, a standard Orthogonal Least Squares (OLS) procedure is utilized to greedily select the partial orthogonal basis which are the first \(d+1\) relevant features for diagnosis of chronic disease according to the established high-level feature set \({\tilde{F}}^{m}\). Finally, a simple linear classifier \(\varvec{\nu }_{m}\) with input data \(F=\left[ F^{1}\ F^{2}\ \ldots \ F^{m}\right]\) is trained. Therefore, there are M linear classifiers in the output layer. It should be pointed out that each linear classifier \(\varvec{\nu }_{m}\) is trained by a stochastic gradient descent method, which is utilized to solve the \(L_{2}\) regularization problem
where \(\ell _{i}\left( F^{m}_{i\cdot }\cdot \varvec{\nu },y_{i}\right) =\max (0,1-(F^{m}_{i\cdot }\cdot \varvec{\nu })\cdot y_{i})\) is a hinge loss and \(F^{m}_{i\cdot }\) represents the i-th row of matrix \(F^{m}\); \(\lambda _{m} \in \Lambda\) is the regularization factor. Then combined with the value set \(\Lambda\) of the regularization factor, we check the network performance layer by layer on the verification dataset to find the optimal network layer and the best regularization factor. Finally, an optimal linear classifier \(\varvec{\nu }^{*}\) is obtained by
and the output is this optimal classifier. The purpose of NLPNN is to adaptively find features related to diagnosis from the augmented data that is augmented in terms of its feature space. The detailed process of NLPNN is shown in Algorithm 1, which briefly describes the entire process from the establishment of the network layer to the acquisition of the output layer.
Attention-empowered NLPNN
Some chronic disease datasets exist the class imbalance problem, where sick cases are generally scarce compared to healthy cases. However, the correct diagnosis of the minority sick cases among all cases is vital in a healthcare system. The reason is that the cost of misdiagnosing sick cases is much higher than healthy cases, where the latter only requires further examination and the former carries a life-threatening risk. During the training phase of NLPNN, since the samples of each class in the imbalanced dataset are utilized equally, the trained model tends to bias towards the majority class and ignore the samples (sick cases) in the minority class. Thus, NLPNN does not perform well in dealing with class imbalance problems and causes serious misdiagnosis of minority sick cases. Furthermore, for the early diagnosis of chronic diseases, although we are more concerned with the accurate diagnosis of sick cases, we cannot ignore the overall diagnostic accuracy. To alleviate the class imbalance problem, we empower the cases with attention (i.e., weight) and propose an AEPNN algorithm. AEPNN pays more attention to the cases misdiagnosed by NLPNN by changing the importance of these cases. Motivated by committee-based learning25, AEPNN trains and combines multiple complementary NLPNN to further improve the performance of NLPNN in alleviating the class imbalance problem. The structure of AEPNN is shown in Fig. 1b.
For the implementation of AEPNN, we first assign an identical initial weight \({\mathcal {D}}_{1}({\varvec{x}})=\frac{1}{n}\) to each sample \({\varvec{x}}\) in the training dataset. An NLPNN classifier \(h_{1}\) is trained from the training dataset \({\varvec{D}}_1\) with the initialized weight distribution \({\mathcal {D}}_{1}\) and \(h_{1}\)’s error \(\epsilon _{1}\) is fed back to the training sample, so that the training sample’s distribution is adjusted by \({\mathcal {D}}_{2}({\varvec{x}})\). Then, the second NLPNN classifier \(h_{2}\) is trained from the training dataset \({\varvec{D}}_2\) with the weight distribution \({\mathcal {D}}_{2}\), where the weights of samples misdiagnosed by \(h_{1}\) are increased in \({\mathcal {D}}_{2}\) to make \(h_{2}\) pay more attention to the samples that are misdiagnosed by \(h_{1}\). This process is repeated until \(h_{L}\) is trained after L iterations. Finally, the predicted label is obtained through the weighted combination of all NLPNN classifiers. The main process is shown in Algorithm 2.
Specifically, we denote the true label corresponding to sample \({\varvec{x}}\) as \(f({\varvec{x}})\), and the predicted label obtained by the NLPNN classifier as \(h({\varvec{x}})\). Obviously, the loss function \(\epsilon\) is defined as
where \(p({\varvec{x}})\) represents the probability density function of \({\varvec{x}}\) following the data distribution \({\mathcal {D}}\). However, \(\epsilon\) has poor mathematical properties (non-convex and non-continuous), which makes it very difficult to be solved directly. To optimize the loss function more conveniently, we select a convex and continuously differentiable exponential loss function (11) to replace the loss function (10). Lemma 1 proves that \(\ell _{\exp }(h \mid {\mathcal {D}})\) is the consistent replacement of the loss function \(\epsilon\), which means that (11) can replace (10) to update the weight \({\mathcal {D}}_{l}({\varvec{x}})\) of the sample and the weight \(\alpha _{l}\) of the classifier in Algorithm 2.
Lemma 1
The consistent replacement of the loss function \(\epsilon\) is the exponential loss function
Proof
Please see Appendix 1. \(\square\)
In Algorithm 2, the \(h_1\) is obtained by applying the NLPNN classifier to the initial samples distribution \({\mathcal {D}}_{1}\). When \(h_{l}\) is generated based on distribution \({\mathcal {D}}_{l}\), the weight \(\alpha _{l}\) of the classifier \(h_{l}\) is obtained iteratively by minimize the exponential loss function \(\ell _{\exp }\left( \alpha _{l} h_{l} \mid {\mathcal {D}}_{l}\right)\). From Lemma 2, we know that \(\alpha _{l}=\frac{1}{2} \ln \left( \frac{1-\epsilon _{l}}{\epsilon _{l}}\right)\) is a necessary and sufficient condition for the exponential loss function \(\ell _{\exp }\left( \alpha _{l} h_{l} \mid {\mathcal {D}}_{l}\right)\) to obtain the minimum value. It means that, under the encouragement of the weight \(\alpha _{l}\), the classifier \(h_{l}\) can achieve the best performance on the dataset \({\varvec{D}}_l\) with distribution \({\mathcal {D}}_{l}\).
Lemma 2
The exponential loss function \(\ell _{\exp }\left( \alpha _{l} h_{l} \mid {\mathcal {D}}_{l}\right)\) at \(\alpha ^{*}_{l}h_{l}\) obtain the minimum value, where \(\alpha ^{*}_{l}=\frac{1}{2} \ln \left( \frac{1-\epsilon _{l}}{\epsilon _{l}}\right)\) and \(\epsilon _{l}=\int _{{\varvec{x}} \sim {\mathcal {D}}_{l}} {\mathbb {I}}(f({\varvec{x}}) \ne h_{l}({\varvec{x}})) p({\varvec{x}}) {\mathrm {d}} {\varvec{x}}\).
Proof
Please see Appendix 2. \(\square\)
\(H_{l}\) is the voting result of the first l NLPNN classifier \(\{h_{i}\}_{1}^{l}\) with weights \(\{\alpha _{i}\}_{1}^{l}\), and its error can be corrected by the next classifier \(h_{l+1}\). Ideally, \(h_{l+1}\) can correct all errors of \(H_{l}\) by minimizing the exponential loss \(\ell _{\exp }(H_{l}+h \mid {\mathcal {D}})\). From Lemma 3, all errors of \(H_{l}\) can be corrected by the NLPNN classifier \(h_{l+1}\) which is trained based on the sample weight distribution \({\mathcal {D}}_{l+1}({\varvec{x}})={{\mathcal {D}}_{l}({\varvec{x}}) \frac{e^{-\alpha _{l}f({\varvec{x}}) h_{l}({\varvec{x}})}}{Z_{l}}}\), where \(\frac{1}{Z_{l}}=\frac{\int _{{\varvec{x}} \sim {\mathcal {D}}}\left[ e^{- f({\varvec{x}}) H_{l-1}({\varvec{x}})}\right] p({\varvec{x}}) {\mathrm {d}} {\varvec{x}}}{\int _{{\varvec{x}} \sim {\mathcal {D}}}\left[ e^{- f({\varvec{x}}) H_{l}({\varvec{x}})}\right] p({\varvec{x}}) {\mathrm {d}} {\varvec{x}}}\) is the normalization factor to ensure that \({\mathcal {D}}_{l+1}\) is a distribution.
Lemma 3
Assume that the base classifier \(h_{1}\) is generated based on the data distribution \({\mathcal {D}}_{1}\), and \(\alpha _{i}\) is the weight of the classifier \(h_{i}\), \(H_{l}({\varvec{x}})={\text {sign}}\left( \sum _{i=1}^{l} \alpha _{i} h_{i}({\varvec{x}})\right)\), \(l=1,2,\ldots\), then all the false predictions of \(H_{l}\) can be corrected through the ideal base classifier \(h_{l+1}\), which is generated based on the data distribution
Proof
Please see Appendix 3. \(\square\)
In summary, we iteratively optimize the exponential loss function by introducing two kinds of attention (\({\mathcal {D}}\) and \(\alpha\)) to achieve the superiority of AEPNN on class-imbalanced datasets.
Experimental results
Some DNN models are not suitable for classification tasks with the small-scale dataset due to the over-fitting problem. However, the PNN-based deep learning algorithm performs well for the early diagnosis of chronic diseases with the small-scale dataset, due to its unique network structure. We select five state-of-the-art machine learning algorithms as the baseline algorithms, i.e. SVM44, LR45, KNN46, DT47, and multi-layer perceptron (MLP)48.
Chornic disease datasets
To verify the effectiveness of the proposed algorithm in the early diagnosis of chronic diseases, we select nine public and two private chronic disease datasets for experiments. Nine public chronic disease datasets (i.e., http://archive.ics.uci.edu/ml, https://www.kaggle.com/datasets) include CKD, Pima Indian diabetes dataset (PIMA), CVD, Heart Disease Dataset (Heart), Framingham Heart Disease dataset (Fra_Heart), Hepatitis dataset (Hep), Breast Cancer Wisconsin dataset (BCW) in UCI Machine Learning Repository, Type 2 Diabetes Mellitus Dataset (T2DM) and Gestational Diabetes Mellitus dataset (GDM) in the Tianchi Precision Medicine Competition. They are scarce and precious, but some of them have problems, such as small size, class imbalance, and missing value. Two private chronic disease datasets (Pri_hyper dataset and Pri_diab dataset) are collected from a district in Chongqing, China. The Pri_hyper dataset consists of the health records of hypertensive patients and healthy people. The Pri_diab dataset consists of the health records of diabetic patients and healthy people. The composition details of the selected datasets are listed by Table 1, in which column Datasets is shorthand for the name of the dataset; column Features represents the number of features; column Samples represents the number of samples; column Positive: Negative represents the ratio of the number of positive and negative samples; column Missing shows whether there are missing values in the corresponding dataset. Consistently, we split each chronic disease dataset randomly into a training dataset and testing dataset with 8:2, and maintain the distribution of the class before the split. For baseline algorithms that have to process missing values and regularize data, we fill the missing values with zeros and regularize the data. The implementation of proposed algorithms does not require any other data preprocessing technology.
Evaluation measurements
For the early diagnosis of chronic diseases, the generalization performance can be estimated on the test dataset. In addition to using the area under the receiver operating characteristic curve (AUC) to evaluate the performance of the model, we also selected the following evaluation indicators to evaluate the proposed algorithm:
-
\(Accuracy =\frac{TP+TN}{TP+TN+FP+FN}\) represents the ratio of the number of correctly predicted specific classes to the total number of samples.
-
\(Specificity =\frac{TN}{TN+FP}\) represents the ratio of the number of correctly predicted healthy cases to the total healthy cases.
-
\(Precision=\frac{TP}{TP+FP}\) represents the ratio of the number of correctly predicted sick cases to the total predicted sick case.
-
\(Recall=\frac{TP}{TP+FN}\) represents the ratio of the number of correctly predicted sick cases to the total number of sick cases.
-
\(F1\_score=\frac{2*Precision*Recall}{Precision+Recall} = \frac{2*TP}{N+TP-TN}\) is defined based on the harmonic average of precision and recall.
where TP, FP, TN, and FN represent true positive, false positive, true negative and false negative respectively; N is the total number of samples.
Comparison of performance
We investigate the impact of different network depths \(\Delta\) and regularization factor \(\lambda\) in the NLPNN model for the diagnostic performance of eleven chronic diseases, where \(\Delta \in \{2,3,4,5\}\) (network layer plus output layer) and \(\lambda \in \Lambda =\{10^{-3},10^{-2}, 10^{-1},10^{0},10^{1}\}\). To visually find the most suitable \(\Delta\) and \(\lambda\), we combine them into a binary set \((\Delta ,\lambda )\), and establish a bijection function between \((\Delta ,\lambda )\) and \(\Pi \in \{1,2,\cdots ,20\}\) described in Table 2. We set \(\Pi\) as the horizontal axis to indirectly draw the generalization performance curve of NLPNN with network depth and regularization factor. From Fig. 2, we can see that NLPNN has two advantages in the diagnosis of all chronic diseases, that is, there is no over-fitting phenomenon; the training accuracy is increasing with the increase of the number of network layers (it can be observed that when \(\Pi\)=1,6, 11,...). However, different \(\Pi\) values will affect the performance of the NLPNN algorithm, the impact on different chronic disease datasets is different.
Figure 2a shows that NLPNN can achieve 100% generalization performance on the CKD dataset when \(\Pi =\{1,2\}\). Then, with the increase of \(\Delta\) and the change of \(\lambda\), the test performance decreases somewhat, but both fluctuate within the range of 5%. It means that only a shallow polynomial neural network model can accurately diagnose chronic kidney disease. We can see from Fig. 2b, c, g and k that the \(\Pi\) value has almost no effect for the diagnostic accuracy of diabetes and heart disease. In particular, for the diagnosis of hepatitis B disease (Fig. 2h), although the accuracy of the NLPNN model does not vary greatly, its specificity is unstable with the change of \(\Pi\) value. This reason is that the Hep dataset has only 155 samples and the negative samples only account for 24% of the total samples. In addition, we can find the best output performance \(P^{*}\) of NLPNN and the corresponding value \(\Pi ^{*}\) on eleven chronic disease datasets from the Fig. 2. Therefore, according to Table 2, we can find the network structure \(\Omega ^{*}\) and the regularization factor \(\lambda ^{*}\) when NLPNN achieves the best performance, as shown in Table 3.
The generalization performance comparison of baseline algorithms and NLPNN algorithm on eleven chronic disease datasets are shown in Table 4, which lists the test performance results under the unified standard. In general, the diagnostic accuracy of NLPNN on the eleven chronic disease datasets is better than baseline algorithms. Especially for the diagnosis of chronic kidney disease and breast cancer, NLPNN can achieve a generalization accuracy, recall, and F1_score, of 1.0000, 1.0000, and 1.0000, respectively. In addition, NLPNN also shows significant advantages in the diagnosis of Hepatitis disease, and its generalization accuracy is about 10% better than the baseline algorithms (SVM:0.8000, LR: 0.8333, KNN: 0.8000, DT: 0.8333, MLP: 0.8000).
Figure 3 plots the ROC curves to further compare the performance of the NLPNN algorithm and the baseline algorithms. The AUC value of the proposed algorithm is generally better than baseline algorithms. It is also worth noting that in the diagnosis task of chronic kidney disease and breast cancer, the NLPNN model is an “ideal model” with an AUC value of 1 (Fig. 3a, i).
In this paper, we not only pay attention to the overall accuracy of the model in the diagnosis of chronic diseases but also pay more attention to whether the model can accurately diagnose sick cases (positive samples). That is, we hope that the recall of the model is as high as possible on the premise that the overall accuracy is high. For T2DM, CVD, Fra_Heart, and Pri_diab datasets, we observe that the ratio of the number of correctly predicted sick cases to the total number of sick cases is low, that is, the recall rate is low. The reason is that there is a class imbalance problem in these datasets. To solve this problem, the AEPNN algorithm 2 is proposed in “Diagnostic framework for chronic diseases” section. Because the NLPNN algorithm is a strong classifier, we do not need too many individual classifiers, whose number is equal to the number of iterations. The test performance will change with the increase of the number of training rounds of the NLPNN algorithm. Although the overall diagnostic accuracy decreases slightly, the diagnostic accuracy of sick cases has been significantly improved. We choose the number of iterations corresponding to the maximum value of the difference between the growth rate of recall and the decrease rate of accuracy as the final number of training rounds of the NLPNN algorithm to obtain the best performance. Figures 4, 5, 6, 7 show the performance of the proposed algorithm when applied to the Fra_Heart, T2DM, Pri_diab, and CVD datasets at different iterations of NLPNN, respectively. Comprehensive analysis with Table 1, we can see that the higher the class imbalance ratio of chronic disease data, the more obvious AEPNN improves the recall.
The generalization performance of AEPNN on the Fra_Heart dataset is shown in Fig. 4a. The performance growth rate is calculated based on the number of NLPNN classifiers being one. From Fig. 4b, we observe that the recall has a growth rate of close to 300% when the number of NLPNN classifiers is six, which is chosen as the best number of NLPNN classifiers for the diagnosis of heart disease. The most surprising thing is the performance of AEPNN on the T2DM and Pri_diab datasets. As it can be seen from Figs. 5a and 6a, when the number of NLPNN is greater than four, the recall is significantly improved. When the number of NLPNN reaches ten, the growth rate of the recall approaches 4000% on the T2DM dataset and 6000% on the Pri_diab dataset. We can also know that the growth rate of recall is much higher than the decreased rate of accuracy from Figs. 5b and 6b.
From Fig. 7, we can see that although the performance of AEPNN on the CVD dataset is not significantly improved, the growth rate of recall is still higher than the decreased rate of accuracy. It indicates that the proposed algorithm is effective for the improvement of recall. The advantage it brings is that it can reduce the missed diagnosis rate for sick cases so that more patients with chronic diseases can treat and control the development of the disease in time. We also quantitatively compare the generalization performance of AEPNN and NLPNN algorithms by introducing \(G\_mean = \sqrt{Recall * Specificity}\), which is a powerful indicator to evaluate the classification accuracy for class imbalanced datasets49. From Table 5, we can see that AEPNN can effectively improve \(G\_mean\) by combining multiple NLPNNs. In particular, AEPNN can increase the \(G\_mean\) from 0.0985 to 0.5728 on the T2DM dataset and from 0.0731 to 0.5385 on the Pri_diab dataset by combining ten NLPNNs.
Conclusion
In this paper, we have investigated a universal learning algorithm based on PNN for the early diagnosis of chronic diseases. Five state-of-the-art baseline algorithms are selected to compare with the NLPNN algorithm. Experiment results show that NLPNN achieves the best accuracy on the nine chronic disease datasets. In particular, for the early diagnosis of chronic kidney disease and breast cancer disease, the generalization accuracy, recall, specificity, and AUC value of this model have achieved 1.000, 1.000, 1.000, and 1.000, respectively. Furthermore, an AEPNN algorithm is further proposed to alleviate the class imbalance problem in chronic disease datasets. We aim to increase the probability of the sick cases being accurately diagnosed, that is, to increase the recall value of the model. Experiments on the four chronic disease datasets with class imbalance problems have confirmed the effectiveness of our model. It is noted that the AEPNN model performs best on the Pri_diab dataset with a positive-negative sample ratio of 1:12.78, and the growth rate of its recall is close to 6000%. The proposed algorithm can effectively assist chronic disease experts in quickly screening patients with chronic diseases, and save the cost of further testing for patients. It should be pointed out that although our algorithm performs better on small-scale datasets, the PNN-based model also shows great application potential on large-scale datasets, such as protein-protein interaction prediction and disease diagnosis based on medical images.
In future work, we will further investigate the PNN-based model in disease diagnosis. Although PNN can effectively capture hidden features parameter-free, there is still a problem with how to adaptively select the best-hidden features from the network architecture of PNN to achieve competitive performance. Thus, we consider combining PNN with computational intelligence algorithms (such as monarch butterfly optimization (MBO), earthworm optimization algorithm (EWA), and elephant herding optimization (EHO)) to improve the performance of disease diagnosis.
Data availability
The datasets used and/or analyzed during the current study available from the corresponding author on reasonable request.
References
Yuan, X., Chen, S., Sun, C. & Yuwen, L. A novel class imbalance-oriented polynomial neural network algorithm for disease diagnosis. In Proceedings of IEEE International Conference on Bioinformatics and Biomedicine (BIBM) 2360–2367 (2021).
Organization, W. H. WHO reveals leading causes of death and disability worldwide: 2000–2019. https://www.who.int/news/item/09-12-2020-who-reveals-leading-causes-of-death-and-disability-worldwide-2000-2019.
Souza-Pereira, L., Pombo, N., Ouhbi, S., Felizardo, V. & Garcia, N. Clinical decision support systems for chronic diseases: A systematic literature review. Comput. Methods Progr. Biomed. 195, 105565 (2020).
Alkenani, A. H., Li, Y., Xu, Y. & Zhang, Q. Predicting Alzheimer’s disease from spoken and written language using fusion-based stacked generalization. J. Biomed. Inform. 118, 103803 (2021).
Yuan, X., Chen, S., Yuwen, L., An, S., Mei, S. & Chen, T. An improved SEIR model for reconstructing the dynamic transmission of COVID-19. In Proceedings of IEEE International Conference on Bioinformatics and Biomedicine (BIBM) 2320–2327 (2020).
Guo, Y. et al. A review of wearable and unobtrusive sensing technologies for chronic disease management. Comput. Biol. Med. 129, 104163 (2020).
Higgins, V., Sohaei, D., Diamandis, E. P. & Prassas, I. COVID-19: From an acute to chronic disease? Potential long-term health consequences. Crit. Rev. Clin. Lab. Sci. 58(5), 297–310 (2021).
Iheanacho, I., Zhang, S., King, D., Rizzo, M. & Ismaila, A. S. Economic burden of chronic obstructive pulmonary disease (COPD): A systematic literature review. Int. J. Chronic Obstr. Pulm. Dis. 15, 439 (2020).
For Disease Control, C., Prevention. About Chronic Diseases. https://www.cdc.gov/chronicdisease/about/index.htm.
Pathak, S. et al. Post-structuring radiology reports of breast cancer patients for clinical quality assurance. IEEE/ACM Trans. Comput. Biol. Bioinform. 17(6), 1883–1894 (2019).
Xia, Y., Yao, Z., Ye, Q. & Cheng, N. A dual-modal attention-enhanced deep learning network for quantification of Parkinson’s disease characteristics. IEEE Trans. Neural Syst. Rehabil. Eng. 28(1), 42–51. https://doi.org/10.1109/TNSRE.2019.2946194 (2020).
Zhang, Q., Zhou, J., Zhang, B. & Wu, E. Dsnet: Dual stack network for detecting diabetes mellitus and chronic kidney disease. Inf. Sci. 547, 945–962 (2021).
Ogunleye, A. & Wang, Q.-G. Xgboost model for chronic kidney disease diagnosis. IEEE/ACM Trans. Comput. Biol. Bioinform. 17(6), 2131–2140 (2019).
Dolatabadi, A. D., Khadem, S. E. Z. & Asl, B. M. Automated diagnosis of coronary artery disease (CAD) patients using optimized SVM. Comput. Methods Progr. Biomed. 138, 117–126 (2017).
Xiao, R., Cui, X., Qiao, H., Zheng, X. & Zhang, Y. Early diagnosis model of Alzheimer’s disease based on sparse logistic regression. Multimed. Tools Appl. 80(3), 3969–3980 (2021).
Jabbar, M. Prediction of heart disease using k-nearest neighbor and particle swarm optimization. Biomed. Res. 28(9), 4154–4158 (2017).
Mathan, K., Kumar, P. M., Panchatcharam, P., Manogaran, G. & Varadharajan, R. A novel Gini index decision tree data mining method with neural network classifiers for prediction of heart disease. Des. Automat. Embed. Syst. 22(3), 225–242 (2018).
Shang, H. & Liu, Z.-P. Prioritizing type 2 diabetes genes by weighted PageRank on bilayer heterogeneous networks. IEEE/ACM Trans. Comput. Biol. Bioinform. 18(1), 336–346 (2021).
Asadi, S., Roshan, S. & Kattan, M. W. Random forest swarm optimization-based for heart diseases diagnosis. J. Biomed. Inform. 115, 103690 (2021).
Tama, B. A. & Rhee, K.-H. Tree-based classifier ensembles for early detection method of diabetes: An exploratory study. Artif. Intell. Rev. 51(3), 355–370 (2019).
Li, J. et al. A tongue features fusion approach to predicting prediabetes and diabetes with machine learning. J. Biomed. Inform. 115, 103693 (2021).
Ma, S. et al. Multiple predictively equivalent risk models for handling missing data at time of prediction: With an application in severe hypoglycemia risk prediction for type 2 diabetes. J. Biomed. Inform. 103, 103379 (2020).
Wang, G.-G., Lu, M., Dong, Y.-Q. & Zhao, X.-J. Self-adaptive extreme learning machine. Neural Comput. Appl. 27(2), 291–303 (2016).
Singh, R. et al. Imbalanced breast cancer classification using transfer learning. IEEE/ACM Trans. Comput. Biol. Bioinform. 18(1), 83–93 (2020).
Bader-El-Den, M., Teitei, E. & Perry, T. Biased random forest for dealing with the class imbalance problem. IEEE Trans. Neural Netw. Learn. Syst. 30(7), 2163–2172 (2018).
Cui, Z. et al. Detection of malicious code variants based on deep learning. IEEE Trans. Ind. Inform. 14(7), 3187–3196 (2018).
Yildirim, P. Chronic kidney disease prediction on imbalanced data by multilayer perceptron: Chronic kidney disease prediction. In Proceedings of IEEE 41st Annual Computer Software and Applications Conference (COMPSAC), vol. 2, 193–198 (2017).
Yi, J.-H., Wang, J. & Wang, G.-G. Improved probabilistic neural networks with self-adaptive strategies for transformer fault diagnosis problem. Adv. Mech. Eng. 8(1), 1687814015624832 (2016).
Wang, Y., Qiao, X. & Wang, G.-G. Architecture evolution of convolutional neural network using Monarch butterfly optimization. J. Ambient Intell. Humaniz. Comput. 13(3), 1–15 (2022).
Gurovich, Y. et al. Identifying facial phenotypes of genetic disorders using deep learning. Nat. Med. 25(1), 60–64 (2019).
Feng, S., Zhou, H. & Dong, H. Using deep neural network with small dataset to predict material defects. Mater. Des. 162, 300–310 (2019).
Shi, J., Zheng, X., Li, Y., Zhang, Q. & Ying, S. Multimodal neuroimaging feature learning with multimodal stacked deep polynomial networks for diagnosis of Alzheimer’s disease. IEEE J. Biomed. Health Inform. 22(1), 173–183 (2017).
Lei, H. et al. Protein–protein interactions prediction via multimodal deep polynomial network and regularized extreme learning machine. IEEE J. Biomed. Health Inform. 23(3), 1290–1303 (2018).
Chrysos, G. G., Moschoglou, S., Bouritsas, G., Deng, J., Panagakis, Y. & Zafeiriou, S. P. Deep polynomial neural networks. IEEE Trans. Pattern Anal. Mach. Intell. https://doi.org/10.1109/TPAMI.2021.3058891 (2021).
Livni, R., Shalev-Shwartz, S. & Shamir, O. An algorithm for training polynomial networks. Comput. Sci. 26(18), 4748–4750 (2013).
Heydari, M., Teimouri, M., Heshmati, Z. & Alavinia, S. M. Comparison of various classification algorithms in the diagnosis of type 2 diabetes in Iran. Int. J. Diabetes Dev. Ctries. 36(2), 167–173 (2016).
Khan, A., Uddin, S. & Srinivasan, U. Chronic disease prediction using administrative data and graph theory: The case of type 2 diabetes. Expert Syst. Appl. 136, 230–241 (2019).
Alirezaei, M., Niaki, S. T. A. & Niaki, S. A. A. A bi-objective hybrid optimization algorithm to reduce noise and data dimension in diabetes diagnosis using support vector machines. Expert Syst. Appl. 127, 47–57 (2019).
Sekar, B. D., Dong, M. C., Shi, J. & Hu, X. Y. Fused hierarchical neural networks for cardiovascular disease diagnosis. IEEE Sens. J. 12(3), 644–650 (2011).
Altan, G., Kutlu, Y. & Allahverdi, N. Deep learning on computerized analysis of chronic obstructive pulmonary disease. IEEE J. Biomed. Health Inform. 24(5), 1344–1350 (2019).
Vuttipittayamongkol, P. & Elyan, E. Improved overlap-based undersampling for imbalanced dataset classification with application to epilepsy and Parkinson’s disease. Int. J. Neural Syst. 30(08), 2050043 (2020).
Krawczyk, B. Learning from imbalanced data: Open challenges and future directions. Progr. Artif. Intell. 5(4), 221–232 (2016).
Sleeman, W. C. IV. & Krawczyk, B. Multi-class imbalanced big data classification on spark. Knowl. Based Syst. 212, 106598 (2021).
Pisner, D. A. & Schnyer, D. M. Support vector machine. In Machine Learning, Academic Press. 101–121 (2020).
Nusinovici, S. et al. Logistic regression was as good as machine learning for predicting major chronic diseases. J. Clin. Epidemiol. 122, 56–69 (2020).
Khateeb, N. & Usman, M. Efficient heart disease prediction system using k-nearest neighbor classification technique. In Proceedings of the International Conference on Big Data and Internet of Thing 21–26 (2017).
Cuesta, H. A., Coffman, D. L., Branas, C. & Murphy, H. M. Using decision trees to understand the influence of individual-and neighborhood-level factors on urban diabetes and asthma. Health Place 58, 102119 (2019).
Kirmani, M. M. Heart disease prediction using multilayer perceptron algorithm. Int. J. Adv. Res. Comput. Sci. 8(5), 1169–1172 (2017).
Soltanzadeh, P. & Hashemzadeh, M. RCSMOTE: Range-controlled synthetic minority over-sampling technique for handling the class imbalance problem. Inf. Sci. 542, 92–111 (2021).
Acknowledgements
Part of this work was accepted by the IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Houston, TX, USA (virtually), December 9-12, 2021, which is cited as reference1.
Funding
This work is supported by National Natural Science Foundation of China (No. 61572090), Graduate Research and Innovation Foundation of Chongqing (No. CYB21068), Chongqing Science and Technology Project (No. cstc2018jscx-mszdX0109) and the Fundamental Research Funds for the Central Universities (No. 2020CDJYGRH-YJ04).
Author information
Authors and Affiliations
Contributions
X.Y. drafted and revised the paper. X.Y. and S.C. conceived and designed the analysis and revision of this paper. X.Y. and C.S. designed and developed the framework. C.S. participated in the design and implementation of the experiment. S.C. and L.Y. have access to the dataset and performed data analysis. All the authors commented on the paper writing and the result discussion.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Yuan, X., Chen, S., Sun, C. et al. A novel early diagnostic framework for chronic diseases with class imbalance. Sci Rep 12, 8614 (2022). https://doi.org/10.1038/s41598-022-12574-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-022-12574-x
This article is cited by
-
Ultrasensitive detection of vital biomolecules based on a multi-purpose BioMEMS for Point of care testing: digoxin measurement as a case study
Scientific Reports (2024)
-
An adaptive multi-class imbalanced classification framework based on ensemble methods and deep network
Neural Computing and Applications (2023)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.