Abstract
Atrial fibrillation remains a major cause of morbi-mortality, making mass screening desirable and leading industry to actively develop devices devoted to automatic AF detection. Because there is a tendency toward mobile devices, there is a need for an accurate, rapid method for studying short inter-beat interval time series for real-time automatic medical monitoring. We report a new methodology to efficiently select highly discriminative variables between physiological states, here a normal sinus rhythm or atrial fibrillation. We generate induced variables using the first ten time derivatives of an RR interval time series and formally express a new multivariate metric quantifying their discriminative power to drive state variable selection. When combined with a simple classifier, this new methodology results in 99.9% classification accuracy for 1-min RR interval time series (n = 7,400), with heart rate accelerations and jerks being the most discriminant variables. We show that the RR interval time series can be drastically reduced from 60 s to 3 s, with a classification accuracy of 95.0%. We show that heart rhythm characterization is facilitated by induced variables using time derivatives, which is a generic methodology that is particularly suitable to real-time medical monitoring.
Similar content being viewed by others
Introduction
Despite constant improvements in the management of atrial fibrillation (AF), this arrhythmia remains a major cause of mortality and severe morbidity. The diagnosis of AF is key in primary and secondary prevention of cardiovascular mortality and morbidity. As it is frequently asymptomatic, AF is often unrecognized. Implementation of systematic AF screening programmes is strongly recommended, particularly in at-risk populations1. In recent years, rapid technological development has led to many new commercially available tools intended to automatically detect arrhythmias, from smartphone applications to smart watches and noncontact ECG methods1,2,3,4,5. Most AF detection methods are based on inter-beat interval time series analysis.
Among the existing approaches dedicated to automatic AF detection, attention has been devoted to RR interval time series rather than atrial activity (p-wave detection). The high-intensity noise often present in ECGs may overlap with the p-wave, limiting AF detection. Various univariate and multivariate techniques have been applied to RR interval analysis. For instance, K. Tateno and L. Glass applied the coefficient of variation and density histograms of the RR interval and ΔRR interval as templates for AF detection6. The differences between the standard histograms (benchmark) and a test record were quantitatively evaluated using the Kolmogorov-Smirnov (KS) test and p-value statistical significance. By comparing the test records, they demonstrated that the density histogram of the ΔRR interval is more accurate for most subjects (average sensitivity of 93.2% and average specificity of 96.7%) than RR interval density histograms. However, short-duration AF leads to a low specificity. S. Dash et al. investigated optimal thresholds of the Turning Points Ratio in combination with the Root Mean Square of Successive RR interval Differences and the Shannon Entropy7. Ectopic beats were suppressed before parameter calculation. The presence of AF using this method was considered if given conditions based on the obtained thresholds were satisfied. The calculated sensitivity and specificity were 94.4% and 95.1%, respectively, for the PhysioNet afdb database and 90.2% and 91.2%, respectively, for the MIT-BIH Arrhythmia Database.
X. Zhou and co-workers detected AF based on symbolic dynamics of ΔRR and Shannon entropy8. The best performance in terms of sensitivity and specificity was 96.72% and 95.07%, respectively, for the training set (Long-Term AF Database). For a test dataset containing AF and normal sinus rhythm (MIT-BIH afdb), the sensitivity and specificity were 96.89% and 98.25%, respectively. However, the authors noted that sporadic AF episodes of relatively short duration (e.g., ten seconds) might incur false negative detection and that this may be a potential limitation.
Although the findings are acceptable in terms of sensitivity and specificity, univariate analysis often refers to a unique feature to make decisions. In contrast, multivariate analysis attempts to couple several variables with confirmed pattern recognition methods, which is beneficial for complex problems such as automatic AF detection. J. Park et al. exploited three features extracted from heart rate variability in Poincaré plots as inputs to K-means clustering methods and Support Vector Machines with radial basis function kernel9. The performance of the SVM was evaluated using a leave-one-out cross-validation method on the Paroxysmal Atrial Fibrillation Prediction Challenge Database10 and the Atrial Fibrillation Termination Challenge Database11. The mean sensitivity and specificity were 91.4% and 92.9%, respectively. M. Carrara et al. compared three multivariate techniques, namely, logistic regression, K-nearest neighbours, and random forests, to distinguish between AF, normal sinus rhythm (NSR) and sinus rhythm (SR) with frequent premature beats12. The random forests analysis applied to five features (mean of RR interval segment, standard deviation, coefficient of sample entropy, local dynamics score and detrended fluctuation analysis) had positive predictive values of 97, 98 and 90% for AF, NSR, and SR with premature beats, respectively. However, this classifier assumed atrial flutter to be the same as AF. R. Acharya et al. classified nine cardiac abnormalities including AF and NSR using multilayer perceptron artificial neural networks and the spectral parameters of the autoregressive moving average (ARMA) model as relevant features13. The classifier achieved 83.33% correct classification of the studied diseases.
As in univariate analysis, these RR interval-based variables (and others) do not yield results with 100% accuracy. All univariate and multivariate analyses required 100- to 200-beat sequences and most authors noted that shorter sequences would lead to dramatic loss of accuracy. Because most reported studies focused on variables extracted from a unique source (i.e., RR interval time series), we believe that time-domain, frequency-domain, and non-linear variables14,15,16,17 do not contain exhaustive information of the underlying heart rhythm. By analogy with kinematics, which states that a dynamical state is better characterized by knowing its position, displacement, speed, and acceleration rather than a single position, we hypothesize that the use of induced physiological variables (i.e., time derivatives of RR interval time series) can extend the set to include additional information sources and enable rapid and efficient clinical decision support. This paper presents a novel methodology to improve physiological and clinical state characterization. We focus on the specific example of heart rhythm characterization. However, it should be noted that we do not propose a new classifier tool to challenge those mentioned above. Rather, we suggest how to generate, select and provide new input variables to existing classifier tools to study induced time series (time-derivative RR interval time series) derived from observed time series that are RR interval time series. The position of our work is illustrated in Fig. 1, which highlights how our methodology increases the field of view of a dynamic physiological phenomenon and the information used to characterize it. This should counteract the lack of observation time in daily medical practice that drastically limits the exploitation of physiological time series18, 19 as advances in technology and computerized calculation enable clinicians to work with time series at the bedside20,21,22,23,24 through new real-time mobile devices under development.
Results
The γ-metric quantifies heart rhythm discriminability
An illustrative simple example of group discrimination using the algebraic geometric distance expressed by the γ-metric is shown in Fig. 2. The γ-metric value can be positive (separate groups along the mean-mean axis, Fig. 2a) or negative (overlapping groups along the mean-mean axis, Fig. 2b). The higher the γ-metric value, the easier the group discrimination is (Fig. 2c–f). When applied to heart rhythm discrimination using an increasing number of variables to describe NSR and AF 1-min RR interval time series, the geometric distance in the selected state space between the NSR and AF groups and the corresponding γ-metric increase when using derivatives up to order three, as depicted in Fig. 2c–f. Inclusion of higher orders of derivation leads to a decrease in the γ-metric by almost 50%. The γ-metric can be used to rank the discriminant power of the variables that describe NSR and AF groups and allows for targeting the most relevant ones.
Derived time series improve heart rhythm characterization
Illustrative examples of NSR and AF ECGs are shown in Fig. 3 with their respective RR interval time series (tachograms) and their first- and second-order time-derived time series. Here, the γ-metric is applied for heart rhythm classification as NSR or AF. Table 1 shows the mean ± standard deviation for the mean and standard deviation operators applied to each studied time series for two groups defined according to their medical classification as NSR or AF. All observed differences were significant (p < 0.001). The corresponding γ-values are also reported. The γ-metric shows that derived time series can lead to improved NSR/AF group discrimination and that, for a given derivative order (from 0 to 10), the application of the standard deviation operator to a time series offers more discriminative power (a higher γ-metric value) than the mean operator. This finding is emphasized in Fig. 4a, which shows the γ-values for the mean and standard deviation operators applied to each time series (derivatives up to the 10th order of the initial tachogram) for the specific NSR/AF classification problem of interest. In addition, Fig. 4a shows that, starting from the initial RR interval tachogram, the γ-metric increases as the derivative order increases up to a certain value (the 2nd order, in this case) and then subsequently decreases. The best γ-value was obtained by applying the standard deviation operator to the second-order derivative of the RR time series, which can be understood as a metric that highlights the heart rhythm dynamics (heart rate jerk). This indicates that if only one variable is to be used to classify a new sample as NSR or AF, better classification should be achieved using the standard deviation of the second-order derivative. To this end, Fig. 4b and c show the percentage accuracies obtained with an unregularized logistic regression classifier trained with each of the 22 variables. The variable selected through the best γ 1 results in one of the best discriminative variable selections after an unregularized logistic regression classifier is used. In this example, the best classification is obtained with the standard deviation of the third-order time derivative, but the classification accuracy difference compared to that obtained with the best γ 1 (0.1%) is not statistically significant (only 1 misclassified sample in the validation set).
Multivariate γ-analysis optimizes heart rhythm discrimination and classification
The correlations between the observed and induced variables are shown in Fig. 4d. Derivatives of the initial time series may be useful in a classification problem because these derivatives are not fully correlated with the initial time series (see the first row or the first column of the figure). As the derivative order increases, the correlations with higher-order derivatives become stronger (increasing along the diagonal towards the bottom right of the figure). However, even with a high-order derivative, it is possible to find a derivative of another order that is only partially correlated with it. Thus, using induced variables to increase the dimensionality of the state space may be beneficial for characterization because uncorrelated induced variables may be interpreted as variables that extract different aspects of information contained in the initially observed variable.
Consequently, more than one variable may be used to address classification problems, and γ-analysis can be used to rank induced variables and their combinations according to their respective discriminant power. The γ-value was calculated for each combination of i ∈ {1..22} variables, and the largest γ-value obtained for each number of selected variables i is presented in Fig. 4e. This multivariate γ-analysis indicates that the maximum discriminatory power (highest γ-value) is achieved with three variables. Adding more variables does not further increase the γ-value; this finding indicates the limited usefulness of such additional variables for this classification problem. The details of the selected variables are summarized in Table 2, where the number of variables is limited to 10 for brevity as three variables are sufficient for classification. The highest γ 1, γ 2 and γ 3 values were obtained with the single variable \(\{\sigma (\frac{{d}^{2}RR}{d{t}^{2}})\,\}\), the pair of variables \(\{\sigma (\frac{{d}^{2}RR}{d{t}^{2}}),\,mean(\frac{{d}^{2}RR}{d{t}^{2}})\,\}\) and the triplet of variables \(\{\sigma (\frac{{d}^{2}RR}{d{t}^{2}}),\,mean(\frac{dRR}{dt}),\,\sigma (\frac{dRR}{dt})\,\}\), respectively, where σ denotes the standard deviation operator. A complete study of all pairs of variables shows that classification accuracy from 69.6% to 99.9% is achieved and that the pair of variables selected by the γ-analysis is among the most discriminative, reaching 99.8%.
Applicability to short time series
Because introducing induced state variables to characterize the derived time series was shown to enable better use of the available information than that achieved using the original data, this approach was used to reduce the observation time required to classify the rhythm state. The evolution of the classification accuracy with decreasing observation time (shorter time series) is depicted in Fig. 5. In this figure, the selected γ 1, γ 2 and γ 3 correspond to the highest γ-value when 1, 2 and 3 variables, respectively, are used to describe the state space according to Table 2. In these cases, Fig. 5 shows that, even for observations as short as 5s, the classification accuracy of logistic regression for this database is higher than 95%, allowing for faster classification.
Robustness of the γ-analysis applied to induced RR interval time series
Data acquired through real-time embedded devices are often corrupted by different sources of perturbations such as timing uncertainty on the real peak positions given by peak detector algorithms applied to noisy data or missing data points. Sinus rhythm labelling of heart rhythm does not exclude the presence of premature contractions, which is not necessarily pathological but is often a source of automatic misclassification. Therefore, to address the robustness of the proposed technique, the variable selection was conducted on a corrupted database. The perturbations used to characterize the robustness of the proposed techniques are the peak detection timing uncertainty (TU), the percentage of missing beats (MB), the percentage of premature atrial contractions (PAC) and the percentage of premature ventricular contractions (PVC). The models used to describe these perturbations are detailed in the methods section. Figure 6 represents how the γ-value evolves when the TU, MB, PAC and PVC perturbations are used to corrupt the rhythm states for separation. The γ-analysis adapts the selected variables that are more prone to discriminate the two groups. For brevity, the variables selected are not detailed here but the red dots (representing the highest γ-value for the fixed amount of perturbation) show that the number of variables needed to discriminate the two AF and NSR groups can increase. This is especially visible in Fig. 6b to d, where the number of variables selected by the γ-analysis is between 4 and 7.
Because the variable selection in the γ-analysis depends on the shapes of the groups to discriminate, a generalized variable selection was attempted with the initial AF and NSR databases to which missing beats, PAC and PVC were added. This mixed NSR database is detailed in the methods section. With these two groups selected, we applied an exhaustive γ-analysis to select 1 to 22 variables for these groups. All 22 groups of variables were used as inputs for an unregularized logistic regression classifier. The accuracies obtained with this classifier are shown in Fig. 7 for the four sources of perturbation and for the selection of variables in the γ-analysis. The resulting best γ variable combination was a group of 11 variables. In this figure, the accuracy of the unregularized logistic regression with the best γ 1, the best γ 11 and the γ 22 variable combination is highlighted. The new best γ 3 variable combination is highlighted to compare with the number of variables selected on noiseless databases. The remainder of the curves are not numerated, but the darkness of the curves gradually increases with the number of variables selected (lightest = 1 variable, darkest = 22 variables). Figure 7 shows that despite the introduced perturbations, the γ-analysis allows for selection of a reduced number of variables associated with high classifier accuracies. Notably, as illustrated in Fig. 7a to c, using more variables does not imply a better accuracy, primarily due to overfitting of the training database. Thus, the γ-analysis helps to select a reduced set of variables to prevent this phenomenon.
The γ-analysis was compared to other techniques used to select variables before use as inputs to a classifier. We compared the best γ 11 selection to no selection at all, to the selection of the first one, first two and first three components of the Principal Component Analysis (PCA) and to the variable selected by the Multiple Discriminant Analysis (MDA). For the PCA, only the first three components were considered because they explained 99.85%, 99.97% and 100% of the variance. The comparison of techniques for the selection of variables and the γ-analysis was done by comparing the consequent logistic regression accuracies. The results are depicted in Fig. 8 for the abovementioned types of perturbations. The results of the classification conducted with variables selected by the γ-analysis appear to be generally similar to those conducted with other well-known techniques. When compared to the best classification results (obtained using the MDA), Fig. 8 shows that the reduced set of variables obtained from the γ-analysis results in similar classification accuracy without the need to process and store all 22 variables used by the hyperplane described by the MDA. Notably, with 99.85% of the input set variance explained by the first component, one could use the first PCA component only and obtain very bad classification results.
Discussion
In this paper, we developed a new approach for characterizing physiological states to solve medical diagnostic challenges and tested it on 7,600 1-min heart rate time series. We highlighted the benefits of automatic AF detection using induced variables as accelerations and jerks calculated from the original instantaneous heart rate. We developed a methodology based on the proposed γ-metric to ensure that the selected induced variables offer high discriminative power to distinguish NSR from AF. We showed that the observation time can be dramatically shortened without significant loss of accuracy using a two-variable model. Finally, we documented the robustness of the proposed γ-analysis for noisy data.
The proposed γ-analysis quantifies how the selected induced variables enable better use of the information contained in the observed variables. The positioning of the proposed technique in the automatic decision process (depicted in Fig. 1) is a preliminary step to select a limited number of relevant variables before application of any classification technique: the γ-analysis is a descriptive tool, not a decision tool. Due to the so-called “curse of dimensionality” that arises when a large number of variables are used as the input of a given classifier, one cannot use all of the available variables. Here, we narrowed our study to a limited number of operators (mean and standard deviation), but one can decide to use all of the well-known variables frequently used in heart rate variability (HRV) analysis such as the SDSD, RMSSD, pNN50, LF/HF power ratio, SD1, SD2, and approximate entropy. Combined with an arbitrary order of time derivation, this can lead to a dramatically large number of variables. To avoid overfitting or non-reasonable training time, one should limit the number of variables used as inputs of a classifier. This is what the proposed γ-metric is used for. The potential advantages are twofold: potentially better classification results (overfitting issue) and lower computational and resource loads required to train the classifier and to process the embedded real-time classification. γ-analysis is complementary to other well-known techniques such as Principal Component Analysis (PCA)25 and Multiple Discriminant Analysis (MDA)26 for variable selection. However, whereas PCA attempts to reduce the number of variables through correlation analysis and MDA attempts to find the best linear combinations of variables for discriminating two or more groups, the γ-analysis selects which variables of an initial set extract the information with the highest discriminative power. As in the γ-analysis, the PCA and MDA techniques help the subsequent classifier to avoid overfitting by reducing the number of input variables. However, as the components of the PCA or the factor for the MDA are a linear combination of the original available variables, all of the original available variables must be processed, stored and combined, which is not desirable in real-time embedded devices. The γ-analysis can also be combined with dimension reduction techniques such as PCA and MDA. Moreover, using the covariance matrix to define the ellipsoids, the proposed technique automatically filters out the outliers. As a result, other combinations of variables than the one selected by the γ-analysis may have better classification accuracy on a specific training or validation database but lack classification accuracy when applied to previously unseen data. Furthermore, for computational simplicity, the γ-metric evaluates the distance between two datasets as the distance between the surfaces of the modelled ellipsoids along their mean-mean axis (Fig. 2a and b). This approach requires fewer computations than processing the intersection volume and allows for selection from several non-intersecting ellipsoids, unlike the volume-based approach.
For a state characterization problem, particularly in the case of complex systems, the relations among the variables may be more important than the variables themselves. For example, for obesity detection, an induced variable such as the body mass index (BMI) would be expected to yield better classification results than height and weight, even when used jointly27. Broadly speaking, induced variables do not generate information but rather enable observation of the available information from another point of view. The induced variables used here for heart rhythm classification were time derivatives. Derivatives of variables with respect to other variables measure how they are locally related and express how a change in one variable affects another variable. Because heart rhythm results from several interacting processes, it exhibits non-linear behaviour28. Time derivatives of heart rate or heart period appear to be adapted to catch heart rhythm dynamics that explore accelerations and jerks of the heart rate through the first and second time derivative of RR interval time series. Notably, this approach appears to be very innovative because although several techniques are listed by the Heart Rate Variability Task Force14, no similar approach is mentioned.
Because AF can have dramatic consequences, recommendations promote mass screening of AF1, encouraging industry to actively develop hardware and software devices devoted to automatic AF detection. Because the trend is toward mobile devices that must be compact and low-power, there is a need for a rapid, accurate method of studying short physiological (i.e., RR interval or equivalent) time series to reach real-time automatic medical monitoring (e-Health). Here, we show that it is possible to optimize the use of the time series’ informational content. We outline and validate a methodology that enables study and discrimination of observed variable dynamics. The soundly computed new time series allow for dramatically reducing the observation time without significant loss of accuracy. We illustrate the usefulness of the methodology by investigating the accuracy of an automatic basic heart rhythm classifier, but there are other similar potential applications in the medical domain. At the individual scale, other physiological signals than ECG could be considered such as EEG, providing new approaches to explore the dynamics of brain electrical activity during epileptic seizures. Time series of various successive discrete measures (self-monitored blood glucose concentrations or blood concentrations of tumour markers) could be used for earlier detection of poorly controlled diabetes or cancer recurrence.
Methods
Time series extraction
The time series studied in this paper were obtained from observations of instantaneous heart cycle periods and consisted of successive RR intervals. They were extracted from the online PhysioNet29 website, from the open MIT-BIH Normal Sinus Rhythm (NSR) Database (nsrdb), the MIT-BIH Normal Sinus Rhythm (NSR) RR Interval Database (nsr2db), and the MIT-BIH Atrial Fibrillation (AF) Database (afdb). The length of the time series was set to correspond to heart rhythm observations spanning 1 min. To obtain the illustrative results depicted in Fig. 2c–f, 20 time series from the NSR group and 20 time series from the AF group were randomly extracted from the nsrdb and afdb databases, respectively. To ensure heart rate overlap between NSR and AF group subsets, the mean heart rates of the randomly selected samples were filtered to obtain 10 subjects in each group (NSR and AF) subset with an average heart rate of 80–100 bpm and 10 subjects in each group (sinus tachycardia and AF) subset with an average heart rate of 100–120 bpm. For simplicity, regardless of the underlying heart rate (normal [80–100 bpm] or increased heart rate [tachycardia, 100–120 bpm]), NSR is used in the remainder of the paper as a generic term to differentiate sinus rhythm from atrial fibrillation. PhysioNet describes the selected databases as including “beat annotation files for 54 long-term ECG recordings of subjects in normal sinus rhythm (30 men, aged 28.5 to 76, and 24 women, aged 58 to 73) and 18 long-term ECG recordings of subjects having no significant arrhythmias (5 men, aged 26 to 45, and 13 women, aged 20 to 50)”. Accordingly, the sinus tachycardia sequences randomly extracted from these databases likely correspond to common daily increased heart rate associated with exercise, REM sleep, anxiety, or coffee drinking, as they are not specifically specified by PhysioNet annotations. To obtain the NSR and AF comparison results depicted in Figs 2 and 4, larger subsets were randomly extracted from the nsrdb, nsr2db and afdb databases. In total, 4,500 and 2,900 1-min RR intervals were collected and used for the NSR and AF groups, respectively.
Although PhysioNet is an extremely rich library of physiological databases that are extensively used for academic and industrial purposes, it is difficult to master the data it contains because they are retrospective and the available information about clinical and acquisition conditions is at least partially deficient. The number of subjects that constitute the databases is limited, and their representativeness is questionable. Our preliminary analysis showed that in the NSR group, the variability of the RR interval of each of the 1-min sequences varied from values that were very different from those of the AF group to values that were very close to or overlapping with the AF group. In contrast, in the AF group, some of the 1-min RR interval sequences were more regular than others and some were sufficiently regular to be similar to those in the NSR group. Inter-individual variability is high in biological and medical problems, particularly in heart rhythm dynamics, and thus, we cannot be certain that NSR and AF data files available from PhysioNet are sufficiently representative of their rhythm categories for our purposes. To minimize this limitation, we randomly selected many 1-min RR interval sequences after pooling and shuffling, which introduced both intra- and inter-individual variability.
Induced Time Series Generation
Time derivatives of the time series were obtained using the forward finite-difference formula, defined as
but similar results are expected when the backward and central difference formulas are used. We studied the time series derivatives up to the 10th order. According to formula (1), induced time series should be correlated and, as correlated variables may not be useful in a classification problem, the correlations between the generated variables were investigated. The correlation matrix was obtained by evaluating the correlations between the ith- and jth-order derivatives of the initial RR interval time series of 20 distinct randomly selected 1-min RR interval time series from the NSR group and 20 distinct randomly selected 1-min RR interval time series from the AF group. The resulting absolute values of the Pearson correlation coefficients are represented in line i and column j by means of a specific colour map.
Time series analysis
The results of 2 mathematical operators applied to each extracted and induced time series were used to characterize each time series. For simplicity, these operators were limited to the arithmetic mean and the unbiased sample variance but could have been any of the typical time, frequency, or informational domain operators14. A two-sided unpaired t-test was used to compare the means of the means and standard deviations of the RR interval time series and their first ten time derivatives for 4,500 samples (NSR group) and 2,900 samples (AF group). The p-values of the differences are provided in Table 1, and p < 0.05 was considered statistically significant.
Discriminant variable selection
Contrary to the Principal Component Analysis (PCA)25 that generates a set of linearly uncorrelated variables from the initial variables or the Multiple Discriminant Analysis (MDA)26 that reduces the number of variables by finding linear combinations of the initial variables and selects only the most discriminant, the selection of the relevant variables from all of the potentially discriminant state variables computed (i.e., mean or standard deviation of the original time series or of one of the first ten time derivatives) is performed using a new metric that we proposed called the γ-metric. The γ-metric is defined as
where K is the number of groups (in this paper, K = 2 for the AF and NSR groups) and d k1k2 is a quantity that is proportional to the algebraic distance between the group denoted k 1 and the group denoted k 2 along their mean-mean axis (Details on the γ-metric are provided in the supplementary information document named Appendix A). This metric is used to select the minimum set of variables that maximizes γ M in the space defined by the M selected variables. In this space, the groups to discriminate are modelled by multidimensional ellipsoids whose axes are defined by the eigenvectors of the selected variables covariance matrix (see Fig. 2a and b for a graphical representation of the algebraic distance between two ellipsoids). A univariate γ-analysis was performed by computing the γ-metric value for each potentially discriminant state variable calculated. A multivariate γ-analysis was performed by computing the γ-metric value for all n-variable (n from 1 up to 22) combinations of the generated variables (means and standard deviations of the RR interval time series and each of the first ten derivatives) as reported in Fig. 4e and Table 2.
Heart rhythm classification
The unregularized logistic regression was arbitrarily chosen to separate the study groups (NSR and AF). The classification accuracy was obtained using this classifier in a univariate and multivariate mode. In the univariate mode, the classifier input variable was the result of applying the mean or standard deviation operator to the 1-min RR interval time series or to one of its first ten time derivatives (resulting in 22 different state variables). In the multivariate mode, the classifier input was a combination of 1 to 22 of these computed state variables. For example, in the case of the highest γ 2 or highest γ 3 in Fig. 5, only the operator-and-variable combinations summarized in Table 2 were considered. To emphasize the interest of using the γ-value to select relevant inputs of the classification model, we studied the classification accuracy of all pairs of the 22 variables. To analyse the logistic regression accuracy according to the length of the observed time series, we challenged the classifier with time series corresponding to lengths of 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 15, 20, 25, 30, 40 and 50 s that were generated from the initial 1-min RR interval time series.
All of the logistic regression accuracy results were obtained using a dataset of RR interval time series partitioned into a training subset (80% of the total dataset) and a validation subset (20% of the total dataset) using the Classification Learner tool from the MATLAB30 Statistics and Machine Learning Toolbox.
RR interval perturbations
To test the robustness of the proposed technique, several sources of perturbations were added to the original AF and NSR databases. These perturbations were added to reflect the typical sources of noise or artefacts that are commonly observed in embedded devices. First, timing uncertainty (TU), reflecting the accuracy of the R peak positioning, was represented as uniformly distributed random variables on each R peak. Such randomly distributed variables were generated in a [−T, T] interval where T ∈ {1, 2, 3, 4, 5, 10, 15, 20, 30, 40, 50, 100} ms. Moreover, R peak detectors sometimes suffer from missing detection due to weak beat amplitudes. To model this phenomenon and the signal disruption, a percentage of missing beats (MB) was added to the initial databases. This percentage relates to the total number of beats for a given sample of the database. If the length of the considered samples was L, (u · L) indexes were randomly selected and the corresponding beats were discarded. The indexes were uniformly distributed in [1, L], and the value of u was taken in {1%, 2%, 3%, 4%, 5%, 10%, 15%, 20%, 30%}. Finally, premature cardiac contractions were also modelled as they may interfere with classification of heart rhythm, especially when AF is considered. Premature atrial contractions (PAC) and premature ventricular contractions (PVC) were modelled with the same percentage technique as that used for the missing beats. The model used for PAC is the insertion of a new beat \(\tilde{b}\) between two original beats indexed by i and i + 1 and denoted by b i and b i+1, respectively. \(\tilde{b}\) was inserted at least 200 ms after b i and 200 ms before b i+1, and b i+1 and the subsequent beats were shifted by \(|\tilde{b}-{b}_{i}|\). The model used for PVC is the insertion of a new beat \(\tilde{b}\) between two original beats b i and b i+1. In this case, \(\tilde{b}\) was inserted at least 200 ms after b i and 200 ms before b i+1, and b i+1 was deleted. Although the NSR samples were corrupted by TU, MB, PAC and PVC, the AF samples were only corrupted by TU and MB.
Pooled and noisy databases were formed by merging 25% of original NSR samples, 25% of NSR samples with MU, 25% of NSR samples with PAC and 25% of NSR samples with PVC. The MU percentages used were uniformly chosen from {1%, 2%, 3%, 4%, 5%, 10%, 15%, 20%}, and the PAC and PVC percentages were uniformly chosen from {1%, 2%, 3%, 4%, 5%, 10%, 15%, 20%, 30%}.
References
Ponikowski, P. et al. ESC Guidelines for the diagnosis and treatment of acute and chronic heart failure: The Task Force for the diagnosis and treatment of acute and chronic heart failure of the European Society of Cardiology (ESC). Developed with the special contribution. Eur. J. Heart Fail. 18, 891–975, doi:10.1093/eurjhf/hfs105 (2016).
Carpenter, A. & Frontera, A. Smart-watches: a potential challenger to the implantable loop recorder? Europace 18, 791 LP–793, doi:1093/europace/euv427 (2016).
Chan, P.-H. et al. Diagnostic Performance of a Smartphone-Based Photoplethysmographic Application for Atrial Fibrillation Screening in a Primary Care Setting. J. Am. Heart Assoc. 5, doi:1161/JAHA.116.003428 (2016).
Fung, E. et al. Electrocardiographic patch devices and contemporary wireless cardiac monitoring. Front. Physiol. 6, doi:3389/fphys.2015.00149 (2015).
Lee, J., Nam, Y., McManus, D. D. & Chon, K. H. Time-Varying Coherence Function for Atrial Fibrillation Detection. IEEE Transactions on Biomedical Engineering 60, 2783–2793, doi:1109/tbme.2013.2264721 (2013).
Tateno, K. & Glass, L. Automatic detection of atrial fibrillation using the coefficient of variation and density histograms of RR and ΔRR intervals. Med. Biol. Eng. Comput. 39, 664–671, doi:1007/bf02345439 (2001).
Dash, S., Chon, K. H., Lu, S. & Raeder, E. A. Automatic Real Time Detection of Atrial Fibrillation. Ann. Biomed. Eng. 37, 1701–1709, doi:1007/s10439-009-9740-z (2009).
Zhou, X., Ding, H., Ung, B., Pickwell-MacPherson, E. & Zhang, Y. Automatic online detection of atrial fibrillation based on symbolic dynamics and Shannon entropy. Biomed. Eng. Online 13, 18, doi:1186/1475-925x-13-18 (2014).
Park, J., Lee, S. & Jeon, M. Atrial fibrillation detection by heart rate variability in Poincare plot. Biomed. Eng. Online 8, 38, doi:1186/1475-925x-8-38 (2009).
Moody, G., Goldberger, A., McClennen, S. & Swiryn, S. Predicting the onset of paroxysmal atrial fibrillation: The Computers in Cardiology Challenge 2001. In Computers in Cardiology 2001, 113–116, doi:1109/cic.2001.977604 (2001).
Moody, G. E. Spontaneous termination of atrial fibrillation: a challenge from physionet and computers in cardiology 2004. In Computers in Cardiology 2004, 101–104, doi:1109/cic.2004.1442881 (2004).
Carrara, M. et al. Heart rate dynamics distinguish among atrial fibrillation, normal sinus rhythm and sinus rhythm with frequent ectopy. Physiol. Meas. 36, 1873, doi:1088/0967-3334/36/9/1873 (2015).
Acharya, U. R., Sankaranarayanan, M., Nayak, J., Xiang, C. & Tamura, T. Automatic identification of cardiac health using modeling techniques: A comparative study. Inf. Sci. (Ny). 178, 4571–4582, doi:1016/j.ins.2008.08.006 (2008).
T Force of The European Society of Cardiology and The North American Society of Pacing and Electrophysiology. Heart Rate Variability: Standards of Measurement, Physiological Interpretation, and Clinical Use. Circulation 93, 1043–1065 (1996).
Pomeranz, B. et al. Assessment of autonomic function in humans by heart rate spectral analysis. Am. J. Physiol. Circ. Physiol 248, H151–H153 (1985).
Cooley, R. L. et al. Evidence for a central origin of the low-frequency oscillation in RR-interval variability. Circulation 98, 556–561, doi:1161/01.cir.98.6.556 (1998).
Richman, J. S. & Moorman, J. R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol 278, H2039–H2049 (2000).
Cazelles, B. & Ferriere, R. H. How predictable is chaos? Nature 355, 25–26, doi:1038/355025b0 (1992).
Barahona, M. & Poon, C.-S. Detection of nonlinear dynamics in short, noisy time series. Nature 381, 215–217, doi:1038/381215a0 (1996).
Goldberger, A. L. Non-linear dynamics for clinicians: chaos theory, fractals, and complexity at the bedside. Lancet 347, 1312–1314, doi:1016/s0140-6736(96)90948-4 (1996).
Pincus, S. M. Approximate entropy as a measure of system complexity. Proc. Natl. Acad. Sci. 88, 2297–2301, doi:1073/pnas.88.6.2297 (1991).
Pincus, S. M. Assessing Serial Irregularity and Its Implications for Health. Ann. N.Y. Acad. Sci. 954, 245–267, doi:1111/j.1749-6632.2001.tb02755.x (2001).
Veiga, J. et al. Approximate entropy as a measure of the airflow pattern complexity in asthma. In 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology 2463–2466, doi:1109/IEMBS.2010.5626547 (2010)
Raoufy, M. R. et al. Classification of Asthma Based on Nonlinear Analysis of Breathing Pattern. PLoS One 11, doi:10.1371/journal.pone.0147976 (2016).
Hotelling, H. Analysis of a complex of statistical variables into principal components. J. Educ. Psychol. 24, 417–441, doi:1037/h0071325 (1933).
Hair, J. F. et al. Multivariate data analysis 6 (Pearson Prentice Hall Upper Saddle River, NJ, 2006).
du V. Florey, C. The use and interpretation of ponderal index and other weight-height ratios in epidemiological studies. J. Chronic Dis. 23, 93–103, doi:1016/0021-9681(70)90068-8 (1970).
Goldberger, A. L. & West, B. J. Applications of Nonlinear Dynamics to Clinical Cardiologya. Ann. N.Y. Acad. Sci. 504, 195–213, doi:1111/j.1749-6632.1987.tb48733.x (1987).
Goldberger, A. L. et al. PhysioBank, PhysioToolkit, and PhysioNet: Components of a New Research Resource for Complex Physiologic Signals. Circulation 101, e215–e220, doi:1161/01.cir.101.23.e215 (2000).
The MathWorks, Inc., MATLAB and Statistics and Machine Learning Toolbox, Release, Natick, Massachusetts, United States (2015b).
Acknowledgements
This work is part of the APPRISE/HIT project and was performed with the support of the A*MIDEX project (n° ANR-11-IDEX-0001-02) funded by the “Investissements d’Avenir” program of the French Government, which is managed by the French National Research Agency (ANR).
Author information
Authors and Affiliations
Contributions
S.D. conceptualized the theoretical and applied problems, designed the study, performed the exploratory data processing, analysed the results, wrote the paper and led the research. J.-F.P. conceptualized the theoretical problem and its mathematical solutions, developed the γ-metric concept and performed its mathematical development, performed extensive data processing and the γ-analysis, analysed the results and wrote the paper. Z.H. extracted the time series from the databases and performed preliminary data processing and classification studies. J.-C.D. formulated the medical problem and additional constraints. M.O. advised on data-processing methods. All authors, including R.B. and A.C., discussed the results and commented on the manuscript.
Corresponding authors
Ethics declarations
Competing Interests
The authors declare that they have no competing interests.
Additional information
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic supplementary material
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Pons, JF., Haddi, Z., Deharo, JC. et al. Heart rhythm characterization through induced physiological variables. Sci Rep 7, 5059 (2017). https://doi.org/10.1038/s41598-017-04998-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-017-04998-7
This article is cited by
-
A filter approach for feature selection in classification: application to automatic atrial fibrillation detection in electrocardiogram recordings
BMC Medical Informatics and Decision Making (2021)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.