Next Article in Journal
Prospect of Internet of Medical Things: A Review on Security Requirements and Solutions
Next Article in Special Issue
Novel Design of a Multimodal Technology-Based Smart Stethoscope for Personal Cardiovascular Health Monitoring
Previous Article in Journal
Application of PSO-BPNN-PID Controller in Nutrient Solution EC Precise Control System: Applied Research
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quantitative Analysis Using Consecutive Time Window for Unobtrusive Atrial Fibrillation Detection Based on Ballistocardiogram Signal

1
College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110819, China
2
College of Medicine and Biological Information Engineering, Eindhoven University of Technology, 5600 MB Eindhoven, The Netherlands
3
BOBO Technology, Hangzhou 310000, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(15), 5516; https://doi.org/10.3390/s22155516
Submission received: 19 June 2022 / Revised: 19 July 2022 / Accepted: 19 July 2022 / Published: 24 July 2022

Abstract

:
Atrial fibrillation (AF) is the most common clinically significant arrhythmia; therefore, AF detection is crucial. Here, we propose a novel feature extraction method to improve AF detection performance using a ballistocardiogram (BCG), which is a weak vibration signal on the body surface transmitted by the cardiogenic force. In this paper, continuous time windows (CTWs) are added to each BCG segment and recurrence quantification analysis (RQA) features are extracted from each time window. Then, the number of CTWs is discussed and the combined features from multiple time windows are ranked, which finally constitute the CTW–RQA features. As validation, the CTW–RQA features are extracted from 4000 BCG segments of 59 subjects, which are compared with classical time and time-frequency features and up-to-date energy features. The accuracy of the proposed feature is superior, and three types of features are fused to obtain the highest accuracy of 95.63%. To evaluate the importance of the proposed feature, the fusion features are ranked using a chi-square test. CTW–RQA features account for 60% of the first 10 fusion features and 65% of the first 17 fusion features. It follows that the proposed CTW–RQA features effectively supplement the existing BCG features for AF detection.

1. Introduction

Atrial fibrillation (AF) is a common type of arrhythmia, of which the incidence is increasing year by year [1,2]. There are currently 335 million individuals suffering from AF worldwide, with an overall prevalence rate of 2.9% [3,4]. It is particularly important to study the daily real-time and continuous monitoring and diagnosis of AF. At present, AF is widely detected using electrocardiograms (ECGs), the current gold standard for diagnosis [5,6,7,8]. However, ECG-based AF detection has been restricted by the need to attach electrodes on the body surface. Although wearable devices such as smartwatches provide a convenient way of collecting ECG data for daily monitoring, the principle of ECG detection is based on electrophysiology, which is unable to characterize the dynamic changes of the heart activity [9]. Therefore, unobtrusive methods based on cardiac dynamics have been brought into focus, including ballistocardiogram (BCG), seismocardiogram (SCG), and photoplethysmogram (PPG) methods [10,11,12].
Considering that the waveform components of PPG are less complex than those of SCG and BCG, SCG sensors need to be attached to the chest, so the AF detection method based on BCG is widely studied, which reflects the status of the larger cardiovascular system [13,14]. BCG is a noninvasive technique for recording the weak vibration signal on the surface of the body transmitted by the cardio-dynamic force [15]. Over the past decade, the development of BCG-based AF diagnosis through machine learning (ML) and deep learning (DL) has been increasingly investigated [16]. Bruser et al. extracted 17 time and time-frequency features of BCG signals from 10 patients, which were split into 856 nonoverlapping epochs with a length of 30 s each, and applied seven ML algorithms to classify AF, sinus rhythm (SR), and body movement [17]. Yu et al. recorded 12 AF patients’ data, and each BCG recording was split into nonoverlapping 30-s epochs, from which 12 power spectrum features were extracted. In addition, three ML classifiers were applied to distinguish AF and non-atrial-fibrillation (NAF) [18]. Wen et al. collected BCG signals from 37 subjects, which were split into 2915 1-min segments without overlap. Further, 16 energy features were extracted to classify AF and SR by means of five ML algorithms [19]. To substitute the feature extraction, the DL method has been applied to detect AF recently. In our previous study, the Bi-directional Long Short-Term Memory (Bi-LSTM) and phase space reconstruction (PSR) algorithm were employed to represent the 1-D morphology feature and 2-D rhythm feature of BCG signals, respectively, which were integrated by means of a CNN network to improve the robustness of AF detection [20]. In summary, the BCG-based AF detection method generally includes three main procedures: signal acquisition and preprocessing, feature extraction and selection, and classification and performance evaluation. Thereinto, the feature extraction is crucial, including the time feature, frequency feature, time-frequency feature, and non-linear feature.
The recurrence plot (RP) as a non-linear feature was first formally proposed by Eckmann et al. in 1987 [21], which maps a multi-dimensional nonlinear dynamic system by a two-dimensional graph. Recurrence quantification analysis (RQA) is an effective nonlinear tool, which is used to represent the non-stationary characteristics of signals and extract the nonlinear characteristic parameters of waveforms [22]. Currently, RP and RQA have been successfully applied in ECG-based arrhythmia detection. Sun et al. effectively predicted spontaneous termination of AF based on the structure and quantification of RP via ECG signals [23]. Mathunjwa et al. utilized ECG-based RP diagrams and convolutional neural networks (CNNs) to classify arrhythmias [24]. Chen et al. investigated cardiovascular activity changes under exposure to low-frequency noise for various noise intensities by using RP analysis of heart rate variability (HRV) estimation [25].
Considering that BCG has the same rhythm as ECG, we reconstruct the RP of the BCG signal and propose novel RQA features using continuous time windows (CTWs) to improve the accuracy of AF classification effectively in this study. In addition, we collect 4000 BCG segments from 59 subjects, which is more than in previous studies. As a comparison, the classical time and time-frequency features [17] and up-to-date energy features [19] are extracted using the BCG dataset of this study. From the contrast results, the AF classification accuracy of the proposed continuous time window RQA (CTW–RQA) features is superior to the accuracy of the other two existing features. Moreover, the CTW–RQA features are combined with the two types of existing features to constitute the fusion features, which achieve the optimal classification accuracy of 95.63%. To evaluate the importance of the proposed feature, the fusion features are ranked, and the CTW–RQA features account for 60% of the first 10 fusion features and 65% of the first 17 fusion features. It follows that the proposed CTW–RQA features effectively supplement the existing BCG features for AF detection.
The remainder of the paper is organized as follows. The Methods section elaborates the proposed approach including the signal acquisition and preprocessing, RP reconstruction, and CTW–RQA feature extraction and feature fusion. The experimental results are shown in the Results. In the Discussion, the experimental results are analyzed comprehensively. Finally, we summarize the drawbacks and future work in the Conclusions.

2. Methods

The framework of the proposed method is shown in Figure 1.

2.1. Signal Acquisition and Preprocessing

BCG is a non-intrusive measurement of the vibration generated from the human heartbeat and arterial aortic blood circulation, which has the same rhythm as an ECG. In this study, BCG signals were recorded by means of the acquisition system that consisted of polyvinylidene fluoride (PVDF) and data acquisition hardware. The PVDF sensor (W × H: 30 cm × 60 cm, piezoelectric constant > 2 × 10−11 C/N) was placed on the top of a regular bed mattress and located underneath the subjects’ thorax with a sampling rate of 125 Hz (see 2.1 of Figure 1). During the recording process, the raw BCG waveform was amplified, and a Butterworth band-pass filter [17,26] was designed with a passband frequency of 0.7 Hz to 10 Hz to remove the high-frequency noise and low-frequency respiratory components and the motion artifacts, which was aimed at achieving a pure BCG signal. The ECG signal was collected by a CT-08S Holter Recorder at a sampling rate of 200 Hz. In order to address the problem of different sampling rates with BCG signals, the ECG signal was down-sampled to 125 Hz based on synchronized time stamps.
In total, 59 volunteers with paroxysmal AF, aged between 27 to 93 years, participated in this study; there were 34 males and 25 females. The BCG signal of each subject was recorded from 8 pm to 8 am in a lying position. Medical experts manually labeled the BCG signal as AF periods and NAF periods, taking synchronized ECG signals as the gold standard. AF periods and NAF periods were split to 24-s BCG segments, and 2000 AF segments and 2000 NAF segments were achieved as the BCG dataset. For AF classification, 80% of the BCG data were applied to train the classifiers, and the remaining 20% were recognized as independent testing data. In order to ensure the fairness of the experiments, all segments were collected from 59 subjects as evenly as possible. Additionally, the training and testing datasets were derived from different subjects to avoid overfitting.

2.2. RP Reconstruction

Recurrence is a fundamental property of dynamic systems, which can be exploited to characterize the system’s behavior in phase space [27]. A powerful tool for their visualization and analysis, called RP, was introduced in the late 1980s by Eckmann, which reveals all the times when the phase space trajectory of the dynamic system visits roughly the same area in the phase space [21]. Cardiac activity is a dynamic system, so RP is used for nonlinear analysis of the BCG rhythm in this study. Figure 2 shows the BCG reconstruction process.
The basis of RP reconstruction is to reconstruct the phase space. Takens theorem states that a phase space can be reconstructed by the system component after selecting the appropriate delay time τ and embedding dimension m for the time series, which contains all of the information of the original time series [28].
Set the BCG time series { u 1 , u 2 , , u n }, and then the m−dimensional BCG reconstructed vector x i is defined as Equation (1).
x i = [ u i , u i + τ , , u i + ( m 1 ) τ ] ( i = 1 , 2 , , n ( m 1 ) τ )
where m is the embedding dimension and τ is the delay time, i = 1, 2, …; n−(m−1) τ. In this study, the mutual information method is used to determine the parameter τ, and the pseudo-neighborhood method is used to determine the parameter m, which have been discussed in literature [20].
As shown in Figure 2, the distance between the points i and j in the reconstructed phase space can be calculated with Equation (2).
S i j = x i x j ( i = 1 , , n ( m 1 ) τ , j = 1 , , n ( m 1 ) τ ,   x i m )
The RP is a collection of time pairs at the same position in the phase space trajectory in the two-dimensional time domain as shown in Equations (3) and (4).
R i , j m ε i = Θ ( ε i S i j ) =   Θ ( z i j ) ( i , j = 1 N )
Θ ( z i j ) = { 0           i f   z i j > 0 1           otherwise
where the R i , j m ε i is the recursion value, and z i j represents the difference between threshold ε i and S i j .
Equation (4) shows that for the distance between the m-dimensional trajectory of time j and the m-dimensional trajectory of time i, a black point is placed at the coordinates (i, j), otherwise, a white point is placed. A key parameter in the analysis is the threshold ε i , which is chosen as 10% of the maximum phase space diameter based on the theory of the RP [27,29,30,31]. The thresholded recurrence plots (TRPs) of AF and NAF are drawn in Figure 3a,b. In another case, without setting the threshold, the value of point (i, j) in the image matrix is calculated as the Euclidean distance between point i and point j in the phase space. The unthresholded recurrence plots (UTRPs) of AF and NAF are shown in Figure 3c,d.

2.3. CTW–RQA Feature Extraction

RQA defines new measures of complexity by using geometrical structures of RP. According to the characteristic of phase space trajectories, recursive graphs contain typical small-scale structures such as single points, diagonals, vertical lines, and horizontal lines, which constitute the large-scale texture structures. RQA quantifies the small-scale and the large-scale texture structures [31]. In this paper, n consecutive time windows are designed to divide each BCG segment, which refine the region of interest for the original signal. Afterwards, k RQA features are extracted in each time window to constitude the CTW–RQA feature. The feature extraction process is shown in Figure 4.
In Figure 4, Wi are the time windows to be analyzed (i = 1, 2, …, n). Wij refers to the jth RQA feature extracted from the ith time window (j = 1, 2, …, k), k is the number of RQA features. Mi represents the feature vector composed of k RQA features in the ith time window, and the CTW–RQA feature of each BCG segment is defined in Equation (5).
CTW–RQA = [M1,M2, …,Mn]
In this study, k is selected as 13, and the 13 RQA features are shown in Table 1, which are denoted as RQA1, RQA2, …, RQA13.
In Table 1, RR is the percentage of recurrence points in an RP; DET is the percentage of recurrence points that form vertical lines, where the P(l) is the histogram of the lengths l of the diagonal lines; LAM is the percentage of recurrence points that form vertical lines, where P(v) is the histogram of the lengths v of the vertical lines; RATIO is the ratio between DET and RR; L is the expression of the average length of the diagonal lines; TT is the percentage of the average length of the vertical lines; L m a x is the expression of the length of the longest diagonal line; Vmax is the expression of the length of the longest vertical line; DIV is the reciprocal of L m a x ; ENTR is the percentage of the Shannon entropy of the probability distribution of the diagonal line lengths P(l); TREND is the percentage decrease of the RP towards its edges; CLUST is the percentage of the ratio of the number of closed triplets to the number of all triplets; W V m a x is the percentage of the length of the white vertical line.

2.4. Feature Fusion

In order to validate the characterization performance of the proposed features, 17 time and time-frequency features and 16 energy features are extracted in this section. And the three types of features are ranked, and finally the fusion features and the ranked features are derived to improve the accuracy of AF detection.

2.4.1. Time and Time-Frequency Features and Energy Feature Extraction

In the literature [17], 6 time features and 11 time-frequency features were extracted from BCG signals, which were fed into 7 popular ML classifiers to detect AF. The definitions of the 17 time and time-frequency features are shown in Table 2, which are denoted as TTF1, TTF2, …, TTF17.
In Table 2, p p 10 ( l ) is defined as the peak-to-peak amplitude of ten segments that is split from 24-s BCG segments. S [ f , t ] represents the result of the energy spectral density calculation of data by defining a 5-s window, S ( f ) represents the result of the time mean of S [ f , t ] , f peak   represents the distance between peaks of S [ f , t ] , and w peak   [ k ] represents the distance from the kth peak to the average peak height.
In the literature [19], BCG signals were transformed into BCG energy signals in order to highlight the features of AF and NAF, and four new data sequences representing different characteristics of the BCG energy signals were generated. The mean value, variance, skewness, and kurtosis of the four data sequences were calculated, and 16 energy features were extracted for each BCG segment. Five ML algorithms were used to distinguish AF and NAF. The definitions of the 16 energy features are shown in Table 3, which are denoted as E1, E2, …, E16.
In Table 3, P( i ) values are sample points that correspond to peaks, where i indicates the ith peak in the segment; T( i ) comprises the coordinates of troughs; DA(i) denotes the relative difference of the peak amplitude; RT( i ) stores the relative value of the trough amplitude; and BP( i ) was defined to denote the number of burrs between P( i ) and P( i + 1 ).

2.4.2. Feature Ranking and Selection

In order to evaluate the features proposed, four types of features are ranked using the Fisher Score, chi-square test, minimum redundancy–maximum relevance (MRMR) and SHapley Additive exPlanations (SHAP) algorithms.
The Fisher Score is defined as the ratio of inter-class variance to intra-class variance [32], which is calculated as Equations (6)–(8). It can be deduced that when the inter-class variance is larger and the intra-class variance is smaller, the Fisher Score is larger. Therefore, higher ranked features are more discriminatory theoretically.
S B ( k ) = i = 1 C n i n m i ( k ) m ( k ) 2
S w ( k ) = 1 n i = 1 C x w i x ( k ) m i ( k ) 2
J fisher   ( k ) = S B ( k ) S w ( k )
where x ( k ) represents the value of sample x for the kth feature, m i ( k ) represents the mean of the values of class i samples for the kth feature, and m ( k ) represents the mean of the values of all categories of samples for the kth feature. S B ( k ) is defined as the inter-class variance of the kth feature in the data set; S w ( k ) is defined the intra-class variance of the kth feature in the data set; and J fisher   ( k ) defines the Fisher Score of the kth feature in the data set.
The chi-square test is a common hypothesis testing method based on the χ2 distribution of the test statistic. Its null hypothesis H0 is that the observed frequency does not differ from the expected frequency. The basic idea of this test is as follows: assume that H0 is established; then, calculate the χ2 value based on this premise, which represents the degree of deviation between the observed value and the theoretical value. According to the χ2 distribution and degrees of freedom, the probability p of obtaining the current statistic and more extreme cases can be determined under the condition that the H0 hypothesis holds. The chi-square test checks whether each feature is independent of the label. A small p-value for the test statistic indicates that the corresponding feature is dependent on the label, proving that the feature is important [33]. To amplify the difference between features, importance scores are proposed, as shown in Equation (9).
Importance   scores = lg ( p )
The MRMR algorithm finds an optimal set of features that are maximally different from each other and that can effectively represent the label variable [34]. The algorithm calculates the mutual information between features and the mutual information between features and labels to quantify redundancy and correlation. The features are ranked according to the criteria of minimizing the redundancy of the feature set and maximizing the correlation between the feature set and the label variable.
The SHAP algorithm can interpret each feature’s importance to predictions [35]. The SHAP value has been used to guide feature selection [36], which explains the deviation of the prediction for the query sample from the average prediction. In this algorithm, the model needs to be retrained on all feature subsets S ⊆ F, where F represents all features. It assigns an importance value to each feature that represents the effect on the model prediction when including that feature. To compute this effect, a model f S   { i } is trained with that feature present, and another model f S is trained with the feature withheld. Then, a comparison is conducted between the two prediction models with respect to the current input f S   { i } ( x S   { i } ) f S ( x S ) to calculate the SHAP values, where x S represents the input features in the value set S [35]. In this study, each BCG segment is used as x S , and the average of the absolute values of the resulting SHAP values is used to draw a SHAP summary plot and is applied to feature ranking.

2.4.3. Fusion Feature Extraction

Based on the ranking results of the CTW–RQA feature, the first 13 features are selected, which is consistent with the dimensionality of RQA features without CTW. Afterwards, 17 time and time-frequency features, 16 energy features, and 13 ranked CTW–RQA features are combined as 46 fusion features, which provide abundant characterization information. However, there must be irrelevant or redundant information in the fused features. Therefore, 46 fusion features are ranked holistically, and the first 13, 16, and 17 features are selected, which are denoted as the Top 13, Top 16, and Top 17 ranked features. The selected dimensions are consistent with the time and time-frequency features, energy features, and ranked CTW–RQA features. The purpose is to compare the classification performance of ranked features with each type of independent feature in the same dimension. Figure 5 demonstrates the extraction process of the proposed fusion features and ranked features.

3. Results

3.1. AF Detection Based on the RP Diagram

TRP and UTRP are reconstructed from 4000 BCG segments, which are fed into the designed CNN directly. The network mainly consists of eight convolution layers, four pooling layers, four dropout layers, one flatten layer, and one full connection layer and outputs the result of dichotomy, which is successfully performed for AF detection from the BCG signal [20]. The accuracy (ACC), sensitivity (SEN), precision (PRE), and specificity (SPE) are calculated to estimate the AF classification performance, which is shown in Table 4.
From Table 4, the classification performance of UTRP is superior to TRP. However, the classification accuracy of UTRP and TRP is less than 80%, which cannot keep up with the demand of AF screening in routine life.

3.2. AF Detection Based on CTW–RQA Features

Based on Table 1, 13 RQA features are extracted from each 24-s BCG segment, and five ML classifiers are used for training and testing, including K-Nearest Neighbors (KNN), Naive Bayes (NB), Ensemble Learning (ENS), Random Forest (RF), and a Decision Tree (DT) [19,37,38]. The AF detection performance is shown in Table A1 of the Appendix A. As a comparison, CTW–RQA features based on three different time window lengths are extracted, and n is set to 6, 3, and 2. The same ML classifiers are used to detect AF, and the performance is shown in Table A2, Table A3 and Table A4 of the Appendix A. From Table A1, Table A2, Table A3 and Table A4, RF is the optimal classifier, so Table 5 illustrates the AF detection performance based on RQA features and CTW–RQA features by means of RF.
As shown in Table 5, the accuracy of the RQA features and CTW–RQA features is higher than that of the CNN based on UTRP in Table 4. When n is selected as 3, the CTW–RQA features obtain the highest classification accuracy of 89.38%, which is superior to the RQA features without CTW.

3.3. AF Detection Based on Time and Time-Frequency Features and Energy Features

Based on Table 2, 17 time and time-frequency features (6 time domains and 11 time-frequency domains) are extracted from the 4000 BCG segments in this study. In addition, the 4000 BCG segments are transformed into BCG energy signals to extract 16 energy features based on Table 3. The same five ML algorithms are employed to classify AF and NAF. The classification performance is shown in Table A5 and Table A6 of the Appendix A.
From Table A5 and Table A6, RF achieves the optimal classification performance. Therefore, RF is applied to evaluate two types of features with CTW. As a comparison, n is also selected as 6, 3, and 2. The AF detection performance is shown in Table 6 and Table 7.
From Table 6 and Table 7, the time and time-frequency features and energy features without CTW exhibit a superior classification performance and are suitable for feature fusion.

3.4. AF Detection Based on Fusion Features

From Table 5, n is selected as 3. According to Method Section 2.4.2, the Fisher Score, chi-square test, MRMR, and SHAP algorithms are used for feature ranking. The calculated results of 13 RQA features based on four feature ranking algorithms are shown in Table 8.
According to Figure 5, the 39 CTW–RQA features are reduced to 13 CTW–RQA features based on Table 8. Then, 46 fusion features, composed of 13 CTW–RQA features, 17 time and time-frequency features, and 16 energy features, are fed into five ML classifiers to detect AF. The results are shown in Table 9.
As shown in Table 9, RF achieves the optimal classification accuracy, which is 95.63%.

3.5. AF Detection Based on Ranked Features

From Table 9, the classification performance of the fusion features has been obviously improved. However, excessive feature dimensionality may affect the classification efficiency, so 46 fusion features are ranked by the same four feature ranking algorithms. Figure 6 shows the ranking results of the fusion features by the chi-square test method, and Figure A1 in the Appendix A shows the ranking results of the fusion features by the SHAP method.
According to Method Section 2.4.3, the Top 13, Top 16, and Top 17 ranked features are selected based on the ranking results of fusion features, which are fed into the optimal RF classifier. The AF detection accuracy based on the ranked features is shown in Table 10.
From Table 10, the ranked features selected by the chi-square test have the highest classification accuracy. The Top 17 ranked features have a comparable classification performance with the 46 fusion features, which validates the effect of feature selection.

4. Discussion

In this study, CTW–RQA features are proposed to improve the accuracy of AF detection based on BCG signals. The main contributions of our work are the following: (1) RP is first applied to detect AF by means of BCG signals. TRP and UTRP are fed into the designed CNN to classify AF and NAF. (2) The CTW–RQA features are proposed for the first time to quantify the AF rhythm characteristic, which are compared with the 17 classical time and time-frequency features and 16 up-to-date energy features using five ML classifiers. The AF classification accuracy of the proposed feature is superior to the other existing features using the BCG dataset of this study. (3) The CTW–RQA features are combined with the two types of existing features to constitute the fusion features. Then, the fusion features are sorted and selected to constitute the ranked features. The 46 fusion features achieve the optimal classification accuracy of 95.63%, and the 17 ranked features result in a decrease of only 0.25%. It follows that the feature ranking and selection are necessary, and the proposed CTW–RQA features effectively supplement the existing BCG features for AF detection. Furthermore, the reasons underlying the experimental results are analyzed in detail.

4.1. Effects of RP and RQA

From Table 4, the AF detection performance based on TRP and UTRP is compared by means of a CNN, and the results show that UTRP has superior classification accuracy, which is 73.25%. This may be because UTRP contains substantially more information from the signal that generated it than TRP [39]. However, the classification accuracy of both TRP and UTRP could not satisfy the requirement of AF screening. This result may be explained as follows: RP is a sparse image, and non-contiguous areas contain key features, which are difficult to extract by the convolution kernel of the CNN [40]. Therefore, RQA is utilized to emphasize textural features of the RP graph more concretely. From Table 5, 13 RQA features are extracted to input into 5 ML classifiers, and the RF classifier has the optimal performance with an accuracy of 83.50%, which is superior to TRP and UTRP. It follows that RQA features characterize the abnormal rhythm information of BCG signals effectively, which improves the accuracy of AF diagnosis. Therefore, RQA features with the ML classifier are taken as the basis for subsequent feature modifications.

4.2. Effect of the Proposed CTW–RQA Features

In Table 5, the AF classification accuracy of the CTW–RQA features with three types of time windows (n = 6, 3, 2) is compared to the accuracy of the RQA features. The CTW–RQA features with 12-s (n = 2) and 8-s (n = 3) time windows exhibit a higher AF detection accuracy than the RQA features without CTW. However, the CTW–RQA features with a 4-s (n = 6) time window exhibit worse performance. It follows that longer time windows are insensitive to small rhythm changes; conversely, shorter time windows may miss occasional rhythm abnormalities. Therefore, the CTW–RQA feature with 8-s time windows (n = 3) achieves the best classification performance.
Similarly, from Table 6 and Table 7, the time and time-frequency features and energy features without CTW show superior classification performance. This is since the time features are extracted to quantify the amplitude, location information of wave peaks, and inter-peaks in the BCG waveform, so time window addition may miss the effective information. Additionally, limited by the uncertainty principle [41], it is understood that a shorter CTW used to capture the signal segment leads to a poor resolution to represent the signal in the frequency domain, which in turn negatively affects the performance of AF detection. For the energy features, the BCG segments need to be transformed into BCG energy signals, which describe the change in energy distribution. Therefore, longer segmentation lengths are suitable to emphasize the distribution of energy. In summary, CTW is not added to the time and time-frequency features and energy features during feature fusion.
Compared to the time and time-frequency features and energy features without CTW, the proposed CTW–RQA features exhibit the optimal AF classification performance based on Table 5, Table 6 and Table 7. This is likely because the two existing features are more dependent on the variability of the waveform, whereas the data used in this study are derived from a larger number of subjects. Therefore, the proposed CTW–RQA features are more suitable for daily AF screening.

4.3. Effect of Fusion Features and Ranked Features

From Table 9, fusion features exhibit an optimal classification accuracy of 95.63%, which is improved significantly. This is because more dimensions and more types of features provide more information in general. Nevertheless, there is redundant information in the fusion features [42]. Therefore, feature ranking is applied to reduce the dimensions of features and to improve the efficiency of the model.
From Table 10, the features selected by the chi-square test achieve the highest accuracy. This is because this method intuitively quantifies the degree of deviation between features and labels [33] and accurately finds the features strongly correlated with the labels, which is suitable for feature selection in large data sets. Based on the results of the chi-square test, the Top 13, Top 16, and Top 17 ranked features with AF detection accuracy are 95.25%, 95.25%, and 95.38%, respectively, which are comparable to the 46 fusion features. Additionally, with the same feature dimension, the classification accuracy of ranked features is higher than 13 CTW–RQA features, 16 energy features, and 17 time and time-frequency features. The Top 17 ranked features reduce the feature dimensions to improve the efficiency of the algorithm, while the classification result is only 0.25% lower than the 46 fusion features. It reflects the superiority of the ranked features, and the feature selection is necessary and effective.
According to the ranking results based on the chi-square test in Figure 6, there are 6 CTW–RQA features from the first 10 features, and CTW–RQA features account for 65% of the first 17 ranked features, verifying the characterization ability of the CTW–RQA features. Moreover, the proposed features supplement the existing BCG features and effectively improve the performance of AF detection.

4.4. Comparison with Existing Methods

For AF detection based on BCG signals, Bruser et al. extracted the time and time-frequency features from 10 subjects, and the best classifier achieved an accuracy of 96% from 856 BCG segments [17]. Yu et al. extracted the time-frequency features from 12 subjects, and the optimized classifier achieved an accuracy of 94.4% from 7816 BCG segments [18]. Wen et al. extracted the energy features from 37 subjects, and the best classifier achieved an accuracy of 94.5% from 2915 BCG segments [19]. In this study, 4000 BCG segments from 59 subjects were collected, which is more diverse. The more diverse range of subjects enabled the proposed method to be more generalizable, which is more relevant to the application environment of AF screening. Additionally, the AF detection accuracy of proposed features is superior to the accuracy of existing features based on the BCG dataset in this study. The optimal classifier achieves an accuracy of 95.63%, which could satisfy the demand of AF diagnosing in routine life. In summary, the proposed features present an advantage in the field of AF detection based on more diverse BCG datasets, which effectively supplement the existing BCG features.
Compared to existing wearable AF detection devices, the superiority of BCG-based AF detection is as follows. Current technologies employed in wearables and evaluated for AF detection are most often based on single-lead ECG or PPG. For ECG-based AF detection, wearable devices include smartwatches/bands [43,44], smartphones [45,46,47], skin patch recorders [45], hand-held devices [46,47], and smart clothing [48]. These monitoring devices are convenient, but they are based on cardiac electrophysiology and lack the analysis of cardio-dynamic changes. For PPG-based AF detection, wearable devices include smartwatches/bands [49,50,51], smartphones [52,53], and cameras [54]. These monitoring devices can provide cardio-dynamic information, but the accuracy is currently affected by motion artefacts, measurement location, skin conditions, ectopic beats, peripheral vascular disease, poor skin contact, and limited battery life [44,55]. In contrast, BCG-based AF detection devices are usually cushions or mattresses [16], which keep individuals unaware of the test and reflect the status of the larger cardiovascular system [13,14].
However, the limitation of BCG-based AF detection is the presence of motion artefacts, thus stationary conditions are always desired. In state-of-the-art research, the BCG dataset is generally collected during sleep, where high signal-to-noise ratio periods cover most of the acquired data. Compared with the techniques using wearable devices, BCG-based AF detection does not introduce an addition burden. For example, some devices demand individuals to place their fingers on the electrodes of a smartphone [46,47]. Moreover, the AF prevalence increases with age, especially for those aged > 65 years, whereas only 4.6% of smartwatch users in the United States are aged > 65 years [56]. Therefore, BCG-based AF detection devices are more suitable for the elderly or bedridden patients.

5. Conclusions

In this study, we reconstruct BCG signals with the RP method and propose CTW–RQA features to optimize the AF classification performance. The experimental results prove that the proposed features are feasible for AF detection using BCG signals. In the future, we will expand the data volume of BCG signals, including multiple postures and more subjects, to verify the universality of the algorithm. Additionally, the data length for each subject will be increased for personalized analysis for specific subjects. For feature selection, more feature ranking methods, such as the SHAP dependence plot, could be applied to evaluate the importance of the BCG features. The CTW–RQA features and feature selection method in this paper could be extended for the diagnosis of other arrhythmia diseases by means of BCG signals, and the influence of respiratory movement could be considered.

Author Contributions

Conceptualization, F.J.; methodology, T.C. and F.J.; software, T.C. and Q.L.; validation T.C. and Q.L.; data curation, B.Z.; writing—original draft preparation, T.C. and F.J.; writing—review and editing, J.Z. and Q.L.; funding acquisition, F.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 61801104, No. 61902058) and the Northeastern University’s 16th session (2022) Innovation Training Program for College Students (No. 221278).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank Bin Yu for his guidance regarding data acquisition and preprocessing. Thanks are extended to Hangzhou Bobo Technology Co.; Ltd. for providing original BCG data for these experiments.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. The ranking results of fusion features based on the SHAP summary plot.
Figure A1. The ranking results of fusion features based on the SHAP summary plot.
Sensors 22 05516 g0a1
Table A1. AF detection performance based on RQA features.
Table A1. AF detection performance based on RQA features.
MethodACCPRESENSPE
KNN56.3859.4140.2572.50
NB62.8866.7851.2574.50
ENS77.8880.3873.7582.00
RF83.5085.6480.5086.50
DT79.6379.7079.5079.75
Table A2. AF detection performance based on CTW–RQA features (n = 6).
Table A2. AF detection performance based on CTW–RQA features (n = 6).
MethodACCPRESENSPE
KNN56.5065.5236.1978.95
NB69.3869.1575.2462.89
ENS75.7578.5474.0577.63
RF81.7583.7480.9582.63
DT72.2572.9275.0069.21
Table A3. AF detection performance based on CTW–RQA features (n = 3).
Table A3. AF detection performance based on CTW–RQA features (n = 3).
MethodACCPRESENSPE
KNN59.2565.5643.1776.15
NB56.6354.2199.0212.05
ENS81.2580.9582.9379.49
RF89.3890.5288.5490.26
DT78.8878.7680.4977.18
Table A4. AF detection performance based on CTW–RQA features (n = 2).
Table A4. AF detection performance based on CTW–RQA features (n = 2).
MethodACCPRESENSPE
KNN63.1367.5247.3178.24
NB61.1355.8597.7026.16
ENS82.1381.4782.1082.15
RF86.1386.4684.9187.29
DT79.7578.7080.3179.22
Table A5. AF detection performance based on time and time-frequency features.
Table A5. AF detection performance based on time and time-frequency features.
MethodACCPRESENSPE
KNN65.8869.2153.7377.37
NB71.6385.2250.3991.73
ENS84.3883.3384.8383.94
RF87.6385.8989.2086.13
DT79.3876.7982.5276.40
Table A6. AF detection performance based on energy features.
Table A6. AF detection performance based on energy features.
MethodACCPRESENSPE
KNN58.3859.0050.7665.76
NB72.1368.3180.9663.55
ENS75.5074.0377.4173.65
RF78.5077.8978.6878.33
DT73.6372.8274.1173.15

References

  1. Miyasaka, Y.; Barnes, M.E.; Gersh, B.J.; Cha, S.S.; Bailey, K.R.; Abhayaratna, W.P.; Seward, J.B.; Tsang, T.S. Secular trends in incidence of atrial fibrillation in Olmsted County, Minnesota, 1980 to 2000, and implications on the projections for future prevalence. Circulation 2006, 114, 119–125. [Google Scholar] [CrossRef] [Green Version]
  2. Go, A.S.; Hylek, E.M.; Phillips, K.A.; Chang, Y.; Henault, L.E.; Selby, J.V.; Singer, D.E. Prevalence of diagnosed atrial fibrillation in adults: National implications for rhythm management and stroke prevention: The AnTicoagulation and Risk Factors in Atrial Fibrillation (ATRIA) Study. Jama 2001, 285, 2370–2375. [Google Scholar] [CrossRef]
  3. Kirchhof, P. The future of atrial fibrillation management: Integrated care and stratified therapy. Lancet 2017, 390, 1873–1887. [Google Scholar] [CrossRef]
  4. Benjamin, E.J.; Muntner, P.; Alonso, A.; Bittencourt, M.S.; Callaway, C.W.; Carson, A.P.; Chamberlain, A.M.; Chang, A.R.; Cheng, S.; Das, S.R.; et al. American Heart Association Council on Epidemiology and Prevention Statistics Committee and Stroke Statistics Subcommittee. Heart Disease and Stroke Statistics-2019 Update: A Report From the American Heart Association. Circulation 2019, 139, e56–e528. [Google Scholar] [CrossRef]
  5. Ladavich, S.; Ghoraani, B. Rate-independent detection of atrial fibrillation by statistical modeling of atrial activity. Biomed. Signal Process. Control 2015, 18, 274–281. [Google Scholar] [CrossRef]
  6. Ledezma, C.A.; Zhou, X.; Rodríguez, B.; Tan, P.; Diaz-Zuccarini, V. A modeling and machine learning approach to ECG feature engineering for the detection of ischemia using pseudo-ECG. PLoS ONE 2019, 14, e0220294. [Google Scholar] [CrossRef] [Green Version]
  7. Sepulveda-Suescun, J.P.; Murillo-Escobar, J.; Urda-Benitez, R.D.; Orrego-Metaute, D.A.; Orozco-Duque, A. Atrial fibrillation detection through heart rate variability using a machine learning approach and Poincare plot features. In Proceedings of the VII Latin American Congress on Biomedical Engineering CLAIB 2016, Bucaramanga, CO, USA, 26–28 October 2016; Springer: Berlin/Heidelberg, Germany, 2017; pp. 565–568. [Google Scholar]
  8. Dash, S.; Chon, K.H.; Lu, S.; Raeder, E.A. Automatic real time detection of atrial fibrillation. Ann. Biomed. Eng. 2009, 37, 1701–1709. [Google Scholar] [CrossRef]
  9. Guidoboni, G.; Sala, L.; Enayati, M.; Sacco, R.; Szopos, M.; Keller, J.M.; Popescu, M.; Despins, L.; Huxley, V.H.; Skubic, M. Cardiovascular function and ballistocardiogram: A relationship interpreted via mathematical modeling. IEEE Trans. Biomed. Eng. 2019, 66, 2906–2917. [Google Scholar] [CrossRef] [Green Version]
  10. Etemadi, M.; Inan, O.T. Wearable ballistocardiogram and seismocardiogram systems for health and performance. J. Appl. Physiol. 2018, 124, 452–461. [Google Scholar] [CrossRef]
  11. di Rienzo, M.; Meriggi, P.; Rizzo, F.; Vaini, E.; Faini, A.; Merati, G.; Parati, G.; Castiglioni, P. A Wearable system for the seismocardiogram assessment in daily life conditions. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 4263–4266. [Google Scholar]
  12. Allen, J. Photoplethysmography and its application in clinical physiological measurement. Physiol. Meas. 2007, 28, R1. [Google Scholar] [CrossRef] [Green Version]
  13. Rajput, J.S.; Sharma, M.; Kumar, T.S.; Acharya, U.R. Automated detection of hypertension using continuous wavelet transform and a deep neural network with Ballistocardiography signals. Int. J. Environ. Res. Public Health. 2022, 19, 4014. [Google Scholar] [CrossRef]
  14. Han, X.; Wu, X.; Wang, J.; Li, H.; Cao, K.; Cao, H.; Zhong, K.; Yang, X. The latest progress and development trend in the research of ballistocardiography (BCG) and seismocardiogram (SCG) in the field of health care. Appl. Sci. 2021, 11, 8896. [Google Scholar] [CrossRef]
  15. Kim, C.S.; Ober, S.L.; McMurtry, M.S.; Finegan, B.A.; Inan, O.T.; Mukkamala, R.; Hahn, J.O. Ballistocardiogram: Mechanism and potential for unobtrusive cardiovascular health monitoring. Sci. Rep. 2016, 6, 31297. [Google Scholar] [CrossRef] [Green Version]
  16. Jiang, F.; Zhou, Y.; Ling, T.; Zhang, Y.; Zhu, Z. Recent Research for Unobtrusive Atrial Fibrillation Detection Methods Based on Cardiac Dynamics Signals: A Survey. Sensors 2021, 21, 3814. [Google Scholar] [CrossRef]
  17. Brueser, C.; Diesel, J.; Zink, M.D.; Winter, S.; Schauerte, P.; Leonhardt, S. Automatic detection of atrial fibrillation in cardiac vibration signals. IEEE J. Biomed. Health Inform. 2012, 17, 162–171. [Google Scholar] [CrossRef]
  18. Yu, B.; Zhang, B.; Xu, L.; Fang, P.; Hu, J. Automatic detection of atrial fibrillation from ballistocardiogram (BCG) using wavelet features and machine learning. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 4322–4325. [Google Scholar]
  19. Wen, X.; Huang, Y.; Wu, X.; Zhang, B. A feasible feature extraction method for atrial fibrillation detection from BCG. IEEE J. Biomed. Health Inform. 2019, 24, 1093–1103. [Google Scholar] [CrossRef]
  20. Jiang, F.; Hong, C.; Cheng, T.; Wang, H.; Xu, B.; Zhang, B. Attention-based multi-scale features fusion for un obtrusive atrial fibrillation detection using ballistocardiogram signal. BioMedical Eng. OnLine 2021, 20, 12. [Google Scholar] [CrossRef]
  21. Eckmann, J.P.; Kamphorst, S.O.; Ruelle, D. Recurrence plots of dynamical systems. World Sci. Ser. Nonlinear Sci. Ser. A 1995, 16, 441–446. [Google Scholar]
  22. Garcia-Ochoa, E. Recurrence plots: A new methodology for electrochemical noise signal analysis. J. Electroanal. Chem. 2020, 114092, 864. [Google Scholar] [CrossRef]
  23. Sun, R.; Wang, Y. Predicting termination of atrial fibrillation based on the structure and quantification of the recurrence plot. Med. Eng. Phys. 2008, 30, 1105–1111. [Google Scholar] [CrossRef]
  24. Mathunjwa, B.M.; Lin, Y.T.; Lin, C.H.; Abbod, M.F.; Shieh, J.S. ECG arrhythmia classification by using a recurrence plot and convolutional neural network. Biomed. Signal Processing Control 2021, 64, 102262. [Google Scholar] [CrossRef]
  25. Chen, S.T.; Chou, C.Y.; Tseng, L.H. Recurrence Plot Analysis of HRV for Exposure to Low-Frequency Noise. Advanced Materials Research; Trans Tech Publication Ltd.: Freienbach, Switzerland, 2014; Volume 1044, pp. 1251–1257. [Google Scholar]
  26. Panula, T.; Hurnanen, T.; Tuominen, J.; Kaisti, M.; Koskinen, J.; Pänkäälä, M.; Koivisto, T. A Wearable Sensor Node for Detecting Atrial Fibrillation Using Real-Time Digital Signal Processing. In Proceedings of the 2018 25th IEEE International Conference on Electronics, Circuits and Systems (ICECS), Bordeaux, France, 9–12 December 2018; pp. 681–684. [Google Scholar]
  27. Marwan, N.; Romano, M.C.; Thiel, M.; Kurths, J. Recurrence plots for the analysis of complex systems. Phys. Rep. 2007, 438, 237–329. [Google Scholar] [CrossRef]
  28. Su, F.; Dong, H. Prediction of ECG Signal Based on TS Fuzzy Model of Phase Space Reconstruction. In Proceedings of the 2019 12th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Quanzhou, China, 19–21 October 2019; pp. 1–6. [Google Scholar]
  29. Matassini, L.; Kantz, H.; Hołyst, J.; Hegger, R. Optimizing of recurrence plots for noise reduction. Phys. Rev. E 2002, 65, 021102. [Google Scholar] [CrossRef] [Green Version]
  30. Schinkel, S.; Dimigen, O.; Marwan, N. Selection of recurrence threshold for signal detection. Eur. Phys. J. Spec. Top. 2008, 164, 45–53. [Google Scholar] [CrossRef]
  31. Marwan, N. A historical review of recurrence plots. Eur. Phys. J. Spec. Top. 2008, 164, 3–12. [Google Scholar] [CrossRef] [Green Version]
  32. Gu, Q.; Li, Z.; Jia, W. Generalized Fisher score for feature selection. In Proceedings of the Twenty-Seventh Conference on Uncertainty in Artificial Intelligence (UAI’11), Barcelona, Spain, 14–17 July 2011; AUAI Press: Arlington, VA, USA, 2011; pp. 266–273. [Google Scholar]
  33. Ikram, S.T.; Cherukuri, A.K. Intrusion detection model using fusion of chi-square feature selection and multi class SVM. J. King Saud Univ.-Comput. Inf. Sci. 2017, 29, 462–472. [Google Scholar]
  34. Ding, C.; Peng, H. Minimum redundancy feature selection from microarray gene expression data. J. Bioinform. Comput. Biol. 2005, 3, 185–205. [Google Scholar] [CrossRef]
  35. Lundberg, S.M.; Lee, S. A Unified Approach to Interpreting Model Predictions. Adv. Neural Inf. Process. Syst. 2017, 30, 4765–4774. [Google Scholar]
  36. Chen, J.; Yuan, S.; Lv, D.; Xiang, Y. A Novel Self-learning Feature Selection Approach Based on Feature Attributions. Expert Syst. Appl. 2021, 183, 115219. [Google Scholar] [CrossRef]
  37. Kotsiantis, S.B.; Zaharakis, I.D.; Pintelas, P.E. Machine learning: A review of classification and combining techniques. Artif. Intell. Rev. 2006, 26, 159–190. [Google Scholar] [CrossRef]
  38. Caruana, R.; Niculescu-Mizil, A. An empirical comparison of supervised learning algorithms. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 161–168. [Google Scholar]
  39. Aloys, S.; Paul, B.; Ralf, P. Robust reconstruction of a signal from its unthresholded recurrence plot subject to disturbances. Phys. Lett. A 2017, 381, 604–615. [Google Scholar]
  40. Zhao, Z.; Zhang, Y.; Comert, Z.; Deng, Y. Computer-Aided Diagnosis System of Fetal Hypoxia Incorporating Recurrence Plot with Convolutional Neural Network. Front. Physiol. 2019, 10, 255. [Google Scholar] [CrossRef] [Green Version]
  41. Li, X.; Bi, G. Systematic analysis of uncertainty principles of the local polynomial Fourier transform. EURASIP J. Adv. Signal Process. 2014, 181. [Google Scholar] [CrossRef] [Green Version]
  42. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature Selection. ACM Comput. Surv. 2018, 50, 1–45. [Google Scholar] [CrossRef] [Green Version]
  43. Inui, T.; Kohno, H.; Kawasaki, Y.; Matsuura, K.; Ueda, H.; Tamura, Y.; Watanabe, M.; Inage, Y.; Yakita, Y.; Wakabayashi, Y.; et al. Use of a smart watch for early detection of paroxysmal atrial fibrillation: Validation study. JMIR Cardio 2020, 4, 14857. [Google Scholar] [CrossRef]
  44. Rajakariar, K.; Koshy, A.N.; Sajeev, J.K.; Nair, S.; Roberts, L.; Teh, A.W. Accuracy of a smartwatch based single-lead electrocardiogram device in detection of atrial fibrillation. Heart 2020, 106, 665–670. [Google Scholar] [CrossRef]
  45. Shao, M.; Zhou, Z.; Bin, G.; Bai, Y.; Wu, S. A wearable electrocardiogram telemonitoring system for atrial fibrillation detection. Sensors 2020, 20, 606. [Google Scholar] [CrossRef] [Green Version]
  46. Xiong, Z.; Stiles, M.K.; Gillis, A.M.; Zhao, J. Enhancing the detection of atrial fibrillation from wearable sensors with neural style transfer and convolutional recurrent networks. Comput. Biol. Med. 2022, 146, e105551. [Google Scholar] [CrossRef]
  47. Jung, S.; Lee, H.A.; Kang, I.S.; Shin, S.H.; Chang, Y.; Woo, S.D.; Park, M.S.; Kim, Y.D.; Nam, H.S.; Heo, J.H.; et al. Clinical implications of atrial fibrillation detection using wearable devices in patients with cryptogenic stroke (CANDLE-AF) trial: Design and rationale. Front. Cardiovasc. Med. 2022, 9, e837958. [Google Scholar] [CrossRef]
  48. Fukuma, N.; Hasumi, E.; Fujiu, K.; Waki, K.; Toyooka, T.; Komuro, I.; Ohe, K. Feasibility of a T-shirt-type wearable electrocardiography monitor for detection of covert atrial fibrillation in young healthy adults. Sci. Rep. 2019, 9, 11768. [Google Scholar] [CrossRef] [Green Version]
  49. Bonomi, A.G.; Schipper, F.; Eerikäinen, L.M.; Margarito, J.; Aarts, R.M.; Babaeizadeh, S.; Morree, H.M.; Dekker, L. Atrial fibrillation detection using photo-plethysmography and acceleration data at the wrist. In Proceedings of the 2016 Computing in Cardiology Conference (CinC), Vancouver, BC, Canada, 11–14 September 2016; pp. 277–280. [Google Scholar]
  50. Bonomi, A.G.; Schipper, F.; Eerikäinen, L.M.; Margarito, J.; van Dinther, R.; Muesch, G.; Morree, H.M.; Aarts, R.M.; Babaeizadeh, S.; McManusand, D.D.; et al. Atrial fibrillation detection using a novel cardiac ambulatory monitor based on photo-plethysmography at the wrist. J. Am. Heart Assoc. 2018, 7, e009351. [Google Scholar] [CrossRef] [Green Version]
  51. Eerikäinen, L.M.; Bonomi, A.G.; Schipper, F.; Dekker, L.R.; de Morree, H.M.; Vullings, R.; Aarts, R.M. Detecting atrial fibrillation and atrial flutter in daily life using photoplethysmography data. IEEE J. Biomed. Health Inform. 2019, 24, 1610–1618. [Google Scholar] [CrossRef]
  52. Krivoshei, L.; Weber, S.; Burkard, T.; Maseli, A.; Brasier, N.; Kühne, M.; Conen, D.; Huebner, T.; Seeck, A.; Eckstein, J. Smart detection of atrial fibrillation. EP Europace 2017, 19, 753–757. [Google Scholar] [CrossRef] [Green Version]
  53. Estrella-Gallego, A.; Vazquez-Briseno, M.; Nieto-Hipolito, J.I.; Gutierrez-Lopez, E.; Sanchez-Lopez, J.D.; García-Berumen, A.; Rivera-Rodriguez, R. Monitoring atrial fibrillation using PPG signals and a smartphone. Int. J. Online Biomed. Eng. 2020, 16, 60–74. [Google Scholar] [CrossRef] [Green Version]
  54. Couderc, J.P.; Kyal, S.; Mestha, L.K.; Xu, B.; Peterson, D.R.; Xia, X.; Hall, B. Detection of atrial fibrillation using contactless facial video monitoring. Heart Rhythm 2015, 12, 195–201. [Google Scholar] [CrossRef]
  55. Duncker, D.; Ding, W.Y.; Etheridge, S.; Noseworthy, P.A.; Veltmann, C.; Yao, X.; Bunch, T.J.; Gupta, D. Smart wearables for cardiac monitoring—real-world use beyond atrial fibrillation. Sensors 2021, 21, 2539. [Google Scholar] [CrossRef]
  56. Nazarian, S.; Lam, K.; Darzi, A.; Ashrafian, H. Diagnostic accuracy of smartwatches for the detection of cardiac arrhythmia: Systematic review and meta-analysis. J. Med. Internet Res. 2021, 23, e28974. [Google Scholar] [CrossRef]
Figure 1. The process of the proposed method in this paper, where 2.1, 2.2, 2.3, and 2.4 represent the corresponding sections in the Methods.
Figure 1. The process of the proposed method in this paper, where 2.1, 2.2, 2.3, and 2.4 represent the corresponding sections in the Methods.
Sensors 22 05516 g001
Figure 2. RP reconstruction schematic diagram: after the BCG signal is transformed into phase space trajectory, RP is obtained through RP reconstruction, where (a) is the waveform of the 24-s BCG segment, (b) is the phase space trajectory of BCG, and (c) is the RP of BCG.
Figure 2. RP reconstruction schematic diagram: after the BCG signal is transformed into phase space trajectory, RP is obtained through RP reconstruction, where (a) is the waveform of the 24-s BCG segment, (b) is the phase space trajectory of BCG, and (c) is the RP of BCG.
Sensors 22 05516 g002
Figure 3. TRP and UTRP of AF and NAF. (a) is the TRP of AF, and (b) is the TRP of NAF; (c) is the UTRP of AF, and (d) is the UTRP of NAF.
Figure 3. TRP and UTRP of AF and NAF. (a) is the TRP of AF, and (b) is the TRP of NAF; (c) is the UTRP of AF, and (d) is the UTRP of NAF.
Sensors 22 05516 g003
Figure 4. The process of CTW–RQA feature extraction.
Figure 4. The process of CTW–RQA feature extraction.
Sensors 22 05516 g004
Figure 5. The extraction process of the proposed fusion features and ranked features.
Figure 5. The extraction process of the proposed fusion features and ranked features.
Sensors 22 05516 g005
Figure 6. The ranking results of fusion features based on the chi-square test.
Figure 6. The ranking results of fusion features based on the chi-square test.
Sensors 22 05516 g006
Table 1. The definition of 13 RQA features.
Table 1. The definition of 13 RQA features.
SymbolsMeasureDefinition
RQA 1Recurrence rate (RR) R R = 1 N 2 sum i , j = 1 N R i , j
RQA 2Determinism (DET) D E T = l l m i n N l P ( l ) l = 1 N l P ( l )
RQA 3Laminarity (LAM) L A M = v v m i n N v P ( v ) v = 1 N v P ( v )
RQA 4Ratio (RATIO) R A T I O = N 2 l l m i n N l P ( l ) ( l 1 N l P ( l ) ) 2
RQA 5Averaged diagonal line length (L) L = l l m i n N l P ( l ) l l m i n N P ( l )
RQA 6Trapping time (TT) T T = v v m i n N v P ( v ) v = v m i n N P ( v )
RQA 7Longest diagonal line ( L m a x ) L m a x = m a x ( { l i ; i = 1 , , N l } )
RQA 8Longest vertical line ( V m a x ) V m a x = m a x ( { v i ; i = 1 , , N v } )
RQA 9Divergence (DIV) D I V = 1 L m a x
RQA 10Entropy (ENTR) E N T R = l l m i n N p ( l ) ln p ( l )
RQA 11 Trend   ( T R E N D ) T R E N D = i = 1 N ( i N ˜ / 2 ) ( R R i R R i ) i = 1 N ( i N ˜ / 2 ) 2
RQA 12Clustering coefficient (CLUST) C L U S T = CTN TN
RQA 13Longest white vertical line ( W V m a x ) W V m a x = m a x ( { w v i ; i = 1 , , N v } )
Table 2. The definition of 17 time and time-frequency features.
Table 2. The definition of 17 time and time-frequency features.
SymbolsMeasureDefinition
TTF1Standard Deviation s t d ( x [ n ] ) = N N 1 m 2 ( x [ n ] )
TTF2Skewness skewness   ( x [ n ] ) = m 3 ( x [ n ] ) m 2 ( x [ n ] ) 3 / 2
TTF3Kurtosis k u r t o s i s ( x [ n ] ) = m 4 ( x [ n ] ) m 2 ( x [ n ] ) 2
TTF4Range p p ( x [ n ] ) = m a x ( x [ n ] ) m i n ( x [ n ] )
TTF5 Ratio   of   p p 10 ( l )   to   the   M e a n max ( p p 10 ( l ) / m e a n ( p p 10 ( l ) ) )
TTF6 Standard   Deviation   of   p p 10 ( l ) std ( p p 10 ( l ) )
TTF7 The   Standard   Deviation   of   S ¯ [ f ] std   ( S ¯ [ f ] )
TTF8 Skewness   of   S ¯ [ f ] skewness   ( S ¯ [ f ] )
TTF9 Kurtosis   of   S ¯ [ f ] kurtosis   ( S ¯ [ f ] )
TTF10 Standard   Deviation   of   f peak   s t d ( Δ f peak   [ k ] )
TTF11 Skewness   of   f peak   skewness   ( Δ f peak   [ k ] )
TTF12 Kurtosis   of   f peak     kurtosis   ( Δ f peak   [ k ] )
TTF13 Standard   Deviation   of   S [ f , k ] s t d ( s t d t ( S [ f , k ] ) )
TTF14 Average   of   w peak   mean ( w peak   [ k ] )
TTF15 Standard   Deviation   of   w peak   std ( w peak   [ k ] )
TTF16Harmonic Drama Frequency max f d k = 1 [ F f b ] log 10 ( S ¯ m 4 ( k f b ) S ¯ [ ( k + 1 2 ) f b ] )
TTF17Kurtosis of Continuous Time Energy Spectral Density t = 1 T 1 kurtosis   ( xcorr ( S [ f , t ] , S [ f , t + 1 ] ) )
Table 3. The definitions of the 16 energy features.
Table 3. The definitions of the 16 energy features.
SymbolsDefinition
E1 Mean   ( PI ( i ) )
E2 Variance   ( PI ( i ) )
E3 Skewness   ( PI ( i ) )
E4 Kurtosis   ( PI ( i ) )
E5 Mean   ( DA ( i ) )
E6 Variance   ( DA ( i ) )
E7 Skewness   ( DA ( i ) )
E8 Kurtosis   ( DA ( i ) )
E9 Mean   ( RT ( i ) )
E10 Variance   ( RT ( i ) )
E11 Skewness   ( RT ( i ) )
E12 Kurtosis   ( RT ( i ) )
E13 Mean   ( BP ( i ) )
E14 Variance   ( BP ( i ) )
E15 Skewness   ( BP ( i ) )
E16 Kurtosis   ( BP ( i ) )
Table 4. AF detection performance based on RP by means of the CNN.
Table 4. AF detection performance based on RP by means of the CNN.
MethodACCPRESENSPE
TRP56.1373.5038.7559.38
UTRP73.2578.7567.7576.12
Table 5. AF detection performance based on RQA features and CTW–RQA features by means of RF.
Table 5. AF detection performance based on RQA features and CTW–RQA features by means of RF.
FeaturesACCPRESENSPE
RQA83.5085.6480.5086.50
4 s CTW–RQA (n = 6)81.7583.7480.9582.63
8 s CTW–RQA (n = 3)89.3890.5288.5490.26
12 s CTW–RQA (n = 2)86.1386.4684.9187.29
Table 6. AF detection performance based on time and time-frequency features with and without CTW by means of RF.
Table 6. AF detection performance based on time and time-frequency features with and without CTW by means of RF.
FeaturesACCPRESENSPE
Without CWT87.6385.8989.2086.13
4 s CTW (n = 6)79.7369.2153.7377.37
8 s CTW (n = 3)84.6385.2250.3991.73
12 s CTW (n = 2)85.3883.3384.8383.94
Table 7. AF detection performance based on energy features with and without CTW by means of RF.
Table 7. AF detection performance based on energy features with and without CTW by means of RF.
FeaturesACCPRESENSPE
Without CWT78.5077.8978.6878.33
4 s CTW (n = 6)68.1565.5236.1978.95
8 s CTW (n = 3)74.6369.1575.2462.89
12 s CTW (n = 2)76.7878.5474.0577.63
Table 8. The calculated results of 13 RQA features based on four feature ranking algorithms.
Table 8. The calculated results of 13 RQA features based on four feature ranking algorithms.
Fisher ScoreMRMR (×10−3)Chi-Square Test ( lg ( p ) )   Mean (|SHAP Value|) (×10−2)
RQA 126.727515.696886.16619.16
RQA 20.006936.166676.78355.83
RQA 30.061426.1460109.64219.26
RQA 40.00677.406079.86174.03
RQA 50.00257.5923126.03310.75
RQA 60.01179.218469.91017.84
RQA 70.04375.9155108.99721.38
RQA 80.051514.022963.50002.99
RQA 90.04458.534683.14691.31
RQA 100.013614.6556114.17491.69
RQA 110.046324.4646102.50032.02
RQA 120.02826.831495.55153.84
RQA 130.028216.0161118.52872.12
Table 9. AF detection performance based on the fusion features.
Table 9. AF detection performance based on the fusion features.
MethodACCPRESENSPE
KNN58.1384.0723.0695.36
NB81.6385.5277.4286.08
ENS91.7595.2988.3595.36
RF95.6395.8495.6395.62
DT84.5085.8283.7485.31
Table 10. AF detection accuracy based on the ranked features.
Table 10. AF detection accuracy based on the ranked features.
MethodTop 13Top 16Top 17
Fisher’s coefficient91.25%89.63%90.88%
MRMR92.88%93.13%93.63%
Chi-square test95.25%95.25%95.38%
SHAP value93.75%94.13%94.63%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cheng, T.; Jiang, F.; Li, Q.; Zeng, J.; Zhang, B. Quantitative Analysis Using Consecutive Time Window for Unobtrusive Atrial Fibrillation Detection Based on Ballistocardiogram Signal. Sensors 2022, 22, 5516. https://doi.org/10.3390/s22155516

AMA Style

Cheng T, Jiang F, Li Q, Zeng J, Zhang B. Quantitative Analysis Using Consecutive Time Window for Unobtrusive Atrial Fibrillation Detection Based on Ballistocardiogram Signal. Sensors. 2022; 22(15):5516. https://doi.org/10.3390/s22155516

Chicago/Turabian Style

Cheng, Tianqing, Fangfang Jiang, Qing Li, Jitao Zeng, and Biyong Zhang. 2022. "Quantitative Analysis Using Consecutive Time Window for Unobtrusive Atrial Fibrillation Detection Based on Ballistocardiogram Signal" Sensors 22, no. 15: 5516. https://doi.org/10.3390/s22155516

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop