Next Article in Journal
CBIR-SAR System Using Stochastic Distance
Next Article in Special Issue
Enhancement in Capacitance of Ionic Type of EAP-Based Strain Sensors
Previous Article in Journal
Data-Driven Insights through Industrial Retrofitting: An Anonymized Dataset with Machine Learning Use Cases
Previous Article in Special Issue
Contactless Technologies, Sensors, and Systems for Cardiac and Respiratory Measurement during Sleep: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Heart Rate Estimation from Facial Image Sequences of a Dual-Modality RGB-NIR Camera †

Department of Electrical Engineering, Center for Innovative Research on Aging Society (CIRAS), and Advanced Institute of Manufacturing with High-Tech Innovations (AIM-HI), National Chung Cheng University, Chia-Yi 621, Taiwan
*
Author to whom correspondence should be addressed.
This paper is an extended version of the paper from 2022 IEEE International Symposium on Circuits and Systems (ISCAS 2022).
Sensors 2023, 23(13), 6079; https://doi.org/10.3390/s23136079
Submission received: 26 May 2023 / Revised: 18 June 2023 / Accepted: 27 June 2023 / Published: 1 July 2023
(This article belongs to the Special Issue Intelligent Health Monitoring Systems Based on Sensor Processing)

Abstract

:
This paper presents an RGB-NIR (Near Infrared) dual-modality technique to analyze the remote photoplethysmogram (rPPG) signal and hence estimate the heart rate (in beats per minute), from a facial image sequence. Our main innovative contribution is the introduction of several denoising techniques such as Modified Amplitude Selective Filtering (MASF), Wavelet Decomposition (WD), and Robust Principal Component Analysis (RPCA), which take advantage of RGB and NIR band characteristics to uncover the rPPG signals effectively through this Independent Component Analysis (ICA)-based algorithm. Two datasets, of which one is the public PURE dataset and the other is the CCUHR dataset built with a popular Intel RealSense D435 RGB-D camera, are adopted in our experiments. Facial video sequences in the two datasets are diverse in nature with normal brightness, under-illumination (i.e., dark), and facial motion. Experimental results show that the proposed method has reached competitive accuracies among the state-of-the-art methods even at a shorter video length. For example, our method achieves MAE = 4.45 bpm (beats per minute) and RMSE = 6.18 bpm for RGB-NIR videos of 10 and 20 s in the CCUHR dataset and MAE = 3.24 bpm and RMSE = 4.1 bpm for RGB videos of 60-s in the PURE dataset. Our system has the advantages of accessible and affordable hardware, simple and fast computations, and wide realistic applications.

1. Introduction

Vital signs monitoring (such as the temperature, heart rate (HR), respiration, blood pressure (BP) [1], pulse rate variability (PRV) [2], etc.) is important for daily care of the elderly or patients. The measuring devices, such as a traditional belt or modern watch, rely on contact with the human body to measure property changes (e.g., photoplethysmography (PPG) or vibration) for physiological parameter inference. These contacting devices take advantage of high accuracy. However, they might not be suitable or preferred for the consideration of inconvenience or the requirement of user-intervention. A remote photoplethysmogram signal (rPPG), reflecting the tiny intensity variation of the skins caused by the heartbeat and following blood flow, hence offers better and more convenient capabilities by removing the demand of physical contact. Additionally, rPPG monitoring provides greater mobility, flexibility, and automation in applications such as robots, cars, or immobile patients [3].
Uncovering an rPPG signal in a facial image sequence, by which the heart rate can be estimated, recently has attracted high attention in research [4,5,6,7,8,9,10,11]. The established methods can be categorized into RGB-based [9], NIR (Near-infrared)-based [10], RGB-NIR fusion [5,6,7], and RGB-NIR-depth fusion methods [8,12]. RGB facial videos are capable of providing up to three channels of intensity information, enabling the extraction of rPPG signals via multivariate analysis techniques where relationships and structures among multivariate measurements are investigated, such as Independent Component Analysis (ICA) [9], Principal Component Analysis (PCA) [13], etc. However, the RGB-based methods are prone to the variations of environmental lighting conditions. The monochrome NIR-based cameras are essentially in the lack of diverse information, but more robust to illumination variations. With the availability of multi-modality camera, research on fusion of multiple information from different sensors (e.g., RGB, NIR [5,6,7], and depth [8]) was well-advanced.
Kado et al. [5] used a mixture of selected facial patches from green- and NIR-channel images for heart rate estimation in both spatial and spectral domains. It was found that accuracy of the RGB-NIR dual-modality is more robust than the RGB single-modality for scenes in low light and with light fluctuations. Later, Kurihara et al. [6] proposed adaptive fusion of RGB and NIR signals by measuring cross-spectral correlations of signals between background and face regions. Recently, Kurihara [7] continued to improve their work with additional motion-robust time-series filters to tackle the difficulty in motion scenario.
Currently, many off-the-shelf multi-modality cameras were on the market. For examples, the 3D RGB-D cameras developed after 2012 by Microsoft (Kinect V1, V2, or Azure) and Intel (e.g., RealSense D400 series) adopted an active-type sensor, meaning that in addition to the passive RGB sensing, NIR light is emitted, reflected, and then received to calculate the depth information. Regev et al. [12] and Yang et al. [14] proposed the capturing of depth video (without RGB information) of a human subject using a Kinect or RealSense camera to estimate the heart rate. On the other hand, Dosso et al. [8] estimated the heart rate based on fusion of three streams (RGB, NIR, and depth) via consensus voting, in contrast to others which used depth for ROI (Region of Interest) extraction or head pose estimation [10,15]. Though depth information was adopted by some of them, the instability in measurement accuracy actually prevents it from accurate heart rate estimation.
Based on the similar goal of RGB-NIR fusion [5,6,7,16], this work investigates the possibility of signal processing based on concatenated RGB-NIR signals from a commercial dual-modality camera. The RealSense D435 model adopts a stereoscopic NIR camera, in addition to an RGB camera, for estimating the depth map by using stereo matching technique. The RGB and one NIR data streams, by ignoring the other NIR and the depth streams, are adopted in this work for estimating the human’s heart rate. Dissimilar to the Kinect V1 camera which adopts coded light technique for NIR emission and the expensive two-plate camera (e.g., JAI AD-130GE) used in [5,6,7] to capture R/G/B/NIR-component signals in spatial synchronization (i.e., optically aligned multi-spectral images), the RealSense D435 camera is in contrast more affordable: compact in size (90 mm in width, see Figure 1), cheaper in price, and more suitable in NIR signal quality, and thus it presents wider applications in such as public health inspection, elder care, etc.
Our contributions mainly come from the reduction of the impacts of poor/dark lighting and facial motions by integrating several denoising techniques such as the Modified Amplitude Selective Filter (MASF), Wavelet Decomposition (WD), and Robust Principal Component Analysis (RPCA). In reality, this might be the first work that fuses RGB-NIR signals for resolving a blind source separation problem. Our method requires only simple computations which usually make applications more practical.
Our experiments were performed on two datasets which show that through efficient denoising steps, a competitive performance among state-of-the-art (SOTA) methods can be achieved. This technical advancement, along with the availability of commercial RGB-NIR dual-modality cameras, will broaden the applications of heart rate estimation, e.g., fitting exercise monitoring in gym, elder care [17], etc.
A preliminary version of this work has been presented in conference [18]. The current work is a substantially extended and improved version of it in the following aspects:
  • The skin segmentation (see Figure 1) in collecting the R/G/B/NIR temporal signals is added so that it will be more robust for motion scenarios.
  • All the experimental results in Table 1, Table 2 and Table 3 are updated.
  • In Table 4, we re-implemented [5] and compared it with our algorithm in RGB-NIR fusion.
  • The sliding-window processing for PURE dataset was modified. The experiments for a 60 s window size are added.
  • The experiments on original PURE dataset (Table 5) are new, and the impacts of compression bit rate for MPEG-4 PURE videos on HR estimation are newly analyzed.
  • The comparison to SOTA methods for original and compressed PURE videos is updated (Table 6).

2. Proposed Methodology

Most of the active RGB-NIR (or, RGB-D) dual-modality cameras, in spite of the much cheaper prices in contrast to the multi-spectral RGB-NIR cameras, have a requirement of spatially aligning RGB/NIR/depth images via homographic transformation. However, misalignments around object boundaries cannot be fully eliminated even after transformation when describing the geometric relationship between two perspective views for non-coplanar 3D points by a simple 2D transform. Even with this possible misalignment, the Intel RealSense camera (D435 model) was still adopted in our work for heart rate estimation by taking advantages of its affordability and popularity but eliminating the impact of the RGB-NIR frame misalignment via face tracking and skin segmentation techniques. Hereafter, we assume non-aligned RGB and NIR images without further notification.
Figure 1 illustrates the flow diagram of our proposed method—Dual-Modality Heart-Rate Estimation (DMHRE).

2.1. ROI Localization and Time Series Signal Formation

For RealSense cameras, there are two kinds of CMOS sensors which are located at a base distance of only a few centimeters and targeted at RGB and NIR sensing band, respectively. Due to different resolutions and FOVs (Field of View) of sensors, RGB and NIR images actually have different sizes and view coverages (see images in Figure 1). We adopt a strategy of not aligning the RGB and NIR images but instead applying face ROI detection separately. Since their corresponding viewpoints are in proximity (at a short base length), this simple arrangement reduces the system complexity and leads to two parallel stream-processing. In this architecture, a co-analysis of RGB and NIR information in temporal axis is conducted in a region-wise, but not pixel-wise, manner, hence eliminating the need of image registration. This probably sacrifices the accuracy to some extent but is surely helpful in providing a simple and fast implementation for practical usage.
Specifically, the Viola–Jones face detector [19] is applied to the first frame of the video sequence. After obtaining a face bounding box with a height H and a width W from the detector for RGB and NIR frame, respectively, the two ROIs (Region of Interest) are defined accordingly (with distinct sizes). For the following frames, a Multiple-Instance Learning (MIL) tracking method (available in OpenCV) [20] is utilized to keep track of the ROIs in each individual stream. However, we focus on skin pixels which reflect the real rPPG signals. Hence, a spatial skin segmentation procedure (based on a LinkNet model [21]) capable of operating on RGB and gray-level channels is performed in individual RGB and NIR ROIs to exclude the non-skin parts (e.g., background, hair, and clothes) for more accurate estimation of the facial masks (i.e., the facial skin part). Finally, pixels in the whole facial mask are spatially averaged to produce a time series of 4 channels (i.e., red, green, blue, and NIR), which can then be arranged to form a 4 × T matrix (T is the number of frames in a temporal window) C for succeeding modules. Though the facial skin areas are actually not uniform in reflecting the HR information (as discussed in [22], which divides the face into 39 anatomical regions), and multiple ROIs (corresponding to patches in face) were used in [23], the whole facial mask obtained was used in this work to calculate the RGB-NIR time series signal for simplicity consideration. In Figure 1, the human’s eyes are blocked simply for privacy consideration (not for computation exclusion).

2.2. Modified Amplitude Selective Filtering (MASF)

Based on the finding that the relative amplitudes (i.e., the DC-normalized signal, AC/DC) of the human pulsatile components are distributed within a lower range (such as [0.0001, 0.002]) for the RGB-channel information, Wang et al. [24] proposed an Amplitude Selective Filter (ASF) to select the RGB frequency components by checking the R-channel spectral amplitude. The principle of the original ASF [24] is to consider any RGB frequency bins whose R-channel amplitudes are higher than the threshold as the noises (mostly due to motion in, e.g., fitness use-cases) and remove them, thus leading to more accurate rPPG signal recovery. However, their method was also vulnerable to the ill-illuminating conditions. In a dark brightness situation, it will be difficult to set up a threshold in amplitude to discriminate the true and noisy HR peaks in power spectrum of the R signals. An example is given in Figure 2a for the blue and red channels, where the noisy peaks are comparable to the HR peak and difficult to discriminate.
While the dependency on RGB color for the rPPG signal will be degraded in the case of low illumination condition, the NIR signal’s spectrum presents more robustness to varying- or under-illuminated conditions. The discrimination between noisy and real HR peaks in NIR’s Power Spectrum Density (PSD) can be observed in Figure 2b, where fewer noisy and larger peak differences are much more distinct. Considering how to overcome both noises due to illumination and motion, we proposed a modified ASF (MASF) algorithm to combat the above challenges.
Comparing the spectra in Figure 2a,b and the observations depicted in [24], better discrimination between the noisy and HR peaks can be achieved by thresholding the NIR spectrum (instead of the R channel in [24]) according to Equation (1) below:
P N I R ( w ) α ,
where α is a pre-determined threshold. That is, frequencies at which the NIR PSD amplitudes are larger than α will be unselected (removed). As the statement in [25,26], the green component has the largest relative PPG contributions, followed by the blue and red components. By inspecting the rPPG example in Figure 2a, the behavior is somewhat similar to the PPG signal, and the blue component contributes a lot of noisy peaks to disturb the extraction of real HR peak. To eliminate the noises, the blue channel is inspected according to a condition (2) below:
P B ( w ) β ( P R ( w ) + P G ( w ) ) ,
where P R ( w ) , P G ( w ) , P B ( w ) , and P N I R ( w ) stand for the amplitudes of the PSD at a given frequency w for R, G, B, and NIR channels, respectively, and β is a constant derived from experimental trials. That is, the blue component P B ( w ) is restricted in magnitudes. It will be considered as noise if the corresponding magnitude is too large. By integrating the above two conditions, frequency ws that do not satisfy either Equation (1) or Equation (2) will not be selected, and their corresponding R/G/B/NIR components in the spectrum will be all suppressed by multiplying with a very small weight. Notice that if zero amplitudes are directly assigned to the suppressed frequencies, the ICA algorithm that follows might not work due to the possibility of (near-) singular covariance matrix [24].
A similar phenomenon has also been demonstrated in a blood-volume-pulse (BVP)-based study [27], where the authors derived distributions of peaks over R, G, and B spectra. In contrast to their goal of improving the motion robustness for application in such as fitness device, our multi-purpose conditions in Equations (1) and (2) consider other factors such as environments and illuminations based on the observations of Figure 2a,b and would be applicable for more rPPG estimations cases. Results after applying Equations (1) and (2) are demonstrated in Figure 2c,d. Clearly, a large portion of noisy peaks have been successfully suppressed which would be helpful to real HR estimation. Notice that the R spectrum in Figure 2b,d is provided for contrastive comparison with the NIR band.
After an Inverse FFT (Fast Fourier Transform) process on the suppressed R/G/B/NIR spectrum, the MASF-filtered observation signals C ^ can be obtained. The MASF in terms of Equations (1) and (2) can be applicable to both single-modality (using Equation (2) only) and dual-modality (using Equations (1) and (2)) HR estimation.
The C ^ signals derived by MASF might be transformed to CIELab color space, but with the luminance channel L being discarded. This will derive a new signal, denoted as M , of a size of 3 × T (i.e., the three components will include a, b, and NIR information). The color conversion is performed based on two considerations. First, the superior performance of HR estimation in CIELab space over the RGB space has been considered in [18,28] for ICA-related algorithms, and the luminance component is less effective in extracting the BVP information [25]. Additionally, the decrease of signal size from 4 × T to 3 × T will help in saving computational loads and speed up the processing in realistic applications (such as fitting devices in gym or elderly caring center).

2.3. Wavelets Decomposition

The wavelet transform (WT) is popularly used in signal analysis which decomposes a signal into both the time and frequency contents. The kernels used in WT jointly consider the characteristics in time and frequency domains, which makes it more powerful in frequency band decomposition and hence in noise filtering/removal. Here, the dyadic stationary wavelet transform (SWT) is adopted, which performs wavelet decomposition in a depth/iterative manner, but skipping the down-sampling procedures [29] between consecutive depth levels. This can avoid the possible reconstruction errors occurring after the inverse transformation process. We execute 1-D SWT for each channel (i.e., R/G/B, or a, b, and NIR) of the M signal. At the first level, M is decomposed into two components: approximation coefficients (AC) which represent the low frequency part and detail coefficients (DC) which represent the high frequency part. For the following levels, the extracted AC signal from the last level is recursively decomposed using the same procedure. By referring to [29], the wavelet type sym4 and two levels are selected in our wavelet decomposition process.
In the above manner, four coefficient parts AC1, DC1, AC2, and DC2 were derived for each channel of M signal. Figure 3 shows such an example of wavelet decomposition. The output signal after WD, denoted as M ^ , hence contains three AC2 components of CIEa, CIEb, and NIR channels, resulting in an M ^ of the same size as M , i.e., 3 × T . The properness of the AC2 components can be based on the following hypothesis. By assuming a spectrum of 2 times of the bandwidth of 4 Hz (i.e., up to 240 bpm), the band of each component after WD will be (1) AC1: 0–4 Hz, DC1: 4–8 Hz and (2) AC2: 0–2 Hz, DC2: 2–4 Hz. Hence, the frequency range of AC2 (0–2 Hz or 0–120 bpm) best matches our requirement in HR estimation. However, this setting might cause larger errors in applications where the bpm is larger than 130 (e.g., in gym fitting exercises). In such an application, DC2 might be a better choice.
Figure 3. WT filtering using 2-level SWT, (top) the input signal and (bottom) the filtered output.
Figure 3. WT filtering using 2-level SWT, (top) the input signal and (bottom) the filtered output.
Sensors 23 06079 g003

2.4. Robust Principle Component Analysis (RPCA)

The goal of PCA in ICA-related processing is to extract major (i.e., of larger energy) and orthonormal components from signals or data. Another functionality of PCA is to reduce dimensionality and extract principal components so that the ICA algorithm that follows can perform efficiently. However, PCA is also known for its vulnerability to noise, which might subsequently degrade the accuracy and efficiency of the ICA analysis. In our work, the noises may be caused by ROI misalignment between the RGB and NIR modalities, wrongly classified background signals in ROIs, etc.
RPCA is a solid tool for ensuring common sparse components among multiple noisy observations. Tulyakov et al. [30] proposed that the temporal smoothness of the HR signal can be modeled as a Matrix Completion (MC) problem, where an unknown low-rank matrix can be recovered to show the most reliable observations from a small set of signals subject to noise or missing data. Their method modeled the HR estimation problem with a complex cost function by considering factors such as temporal smoothing and spatial-temporal masking to exclude large head movement and spontaneous facial expressions. Though an iterative SAMC (Self-Adaptive Matrix Completion) algorithm was proposed to find a low-rank matrix that best approximates the observations and related constraints, it is challenging to converge to the true optimum.
Inspired by Tulyakov’s work [28] and the technique of RPCA (Robust Principal Component Analysis) [31], we would like to model our observations matrix M ^ by
M ^ = L + S
where L stands for a low-rank matrix and S is a complementary part.
In this model, the low-rank matrix L is considered as a “background signal” which is present throughout the video and to be recovered from the highly corrupted measurements M ^ ; the matrix S can be considered as a sparse outlier noise resulted from abrupt motion, illumination change, or region tracking errors. Proposed by [31], the non-convex problem in Equation (3) can be solved by converting it to a convex equation and applying the principal component pursuit (PCP) technique [31] subsequently. The conversion result is as follows:
a r g m i n L L * + λ S 1
subject to
M ^ = L + S ,
where
L * = i σ i ( L )
denotes the nuclear norm of the matrix L, i.e., the sum of singular values of L. S 1 is the l 1 -norm of matrix S, and λ controls the relative proportion of the signal energy that will be transferred to matrix S. The PCP’s main algorithm is based on an Augmented Lagrange Multiplier (ALM) introduced in [32]. The ALM works stably across a large range of problems without the need for parameter tuning. Denoting Y as the Lagrange multiplier matrix and l, as expressed in Equation (6), as the augmented Lagrangian equation, the PCP algorithm iteratively minimizes l with respect to L (by fixing S) in Equation (7) and with respect to S (by fixing L) in Equation (8) and updates Y in Equation (9) based on the residual M ^ L S .
l L , S , Y = L * + Y , M ^ L S + μ 2 M ^ L S F 2
arg min L l L , S , Y = D μ ( M ^ S + μ 1 Y )
arg min S l L , S , Y = S λ μ ( M ^ L + μ 1 Y )
Y = Y + μ M ^ L S
where ‖*‖ denotes the Frobenius norm, D μ X denotes the singular value soft-thresholding operator given by D μ X = U D μ Σ V * for X = U Σ V * , and S λ μ ( x ) is the shrinkage operator equivalent to s g n x m a x ( x λ μ , 0 ) .
The guarantee for PCP solution also scales well with the number of input channels (e.g., more than three) when there are more observations to outline the low-rank and the sparse matrices.

2.5. Independent Component Analysis (ICA) and Fast Fourier Transform (FFT)

The measured rPPG signal can be actually modeled as a combination of real HR signal and other irrelevant noises (e.g., from head motion, facial expression, or environmental illumination), denoted as E, which might also contribute to L in Equation (3). After denoising M ( 3 × T ) and decomposing M ^ ( 3 × T ) to obtain L ( 3 × T ), ICA was often adopted to separate L into the rPPG signal P and other irrelevant sources E. By referring to [9,28], the Joint Approximation Diagonalization of Eigen-matrices (JADE) algorithm [33] was adopted for ICA analysis to derive another 3-component output signals from L, which are then further processed by a customized selection algorithm to identify the best component as P for HR estimation. For each of the 3 output component signals, we locate all the peak positions and calculate the standard deviation (SD) of the peak-to-peak temporal distances. The component signal which has the highest periodicity (i.e., the least SD) will be selected as the one as the rPPG signal P (the other two components are then considered as E). A band-pass filter of order-10, followed by FFT, is then applied to the selected component signal to derive the corresponding spectrum. The position of the largest energy peak in the spectrum is then identified as the HR value.

3. Experimental Results

3.1. Configurations

To evaluate the estimation accuracy of our proposed algorithm, the popular Root-Mean-Square-Error (RMSE) and the Mean Absolute Error (MAE) are used, along with an accuracy assessment defined as follows:
Acc = D S / D T ,
where D S is the number of estimations with errors smaller than 5 beats per minute (bpm) and D T is the total number of estimations. The error threshold of 5 bpm for success actually follows the safety requirements for physiological meters established in ANSI/AAMI EC13-2002.
The datasets used here include the public PURE [34] and the CCUHR [35] datasets. The CCUHR dataset provides 116 dual-modality (unaligned RGB and NIR optics) videos from 22 individuals captured by using an Intel RealSense D435 camera, and the HR ground truths (GTs) were measured simultaneously based on a BIOPAC PPG 100C instrument [36] (with a contacting sensor). All the RGB/NIR videos in the CCUHR dataset have a resolution of 640 × 480 pixels at 30 fps (frames per second) and are divided into two subcategories (with a number of 62 and 54 videos, 116 in total) whose lengths are 10 and 20 s, respectively. These 116 video clips are also categorized into scenarios of non-motion (77, 66.4%, containing good or low illumination without head motion) and motion (39, 33.6%, containing good illumination with facial expression change or medium head rotation). Some examples (localized head ROIs) of the CCUHR dataset are shown in Figure 4.
PURE dataset contains 60 RGB videos, featuring six different motion scenarios (including non-motion and head movements such as talking, rotation, and horizontal transition) from 10 persons, and each video lasts for about 1 min. Additionally, the environmental lightness is lower compared to CCUHR dataset.
All the experiments were conducted on a platform of [email protected] GHz CPU.

3.2. Ablation Experiments on CCUHR Dataset

It can be seen from Figure 4 that no prominent rPPG signals are apparent in the raw R signals, which hence presents some difficulties in estimating HR. As described in Section 2.2 and Section 2.4, there are some parameters (α and β) for MASF and λ for RPCA. Based on our preliminary experiments, shorter signals in principle need a larger α (since a temporal noise contributes to a higher peak in low spectrum for shorter signals). Thus, by experimental experiences, α = 0.004 and 0.003 are chosen for 10 and 20 s signals, respectively. After a coarse-to-fine trial procedure in experiments, a value of β = 1.25 is selected and adopted for all the estimates. For RPCA operation, λ is determined as P 1 / 2 , where P stands for the larger dimension value of the matrix M ^ [31]. For the band-pass filter, the pass band is set to 0.83~2.4 Hz (equivalent to 50~144 bpm). T is set to 300 and 600 for 10 s and 20 s videos, respectively.
In cases of motion videos, inaccurate and varying ROI localizations frequently occur due to imperfect tracking. This can be alleviated by the skin segmentation [21] and non-skin removal after ROI localization so that the refined mask of ROI only captures the human skin part.
Figure 4. Video samples in the CCUHR dataset [35] (eyes are masked for privacy considerations), (a) motion category: facial expression change (bending the brows), (b) motion category: head movement, (c) non-motion category: RGB and NIR frames for dark lighting. The curves represent corresponding red signals over 300 samples (10 s).
Figure 4. Video samples in the CCUHR dataset [35] (eyes are masked for privacy considerations), (a) motion category: facial expression change (bending the brows), (b) motion category: head movement, (c) non-motion category: RGB and NIR frames for dark lighting. The curves represent corresponding red signals over 300 samples (10 s).
Sensors 23 06079 g004
Table 1 shows experimental results for 116 RGB-NIR video clips (of 10 and 20 s) in the CCUHR dataset. The “baseline” represents the technique including only ASF, ICA, and IFFT (Inverse FFT). Subsequently, P1 represents the technique with the addition of WD for denoising; P2 represents the replacement of ASF in P2 by MASF; and P3 represents the full version of our proposed DMHRE algorithm containing all the tools (i.e., plus RPCA) introduced in Section 2. It can be seen that P3 has the best performance by adopting all the denoising techniques.
Table 1. Estimation errors (in bpm) and accuracy on the CCUHR dataset. All methods were implemented with face tracking and skin segmentation in both RGB and NIR streams. CIELab color space is adopted. The best ones are bold-faced.
Table 1. Estimation errors (in bpm) and accuracy on the CCUHR dataset. All methods were implemented with face tracking and skin segmentation in both RGB and NIR streams. CIELab color space is adopted. The best ones are bold-faced.
MethodsRMSE ↓Accuracy ↑MAE ↓SD ↓
Baseline (ASF)9.400.546.876.41
P1 (ASF + WD)7.290.595.364.95
P2 (MASF + WD)6.980.655.064.81
P3 (MASF + WD + RPCA)6.180.664.454.29
Similarly, Table 2 shows a comparison of RGB-NIR fusion against RGB-only for the P3 method operating on individual category (non-motion or motion) of videos of the CCUHR dataset. Two types of RGB-NIR fusion were conducted, which are CIEa-CIEb-NIR and R-G-B-NIR. For the R-G-B-NIR fusion setting, the CIELab color conversion is disabled, and the input matrix M is kept with a size of 4 × T in all the succeeding steps. It is obvious that performances in all metrics were improved substantially by two kinds of RGB-NIR fusion strategy. This proves the capability of NIR information to remove noises. In comparison between different color spaces for fusion, an advantage can be witnessed in using the CIEa-CIEb-NIR setting against the R-G-B-NIR setting for both categories, reaffirming the effectiveness of chrominance channels in terms of hemoglobin’s energy reflection. Noteworthy, using three channels of signals (i.e., CIEa-CIEb-NIR) typically inflicts a lower overhead for the computations (about 14% in our environmental settings).
Table 2. Comparison of RGB-NIR fusion against RGB-only on the CCUHR dataset for the P3 method. The best ones are bold-faced.
Table 2. Comparison of RGB-NIR fusion against RGB-only on the CCUHR dataset for the P3 method. The best ones are bold-faced.
MethodsRMSE ↓Accuracy ↑MAE ↓SD ↓
R-G-B, no motion7.790.585.585.41
R-G-B-NIR, no motion6.420.614.774.29
CIEa-CIEb-NIR, no motion5.260.704.033.37
R-G-B, motion9.150.417.115.77
R-G-B-NIR, motion9.930.496.927.12
CIEa-CIEb-NIR, motion7.690.595.275.60
R-G-B, all8.270.536.095.59
R-G-B-NIR, all7.780.575.495.50
CIEa-CIEb-NIR, all6.180.664.454.29
For practical consideration, it is desirable that a shorter length of video is required for HR estimation. We thus evaluate the performance of P3 on video clips of different lengths, (e.g., 10 and 20 s). The performances are shown in Table 3. As expected, “No motion-20s” has the best performance, while “Motion-20s” is the most challenging one. It is observed that even the video length is reduced from 20 to 10 s and the performance of the “Motion-10s” category is still acceptable. The category of “Motion-20s” achieves the worst result due to inaccurate face tracking for some frames of specific videos (hence resulting in large HR errors). It is thus hypothesized that a specifically designed face tracker (not only the functions [20] provided in OpenCV) also plays an important role in HR estimation, especially concerning motion scenarios.
Table 3. Performances for test videos of different lengths on the CCUHR dataset for the P3 method. The best ones are bold-faced.
Table 3. Performances for test videos of different lengths on the CCUHR dataset for the P3 method. The best ones are bold-faced.
MethodsRMSE ↓Accuracy ↑MAE ↓SD ↓
No motion-10s6.100.614.963.55
No motion-20s3.850.822.792.64
Motion-10s6.090.724.104.50
Motion-20s8.830.486.276.22

3.3. Comparison with the State-of-the-Art Method on the CCUHR Dataset

The RGB-NIR fusion method proposed by [5] for HR estimation is re-implemented for comparison. The modules of face tracker and band-pass filter are the same as in our method, while the landmark detector is based on the Dlib in OpenCV, and the G-G (green-green), G-N (green-NIR), and N-N (NIR-NIR) patch pairs were randomly selected for HR estimation within the area localized by the detected landmarks. Similar to [5], heartrate values contributed to from all patch pairs were fused in a histogram, and the final HR was readout from the peak of the histogram.
Table 4 shows that our DMHRE algorithm (P3 version) actually outperforms [5] on the CCUHR dataset. The higher errors of [5] on the CCUHR dataset might come from the fact that images from different spectral bands (G and NIR) are not optically aligned and have different sizes. The use of smaller patches which are probably misaligned between different times (due to face motion) and different spectral bands (due to RealSense D435 camera) might lead to instability of the extracted temporal signals and hence the HR estimation. In contrast, our algorithm, based on the time series of the whole segmented skin area, presents more stability and hence is promising for wider practical applications.
Table 4. Performances comparison on the CCUHR dataset for our DMHRE method and [5]. Both are based on RGB-NIR fusion.
Table 4. Performances comparison on the CCUHR dataset for our DMHRE method and [5]. Both are based on RGB-NIR fusion.
MethodsRMSE ↓Accuracy ↑MAE ↓SD ↓
No motion (ours)5.130.704.033.16
No motion ([5])22.740.2817.9513.97
Motion (ours)7.260.625.035.24
Motion ([5])30.240.0328.4110.35

3.4. Experiments on the PURE Dataset

In the testing of the PURE dataset, which contains only RGB image information, similar parameters (such as β and λ) are used as in CCUHR, except that a band-pass filter of 0.67–2.4 Hz is used (a wider range). Table 5 shows the estimation errors of our proposed SOTA methods on the PURE dataset. Since the PURE dataset provides only RGB videos, the condition in Equation (1) (i.e., the α threshold) is ignored in our MASF. Additionally, the original RGB color space was used instead of being converted to CIEa and CIEb for combination with NIR components (so that the number of channels of signal will be maintained to be three).
In applying our algorithm, a sliding window with a length of 10 s, 20 s, 30 s, and 60 s is operated (i.e., T = 300, 600, 900, and 1800) with a step size of six frames. In the PURE dataset, the HR GTs were given in a frame basis. The error of the estimated HR in each window will be calculated based on the averaged GTs of the corresponding window. The setting of a six-frame step size is for consideration of the speed of the signal analysis in Figure 1 (i.e., MASF, WD, RPCA, ICA, and FFT) which is about 5.8–9.6 fps (at our i5 platform) so that the whole estimation process can be smooth at 5 fps. The HR errors of all sliding windows are averaged to obtain the MAE and RMSE statistics at each specific window length. We also try to exclude cases of high HR (e.g., >100 bpm) since they are seldom found in applications such as elder health care. In Table 5, it is observed that our best performance occurs at the window size of 60 s (MAE = 3.48 and RMSE = 6.11) if the high Ground Truth (HGT) HRs are considered. By ignoring the cases of HGT HR (about 11% of all the estimations), the performance is better (reduction from MAE = 3.48 to 3.24 and RMSE = 6.11 to 4.1).
The algorithms of [25,37,38] were reimplemented and reported in [39], and those of [25,26] were also re-implemented in [40,41], respectively. Both types of Digital-signal-processing-based (DSP) and recent Deep-learning (DL)-based methods are categorized in Table 5 for references. It was found that the DL-based methods almost outperform the DSP-based methods.
Table 5. Comparison with SOTA methods on the PURE dataset.
Table 5. Comparison with SOTA methods on the PURE dataset.
MethodsYearTypeRMSE ↓MAE ↓
CHROM [25] 1,22013DSP6.8/2.53.82/2.07
LiCVPR [37] 12014DSP30.9628.22
2SR [38] 12015DSP3.062.44
POS [26] 32016DSP10.573.14
NMD-HR [42]2018DSP-8.68
SB-CWT [43]2018DSP7.322.79
Zhao et al. [44]2019DSP4.263.09
ReViSe [45]2022DSP-3.95
DMHRE-10s (ours)2023DSP14.37.73
DMHRE-20s (ours)2023DSP9.414.84
DMHRE-30s (ours)2023DSP9.634.46
DMHRE-60s (ours) w HGT2023DSP6.113.48
DMHRE-60s (ours) w/o HGT2023DSP4.103.24
HR-CNN [39]2018DL2.371.84
PulseGan [40]2021DL4.292.28
EfficientGAN [41]2022DL2.301.83
SA-F [46]2022DL2.832.13
1 re-implemented by [39], 2 re-implemented by [40], 3 re-implemented by [41]. DSP: Digital signal processing-based, DL: Deep-learning-based. HGT: high HR ground truth (>100 bpm).
Table 6. Comparison with SOTA methods on the MPEG-4 PURE video dataset.
Table 6. Comparison with SOTA methods on the MPEG-4 PURE video dataset.
MethodsYearTypeRMSE ↓MAE ↓
CHROM [25] 12013DSP11.366.29
LiCVPR [37] 12014DSP31.128.4
2SR [38] 12015DSP12.815.78
HR-CNN [39]2018DL11.08.72
RhythmNet [47] 22018DL19.6717.5
2-stream CNN [48]2019DL11.819.81
IBIS-CNN [49]2022DL11.999.39
DMHRE-250k-60s
w HGT (ours)
2023DSP14.86.64
DMHRE-250k-60s
w/o HGT (ours)
2023DSP5.393.75
1 re-implemented by [39], 2 re-implemented by [49].
Some of the SOTA methods made experiments on compressed (e.g., MPEG-4) PURE videos. Figure 5 and Figure 6 illustrate the results at different compression bit rates achieved by using “moviepy” package with the “libx264” codec, when the HGT HRs are included or excluded in consideration, respectively. According to the instruction in [39], the compression bit rate they used is at 250 kbps. We conducted experiments by varying the compression bit rates from 250 kbps (about 321× compression ratio) to 1467 kbps (about 56×) and different window sizes from 10 s to 60 s. It can be found that (1) larger compression ratios (or, less bit rate) lead to higher MAE error and (2) a larger window size (e.g., 60 s) leads to a smaller MAE error, while no further improvement is present after the 30-s window length if HGT HRs (>100 bpm) are not excluded. The above two observations reveal that HGT HRs constitute the dominating sources of errors when the facial video is compressed.
Table 6 shows a comparison with SOTA methods for the MPEG-4 compressed PURE video dataset. Our compression bit rate is chosen to be 250 kbps (according to [39]), and the window length is set to 60 s (T = 1800). It was observed that our method even outperforms the DL-based methods. However, it is not confident whether all the methods are compared at the same compression bit rate (the bit rates were not claimed in most of the literatures) since it plays an influential role in evaluation metrics of MAE and RMSE.

4. Discussions

From the comparison in Table 5, even though almost the DL-based methods outperform the DSP-based methods for PURE dataset, the latter ones might require the availability of the GPU (Graphic Processing Unit) as the accelerator in inference, and their performances suffer from significant descent due to different training datasets (for cross-dataset tests, the performances are much degraded).
To be more practical in realistic applications, it is crucial to have a lower requirement on window length T for estimating the accurate HR. From Table 5, in shortening the window length from 60 s to 30 s, 20 s, or even 10 s for our algorithm, the accuracy is descending, to the benefit of higher processing speed and fps rate (not shown in related figures). However, our results still achieve competitive performance, e.g., MAE = 4.84 and RMSE = 9.41 at 20 s. It was understood that accuracy and window length should be kept a tradeoff and the selection of window length and the demand on accuracy depend on respective application.
Another notice is about the frequency resolution for FFT which was addressed in [50]. The bin size, b, in the unit of bpm, for FFT, is defined as b = 60/Ls [50], where Ls is the video length in seconds. This means that for a video of 10 s, the frequency resolution is 6 bpm, and it is 3 bpm for 20 s, 2 bpm for 30 s, and 1 bpm for 60 s. This shows theoretical upper bounds of HR estimation accuracy for different video lengths. Our experimental results in Table 5 match these upper bounds. On the other hand, the results in Table 3 are better than the upper bounds for the “no-motion” cases due to the post-processing we used by performing averaging on several estimations for a sliding window.
For applications that require facial video compression and transmission to remote sites for HR estimation (one of the possibility is for tele-healthcare or tele-medicine), the results in Figure 5 and Figure 6 and Table 6 reveal a limitation since in such a scenario, the tiny high-frequency intensity variation of the skins caused by the heartbeat and following blood flow will be suppressed by the MPEG-4 video compression procedure, thus causing larger estimation errors (especially for HGT HRs which are larger than 100 bpm). Our results in Table 6 show that when HGT HRs are considered, the RMSE is increased significantly, revealing the observation that HR estimation accuracies of the compressed facial videos are limited by the network bandwidth or the compression bitrate.

5. Conclusions

In this work, we have presented a method to estimate HR values from facial video sequence in an accurate and robust manner based on sensors of RGB-NIR dual-modalities, as well as computing techniques of MASF, WD, and RPCA. It is shown that our method, by combining the above filtering and denoising techniques appropriately, is capable of resolving the problem well even with reduced window lengths (e.g., 20–30 s, compared to 60 s which is commonly used in other works). Hence, the demands of a long, static, and well-illuminated video for HR estimation can be eliminated, leading to more practical scenarios. The results will motivate the applications of our research under extreme environments (e.g., in gym fitting exercise and elderly health caring) with the RGB-NIR dual-modality fusion. Even in the absence of NIR information, our proposed method is shown to be capable of achieving competitive accuracies among the SOTA methods due to its good denoising performance.
Our method aims to take advantage of the popular RGB-NIR dual-modality camera on the market (like Intel RealSense D430 series), making our design more promising for wider applications. On the other hand, though DSP-based methods do not own outstanding MAE/RMSE performances as the DL-based methods in terms of some specific datasets, they are of no need in the use of GPU computing resources and have good generalization to different facial image sources (i.e., not highly related to specific training dataset used), which will make them still competitive in reality. A challenge of future work is to further reduce the minimum video length to, for example, 10 or even 5 s and keep similar accuracy so that facial rPPG/HR estimation will be broadened to more realistic applications.

Author Contributions

Conceptualization, W.-N.L.; methodology, D.-Q.L.; software, D.-Q.L. and C.-Y.L. and Y.-S.F.; validation, W.-N.L.; formal analysis, W.-N.L.; resources, W.-N.L.; writing—original draft preparation, W.-N.L. and D.-Q.L.; writing—review and editing, W.-N.L.; supervision, W.-N.L.; funding acquisition, W.-N.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by the Ministry of Education (MOE) and the Ministry of Science and Technology (MOST) in Taiwan (MOST 109-2221-E-194-035-MY3).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was financially supported by the Center for Innovative Research on Aging Society (CIRAS), Advanced Institute of Manufacturing with High-tech Innovations (AIM-HI) from The Featured Areas Research Center Program within the framework of the Higher Education Sprout Project by the Ministry of Education (MOE), as well as the Ministry of Science and Technology (MOST), in Taiwan.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

ACAlternating current
ACApproximation coefficients
ALMAugmented Lagrange Multiplier
BPblood pressure
bpmbeats per minute
bpsbits per second
BSSBlind Source Separation
BVPBlood-volume-pulse
DCDirect current
DCDetail coefficients
DLDeep learning
DMHREDual-Modality Heart-Rate Estimation
DSPDigital signal processing
FFTFast Fourier Transform
FOVField of View
fpsframes per second
GPUGraphic Processing Unit
GTGround truth
HGThigh ground truth
HRheart rate
ICAIndependent Component Analysis
IFFT Inverse Fast Fourier Transform
JADEJoint Approximation Diagonalization of Eigen-matrices
MAEMean absolute error
MASF Modified Amplitude Selective Filtering
MCMatrix Completion
MILMultiple-Instance Learning
NIRNear Infrared
PCPPrincipal component pursuit
PRVpulse rate variability
PSDPower Spectrum Density
RMSERoot mean square error
ROIRegion of Interest
RPCARobust Principal Component Analysis
rPPGremote photoplethysmogram
SAMCSelf-Adaptive Matrix Completion
SOTAState-of-the-art
SWTStationary wavelets transform
w, w/owith, without
WDWavelet decomposition
WTWavelets transform

References

  1. Chen, Y.H.; Zhuang, J.L.; Li, B.; Zhang, Y. Remote Blood Pressure Estimation via the Spatiotemporal Mapping of Facial Videos. Sensors 2023, 23, 2963. [Google Scholar] [CrossRef] [PubMed]
  2. Yu, S.G.; Kim, S.E.; Kim, N.H.; Suh, K.H.; Lee, E.C. Pulse Rate Variability Analysis Using Remote Photoplethysmography Signals. Sensors 2021, 21, 6241. [Google Scholar] [CrossRef]
  3. Chiang, J.C.; Lie, W.N.; Huang, H.C.; Chen, K.T.; Liang, J.Y.; Lo, Y.C.; Huang, W.H. Posture Monitoring for Health Care of Bedridden Elderly Patients Using 3D Human Skeleton Analysis via Machine Learning Approach. Appl. Sci. 2022, 12, 3087. [Google Scholar] [CrossRef]
  4. Jeanne, V.; Asselman, M.; Brinker, B.D.; Bulut, M. Camera-based Heart Rate Monitoring in Highly Dynamic Light Conditions. In Proceedings of the International Conference on Connected Vehicles and Expo (ICCVE), Las Vegas, NV, USA, 2–6 December 2013. [Google Scholar]
  5. Kado, S.; Monno, Y.; Moriwaki, K.; Yoshizaki, K.; Tanaka, M.; Okutomi, M. Remote Heart Rate Measurement from RGB-NIR Video Based on Spatial and Spectral Face Patch Selection. In Proceedings of the 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018. [Google Scholar]
  6. Kurihara, K.; Sugimura, D.; Hamamoto, T. Adaptive Fusion of RGB/NIR Signals Based on Face/Background Cross-Spectral Analysis for Heart Rate Estimation. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019. [Google Scholar]
  7. Kurihara, K.; Sugimura, D.; Hamamoto, T. Non-Contact Heart Rate Estimation via Adaptive RGB/NIR Signal Fusion. IEEE Trans. Image Process. 2021, 30, 6528–6543. [Google Scholar] [CrossRef] [PubMed]
  8. Dosso, Y.S.; Bekele, A.; Green, J.R. Eulerian Magnification of Multi-Modal RGB-D Video for Heart Rate Estimation. In Proceedings of the IEEE International Symposium on Medical Measurements and Applications (MeMeA), Rome, Italy, 11–13 June 2018. [Google Scholar]
  9. Poh, M.Z.; McDuff, D.J.; Picard, R.W. Advancements in Noncontact, Multiparameter Physiological Measurements Using a Webcam. IEEE Trans. Biomed. Eng. 2010, 58, 7–11. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Chen, J.; Chang, Z.; Qiu, Q.; Li, X.; Sapiro, G.; Bronstein, A.; Pietikäinen, M. RealSense = Real Heart Rate: Illumination Invariant Heart Rate Estimation from Videos. In Proceedings of the 2016 Sixth International Conference on Image Processing Theory, Tools and Applications (IPTA), Oulu, Finland, 12–15 December 2016. [Google Scholar]
  11. Kado, S.; Monno, Y.; Yoshizaki, K.; Tanaka, M.; Okutomi, M. Spatial-Spectral-Temporal Fusion for Remote Heart Rate Estimation. IEEE Sens. J. 2020, 20, 11688–11697. [Google Scholar] [CrossRef]
  12. Regev, N.; Wulich, D. Multi-Modal, Remote Breathing Monitor. Sensors 2020, 20, 1229. [Google Scholar] [CrossRef] [Green Version]
  13. Lewandowska, M.; Rumiński, J.; Kocejko, T.; Nowak, J. Measuring Pulse Rate with a Webcam—A Non-contact Method for Evaluating Cardiac Activity. In Proceedings of the Federated Conference on Computer Science and Information Systems (FedCSIS), Szczecin, Poland, 18–21 September 2011. [Google Scholar]
  14. Yang, C.; Cheung, G.; Stankovic, V. Estimating Heart Rate and Rhythm via 3D Motion Tracking in Depth Video. IEEE Trans. Multimed. 2017, 19, 1625–1636. [Google Scholar] [CrossRef] [Green Version]
  15. Nagamatsu, G.; Nowara, E.M.; Pai, A.; Veeraraghavan, A.; Kawasaki, H. PPG3D: Does 3D Head Tracking Improve Camera-based PPG Estimation? In Proceedings of the 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC). Montreal, QC, Canada, 20–24 July 2020. [Google Scholar]
  16. Cheng, J.; Wang, P.; Song, R.; Liu, Y.; Li, C.; Liu, Y.; Chen, X. Remote Heart Rate Measurement from Near-Infrared Videos Based on Joint Blind Source Separation with Delay-Coordinate Transformation. IEEE Trans Instrum. Meas. 2020, 70, 1–13. [Google Scholar] [CrossRef]
  17. Morawski, I.; Lie, W.N.; Aing, L.; Chiang, J.C.; Chen, K.T. Deep-Learning Technique for Risk-Based Action Prediction Using Extremely Low-Resolution Thermopile Sensor Array. IEEE Trans. Circuits Syst. Video Technol. 2022. [Google Scholar] [CrossRef]
  18. Le, D.Q.; Chiang, J.C.; Lie, W.N. Remote PPG Estimation from RGB-NIR Facial Image Sequence for Heart Rate Estimation. In Proceedings of the 2022 IEEE International Symposium on Circuits and Systems (ISCAS), Austin, TX, USA, 27 May 2022; 2022. [Google Scholar]
  19. Viola, P.; Jones, M. Rapid Object Detection Using a Boosted Cascade of Simple Features. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Kauai, HI, USA, 8–14 December 2001. [Google Scholar]
  20. Babenko, B.; Yang, M.H.; Belongie, S. Visual Tracking with Online Multiple Instance Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  21. Nasir6. Available online: https://github.com/nasir6/face-segmentation (accessed on 1 May 2023).
  22. Kim, D.Y.; Lee, K.; Sohn, C.B. Assessment of ROI Selection for Facial Video-Based rPPG. Sensors 2021, 21, 7923. [Google Scholar] [CrossRef]
  23. Molinaro, N.; Schena, E.; Silvestri, S.; Massaroni, C. Multi-ROI Spectral Approach for the Continuous Remote Cardio-Respiratory Monitoring from Mobile Device Built-In Cameras. Sensors 2022, 22, 2539. [Google Scholar] [CrossRef]
  24. Wang, W.; Brinker, A.C.D.; Stuijk, S.; Haan, G.D. Amplitude-selective filtering for remote-PPG. Biomed Opt. Express 2017, 8, 1965–1980. [Google Scholar] [CrossRef] [Green Version]
  25. Haan, G.D.; Jeanne, V. Robust Pulse Rate from Chrominance-Based rPPG. IEEE Trans. Biomed. Eng. 2013, 60, 2878–2886. [Google Scholar] [CrossRef]
  26. Wang, W.; Brinker, A.C.D.; Stuijk, S.; Haan, G.D. Algorithmic Principles of Remote PPG. IEEE Trans. Biomed. Eng. 2017, 64, 1479–1491. [Google Scholar] [CrossRef] [Green Version]
  27. Haan, G.D.; Leest, A.V. Improved Motion Robustness of Remote-PPG by Using the Blood Volume Pulse Signature. Physiol. Meas. 2014, 35, 1913–1926. [Google Scholar] [CrossRef] [Green Version]
  28. Le, D.Q.; Lie, W.N.; Nguyen, Q.N.; Nguyen, T.T.A. Heart Rate Estimation Based on Facial Image Sequence. In Proceedings of the 5th International Conference on Green Technology and Sustainable Development (GTSD), Ho Chi Minh City, Vietnam, 27–28 November 2020. [Google Scholar]
  29. Smital, L.; Vítek, M.; Kozumplík, J.; Provazník, I. Adaptive Wavelet Wiener Filtering of ECG Signals. IEEE Trans. Biomed. Eng. 2012, 60, 437–445. [Google Scholar] [CrossRef]
  30. Tulyakov, S.; Alameda-Pineda, X.; Ricci, E.; Yin, L.; Cohn, J.F.; Sebe, N. Self-Adaptive Matrix Completion for Heart Rate Estimation from Face Videos under Realistic Conditions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  31. Candés, E.J.; Li, X.; Ma, Y.W.; Wright, J. Robust Principal Component Analysis? J. ACM 2011, 58, 1–37. [Google Scholar] [CrossRef]
  32. Yuan, X.; Yang, J.F. Sparse and Low Rank Matrix Decomposition via Alternating Direction Method. Pac. J. Optim. 2009, 9, 1–11. [Google Scholar]
  33. Cardoso, J.F. High-Order Contrasts for Independent Component Analysis. Neural Comput. 1999, 11, 157–192. [Google Scholar] [CrossRef]
  34. Stricker, R.; Müller, S.; Gross, H.M. Non-contact Video-based Pulse Rate Measurement on a Mobile Service Robot. In Proceedings of the 23rd IEEE International Symposium on Robot and Human Interactive Communication, Edinburgh, UK, 25–29 August 2014. [Google Scholar]
  35. Visit the CCUHR Dataset. Available online: https://ccu365-my.sharepoint.com/:u:/g/personal/ieewnl_office365_ccu_edu_tw/EcBZOCHSeKNMi1gMGkREqk8BnHcT7pjpnzCPprCsHekSUA?e=aKFjyn (accessed on 1 May 2023).
  36. “BIOPAC’s Pulse Plethysmogram Amplifier”, BIOPAC System, Inc., [Online]. Available online: https://www.biopac.com/product/pulse-plethysmogram-amplifier/ (accessed on 1 May 2023).
  37. Li, X.; Chen, J.; Zhao, G.; Pietikäinen, M. Remote Heart Rate Measurement from Face Videos Under Realistic Situations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  38. Wang, W.; Stuijk, S.; Haan, G.D. A Novel Algorithm for Remote Photoplethysmography: Spatial Subspace Rotation. IEEE Trans. Biomed. Eng. 2016, 63, 1974–1984. [Google Scholar] [CrossRef] [PubMed]
  39. Spetlik, R.; Franc, V.; Cech, J.; Matas, J. Visual Heart Rate Estimation with Convolutional Neural Network. In Proceedings of the 29th British Machine Vision Conference, Newcastle, UK, 3–6 September 2018. [Google Scholar]
  40. Song, R.; Chen, H.; Cheng, J.; Li, C.; Liu, Y.; Chen, X. PulseGAN: Learning to Generate Realistic Pulse Waveforms in Remote Photoplethys-mography. IEEE J. Biomed. Health Inform. 2021, 25, 1373–1384. [Google Scholar] [CrossRef] [PubMed]
  41. Comas, J.; Ruiz, A.; Sukno, F. Efficient Remote Photoplethysmography with Temporal Derivative Modules and Time-Shift Invariant Loss. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022. [Google Scholar]
  42. Demirezen, H.; Erdem, C.E. Remote Photoplethysmography Using Nonlinear Mode Decomposition. In Proceedings of the 2018 IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 1060–1064. [Google Scholar]
  43. Finžgar, M.; Podraj, P. A Wavelet-based Decomposition Method for a Robust Extraction of Pulse Rate Signal from Video Recordings. Peer J. 2018, 6, e5859. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Zhao, C.; Mei, P.; Xu, S.; Li, Y.Q.; Feng, Y. Performance Evaluation of Visual Object Detection and Tracking Algorithms Used in Remote Photoplethysmography. In Proceedings of the 2019 IEEE International Conference on Computer Vision (ICCV) Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  45. Qiao, D.; Ayesha, A.H.; Zulkernine, F.; Masroor, R.; Jaffar, N. ReViSe: Remote Vital Signs Measurement Using Smartphone Camera. IEEE Access 2022, 10, 131656–131670. [Google Scholar] [CrossRef]
  46. Wu, B.F.; Wu, Y.C.; Chou, Y.W. A Compensation Network with Error Mapping for Robust Remote Photoplethysmography in Noise-Heavy Conditions. IEEE Trans. Instrum. Meas. 2022, 71, 1–11. [Google Scholar] [CrossRef]
  47. Niu, X.; Han, H.; Shan, S.; Chen, X. VIPL-HR: A Multi-modal Database for Pulse Estimation from Less-constrained Face Video. In Proceedings of the Asian Conference on Computer Vision (ACCV), Perth, Australia, 2–6 December 2018; pp. 562–576. [Google Scholar]
  48. Wang, Z.K.; Kao, Y.; Hsu, C.T. Vision-Based Heart Rate Estimation Via A Two-Stream CNN. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019. [Google Scholar]
  49. Napolean, Y.; Marwade, A.; Tomen, N.; Alkemade, P.; Eijsvogels, T.; van Gemert, J. Heart Rate Estimation in Intense Exercise Videos. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022. [Google Scholar]
  50. Lingvall, K.U. Remote Heart Rate Estimation by Evaluating Measurements from Multiple Signals. Master’s Thesis, School of Computer Science and Communication, KTH Royal Institute of Technology, Stockholm, Sweden, 2017. [Google Scholar]
Figure 1. Flow diagram of the proposed DMHRE method.
Figure 1. Flow diagram of the proposed DMHRE method.
Sensors 23 06079 g001
Figure 2. Proposed MASF algorithm, (a) PSD of original RGB spectra, (b) PSD of original NIR and R spectra, (c) PSD of MASF-filtered RGB spectra subject to Equation (2), (d) PSD of MASF-filtered NIR and R spectra subject to Equation (1).
Figure 2. Proposed MASF algorithm, (a) PSD of original RGB spectra, (b) PSD of original NIR and R spectra, (c) PSD of MASF-filtered RGB spectra subject to Equation (2), (d) PSD of MASF-filtered NIR and R spectra subject to Equation (1).
Sensors 23 06079 g002aSensors 23 06079 g002b
Figure 5. MAE performances for the PURE dataset at different compression bit rates (from 250 kbps to 1467 kbps), considering the HGT HR cases.
Figure 5. MAE performances for the PURE dataset at different compression bit rates (from 250 kbps to 1467 kbps), considering the HGT HR cases.
Sensors 23 06079 g005
Figure 6. MAE performances for the PURE dataset at different compression bit rates (from 250 kbps to 1467 kbps), ignoring the HGT HR cases.
Figure 6. MAE performances for the PURE dataset at different compression bit rates (from 250 kbps to 1467 kbps), ignoring the HGT HR cases.
Sensors 23 06079 g006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lie, W.-N.; Le, D.-Q.; Lai, C.-Y.; Fang, Y.-S. Heart Rate Estimation from Facial Image Sequences of a Dual-Modality RGB-NIR Camera. Sensors 2023, 23, 6079. https://doi.org/10.3390/s23136079

AMA Style

Lie W-N, Le D-Q, Lai C-Y, Fang Y-S. Heart Rate Estimation from Facial Image Sequences of a Dual-Modality RGB-NIR Camera. Sensors. 2023; 23(13):6079. https://doi.org/10.3390/s23136079

Chicago/Turabian Style

Lie, Wen-Nung, Dao-Quang Le, Chun-Yu Lai, and Yu-Shin Fang. 2023. "Heart Rate Estimation from Facial Image Sequences of a Dual-Modality RGB-NIR Camera" Sensors 23, no. 13: 6079. https://doi.org/10.3390/s23136079

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop