Literature DB >> 26822804

Measurement of heart rate variability using off-the-shelf smart phones.

Ren-You Huang1, Lan-Rong Dung2.   

Abstract

BACKGROUND: The cardiac parameters, such as heart rate (HR) and heart rate variability (HRV), are very important physiological data for daily healthcare. Recently, the camera-based photoplethysmography techniques have been proposed for HR measurement. These techniques allow us to estimate the HR contactlessly with low-cost camera. However, the previous works showed limit success for estimating HRV because the R-R intervals, the primary data for HRV calculation, are sensitive to noise and artifacts.
METHODS: This paper proposed a non-contact method to extract the blood volume pulse signal using a chrominance-based method followed by a proposed CWT-based denoising technique. The R-R intervals can then be obtained by finding the peaks in the denoised signal. In this paper, we taped 12 video clips using the frontal camera of a smart phone with different scenarios to make comparisons among our method and the other alternatives using the absolute errors between the estimated HRV metrics and the ones obtained by an ECG-accurate chest band.
RESULTS: As shown in experiments, our algorithm can greatly reduce absolute errors of HRV metrics comparing with the related works using RGB color signals. The mean of absolute errors of HRV metrics from our method is only 3.53 ms for the static-subject video clips.
CONCLUSIONS: The proposed camera-based method is able to produce reliable HRV metrics which are close to the ones measured by contact devices under different conditions. Thus, our method can be used for remote health monitoring in a convenient and comfortable way.

Entities:  

Mesh:

Year:  2016        PMID: 26822804      PMCID: PMC4731953          DOI: 10.1186/s12938-016-0127-8

Source DB:  PubMed          Journal:  Biomed Eng Online        ISSN: 1475-925X            Impact factor:   2.819


Background

Heart rate variability (HRV) provides useful physiological parameters according to the beat-to-beat intervals (R–R intervals, RRI) obtained from cardiac pulse signals. Some publications have validated that abnormality of HRV are related to some cardiological and noncardiological diseases, e.g., myocardial infarction, diabetic neuropathy, and myocardial dysfunction [1]. The conventional way to obtain the cardiac pulse is using electrocardiography (ECG) to sense the electrical activity of heart over a period of time by electrodes attached to the surface of human skin. ECG provides clean and accurate pulse signals, however, it is prone to be interfered by electrical activity produced by skeletal muscles near the electrodes. Another way to obtain the cardiac pulse is photoplethysmography (PPG), which was first described in 1930s [2]. PPG detects the optical absorption variations of the human skin due to the blood volume variations. Both ECG and PPG need to contact the human skin, which are not suitable for the cases of extreme sensitivity, e.g., neonates, skindamaged patients, or when the non-contact property is required (surveillance, fitness, etc.). Peng et al. [3] proposed an alternative method for extracting the PPG signal through the smart-phone camera followed by computing the HRV. However, this method still requires the subjects to put their finger on the smart-phone camera and keep themselves static, which has similar disadvantages as the traditional PPG device. Recent works have shown that cardiac pulse rate can be measured in a non-contact way, which is also known as remote-PPG (rPPG) [4-12]. These works obtain pulse signals under ambient light conditions with only one camera, which are low-cost, simple, and effective. The main idea of rPPG is that the blood volume variations can be captured during video recording. The earlier works [4, 5] first obtain the mean intensity of skin region and perform frequency analysis (Fourier or wavelet transform) to estimate the pulse rate. Recent works [6-12] estimate the pulse rate using a regular color video camera. The first step of these methods are locating the region of interest by manual selection or automatic face detection, followed by different analysis algorithms to extract the pulse signals, e.g., difference of RGB [6], source separation [7-9], chrominance [10, 12], motion magnification [11]. Poh et al. [7] proposed an algorithm for heart rate (HR) measurement. They first detected the face every frame and extracted the mean RGB color values to form a three-dimensional time series. Then they applied independent component analysis (ICA) [13, 14] to separate the independent sources from these RGB signals which may contain the pulse signal, followed by FFT and select the frequency with maximum amplitude in the spectral of the component which has highest peak as the HR. Later on, the authors proposed a similar method in [8] to extract the R–R intervals by finding the peaks of the pulse signal. The peaks of pulse signal are treated as the R wave of ECG signal, and the peak intervals are treated as R–R intervals. Alternatively, one may apply PCA, as shown in [9], instead of ICA to separate the pulse signal from RGB time series. Wu et al. [11] proposed an Eulerian-based motion magnification to magnify the subtle motions or color changes in temporal domain using Laplacian pyramid. This method is able to obtain a clean pulse signal if the subjects are almost static. Haan and Jeanne [10] proposed a chrominance-based remote PPG (we denote “C-rPPG” in the rest of this paper) which takes different factors into account to form the color model captured by camera. Given the pulsatility as a function of wavelength exhibits a strong peak in green and the dips in red [15, 16]. To exploit this fact and to reduce the specular reflection problem mentioned in [17], they proposed a model using difference of wighted color channels to obtain chrominance signals. This method is robust to different skin-tone and adaptive to non-white illumination. Moreover, the authors showed the impressive results of HR estimation for the scenario with the subjects exercising on stationary bike. Wang et al. [12] proposed an algorithm exploiting the spatial redundancy of image sensor and the idea of chrominance to improve the robustness to motions. In the indoor scenes, the lighting sources are usually on the top of subjects, i.e., on ceiling. The color intensities or brightness of the skin captured by camera are various at different positions. Different angles between the lighting sources and camera also result in intensity variations. These periodical or non-periodical variations will produce artifacts which severely influence most of the rPPG algorithms. The methods based on source separation [7-9] may separate the artifacts rather than true pulse signal. Nevertheless, the component with the highest spectral amplitude is not necessary to be the true pulse signal. The Eulerian motion magnification [11] requires the subject as stationary as possible; otherwise, the motion in the specific frequency band will be magnified accordingly. Hence the motion magnification is not appropriate for general scenarios. The C-rPPG [10] improved the robustness to motions and has much better performance in HR estimation with non-static subjects. However, we found that there exist noises and artifacts in the C-rPPG signal which produce false peaks and severely influence the accuracy of R–R intervals. The paper [12] further improves the motion-robustness of the C-rPPG algorithm by adaptively combining local PPG-signals, and improves the SNR using an adaptive band-pass filter. This more elaborated C-rPPG concept, however, leads to an increase in computational complexity, which we consider less attractive for a mobile platform. This paper proposed a non-contact method to estimate accurate HRV metrics from 30 fps video clips captured by frontal camera of off-the-shelf smart phones. The face of subject is located every frame followed by averaging all the skin pixels to form the RGB time series. The RGB time series are then used to compute the pulse signal by C-rPPG [10] algorithm. We proposed a denoising method based on continuous wavelet transform (CWT) to increase the robustness to interferences. The R–R intervals are obtained by computing the intervals of successive peaks in the denoised signal. To demonstrate the performance of HRV measurement, we taped a 12 video clips for different scenarios with static and non-static subjects and used an ECG-accurate chest band to obtain the R–R intervals as the ground truth. Comparing with existed approaches, the absolute errors of HRV metrics generated by proposed approach is relatively low. For the video clips with static subjects, the mean of absolute errors of the HRV metrics obtained by our method is only 3.53 ms.

Methods

Overview

The processing flow of the proposed algorithm Figure 1 shows the overall flow chart of the proposed algorithm. For each frame of the video, we first detect the face based on the nose positions to increase the robustness to non-frontal faces. Once the face is located, we perform skin detection in YCbCr color space followed by averaged the RGB channels of the skin pixels in the face region to form the time series. To increase the fineness of time grid, we upsample the time series by a factor of eight, i.e., the sampling rate is from 30 to 240 Hz. After data acquisition stage, we compute the C-rPPG signal to extract the pulse signal. Next, we perform our denoising technique based on CWT. Finally, the peaks in the denoised signal are detected to compute the R–R intervals.
Fig. 1

The processing flow of the proposed algorithm

Data acquisition

First of all, we locate the face in every frame to extract the color signals. There are plenty of face detection works and surveys [18-21]. For simplicity and convenience, one may apply the face detector proposed by Viola and Jones [18] which is effective and efficient to locate the faces in frames. However, the face detector will fail if the faces in video are non-frontal. We found that detecting the nose is more stable than detecting the face, thus we can exploit the nose position to derive appropriate face region. We use the object detection toolbox (vision.CascadeObjectDetector) built in MATLAB to detect the nose in every frame. The region of the face can be determined bywhere w and h are the width and height, (x, y) is the top left coordinate of the bounding box. The subscripts n and f represent the “nose” and “face”. Next, we use a simple skin color detection to ensure the processed data are obtained from skin pixels. There are lots of works one may refer for the skin detection, e.g., the method proposed in [22]. We only take into account the Cb and Cr components to detect the skin color. A pixel is classified as skin pixel if it satisfies the following conditions:After skin detection, we then record the averaged RGB values of skin pixels in the ROI to form the time series. Finally, we upscale the time series by a factor of 8.

Computing C-rPPG

Inspired by previous works, we apply the chrominance-based method in our algorithm due to the better performance for extracting the real pulse signals instead of false ones. We apply the model proposed in [10] which is briefly reviewed in the following. In [10], the intensity of a given pixel in i-th frame in color channel registered by the camera is modeled aswhere is the intensity of the light source integrated over the exposure time of the camera, is the stationary part of the reflection coefficient of the skin, is the zero-mean time-varying fraction caused by the pulsation of the blood volume, and is the additive specular reflection contribution. The RGB data are normalized using the following formulawhere is a moving average centered around frame index i. The chrominance signals are defined as followsFinally, the pulse signals can be extracted bywithwhere is standard deviation of the signals, the signals with subscript f represent their band-pass filtered versions. We can further rewrite (6) as follows

CWT-based denoising method

The CWT transforms a time series to a time-frequency representation and has been used to denoise the PPG signals in some works [23-25]. The CWT uses inner product to measure the similarity between a signal and a specific analysis function, which outperforms the Fourier Transform and the short term Fourier Transform since the CWT can detect rapid changes in frequency due to the multi-scale representation. We will briefly review the theory of CWT and describe our denoising method based on CWT in the following. The CWT convolves a signal x(t) with child wavelets which represent scaled and translated versions of mother wavelet , represents similarity between the signal x(t) and a child wavelet scaled by s and translated by , which is define as follows:There are many standard mother wavelets available in CWT literature. We selected the Morlet wavelet in our algorithm since it has been used to analysis PPG signals in [25]. The signal can be reconstructed from the wavelet transform by the inverse formula of (11).where is the admissible constant of wavelet transform. Let denoted as Fourier version of , the admissible constant is defined as follows:One may reserve the coefficients of specific scales corresponding to the observed frequency band (0.75, 4) Hz [(45, 240) bpm] and set the others to zero followed by inverse transform, which is equivalent to bandpass filtering (we call it “CWT-BP”). However, the motion artifacts are usually in the same frequency band, hence there will be false peaks produced by motion artifacts in the reconstructed signal. Assume the pulse signal is the most significant component of the C-rPPG signal, our goal is to select a representative scale to reconstruct the pulse signal. We computed the summation of the magnitude of CWT coefficients in the same scale within a time interval, followed by selecting the scale with maximal value of the summation, i.e.,where is the optimal scale to reconstruct the pulse signal. The CWT coefficients belonging to the scale are reserved and the others are set to zero. In practice, we should take into account computation efficiency, thus we divide the CWT coefficients into non-overlapping time intervals with length T (seconds) and select the representative scales for every time interval. Another factor we should take into account is the non-stationary property of cardiac activity, hence the value of T should be carefully selected. Choosing smaller T is able to catch up to variation of cardiac activity but is less robust to the strong interference such as motion artifacts, and vice versa. Here we suggest that one can set T in the range of 10–30 (seconds). After selecting the optimal scales for every time interval, the pulse signal is reconstructed by inverse CWT. We denoted this method as “CWT-MAX” in the following. Examples for using CWT to detrend and denoise. a The original C-rPPG signal. b The CWT coefficients of the original signal. Note that the black solid line denotes the representative frequency (scales) of pulse signal computed by (13). c The zoomed-in part of original signal. d The zoomed-in part of signal denoised by CWT-BP. e The zoomed-in part of signal denoised by CWT-MAX Figure 2 shows an example to demonstrate our approach. We can observe that the original signal shown in Fig. 2c is noisy and with many false peaks. After applying CWT to the original signal, we can obtain the CWT coefficients as shown in Fig. 2b. The black line represents the coefficients of optimal scales of every time interval. The CWT-BP reserved all the coefficients in the observed band [(0.75, 4) Hz] and set the others to zero which may smooth and denoise the original signal as shown in Fig. 2d; however, it still retained some false peaks which may degrade the accuracy of R–R intervals. On the contrary, the CWT-MAX only reserved the coefficients of representative scales of each time interval and set the others to zero. The signal reconstructed by CWT-MAX is much cleaner, as shown in Fig. 2e. Therefore, we applied the CWT-MAX in our algorithm.
Fig. 2

Examples for using CWT to detrend and denoise. a The original C-rPPG signal. b The CWT coefficients of the original signal. Note that the black solid line denotes the representative frequency (scales) of pulse signal computed by (13). c The zoomed-in part of original signal. d The zoomed-in part of signal denoised by CWT-BP. e The zoomed-in part of signal denoised by CWT-MAX

Peak detection and R–R intervals

R–R intervals of the example in Fig. 2. The red dot line is the R–R intervals measured by an ECG-accurate chest band. The blue dot line is the R–R intervals computed by our method After CWT denoising, the proposed approach then detect the peaks in the denoised pulse signal to compute the R–R intervals. One may simply use the findpeaks function built in MATLAB, or use the customized peak-finding algorithm. Since the CWT-denoised signal is almost noise-free, the selection of peak-finding algorithms does not play a crucial role to our results. After peak detection, let be the time instance of k-th peak in the signal, the R–R intervals can be calculated byand generally in the unit of millisecond (ms). Figure 3 shows the R–R intervals of the example in Fig. 2 computed by our method and the R–R intervals measured by an ECG-accurate chest band.
Fig. 3

R–R intervals of the example in Fig. 2. The red dot line is the R–R intervals measured by an ECG-accurate chest band. The blue dot line is the R–R intervals computed by our method

Experimental setup

Descriptions of the video clips in each category We totally taped 12 video clips with one minute long to evaluate the performance of R–R intervals extraction. The clips are taped by frontal camera of a smart phone (Sony Xperia Z1) with 30 fps frame rate and size of 640 × 480. We simulated the scenario that the video recorded the subjects when they were using their smart phone or tablet to monitoring their cardiac physiology. These video clips are classified into four categories, which are “static subjects”, “static subject with makeup”, “occasional motion”, and “frequent motion”, respectively. Each of the category has 2–4 clips with different subjects or slightly different conditions. Note that the word “static” here means the subjects kept their bodies static but slight movements (e.g., talking, facial expression, slight shaking) are allowable. The detail descriptions of the video clips are listed in the Table 1.
Table 1

Descriptions of the video clips in each category

CategoriesClip namesDescriptions
Static subjectsStatic_1The subject kept the body relaxed and static
Static_2There was desk light illuminated on the face
Static_3The subject kept smile during video recording
Static_4The subject kept making facial expression
Static subject with makeupstatic_M1The subject put the CC cream on her face
Static_M2The subject in static_M1 put additional powder foundation on her face
Static_M3The subject in static_M2 put additional blush powder on her face
Occasional motionMotion_O1The subject shook the head three times
Motion_O2The subject moved away from camera and then moved back
Motion_O3The subject turned the head, talking, then turned back, twice
Frequent motionMotion_F1The subject kept shaking his head
Motion_F2The subject rotated his head, move the body, or moved the camera several times
Descriptions of all the subjects We have totally six subjects in the 22–28 age range involved in the experiments. There are four subjects (two males and two females) in “static subjects” category, one female subject in “static subject with makeup” category, and one male subject in both “occasional motion” and “frequent motion” categories, respectively. For more details of the subjects, please see the Table 2. The makeups we used in the experiments are, CC cream (CLINIQUE Molsture Surge CC cream hydrating colour corrector broad spectrum SPF30), powder foundation (DiorSnow Sublissime SPF30 PA+++), and blush powder (Christian Dior Diorshow Powder Backstage Makeup Color in a flash loose powder 0.17oz/5g 003 Catwalk Pink), respectively. This study had received approval by China Medical University and Hospital Research Ethics Committee. All the subjects have signed an informed consent allowing the authors to publish their HRV data.
Table 2

Descriptions of all the subjects

SubjectsGendersAgesThe corresponding video clips
Subject_1Male22Static_1
Subject_2Female24Static_2
Subject_3Male26Static_3
Subject_4Male25Static_4
Subject_5Female23Static_M1, static_M2, and static_M3
Subject_6Male28Motion_O1, motion_O2, motion_O3, motion_F1, and motion_F2
We also used an ECG-accurate chest band (R1 Blue Comfortex+, made by Sigma sport) during video recording to obtain the ground truth of cardiac activity and exported the R–R intervals for the following comparisons. Because we aimed at the implementations suitable for smart phone applications, thus we only compared the proposed method with the algorithms which have similar computational cost. We will compare the performance of our algorithm with the ICA-based method [8] and original C-rPPG [10]. Since the authors did not release the source codes, we have tried our best to implement the algorithms described in their papers. We implemented the same bandpass filter in [8] for the original C-rPPG signal. For fair comparisons, we applied the same peak detection function (findpeaks) built in MATLAB to all the methods in the following experiments. Note that we do not further process the R–R intervals no matter they are reasonable or not. All the algorithms are implemented in MATLAB code.

Results and discussions

Quantitative evaluation

To show the accuracy of HRV estimation, this paper make comparisons with existed works by using well-known HRV metrics [26]. The scatter plot of R–R intervals is usually a good tool to show the relationship between and and thereby evaluate HRV. The scatter plot is a 2-D plot, in which the calculated eigenvalues are useful in the following comparisons. The square root of an eigenvalue describes the standard deviation along the direction of corresponding eigenvector. In this paper we denote SD1 as the square root of the smallest eigenvalue and SD2 as the other one in our HRV comparisons. The time-domain HRV metrics used here are: the standard deviation of R–R intervals (SDNN), root mean square of successive differences (RMSSD), standard deviation of successive differences (SDSD). All the HRV metrics mentioned above are in the unit of millisecond (ms).

Results and discussions of each category

Static subjects

The HRV metrics estimated by different methods in the “static subjects” category The HRV metrics of the “static subjects” clips estimated by the chest band and the different methods are listed in Table 3. Generally, the pulse signal extracted by C-rPPG [10] has better performance than the one extracted by ICA [8]. Our method inherited from C-rPPG and the HRV metrics are very close to the ones measured by chest band (see the absolute errors). Ideally, the clips with static subjects have no motion artifacts. However, as mentioned above, the HRV metrics are computed by R–R intervals which are very sensitive to the false peaks in the noisy signals. The proposed CWT-based denoising method removes the most interferences; hence, the R–R intervals are reliable and close to the ground truth even the subjects keep making facial expressions.
Table 3

The HRV metrics estimated by different methods in the “static subjects” category

Clip namesHRV metrics (ms)Chest bandICA [8]C-rPPG [10]Our method
Est.Est.Abs. errorEst.Abs. errorEst.Abs. error
Static_1SD122.0268.8946.8761.7739.7524.892.87
SD265.1583.1117.9694.7729.6264.770.38
SDNN49.2877.6928.4183.8734.5948.790.49
RMSSD30.8196.3165.5087.3056.4934.633.82
SDSD31.2197.8866.6788.5957.3835.203.99
Static_2SD124.5878.5954.0138.7314.1518.855.73
SD272.8090.6617.8679.036.2369.403.40
SDNN54.0385.5631.5362.358.3250.753.28
RMSSD34.29109.8075.5154.0419.7526.327.97
SDSD34.76111.3476.5854.7720.0126.668.10
Static_3SD122.3789.1366.7662.3139.9418.793.58
SD262.23101.1638.9394.8132.5859.922.31
SDNN47.0995.7048.6184.6837.5944.512.58
RMSSD31.25140.83109.5887.5956.3426.165.09
SDSD31.67143.04111.3788.9257.2526.545.13
Static_4SD118.6299.7981.1778.7860.1618.930.31
SD238.41107.3168.9091.6453.2341.152.74
SDNN30.21104.9174.7086.2956.0831.911.70
RMSSD26.01139.95113.94127.83101.8226.430.42
SDSD26.34141.69115.35129.47103.1326.770.43

Static subjects with makeup

The HRV metrics estimated by different methods in the “static subjects with makeup” category In this category, we made experiments on the cases which the subjects had different kind of makeup on her face. Table 4 shows the results of HRV estimated by the different methods. The ICA-based method and C-rPPG deviated from the ground truth while our method still got much lower errors. In these experiments, we can observe that the makeup may interfere the performance of the pulse signal extraction. However, this interference will not degrade the results of our method because our technique can successfully remove noises and artifacts.
Table 4

The HRV metrics estimated by different methods in the “static subjects with makeup” category

Clip namesHRV metrics (ms)Chest bandICA [8]C-rPPG [10]Our method
Est.Est.Abs. errorEst.Abs. errorEst.Abs. error
Static_M1SD114.95101.1686.2197.8982.9419.594.64
SD248.17109.1660.99127.4879.3152.624.45
SDNN35.66106.3970.73114.0378.3739.623.96
RMSSD20.85141.85121.00136.41115.5627.286.43
SDSD21.14143.76122.62138.52117.3827.716.57
Static_M2SD113.9749.9335.9654.7140.7413.940.03
SD265.3474.359.0176.7211.3868.563.22
SDNN47.1464.4217.2867.8120.6750.213.07
RMSSD19.5069.7750.2776.4556.9519.590.09
SDSD19.7570.7751.0277.5257.7719.720.03
Static_M3SD120.9671.5650.679.6258.6623.952.99
SD286.84127.9041.06135.2648.4292.185.34
SDNN63.79104.0740.28114.1750.3867.583.79
RMSSD29.3499.6670.32111.3081.9633.434.09
SDSD29.73101.2071.47112.8883.1533.924.19

Occasional motion

The illumination changes of the face in the “motion_O3” clip The HRV metrics estimated by different methods in the “occasional motion” category The clips in this category are that the subjects moved his/her body or head less than three times, just like the regular motions we make in daily-life. Table 5 shows the results of this category. Our method only severely deviated from the ground truth in the “motion_O3”. To explain the result, we computed the averaged illumination (grayscale) on the face of the “motion_O3”, as shown in Fig. 4. We found that the illumination changes significantly due to the auto-exposure function of camera. The camera changed the exposure automatically when the subject turned the head, while the other two clips (“motion_O1” and “motion_O2”) have no such illumination changes. Therefore, our method still obtained the HRV metrics close to the ground truth in “motion_O1” and “motion_O2”.
Table 5

The HRV metrics estimated by different methods in the “occasional motion” category

Clip namesHRV metrics (ms)Chest bandICA [8]C-rPPG [10]Our method
Est.Est.Abs. errorEst.Abs. errorEst.Abs. error
Motion_O1SD127.7750.3922.6262.0634.2920.447.33
SD259.8268.478.6589.4329.6156.383.44
SDNN46.9460.0313.0978.8231.8842.084.86
RMSSD38.7670.1931.4386.9148.1528.4710.29
SDSD39.2971.2631.9788.1648.8728.9110.38
Motion_O2SD120.8487.9567.1165.9145.0729.208.36
SD272.9391.7218.7987.7814.8572.570.36
SDNN53.2989.6036.3178.2824.9955.372.08
RMSSD29.07122.6593.5891.9962.9240.7411.67
SDSD29.47124.3994.9293.2663.7941.3111.84
Motion_O3SD124.60123.3298.72103.6379.0360.6336.03
SD262.16206.86144.70152.9090.7490.5428.38
SDNN47.63170.02122.39131.7384.1076.5028.87
RMSSD34.35171.91137.56144.61110.2684.3950.04
SDSD34.83174.41139.58146.71111.8885.7450.91
Fig. 4

The illumination changes of the face in the “motion_O3” clip

Frequent motion

The face positions and illumination in “motion_F1” clip. a The face position (x-axis). b The face position (y-axis). c The illumination (grayscale) of the face The HRV metrics estimated by different methods in the “frequent motion” category Table 6 shows the HRV metrics of the “frequent motion” video clips. This category is extremely challenging since the subjects kept making movements during the video recording. Both the ICA [8] and C-rPPG [10] severely deviated from the ground truth. Although our method was interfered by the large motion artifacts, we still obtained a reasonable HRV metrics close to ground truth. Figure 5 shows the face positions and the illumination changes on the face of “motion_F1”. The face positions changes periodically due to the continuously shaking of the head. We can observe that the illumination changed with the motion of face rather than exposure changes. The results have shown that our method can deal with the motion artifacts even the subject kept shaking his head during the video recording if the exposure of camera is almost fixed.
Table 6

The HRV metrics estimated by different methods in the “frequent motion” category

Clip namesHRV metrics (ms)Chest bandICA [8]C-rPPG [10]Our method
Est.Est.Abs. errorEst.Abs. errorEst.Abs. error
Motion_F1SD119.78177.67157.8989.0569.2727.057.27
SD248.73271.01222.28133.7084.9758.099.36
SDNN37.76227.73189.97112.8775.1145.047.28
RMSSD27.72247.65219.93124.2296.5037.7210.00
SDSD28.03251.26223.23125.9497.9138.2610.23
Motion_F2SD130.89149.01118.12130.1199.2241.6010.71
SD295.04255.18160.14164.6969.6599.604.56
SDNN71.15207.28136.13147.2876.1375.974.82
RMSSD43.12207.25164.13180.92137.8057.8514.73
SDSD43.73210.73167.00184.00140.2758.8415.11
Fig. 5

The face positions and illumination in “motion_F1” clip. a The face position (x-axis). b The face position (y-axis). c The illumination (grayscale) of the face

Conclusion

In this paper, we have analyzed the problems of camera-based PPG and proposed an algorithm to extract accurate R–R intervals using 30 fps camera. We first extract the pulse signal using the chrominance-based method (C-rPPG) followed by a denoising method based on the CWT. The R–R intervals are computed by finding the peaks in the denoised signals. The experimental video clips were recorded by a frontal camera of smart phone (Sony Xperia Z1) held by the subjects in different situations. The experiments have shown that our method is able to extract much more accurate results than the related works. The mean of absolute errors of HRV metrics obtained by our method is only 3.53 ms in the “Static subjects” and “Static subjects with makeup” categories. This shows the potential of our method for remote health monitoring of patients, which can be done by an easy and comfortable way in daily-life. Note that the measurements of HRV for clinical use should conform to professional recommendations (e.g., [1, 26]), and our method might not meet those requirements. However, it can be useful for informal applications; for instance, monitoring the physiological status of the tablet users and giving warnings to the users who may have some potential healthy problems. Although the proposed method is able to alleviate the interference of motion artifacts, we still have room for improvement to deal with the artifacts made by the significantly changes of exposure due to auto-exposure function of camera. In addition, for a proof of concept, this paper validates our work with six subjects which might not be enough to show convincing, statistically significant evidence of efficacy. In the future, we will aim to refine our algorithm for being robust to artifacts generated by built-in functions in smart phones, and conduct the experiments with larger number of subjects.
  10 in total

Review 1.  Insights from the study of heart rate variability.

Authors:  P K Stein; R E Kleiger
Journal:  Annu Rev Med       Date:  1999       Impact factor: 13.739

2.  Advancements in noncontact, multiparameter physiological measurements using a webcam.

Authors:  Ming-Zher Poh; Daniel J McDuff; Rosalind W Picard
Journal:  IEEE Trans Biomed Eng       Date:  2010-10-14       Impact factor: 4.538

3.  Heart rate measurement based on a time-lapse image.

Authors:  Chihiro Takano; Yuji Ohta
Journal:  Med Eng Phys       Date:  2006-10-30       Impact factor: 2.242

4.  Robust pulse rate from chrominance-based rPPG.

Authors:  Gerard de Haan; Vincent Jeanne
Journal:  IEEE Trans Biomed Eng       Date:  2013-06-04       Impact factor: 4.538

5.  High-order contrasts for independent component analysis.

Authors:  J F Cardoso
Journal:  Neural Comput       Date:  1999-01-01       Impact factor: 2.026

6.  Heart rate variability. Standards of measurement, physiological interpretation, and clinical use. Task Force of the European Society of Cardiology and the North American Society of Pacing and Electrophysiology.

Authors: 
Journal:  Eur Heart J       Date:  1996-03       Impact factor: 29.983

7.  Exploiting spatial redundancy of image sensor for motion robust rPPG.

Authors:  Wenjin Wang; Sander Stuijk; Gerard de Haan
Journal:  IEEE Trans Biomed Eng       Date:  2014-09-08       Impact factor: 4.538

8.  Non-contact, automated cardiac pulse measurements using video imaging and blind source separation.

Authors:  Ming-Zher Poh; Daniel J McDuff; Rosalind W Picard
Journal:  Opt Express       Date:  2010-05-10       Impact factor: 3.894

9.  Remote plethysmographic imaging using ambient light.

Authors:  Wim Verkruysse; Lars O Svaasand; J Stuart Nelson
Journal:  Opt Express       Date:  2008-12-22       Impact factor: 3.894

10.  Extraction of heart rate variability from smartphone photoplethysmograms.

Authors:  Rong-Chao Peng; Xiao-Lin Zhou; Wan-Hua Lin; Yuan-Ting Zhang
Journal:  Comput Math Methods Med       Date:  2015-01-12       Impact factor: 2.238

  10 in total
  11 in total

1.  Two-dimensional XD-GRASP provides better image quality than conventional 2D cardiac cine MRI for patients who cannot suspend respiration.

Authors:  Eve Piekarski; Teodora Chitiboi; Rebecca Ramb; Larry A Latson; Puneet Bhatla; Li Feng; Leon Axel
Journal:  MAGMA       Date:  2017-10-24       Impact factor: 2.310

2.  Passive Sensing of Health Outcomes Through Smartphones: Systematic Review of Current Solutions and Possible Limitations.

Authors:  Alina Trifan; Maryse Oliveira; José Luís Oliveira
Journal:  JMIR Mhealth Uhealth       Date:  2019-08-23       Impact factor: 4.773

Review 3.  The Current State of Mobile Phone Apps for Monitoring Heart Rate, Heart Rate Variability, and Atrial Fibrillation: Narrative Review.

Authors:  Ka Hou Christien Li; Francesca Anne White; Gary Tse; Timothy Tipoe; Tong Liu; Martin Cs Wong; Aaron Jesuthasan; Adrian Baranchuk; Bryan P Yan
Journal:  JMIR Mhealth Uhealth       Date:  2019-02-15       Impact factor: 4.773

Review 4.  A Broader Look: Camera-Based Vital Sign Estimation across the Spectrum.

Authors:  Christoph Hoog Antink; Simon Lyra; Michael Paul; Xinchi Yu; Steffen Leonhardt
Journal:  Yearb Med Inform       Date:  2019-08-16

5.  Feasibility of assessing ultra-short-term pulse rate variability from video recordings.

Authors:  Miha Finžgar; Primož Podržaj
Journal:  PeerJ       Date:  2020-01-07       Impact factor: 2.984

Review 6.  The Role of Heart Rate Variability in the Future of Remote Digital Biomarkers.

Authors:  Andrew P Owens
Journal:  Front Neurosci       Date:  2020-11-13       Impact factor: 4.677

Review 7.  Effectiveness of consumer-grade contactless vital signs monitors: a systematic review and meta-analysis.

Authors:  Chi Pham; Khashayar Poorzargar; Mahesh Nagappa; Aparna Saripella; Matteo Parotto; Marina Englesakis; Kang Lee; Frances Chung
Journal:  J Clin Monit Comput       Date:  2021-07-09       Impact factor: 1.977

8.  Video pulse rate variability analysis in stationary and motion conditions.

Authors:  Angel Melchor Rodríguez; J Ramos-Castro
Journal:  Biomed Eng Online       Date:  2018-01-29       Impact factor: 2.819

9.  Enhancing the Robustness of Smartphone Photoplethysmography: A Signal Quality Index Approach.

Authors:  Ivan Liu; Shiguang Ni; Kaiping Peng
Journal:  Sensors (Basel)       Date:  2020-03-30       Impact factor: 3.576

Review 10.  Smartphone Apps Using Photoplethysmography for Heart Rate Monitoring: Meta-Analysis.

Authors:  Benjamin De Ridder; Bart Van Rompaey; Jarl K Kampen; Steven Haine; Tinne Dilles
Journal:  JMIR Cardio       Date:  2018-02-27
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.