Literature DB >> 33020401

Generalized Deep Neural Network Model for Cuffless Blood Pressure Estimation with Photoplethysmogram Signal Only.

Yan-Cheng Hsu1, Yung-Hui Li2, Ching-Chun Chang3, Latifa Nabila Harfiya2.   

Abstract

Due to the growing public awareness of cardiovascular disease (CVD), blood pressure (BP) estimation models have been developed based on physiological parameters extracted from both electrocardiograms (ECGs) and photoplethysmograms (PPGs). Still, in order to enhance the usability as well as reduce the sensor cost, researchers endeavor to establish a generalized BP estimation model using only PPG signals. In this paper, we propose a deep neural network model capable of extracting 32 features exclusively from PPG signals for BP estimation. The effectiveness and accuracy of our proposed model was evaluated by the root mean square error (RMSE), mean absolute error (MAE), the Association for the Advancement of Medical Instrumentation (AAMI) standard and the British Hypertension Society (BHS) standard. Experimental results showed that the RMSEs in systolic blood pressure (SBP) and diastolic blood pressure (DBP) are 4.643 mmHg and 3.307 mmHg, respectively, across 9000 subjects, with 80.63% of absolute errors among estimated SBP records lower than 5 mmHg and 90.19% of absolute errors among estimated DBP records lower than 5 mmHg. We demonstrated that our proposed model has remarkably high accuracy on the largest BP database found in the literature, which shows its effectiveness compared to some prior works.

Entities:  

Keywords:  artificial neural network; cardiovascular disease (CVD) prevention; photoplethysmogram (PPG), cuffless blood pressure (BP) estimation; wearable biomedical applications

Mesh:

Year:  2020        PMID: 33020401      PMCID: PMC7582614          DOI: 10.3390/s20195668

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


1. Introduction

According to a statistical report from the World Health Organization (WHO), cardiovascular disease (CVD) is the leading cause of death worldwide, with an estimated 17.9 million people dying from CVD in 2016, representing 31% of global deaths [1]. Early detection and treatment could effectively reduce the incidence and mortality rates. As a result, there is an urgent need for efficient and reliable means of managing cardiovascular risk factors, such as diabetes, hypertension or hyperlipidemia. Blood pressure (BP) is considered to be one of the most important contributory risk factors and, therefore, real-time monitoring of BP plays a crucial role in saving people from premature death caused by CVD. The most common automated BP measurement devices are cuff based, as shown in Figure 1. They take about one to two minutes to produce one set of diastolic blood pressure (DBP) and systolic blood pressure (SBP) measurements before making another measurement. This type of measurement can be time-consuming and is often inaccurate [2]. In view of these issues, some neural network-based regression models were developed and they shortened the time interval of BP measurement from 1–2 min to less than 10 s. The accuracy of these regression models also met the criteria (i.e., a protocol of requirements for the evaluation of BP measuring devices) defined and recommended by the British Hypertension Society (BHS) [3,4]. While these models produced satisfactory performance in terms of operation time and accuracy for real-time BP estimation, they may not be practical solutions. The key reason is that these models are required to estimate physiological parameters from electrocardiogram (ECG) and photoplethysmogram (PPG) signals, implying that physiological parameter extraction from two different sensors is needed, and this solution incurs substantial cost.
Figure 1

Traditional blood pressure measurement device.

The other reason is from the basic theory of the classical method of extracting physiological parameters from ECGs and PPGs, which relies heavily upon the theory based on pulse wave velocity (PWV). PWV is the velocity of pressure pulse initiated by the heartbeat, propagating through arteries, similar to a pipe with elastic walls. PWV has been proved to be highly related to BP and their correlation can be represented as [5,6]: where , , and denote the radius of the artery, the thickness of the artery, the elastic modulus of the arterial wall and the density of blood in of the artery, respectively. There are several existing approaches that can calculate PWV and, among them, the most widely used one for PWV calculation is pulse wave transit time, commonly referred to as pulse transit time (PTT). The relation between PWV and PTT can be represented as follows [7]: where PTT is the time interval between a pulse wave being detected by two sensors and is the distance between the sensors on the artery. In (1), the elastic modulus is assumed as a constant when in fact the value of in the artery is testified to be exponentially escalated with the blood pressure, as follows [8]: where denotes the elastic modulus at 0 mmHg (the unit of blood pressure) and is a parameter larger than zero that is closely related to arterial stiffness. The stiffer the artery, the greater the value of . We can find a nonlinear relationship between blood pressure and PTT after we substitute (1) and (2) into (3) to be Though it may seem easy from the theoretical perspective, it would be inconvenient and almost impossible to use (4) directly since there is no way to get all the person-dependent variables in a short period of time. The other approach is done by extracting a set of representative time indices, including PTT (p), PTT (d) and PTT (f), as shown in Figure 2, from the relative location between the PPG and ECG signals [9,10]. However, it is still a very challenging task since the ECG waveform, in particular, has higher variability [7] and its accuracy is still limited for clinical uses [11].
Figure 2

The expedient solution from engineers, including pulse transit time (PTT) (p), PTT (d) and PTT (f).

Notwithstanding that devices such as, but not limited to, wearable devices that have the capability of recording both ECG and PPG signals are now thriving in the market, these are still highly priced and prevent users from having direct access to the data [12]. On the other hand, PPG sensors have been largely applied in wearable devices. They are popular as a low-cost but robust technology with full accessibility. Prior works have emphasized the relevance of a more detailed study of PPG signals only [13,14,15,16,17,18]. Hence, a ECG-free BP estimation model is preferable to improve the usability and to reduce the cost since the devices will no longer need additional biosensors for detecting ECG signals.

2. Literature Review

PPG is a non-invasive technique for measuring changes in blood volume due to the blood pulsatile nature of microvascular tissue under the skin [19]. The characteristics of the PPG waveform, along with its derivatives, have been discussed in [16,20,21]. Here, we can conclude that taking its first and second derivatives significantly helps in detecting the informative features in the PPG waveform. From a biomedical application perspective, [13] shows that PPG is an effective technique for diagnosing several CVDs and is able to be utilized in new medical tools such as the Internet of Things and biosensors. The clinical applicability of PPG is also verified in [18], which tried to distinguish individuals with congestive heart failure from healthy individuals by applying the concept of natural time analysis (NTA). NTA is applied to analyze a phase change or critical point in a complex system such as the human heart. The results obtained by PPG demonstrate a comparable value of accuracy to the results obtained by ECG. For a more specific BP estimation task, we can basically divide them into two approaches, feature-based and whole-based methods. In [14], five different features, which consist of the pulse area, pulse rising time, pulse width at 25% of pulse height, pulse width at 50% of pulse height and pulse width at 75% of pulse height, were extracted from a PPG segment. Machine learning methods, such as multiple linear regression (MLR), support vector machine (SVM) and regression tree, were then utilized for training and testing the data for estimating both DBP and SBP values, with the best overall accuracy being achieved using the regression tree. Another work, [22], uses several spectral and morphological features, such as systolic upstroke time and diastolic time. Using artificial neural network (ANN) architecture for fitting the features to simultaneously estimate the DBP and SBP, this method reduces the error from the other methods used as comparisons, such as linear regression and regression support vector machine (RSVM). On the contrary, the whole PPG waveform segment was extracted and used as the input of deep learning models in [15,17]. Both models comprised a convolutional neural network (CNN) and its modification to capture the spatial features of the waveforms. Both models achieved impressive accuracy, with a relatively low distribution of error as well.

3. Materials and Methods

The main flow of the proposed model is illustrated in Figure 3, which was composed of preprocessing, a feature extractor and a deep neural network predictor. Since the raw PPG might contain noise and long sequences, preprocessing is necessary for further feature extraction processes. In this section, a detailed explanation about each part is presented, with a summary introduced as follows:
Figure 3

System overview of the proposed model and zoomed in view of fully connected neural network.

This part comprises signal smoothing of raw PPG data and the removal of abnormal data following standard procedures suggested by [23]. Next, we partition the preprocessed PPG into an approximately 2.17 million heart cycles. Features from the preprocessed data are further extracted and selected as the input set. We feed the feature set into a deep neural network predictor, which consists of five fully connected layers, and each layer contains 2000+ units of fully connected perceptrons, responsible for predicting BP in each heart cycle from 32 extracted physiological parameters.

3.1. Data Source

For every data-driven neural network application, the data themselves mean everything, affecting regression models from every perspective. Specifically, the diversity of data affects how generalized a model is and, the more variability the data has, the more generalized the model that can be trained. Based on this idea, the Multi-parameter Intelligent Monitoring in Intensive Care (MIMIC) II online waveform database, which has been refined and prescreened in the literature [9], and contains 12,000 data instances indicating unique subject records and an estimated more than 4 million heart cycles, was used in this study. Among this enormous dataset, we preselected 9000 data instances for training, validating and testing our deep neural network predictor.

3.2. PPG Raw Data Preprocessing

Prior to the actual process of estimating the blood pressure using PPG signal exclusively, we conduct a preprocessing to enhance the quality of the PPG. The raw PPG data preprocessing consists of four main steps, including noise removal, normalization, feature point detection, and partitioning. The detailed explanation of each step is presented as follows: Noise removal: Fast Fourier transform (FFT) is applied to every PPG data segment to convert it from its time domain into the frequency domain. Let , represent the PPG, and the FFT of x[n] is denoted as . We remove the frequency components that are lower than 0 Hz or higher than 8 Hz by turning off those frequency components, as follows By removing this range of frequency, we aim at removing noise and the baseline wander. The PPG signal can then be restored into the time domain with inverse FFT (IFFT). Normalization and 1st and 2nd derivative of PPG calculation (denoted as “dPPG” and “sdPPG”): All the raw values of PPG are positive, so min–max normalization is applied to every PPG data segment. The equation of min–max normalization can be represented as (5): where x are data points in each PPG data segment {X} and and are the minimum and maximum values, respectively, in each PPG data segment {X}. After min–max normalization, the values of every PPG data segment are within the range [0 1] and dPPG and sdPPG (1st and 2nd derivative of waves of PPG) are calculated at the same time. Feature point detection: Before feature extraction, a few points should be marked and detected in every cycle of the heartbeat for every signal (PPG, dPPG and sdPPG) for cycle segmentation and alignment. Firstly, the systolic peaks of PPG waves of each heart cycle are marked by taking advantage of an algorithm mentioned in [24]. The correctness and validity of the systolic peak detection algorithm is of vital importance because the rest of the feature point detection algorithm is based on it. Secondly, the onset and offset valley points of PPG are detected by finding the minimum between two consecutive systolic peaks. Thirdly, with the valley points of PPG found, the location with the maximal and minimal slope values of PPG and dPPG can easily be derived by computing their gradients. Fourthly, the dicrotic notch points of PPG are detected by finding the secondary peaks of the sdPPG contour [20]. An example set of waveforms is shown in Figure 4.
Figure 4

The results of feature point detection; (a) Photoplethysmogram (PPG), (b) 1st derivative of PPG (dPPG) and (c) 2nd derivative of PPG (sdPPG).

Partitioning and abnormal cycle removal: After feature points are located, each PPG data segment and its corresponding dPPG and sdPPG waves are partitioned into fragments by reserving each PPG data segment from one valley point of PPG to the next consecutive valley point of PPG. Abnormal heart cycles are also removed following the criteria mentioned in [23]. After abnormal cycle removal is done, the histograms of distribution of SBP and DBP are plotted, as seen in Figure 5, and approximately 2.17 million PPG, dPPG and sdPPG data fragments of heart cycles are obtained.
Figure 5

Histograms of distribution of blood pressure; (a) distribution of systolic blood pressure (SBP) (b) distribution of diastolic blood pressure (DBP).

3.3. Feature Extraction and Selection Index

3.3.1. Feature Extraction

The candidate features are the 65 features proposed in past studies [10,16,25,26,27,28], which are reported to be highly related to blood pressure estimation [11]. Among them, we select 59 features, including hr, t1, t2, t3, t4, t5, t6, t7, t8, AS, dAS, sdAS, DS, dDS, sdDS, S1, S2, AA, dAA, sdAA, DA, dDA, sdDA, RAAD, dRAAD, sdRAAD, PI, dPI, sdPI, dVI, sdVI, AID, dAID, sdAID, dDID, sdDID, PIR, dPIR, sdPIR, dRIPV, sdRIPV, AT, dAT, sdAT, DT, dDT, sdDT, dTVO, sdTVO, Slope_a, S3, S4, RtArea, NI, AI, AI1, RSD, RSC and RDC. All the definitions of the 59 features are listed in Table 1. The extracted features are first standardized to value [−1 1] using Z-score normalization, as shown in (7), for each feature: where y are the elements in each feature and and are the mean value and standard deviation of each feature, respectively. Although the authors of [16] observed a phenomenon that the fluctuation in BP led to conspicuous changes in these 59 features, seemingly unveiling the close correlation between BP and these 59 features in their dataset, whether the same phenomenon will happen again in our experimental data source is still unclear and unpredictable due to the fact that the database we use is not only different but is also more diverse than the dataset mentioned in [16]. Consequently, statistical experiments on our dataset are recomputed and, moreover, an index is introduced to evaluate the degree of correlation between each feature and BP.
Table 1

Part of the PPG feature definitions and corresponding s are computed and listed in this table and all the definitions and denotations are in reference to past studies [10,16,25,26,27,28].

DenotationFeatureDefinition of Feature σ
η1 hrHeart rate# 0.01281 *
η2 t1Time interval of S1 (as seen in Figure 4)-
η3 t2Time interval of S2 (as seen in Figure 4)0.01653
η4 t3Time interval of S3 (as seen in Figure 4)0.01637
η5 t4Time interval of S4 (as seen in Figure 4)0.01604
η6 t5Time interval of dAA (as seen in Figure 4)0.01634
η7 t6Time interval of sdDA (as seen in Figure 4)0.01628
η8 t7Time interval of sdAA (as seen in Figure 4)0.01673
η9 t8Time interval of sdDA (as seen in Figure 4)0.01628
η10 ASAscending slope of PPG (slope from onset point to maximum peak)# 0.01276 *
η11 dASAscending slope of dPPG0.01455
η12 sdASAscending slope of sdPPG# 0.01405
η13 DSDescending slope of PPG (slope from maximum peak to offset point)# 0.01405 *
η14 dDSDescending slope of dPPG0.01646
η15 sdDSDescending slope of sdPPG# 0.01298
η16 S1Area under PPG curve between onset point and maximum slope point (as seen in Figure 4)0.01556 *
η17 S2Area under PPG curve between maximum slope point and maximum peak (as seen in Figure 4)0.01411 *
η18 AAAscending area of PPG (as seen in Figure 4)# 0.01381 *
η18 AAAscending area of PPG (as seen in Figure 4)# 0.01381 *
η19 dAAAscending area of dPPG (as seen in Figure 4)# 0.01255 *
η20 sdAAAscending area of sdPPG (as seen in Figure 4)# 0.01298 *
η21 DADescending area of PPG (as seen in Figure 4)# 0.01232 *
η22 dDADescending area of dPPG (as seen in Figure 4)# 0.01228 *
η23 sdDADescending area of sdPPG (as seen in Figure 4)# 0.01265 *
η24 RAADRatio of ascending area to descending area, AA/DA-
η25 dRAADdAA/dDA-
η26 sdRAADsdAA/sdDA-
η27 PIPeak intensity of PPG# 0.01261 *
η28 dPIPeak intensity of dPPG# 0.01313 *
η29 sdPIPeak intensity of sdPPG# 0.01305 *
η30 dVIValley intensity of dPPG# 0.01296 *
η31 sdVIValley intensity of sdPPG# 0.01299 *
η32 AIDAscending intensity difference of PPG, intensity difference between maximum peak and onset point # 0.01324 *
η33 dAIDAscending intensity difference of dPPG, intensity difference between maximum peak and onset point # 0.01311 *
η34 sdAIDAscending intensity difference of sdPPG, intensity difference between maximum peak and onset point # 0.01305 *
η35 dDIDDescending intensity difference of dPPG, intensity difference between offset point and maximum peak # 0.01322 *
η36 sdDIDDescending intensity difference of sdPPG, intensity difference between offset point and maximum peak # 0.01310 *
η37 PIRPeak intensity ratio of PPG, ratio of maximum peak intensity to onset intensity -
η38 dPIRPeak intensity ratio of dPPG, ratio of maximum peak intensity to onset intensity-
η39 sdPIRPeak intensity ratio of sdPPG, ratio of maximum peak intensity to onset intensity-
η40 dRIPVRatio of maximum peak intensity to valley intensity of dPPG# 0.01305 *
η41 sdRIPVRatio of maximum peak intensity to valley intensity of sdPPG# 0.01350 *
η42 ATAscending time interval of PPG # 0.01348 *
η43 dATAscending time interval of sPPG0.01634
η44 sdATAscending time interval of sdPPG0.01673
η45 DTDescending time interval of PPG 0.01490
η46 dDTDescending time interval of dPPG 0.01628
η47 sdDTDescending time interval of sdPPG 0.01628
η48 dTVOTime interval between valley point and offset point of dPPG0.01569
η49 sdTVOTime interval between valley point and offset point of sdPPG0.01438
η50 Slope_aSlope from maximum peak to dicrotic notch of PPG# 0.01308 *
η51 S3Area under PPG curve between maximum peak and dicrotic notch (as seen in Figure 4)# 0.01333 *
η52 S4Area under PPG curve between dicrotic notch and offset point (as seen in Figure 4)# 0.01323 *
η53 RtAreaRatio of systolic area to diastolic area, (S1 + S2 + S3)/S4 (as seen in Figure 4)-
η54 NIDicrotic notch intensity# 0.01230 *
η55 AIAugmentation index = NI/PI# 0.01277 *
η56 AI1Augmentation index 1 = (PI − NI)/PI# 0.01274 *
η57 RSDRatio of systolic duration to diastolic duration, (t1 + t2 + t3)/t4 # 0.01405 *
η58 RSCRatio of diastolic duration to cardiac cycle, t4/(t1 + t2 + t3 +t4)# 0.01286 *
η59 RDCRatio of systolic duration to cardiac cycle, (t1 + t2 + t3)/(t1 + t2 + t3 + t4)0.01611

“#” indicates that a value is one of the first 32 low values of the features. “*” indicates one of the selected 32 features. “-” suggests that the value is too large to be considered.

3.3.2. Selection Index

An index is introduced to ensure the validity of all selected features, and is defined as: where: f = probability mass function of standardized target feature, its estimated precision is down to k decimal places. (u) = probability density function of standard normal distribution. = . = 0, the mean of the standardized target feature. = C, where is the standard deviation of the target feature (in the case of standardized features, and are equal to 0 and 1) and C is an integer. For evaluation, the definition of the values of features ranging from – C to + C is used. Figure 5 shows the histogram distribution of SBP and DBP values in our dataset. From Figure 5, it seems that the distributions of SBP and DBP are close to the normal distribution and, as a result, is designed to check the degree of similarity between the standard normal distribution and the distributions of each standardized feature. The smaller the value of , the higher the degree of similarity. In fact, the basic concept of is to compute the mean absolute error (MAE) between the standard normal distribution and the distributions of each standardized feature within u =. In our experiments, firstly, while computing our standardized features are all computed to one decimal place. So, in our case, k = 1. In addition, since we plan to evaluate the similarity using a value within u =, in our case, C = 3. All the values of of the 59 features mentioned above are computed and listed in Table 1.

3.4. Deep Neural Network Predictor

After feature selection is done, the optimal feature set considered to be highly correlated to BP is obtained. The next step is to train a machine learning model which is able to predict the SBP and DBP values accurately given the selected features. In this work, we use a fully connected deep neural network regressor for this goal. As shown in Figure 3, our model is composed of multiple fully connected layers with activation function “ReLU”. Between input layers and output layers, each hidden layer contains 2048, 4096, 8192 and 2048 fully connected neurons, respectively.

3.4.1. Introduction to Fully Connected Neural Network

As a matter of fact, a brain of a human comprises billions of neurons connected each other with synapses, and each neuron communicates through electrical currents. A special kind of machine learning model, called a neural network (NN), was proposed a long time ago to mimic the behaviors of neurons. A generic NN consists of perceptrons, mimicking the function of biological neurons, and an interconnected layered structure that connects every perceptron in one layer to another. Each perceptron contains a weighted vector W and a bias b, as seen in Figure 3, whose value gets updated iteratively during the training process. The correlation between the input and output of a fundamental perceptron can be formulated as (9): where is a transpose input vector of a perceptron, O() is an output value of a perceptron, act() represents an activation function, and, in our case, activation function “ReLU” is applied in our deep neural network models. If every perceptron in one layer is connected to every perceptron in the next layer, such an NN is called a “fully connected neural network”, as seen in Figure 6.
Figure 6

The deep neural network (DNN) architecture for the proposed method. There are four hidden layers, denoted as H1, …, H4. The numbers of neurons for H1, H2, H3 and H4 are 2048, 4096, 8192 and 2048, respectively.

3.4.2. Neural Network Selection

As the development of artificial intelligence has evolved, more and more different kinds of neural networks, such as fully connected networks, convolutional neural networks, and recurrent neural networks, have been proposed to tackle different kinds of problems. Among them, long short-term memory (LSTM) and fully connected neural networks are the most commonly applied regressors for building BP estimation models. Su et al. [29] constructed an LSTM-based model with high accuracy across 84 subjects by extracting classical PTT-related features mentioned in Section 1. On the other hand, Kurylyak et al. [30] used a fully connected neural network to build up a valid model across 15,000 cardiac cycles by utilizing temporal features extracted from PPG segments. The two different kinds of models actually have their pros and cons. In this paper, we decided to adopt the fully connected neural network as our regressor since it is easier to be implemented in wearable devices. The model structure is clean and easier to understand compared to LSTM, which enables software engineers to transfer and deploy the code to wearable devices. Another advantage of a fully connected network for BP estimation is that it takes inputs from only one cardiac cycle to estimate BP. On the other hand, the LSTM model usually takes inputs from several cardiac cycles before it outputs BP values, which causes a time delay when dealing with patients with a critical situation of CVD.

4. Experiments and Results

4.1. Feature Point Detection and Abnormal Cycle Removal

Following the methodologies mentioned in Section 3.3, one example of a result of a PPG and its corresponding dPPG and sdPPG marked with feature points is demonstrated in Figure 7 and the validity of the algorithm that helps us locate the dicrotic notch in every cardiac cycle by finding the corresponding secondary peaks of the sdPPG contour signal [20] is strengthened and verified by our statistical results after experimenting with normalized notch intensity across more than 2.17 million cycles in our dataset. The distribution of normalized notch intensity is shown in Figure 8. After feature point detection, partitioning and abnormal cycle removal are done, approximately 2,176,188 data fragments of the PPG, dPPG and sdPPG of a single cardiac heart cycle are obtained and the distributions of the corresponding SBP and DBP values are shown in Figure 5. One of the results of feature point detection is demonstrated in Figure 7.
Figure 7

Examples of the results of feature point detection.

Figure 8

Distribution of normalized notch intensity.

4.2. Characteristic Features of Cardiac Cycles

With the help of the s computed and listed in Table 1, the process of feature selection becomes easier. The criterion for feature selection is to choose the features with the lowest . Indeed, 30 out of 32 selected features, including hr, AS, DS, AA, dAA, sdAA, DA, dDA, sdDA, PI, dPI, sdPI, dVI, sdVI, AID, dAID, sdAID, dDID, sdDID, dRIPV, sdRIPV, AT, Slope_a, S3, S4, NI, AI, AI1, RSD and RSC, are selected from the first 32 features with the lowest values, while the remaining two of the 32 selected features, which are S1 and S2, are not. The reasons why we include these two features are that their are relatively low and they are commonly used features that were reported to be highly related to BP in the literature. Most importantly, the reason why sdAS and sdDS, which are on the list of the first 32 features with the lowest values, are not selected is because that the performance of models whose input features contain sdAS and sdDS is worse than the performance of those whose inputs contain S1 and S2. Finally, a set of features (), including hr, AS, DS, AA, dAA, sdAA, DA, dDA, sdDA, PI, dPI, sdPI, dVI, sdVI, AID, dAID, sdAID, dDID, sdDID, dRIPV, sdRIPV, AT, Slope_a, S3, S4, NI, AI, AI1, RSD, RSC, S1 and S2, is selected as our final feature set. Figure 9 shows the distributions of the values of the first two features, dDA(a) and NI(b), and the last two features, sdAT(c) and t7(d), which have the lowest values. Figure 10 shows the distributions of the values of features S1(a), S2(b), sdAS(c) and sdDS(d).
Figure 9

(a) Distribution of dDA, (b) distribution of NI, (c) distribution of sdAT and (d) distribution of t7.

Figure 10

(a) Distribution of S1, (b) distribution of S2, (c) distribution of sdAS and (d) distribution of sdDS.

4.3. Model of Deep Neural Network Predictor

Despite the fact that models built from LSTM units are the most frequently and widely applied for time series-related problems, in this study, we choose a different approach. We use a fully connected neural network, which is much simpler in terms of the number of parameters compared to an LSTM unit, to build up our core deep fully connected neural network, serving as a predictor of BP. We conduct tests on several models to determine the optimal number of hidden layers and neurons, following the approach in [30]. Finally, our model is introduced as follows. Our model, as shown in Figure 3, is a six-layered structure of a fully connected neural network and the dimensions of the input layer are 1 × 32, which represents the features extracted from a single cardiac cycle. The numbers of hidden nodes in each layer are 2048, 4096, 8192 and 2048 and every node is fully connected to all nodes in the next layer. The activation function we use is ReLU, and the output layer has dimensions of 1 × 2, which are the estimated SBP and DBP, respectively. Before the training of our proposed deep neural network predictor, the selected feature set () is split into three parts, and each part contains 70%, 20% and 10% of the data, which serve as training, testing and validation datasets, respectively. As for the training process, a gradient descent optimizer, called “Nadam” [31], is applied to update all the variables, including w and b, in the model. In every epoch, the root mean square error (RMSE) and MAE are measured as loss functions for every 512 batches. For software implementation, we use Keras [32] to build the DNN model, “numpy” toolkits [33] for signal preprocessing and the “Heartpy” toolkit [24] for cardiac cycle segmentation.

4.4. Performance of Proposed Model

There are several mathematical methods and indices that are used to evaluate the validity of the regression model. Among them, the RMSE and MAE between the ground truth of BP and the estimated BP are the most widely used to gauge the performance of a BP estimation model. The definitions of the RMSE and MAE are shown in (10) and (11), respectively [34]: and where N is the number of total BP samples (SBP or DBP) to be evaluated and and are the ground truth BP (SBP or DBP) and estimated BP (SBP or DBP), respectively. The performance of our proposed model is assessed by the standards established by the Association for the Advancement of Medical Instrumentation (AAMI) [35] and the British Hypertension Society [4]. Additionally, two classical statistical approaches to evaluate a regression model, which are Bland–Altman analysis and Pearson’s correlation analysis, are conducted to evaluate our proposed model. Most importantly, at the end, the RMSE is computed to compare the performance of our work with others and the results will be further discussed and elaborated in the following sub-sections. Figure 11 shows the distribution of absolute error across 2,176,188 records of SBP and DBP.
Figure 11

(a) Distribution of absolute error across 2,176,188 records of SBP and (b) distribution of absolute error across 2,176,188 records of DBP.

4.4.1. Performance Evaluation by AAMI Standards

An article by the Association for the Advancement of Medical Instrumentation (AAMI) [35] suggests that the average and standard deviation of error among numerous measurements of SBP and DBP should not be larger than 5 mmHg and 8 mmHg, respectively. Fortunately, our proposed deep neural network model fulfills the criteria suggested by the AAMI with averages and standard deviations equal to 3.21 mmHg and 3.35 mmHg for SBP and 2.23 mmHg and 2.44 mmHg for DBP across 2,176,188 records of SBP and DBP.

4.4.2. Performance Evaluation by BHS Standards

Table 2 shows the BHS standard for BP measuring devices and the performance of our model. From Table 2, the performances of our deep neural network estimator for both SBP and DBP satisfy grade A of the BHS standards, with 80.63% of error lower than 5 mmHg, 95.86% of error lower than 10 mmHg and 98.78% lower than 15 mmHg for SBP and 90.19% of error lower than 5 mmHg, 98.29% of error lower than 10 mmHg and 99.59% lower than 15 mmHg for DBP.
Table 2

The standards of the British Hypertension Society (BHS) for BP measuring devices and the performance of our model.

Error ≤ 5 mmHgError ≤ 10 mmHgError ≤ 15 mmHg
BHS [4]Grade A60%85%95%
Grade B50%75%90%
Grade C40%65%85%
Our ModelSBP80.63%95.86%98.78%
DBP90.19%98.29%99.59%

4.4.3. Pearson’s Correlation and Bland–Altman Analysis

Pearson’s correlation analysis is one of the most popular methods used to evaluate the validity of a regression model by computing Pearson’s correlation coefficient r, whose value ranges from -1 to 1 and it can be formulated as [36]: where is the number of points (, ) on the plot and and are means and standard deviations, respectively. The basic concept of r is to measure the degree of correlation between two signals, and . In our case, if our proposed model is perfect and error free, then the Pearson’s correlation coefficient of our model should be equal to 1. From Figure 11, the results show that is equal to 0.977 between the ground truth of SBP and the estimated SBP and that is equal to 0.947 between the ground truth of DBP and the estimated DBP across 2,176,188 records in our dataset, revealing the extremely high correlation between estimated BP and the ground truth of BP. Figure 12 shows the Pearson’s correlation analysis results for the estimation error for SBP and DBP.
Figure 12

(a) Pearson’s correlation analysis results for error across 2,176,188 records of SBP and (b) Pearson’s correlation analysis results for error across 2,176,188 records of DBP.

A Bland–Altman plot [37] is used in analyzing the agreement of two different arrays. It is another way to test the difference between estimated BP and the ground truth of BP in our case. Figure 12 show the Bland–Altman analysis results for SBP and DBP. From Figure 12a, there are two horizontal lines, which are and , respectively, forming a range called limits of agreement (LOA), and 95% of data points are in the range. From Figure 12, the LOA for errors of SBP is (−9.38 8.76) mmHg. On the other hand, the LOA for errors of DBP is (−5.97 6.87) across 2,176,188 records, which confirms the accuracy of our proposed model. Figure 13 shows the Bland–Altman analysis results for the estimation error for SBP and DBP.
Figure 13

(a) Bland–Altman plot for error of SBP across 2,176,188 records and (b) Bland–Altman plot for error of DBP across 2,176,188 records.

4.4.4. Comparison with Other Works

To be honest, it is extremely difficult for us to do a fair comparison with prior works for the following reasons. First, most of the existing models used both ECG and PPG as inputs of the models. Second, even if the inputs of the models were PPG only, it is still difficult to compare them, since the datasets used in different studies may be different. Last, but not least, even if the model to be compared uses the same dataset as our model and also takes only PPG as input, the number of subjects (and cardiac cycles) used for training and testing, which will hugely affect the degree of generalization of a model, may still be different. However, despite all the impediments, it is still necessary to compare our proposed model to other works owing to the fact that through the comparison, we will be able to understand and judge our own model better. The comparison results are shown in Table 3.
Table 3

Comparison of different models using PPG only as input for BP estimation.

ResearchersDatasetInputPerformance
Mousavi et al. [38]MIMIC II (441 subjects)PPGBHS standard:Grade A for DBP and Grade D for SBPAAMI:only the results of DBP satisfy the standardsMAE, RMSE: not mentioned in the paper
Slapnivcar et al. [39]MIMIC II (510 subjects)PPGMAE:DBP = 9.43 mmHg, SBP = 6.88 mmHgRMSE: not mentioned in the paper
Ibtehaz and Rahman [15]MIMIC II (942 subjects)PPGBHS standard:Grade A for DBP and Grade B for SBPAAMI:the results of both DBP and SBP satisfy the standardsMAE:DBP = 3.45 mmHg, SBP = 5.73 mmHgRMSE: not mentioned in the paper
Our proposed modelMIMIC II (2,176,188 records of BP in total)PPGBHS standard:Grade A for both DBP and SBPAAMI:the results of both DBP and SBP satisfy the standardsMAE:DBP = 2.23 mmHg, SBP = 3.21 mmHgRMSE:DBP = 3.21 mmHg, SBP = 4.63 mmHg
From Table 3, it is clear that in terms of the accuracy and scale of the experiments, our proposed deep neural network model is one of the best so far and, hence, the generalized BP estimation model with PPG signals only has been achieved.

5. Conclusions and Future Works

In summary, in this study, we propose a fully connected DNN model to estimate SBP and DBP, from a PPG signal only. We perform feature selection based on big data analysis using 9000 subjects, 2,176,188 records of BP in total and 32 optimal features selected based on the proposed selection index. Finally, our proposed model reaches BHS Grade A and satisfies the AAMI standard. The MAE is as low as 3.21 mmHg and 2.23 mmHg for SBP and DBP, and the RMSE is as low as 4.63 mmHg and 3.21 mmHg for SBP and DBP, which outperforms all existing works using the same dataset (MIMIC II). For future works, we plan to apply other RNN-related models to see if the MAE and RMSE can be further decreased. A sequence to sequence model is also a promising direction for this topic. Phase changes in blood pressure can happen under various influences. Thus, further studies should try to involve the correlation between BP estimation and natural time analysis. Additionally, we would like to implement our model in wearable devices to help people at risk of high blood pressure to monitor their BP continuously in their daily activities. Hence, studies on data from people outside hospital should be conducted.
  17 in total

1.  Continuous Cuffless Blood Pressure Estimation Using Pulse Transit Time and Photoplethysmogram Intensity Ratio.

Authors:  Xiao-Rong Ding; Yuan-Ting Zhang; Jing Liu; Wen-Xuan Dai; Hon Ki Tsang
Journal:  IEEE Trans Biomed Eng       Date:  2015-09-22       Impact factor: 4.538

2.  Measurements of Young's modulus of elasticity of the canine aorta with ultrasound.

Authors:  D J Hughes; C F Babbs; L A Geddes; J D Bourland
Journal:  Ultrason Imaging       Date:  1979-10       Impact factor: 1.578

Review 3.  Photoplethysmography Revisited: From Contact to Noncontact, From Point to Imaging.

Authors:  Yu Sun; Nitish Thakor
Journal:  IEEE Trans Biomed Eng       Date:  2015-09-15       Impact factor: 4.538

4.  The British Hypertension Society protocol for the evaluation of automated and semi-automated blood pressure measuring devices with special reference to ambulatory systems.

Authors:  E O'Brien; J Petrie; W Littler; M de Swiet; P L Padfield; K O'Malley; M Jamieson; D Altman; M Bland; N Atkins
Journal:  J Hypertens       Date:  1990-07       Impact factor: 4.844

5.  Characters available in photoplethysmogram for blood pressure estimation: beyond the pulse transit time.

Authors:  Yanjun Li; Zengli Wang; Lin Zhang; Xianglin Yang; Jinzhong Song
Journal:  Australas Phys Eng Sci Med       Date:  2014-04-11       Impact factor: 1.430

6.  Investigating the physiological mechanisms of the photoplethysmogram features for blood pressure estimation.

Authors:  Wan-Hua Lin; Xiangxin Li; Yuanheng Li; Guanglin Li; Fei Chen
Journal:  Physiol Meas       Date:  2020-05-07       Impact factor: 2.833

7.  Statistics review 7: Correlation and regression.

Authors:  Viv Bewick; Liz Cheek; Jonathan Ball
Journal:  Crit Care       Date:  2003-11-05       Impact factor: 9.097

8.  Toward Generating More Diagnostic Features from Photoplethysmogram Waveforms.

Authors:  Mohamed Elgendi; Yongbo Liang; Rabab Ward
Journal:  Diseases       Date:  2018-03-11

9.  Blood Pressure Estimation Using Photoplethysmography Only: Comparison between Different Machine Learning Approaches.

Authors:  Syed Ghufran Khalid; Jufen Zhang; Fei Chen; Dingchang Zheng
Journal:  J Healthc Eng       Date:  2018-10-23       Impact factor: 2.682

Review 10.  Advances in Photopletysmography Signal Analysis for Biomedical Applications.

Authors:  Jermana L Moraes; Matheus X Rocha; Glauber G Vasconcelos; José E Vasconcelos Filho; Victor Hugo C de Albuquerque; Auzuir R Alexandria
Journal:  Sensors (Basel)       Date:  2018-06-09       Impact factor: 3.576

View more
  2 in total

1.  A Shallow U-Net Architecture for Reliably Predicting Blood Pressure (BP) from Photoplethysmogram (PPG) and Electrocardiogram (ECG) Signals.

Authors:  Sakib Mahmud; Nabil Ibtehaz; Amith Khandakar; Anas M Tahir; Tawsifur Rahman; Khandaker Reajul Islam; Md Shafayet Hossain; M Sohel Rahman; Farayi Musharavati; Mohamed Arselene Ayari; Mohammad Tariqul Islam; Muhammad E H Chowdhury
Journal:  Sensors (Basel)       Date:  2022-01-25       Impact factor: 3.576

2.  Subject-Based Model for Reconstructing Arterial Blood Pressure from Photoplethysmogram.

Authors:  Qunfeng Tang; Zhencheng Chen; Rabab Ward; Carlo Menon; Mohamed Elgendi
Journal:  Bioengineering (Basel)       Date:  2022-08-18
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.