Abstract

Objective. Measurement and monitoring of blood pressure are of great importance for preventing diseases such as cardiovascular and stroke caused by hypertension. Therefore, there is a need for advanced artificial intelligence-based systolic and diastolic blood pressure systems with a new technological infrastructure with a noninvasive process. The study is aimed at determining the minimum ECG time required for calculating systolic and diastolic blood pressure based on the Electrocardiography (ECG) signal. Methodology. The study includes ECG recordings of five individuals taken from the IEEE database, measured during daily activity. For the study, each signal was divided into epochs of 2-4-6-8-10-12-14-16-18-20 seconds. Twenty-five features were extracted from each epoched signal. The dimension of the dataset was reduced by using Spearman’s feature selection algorithm. Analysis based on metrics was carried out by applying machine learning algorithms to the obtained dataset. Gaussian process regression exponential (GPR) machine learning algorithm was preferred because it is easy to integrate into embedded systems. Results. The MAPE estimation performance values for diastolic and systolic blood pressure values for 16-second epochs were 2.44 mmHg and 1.92 mmHg, respectively. Conclusion. According to the study results, it is evaluated that systolic and diastolic blood pressure values can be calculated with a high-performance ratio with 16-second ECG signals.

1. Introduction

1.1. Background and Motivation

The state of blood pressure higher than usual is called hypertension [1]. Hypertension is a risk factor for many cardiovascular diseases such as stroke, renal failure, and heart attack that affect many people today [24]. Hypertension is responsible for one out of every two deaths in the world [3]. However, with BP data from 1.7 million people in 31 provinces of China, many people with or without hypertension or advanced disease have been identified, and they have developed methodologies for awareness, diagnosis, and control of the disease [5]. The level of importance is emphasized in the study conducted in a country with a significant population such as China. However, continuous blood pressure monitoring is essential for diagnosis and treatment.

BP monitoring is an essential factor in the early diagnosis and treatment of hypertension [2, 3]. When the heart muscle contracts, systolic blood pressure occurs, and diastolic blood pressure occurs when it relaxes. In other words, high BP is an indicator of the pressure exerted by the blood on the blood vessels [1, 3, 6]. SBP is about 80 mmHg, and DBP is about 120 mmHg. Hypertension is defined as above 140 mmHg SBP and 90 mmHg DBP. If the BP is above average, the heart starts to work harder. An imbalanced heart causes swelling in the vessels, blindness, and heart failure causes many diseases [3]. Since the heart meets the nutritional needs of all organs, imbalances in the heart indirectly affect other organs. In this case, hypertension affects many organs, especially the brain and kidneys [13].

Although the gold standard is mercury methods in BP monitoring, many methods have been developed today [3, 7]. Due to the toxic effects of mercury, reliable electronic devices are recommended by the World Health Organization [3].

According to the World Health Organization, the hypertension diagnosis algorithm is as follows [3]. (1) A few days of registration should be taken. (2) Measurements should be made twice a day, in the morning and the evening. (3) Two consecutive measurements are taken with at least one minute between them in each measurement. The average value of the measurements is taken. (4) Measurements are taken in the sitting position. Although a few daily measurements are considered sufficient by the World Health Organization, a one-week follow-up is recommended in some studies [8].

Studies show that drug treatments can control hypertension [1, 9]. In patients with hypertension, the disease and drug treatment follow-up are vital. For this reason, it is recommended that patients follow their treatments meticulously [1, 3, 9].

In the past, BP measurements were made in health centers since devices for BP measurements were not widely used. However, today, with the widespread use of BP measuring devices, the possibilities of measuring at home have increased. The most important feature of the devices was that individuals could measure without the need for technical knowledge [7, 10]. Although the devices are easy to use, measurement accuracy is vital. For this reason, it is recommended to prefer verified devices [10].

Each device can use a different measurement technique and location [7]. However, the World Health Organization recommends measurement from the upper arm in a sitting position [3]. It is known that upper arm measurement produces more accurate results than wrist measurement [3]. In patients who cannot be in every position, such as pregnant, there is a significant change in the results because the measurement standard is exceeded [11].

Although BP monitors take measurements from different places and calculate BP with different methods, they are subject to the same design protocols. There are many protocols in the literature [10]. The aim of all of them is to develop a quality measuring system. The International Organization for Standardization (ISO) [12], the European Society of Hypertension International Protocol (ESH-IP) [1315], and the Association for the Advancement of Medical Instrumentation (AAMI) standard [16] are just a few of them.

Noninvasive brachial BP measurement is a primary method in diagnosing hypertension diseases [17, 18]. However, this measurement creates significant problems for public health. Every year, in American states, it has been determined that the BP measurement of approximately one million people is measured above or below normal by 5 mmHg. With the resulting extra costs, people have suffered the adverse effects of the wrong treatment [17]. For this reason, it is essential to determine the correct BP measurement method. There are many BP estimation methods in the literature, such as auscultatory methods, plethysmography, tonometry, and oscillometric methods [7, 10]. Although cuffed devices have been produced for many years, the new target is to measure BP without causing discomfort to the patient [10, 19]. Noninvasive methods, including PPG- and electrocardiography- (ECG-) based signal processing and artificial intelligence, are promising BP measurement methods [10, 19, 20].

Many methods with invasive and noninvasive structures are used in BP measurements. Invasive measurement systems often cause problems such as incorrect measurements and loss of time. However, noninvasive models cause many problems, such as ease of use and hardware and software costs. If it is mentioned in a few articles in the literature that these problems are seen, the initial states of the model designed with a neural network using a genetic algorithm were tried to be optimized, but this caused problems in terms of time and hardware cost and caused only a tiny change in the accuracy rate. While the impact of BP measurements on human life is of such high importance, developing these systems is of great importance for researchers. For BP measurement models, each study applies a different methodology. This study takes its place with a different approach between these studies, which differ in their superiority to each other. A study on the determination of blood pressure measurement time has not been found so far. For this reason, the study presents an original model. Instead of complex and sophisticated mathematical expressions seen in deep learning algorithms, a model based on machine learning techniques has been developed. In addition, the findings obtained in the literature review show that the designed model has a higher accuracy rate than other models.

Signal processing, spectrum segmentation, feature extraction applications, and morphological filters are used in many fields. In this study, signal segmentation and feature selection were performed. These processes occur not only in fields such as biomedicine, medicine, and image processing but also in engineering and commerce. It has applications with various methodologies in different fields. Tabatabaei et al. performed signal analysis with acoustic emission method (AEM) to detect defects on angular contact bearings. The authors applied feature extraction with empirical mode decomposition (EMD) algorithm on the signals obtained with AEM [21]. EMD is used to analyze nonlinear and nonstationary signals by separating them into components with different resolutions [22]. Using this method, the authors extracted intrinsic mode functions from the signals as if they were extracting time and frequency domain properties, and by transforming these functions into an analytical expression with Hilbert transform, they created a model that can detect defects on bearings. However, the same application process has been applied by different authors in classifying electroencephalography (EEG) signals as seizure and nonseizure [23]. EEG signals are decomposed into intrinsic mode functions by the EMD method. These functions were converted into an analytical expression and extracted from the bandwidth properties of the signals. Classification based on ML techniques was created at the end of the model. Two articles with similar content applied the exact solution to different problems. In yet another article, different domain properties were used to detect defects on angular contact bearings, which is one of the crucial subgroups of bearings, and as a result, the model that would make the best defect detection was realized by using time-domain properties [24]. Zhang and Wang extracted 21-time domain features for BP detection from PPG signals [25]. They provided the size optimization with mean impact value (MIV). By calculating the effect factor on the signals of each feature, they extracted 8 features with a shallow impact. He created a new model by optimizing the initial conditions of the neural network, the last used ML technique, with the genetic algorithm. This article, which has the same research subject, has created its model with a specific approach [25]. However, considering that the amount of features used is high, the use of optimization algorithms causes a minimal change in the accuracy of blood pressure detection, and it is thought that it will be expensive in terms of hardware cost. It is concluded that it is not a good model.

The development of machine learning algorithms makes it possible to create innovative models for the same problem. The article’s methodology is presented step by step by addressing such processes. First, signal processing is applied so that the signals can be used efficiently. There are many studies based on signal processing in the literature. These are frequently preferred because they are noninvasive methods [2628]. Although its measurement is relatively difficult compared to PPG, some studies use oscillometric waveforms [29, 30], auscultatory and oscillometric waveforms [31], and peripheral signals [26]. In addition to these, speech-based measurement methods are also available [32]. These studies applied signal processing processes, and machine learning algorithms were used in BP estimation according to the need. PPG and ECG are a group of biomedical signals that are very easy to measure. ECG can be easily measured on bedside monitors and Holter devices. On the other hand, PPG can now be measured even on intelligent wristbands. For this reason, interest in ECG- and PPG-based BP estimation studies has increased in the literature [10, 19, 20]. ECG signals were preferred in this study because they are easy to measure.

Signal processing studies include digital filtering, feature extraction, feature selection, and machine learning-based regression steps [20, 27, 33]. In the case of using deep learning, feature extraction and selection operations are performed by deep learning [34]. The success of the signal processing process depends on the designed steps. The feature extraction steps often focuses on the formal properties of the signals [2931]. In the event of deformities, the system’s collapse is expected. However, examining the statistical properties of the signal instead of the formal properties will reduce the errors and help catch the overlooked information [3537]. In this study, 25 statistical-based features were extracted instead of the standard features used in the literature.

Although deep learning methods include feature extraction and selection steps, the training periods are pretty long [20, 28, 34]. Compared to classical machine learning algorithms, the success rate of deep learning is relatively high. This study proposes a method based on classical machine learning algorithms by optimizing signal processing processes. This study’s Gaussian process regression (GPR), regression tree ensembles, and regression trees were preferred because of their high-performance [3840].

Feature selection algorithms are generally not preferred in BP estimation studies [26, 27, 31]. In studies where classical machine learning algorithms are preferred, model performance is increased using feature selection algorithms [4143]. The Spearman correlation coefficient-based feature selection algorithm was preferred in this study due to its high performance.

In the literature, studies on BP estimation seem to be one step ahead of learning-based studies [20, 28, 32]. Model performance values in BP estimation models developed with ECG and PPG signals are , , and [20, 33, 34]. In a study with auscultatory and oscillometric waveforms, the model performance was and .

Although models have been developed in the literature, there is no information about how many seconds the models can measure. This study is aimed at determining the minimum duration of ECG signal required to estimate SBP and DBP with ECG. Firstly, ten different datasets were created in the study by dividing the ECG signal into epochs of 2, 4, 6, 8, 10, 12, 14, 16, 18, and 20 seconds. Then, 25 features were extracted from each epoch in the time domain. With the help of the Spearman feature selection algorithm, relevant features were selected, and BP values were estimated with the help of machine learning algorithms.

1.2. Literature Review

Machine learning, the application of artificial intelligence, creates a paradigm shift in medicine with its features in pathological diagnosis, patient monitoring, and helping treatment [44]. Numerous studies are being conducted on the relationship between biomedical signals and blood pressure for the appropriate and timely treatment of hypertension using machine learning algorithms [45]. A typical biomedical signal processing system includes the biological system of interest, the sensors used to capture the activity of the biomedical system, and the process of extracting the appropriate methodology to analyze the signals and extract the desired information from the relevant signal. The biological signal examined in this study is the ECG signal, which shows the heart’s electrical activities.

Recent technological advances have made wearable biosensors suitable for daily use. Wearable biosensors provide an opportunity for real-time monitoring of vital human signs, providing timely feedback, and providing early diagnosis and treatment possibilities [46, 47]. Unlike conventional BP sensors, which are subject to a specific measurement procedure, modern wearable biosensors monitor the relevant signals all day long and do not create a burden other than wearing the device. Reliably receiving these signals from the human body and collecting the received signal data brings along important research [4850]. Securely collecting, verifying, and transporting electronic data is achieved through the integration of the internet of things (IoT) and artificial intelligence technology [5153]. The difficulty of effectively guaranteeing the quality of IoT equipment brings with it the difficulty of ensuring the reliability and accuracy of the data [54, 55]. This situation is of great importance in the accuracy of algorithms based on ML techniques. Ahamed and Farid states that datasets produced with IoT terminals cannot fully cover medical scenarios, and intelligent diagnosis will be significantly reduced [56]. However, advances are currently being made in IoT, allowing high-reliability acquisition of medical image data and multiwaveform data [57]. Good results are obtained from such devices in real-life conditions. Superficial temporal artery tonometry-based device [58], PPG optical sensor [59], ARTSENS (Arterial Stiffness Assessment for Noninvasive Scanning) pressure for brachial arterial pressure [60], a BP estimator based on the principle of volume compensation citeTanaka2007, and the Modulated Magnetic Blood Signature mechanism [61] noninvasive are some of the measurement systems developed for BP monitoring.

Most research on blood pressure estimation uses either electrocardiogram and photoplethysmograph signals or a combination. While this causes more problems, it also brings with it the need for equipment. Taking PPG signal measurements requires the use of many techniques [6265]. Proper use of these techniques requires accurate measurement of PPG signals. Therefore, the use of PPG signals for BP measurement is not a correct option [66, 67]. This study, on the other hand, is aimed at estimating BP with the dataset created by extracting statistical features from ECG signals.

1.3. Aims and Contributions

In real-life situations, in vehicles, at home, or in hospitals, BP can only be measured with a stand-alone BP device. On the other hand, modern telemedicine allows the development of biosensors in electrodes attached to the patient’s chest, allowing the measurement of BP to be obtained. The proposed method offers usage areas ranging from clinical situations to military environments with wearable sensor technology. In addition, a suitable methodology has been developed to reduce the need to connect various sensors to the human body. The developed model is aimed at determining the SBP and DBP estimation time based on artificial intelligence-based ECG signals. Although there are many studies in the literature that detect blood pressure with ECG, there has been no study on how long blood pressure can be measured.

2. Methodology

The basic approach to providing BP estimation is shown as a flow diagram in Figure 1. The process given in the flow chart was applied step by step. First, the ECG signals are sourced from the IEEE open-source data-sharing platform [68, 69]. In the received dataset, epoching was applied to estimate BP in the shortest possible time frame for ECG signals. After this step, the time domain features of each ECG signal are taken for feature extraction. Derived features are used as inputs for various machine learning algorithms. Finally, a suitable feature selection algorithm was applied to the model to increase the performance. The described process will be discussed in detail in the following sections.

2.1. Data Preprocessing
2.1.1. Data Acquisition

In this study, the open-source dataset on the IEEE database sharing platform was used [68, 69]. In this dataset, there are ECG, PPG, and BP records. These records were obtained from five young, healthy individuals (one female and four male) who did not have peripheral or cardiovascular disease and ranged from sedentary to regular activity levels. The dataset was created by taking records for six and a half hours each day for three days. Example ECG signals of these individuals are given in Figure 2.

2.1.2. Epoching

The ECG signal with a sampling frequency of 64 Hz was split into epochs of 2-4-6-8-10-12-14-16-18-20 seconds to generate the BP prediction model. SBP and DBP signals were obtained for each period. Example ECG and BP signals of the 4-second epoch are given in Figure 3.

Figure 3 contains simultaneous BP signal for 4-second ECG recording. The BP signal has four maximum and minimum points for 4 seconds. The minimum points of the BP signal represent DBP, while the maximum points represent SBP. DBP and SBP values are calculated over these points. SBP corresponds to the average of the maximum points and DBP to the average of the minimum points. If Figure 3 is examined in detail, it will be observed that there are similar characteristics of ECG and BP signals.

2.2. Feature Extraction

The five simultaneous ECG signals obtained must have certain inputs for the machine learning algorithms to be applied. These inputs are descriptive parameters often used in statistical science. Descriptive parameters include information such as standard deviation, central moment, IQR, and 25 in total. A total of 125 features were extracted from five signals (Table 1). Here, feature extraction is aimed at obtaining ECG signal information with the help of different parameters by preventing information loss.

2.3. Feature Selection Algorithm

In other words, feature selection, called size optimization, is an algorithm that eliminates irrelevant features in the dataset [32]. Feature selection is applied to increase the performance of machine learning algorithms and reduce the size of the dataset containing the feature entries and the computational load [27]. Many parameters are used to calculate the level of relationship between features in statistics. This study used a Spearman correlation coefficient-based feature selection algorithm for feature selection.

2.3.1. Spearman Correlation Coefficients

Spearman’s correlation coefficient is a statistical method used to express the level of correlation of features in a dataset with the label (SBP or DBP). It takes values between -1 and +1 [70]. While +1 indicates a perfect fit between the data, -1 indicates a negative perfect fit. 0 indicates that there is no relationship level. Accordingly, the level of relationship between SBP and DBP values and 125 features is presented in Table 2.

2.4. Machine Learning Algorithms

The study’s machine learning algorithms are ensemble bagged tree, fine tree, and Gaussian process regression. These models have good performance, simple structure, and widespread use in regression problems [43]. The top three algorithms with the best performance were selected for use in a software environment with extensive machine learning algorithms. Performance monitoring of each epoched signal has been made for the selected ML techniques. The explanation based on the working principle of these algorithms is presented in detail in the following sections.

2.4.1. Ensemble Bagged Tree: Prediction of SBP and DBP

Classification and regression trees (CART) is a machine learning technique developed by Breiman et al. (1984) [71]. The change in training sampling in the dataset causes an imbalance in the technique in question [72]. For this reason, the ensemble technique was preferred in the study. In addition, the nonlinear structure of the output values of the study data has been a good reason for the use of this method. It was created by combining several decision tree structures [73] to increase the performance value derived from a single decision tree structure. Ensemble has three methods: bagging, random forests, and boosting [72]. EBT was used to estimate SBP and DBP in the study. This model was designed based on a statistical method called boostrap [71, 74].

In bagging, many bootstrap samples are taken from the ECG signal training dataset. The regression model for each sample was carried out, the generated bootstrap samples were combined, and the model took its final form. The bagging predictor is determined by taking the average of this regression model. It is desired that the mean value is low, and the variance is high. For this reason, it is aimed that the algorithm structure that will perform the SBP and DBP estimation will work with a high accuracy rate. Accordingly, the performance of EBT on each epoched signal was determined.

2.4.2. Fine Tree: Prediction of SBP and DBP

There are many original structures developed for CART [75, 76]. EBT, one of these structures, is explained in detail. The decision tree structure examined in the EBT is implemented using a single decision tree for this model. The model is designed by performing depth control with the split number of the decision tree. In addition, the decision tree is divided into fine, medium, and coarse according to the performance values (from best to worst) for certain split numbers in the software environment. A FT algorithm has been applied for the model design, which exhibits a reasonable prediction accuracy rate with its fit to the dataset.

Empirical evidence predicts that a correct, step-by-step decision tree is faster than a model in which the entire training dataset is tested and constructed. However, the final model of an iteratively designed tree cannot be reached without using the entire training set [77]. In addition to these contradictions, the model design must be clear, simple, and have a high accuracy rate. This study achieved the desired targets by keeping the number of splits to a minimum. SBP and DBP prediction models were created for each epoched signal. The highest performance ratio was obtained in the eleventh feature group. However, FT was observed to be low compared to the performances of other used EBT and GPR models. This can be explained as the dataset used does not fit well with the algorithm.

2.4.3. Gaussian Process Regression: Prediction of SBP and DBP

Gaussian process regression (GPR), a Bayesian method, is a robust algorithm used for nonlinear regression models. Since the input parameters are nonlinear and more than one, SBP and DBP estimation were performed with this algorithm.

A Gaussian operation can be defined by its mean (1) and covariance function (2) [30]. The GPR covariance of the input variables is named as kernel or covariance function. The use of mean function and kernel function together refers to GPR.

2.5. Performance Evaluation Criteria

Mean absolute percentage error (MAPE-Equation (3)), mean absolute deviation (MAD-Equation (4)), standard error (SE-Equation (5)), mean squared error (MSE)-equation refMSE), root mean square error (RMSE-Equation (7)), , and are used in 7 parameters [43].

The dataset for training and testing is divided as specified in Table 3. The performance value of the model was evaluated for both split datasets.

3. Results

The study mainly aims to create models that can detect SBP and DBP estimation as soon as possible based on machine learning algorithms with ECG signals. In line with this goal, the application was carried out step by step according to the flow diagram given in Figure 1. First, the IEEE open-source data-sharing platform obtained a PPG, ECG, and BP information dataset. ECG and BP data from five individuals were used in the total. Afterward, the signals were epoched up to 20 at intervals of two seconds, starting from 2 for the epoching process. Since the dataset does not have input information, 25 feature extractions were made in the time domain. In order to achieve size optimization, feature selection was made in the next step. Performance evaluation based on the three best-performing machine learning algorithms was conducted for 11 selected feature sets. Finally, the algorithm that performs the estimation of SBP and DBP with the best performance ratio in the shortest time is expressed both graphically and numerically.

In the first part of the study, SBP models were prepared for each epochated signal (Tables 413). Twenty-five feature extractions were performed as ECG signal input information. Feature selection was carried out to reduce the size of the feature vector and get rid of the workload of unnecessary features. The correlation of 25 features was calculated by Spearman’s method, and they were ranked from the highest correlation level to the lowest correlation level. Feature vectors were selected at 5% intervals according to the order. Initially, 5% of 25 features are taken (rounded to integers) to 1, 10% to 3, 15% to the fourth feature vector, up to 50%, 13. It was continued in the same way, and finally, the feature vector of the 25th column was obtained by taking 100%. Eleven groups were created with the ECG signal, and the performance evaluation table of 33 models was produced. According to this table, the best model is determined by dark color. EBT, the model created using all feature vectors (11 items), was the best performing algorithm with mmHg and  mmHg SBP values (Table 4).

The working process continued in the same way for the other epoched signals. For the 4-second epoch, SBP estimation models were created by evaluating the performance of each algorithm. A total of 11 groups were created with the ECG signal, and a performance evaluation table of 33 models was produced (Table 5). For the model constructed using all feature vectors, the EBT was the best performing algorithm with SBP values of  mmHg and  mmHg. There are 25 feature vectors for the 6-second epoched ECG signal. The best-performing columns of these feature vectors were taken at 5% intervals, and performance evaluation was made for 11 features. For the 11th feature group, the best performing algorithm with  mmHg and  mmHg SBP values were determined as EBT (Table 6). The same operations were performed for each epoched signal. The 11th feature vector per second was the best performing group. Accordingly,  mmHg and  mmHg SBP values were determined for the 8-second epoch with the GPR algorithm (Table 7). According to the 10-second epoched ECG signal,  mmHg and  mmHg SBP values were calculated using the EBT algorithm. Gradually improving performance values were calculated as  mmHg and  mmHg for the 12-second epoched ECG signal (Table 8). Considering the increasing epoching times, the algorithm’s best performance has varied each time. For the epoching times where the increase was observed, the GPR algorithm increased the value while it caused a decrease in the value. For the 14-second epoch, SBP were obtained as  mmHg and  mmHg (Table 10). For the 16-second epoch, SBP values were obtained as  mmHg and  mmHg (Table 11). For the 18-second epoch, SBP values were obtained as  mmHg and  mmHg (Table 12). For the 20-second epoch, SBP values were obtained as  mmHg and  mmHg (Table 13).

While the value did not change much before the 16th second, the value decreased. After the 16th second, the value decreases, while the value decreases. should be as low as possible and as high. According to all these, it is evaluated that using the ECG signal and the GPR algorithm, and BP detection can be performed in 16 seconds in the minimum desired time.

The tabular models defined for SBP apply to DBP. The tables for the whole process are modeled separately (Tables 1423). Models using all feature groups obtained the best performance value for each epoched period. Accordingly, DBP of  mmHg and  mmHg for 2 seconds were obtained using the EBT algorithm (Table 14). DBP values of  mmHg and  mmHg for the 4 seconds were obtained using the EBT algorithm (Table 15). DBP values of  mmHg and  mmHg for 6 seconds were obtained using the EBT algorithm (Table 16). DBP values of  mmHg and  mmHg for 8 seconds were obtained using the EBT algorithm (Table 17). DBP values of  mmHg and  mmHg for 10 seconds were obtained using the EBT algorithm (Table 18). DBP values of  mmHg and  mmHg for the 12 seconds were obtained using the GP algorithm (Table 19). DBP values of  mmHg and  mmHg for 14 seconds were obtained using the EBT algorithm (Table 20). DBP values of  mmHg and  mmHg for 16 seconds were obtained using the GPR algorithm (Table 21). DBP values  mmHg and  mmHg for the 18 seconds were obtained using the EBT algorithm (Table 22). DBP values of  mmHg and  mmHg for the 20 seconds were obtained using the EBT algorithm (Table 23).

The performance tables of the models created for DBP prediction should be carefully examined. For each increasing epoch time, both decreases and increases in were observed together, and the lowest value was obtained in Table 23. Accordingly, considering the value, the maximum value was obtained in Tables 20 and 21. It is desirable for to be as low as possible and high. According to all these, it is evaluated that DBP detection can be performed in 16 seconds using the ECG signal and the GPR algorithm in the minimum desired time. A 0.07 drop-in for 20 seconds is not very significant. For this reason, 16 seconds is considered suitable for time determination. Also, the summarized pattern table for the whole process is given in Table 24. In addition to all these, the GPR algorithm is evaluated to be appropriate for SBP and DBP time detection since it fits well with the ECG-based dataset.

Bland-Altman plots were prepared for the proposed models (Figure 4). It was determined that the resulting error rates of the models were close to zero, and the correlation values were (Table 23). The difference between the actual and predicted values is expected to be zero. Each scatter being close to zero indicates good performance.

The best findings obtained in the study were compared with literature studies (Table 23). The findings show that the proposed models are compatible with the literature and are a step forward in performance and low processing time.

4. Discussions

Since hypertension is known as a silent killer, patients with hypertension should be followed constantly [1, 3]. In order to keep hypertension under control, patients are expected to strictly comply with drug treatments [3]. There is a need for new technologies that can be used without the need for technical knowledge for continuous monitoring of hypertension at home [2, 3]. This study is aimed at calculating the ECG signal and machine learning-based minimum BP time response in hypertension patients. First, ECG signal data of five individuals were collected in the study. The signals were then divided into periods of 2-4-6-8-10-12-14-16-18-20 seconds. 25 statistical features in the time domain were extracted from each epoched signal. The feature selection algorithm is used to reduce the model’s unnecessary workload and provide size optimization. With EBT, GPR, and FT algorithms, selected features were used for BP estimation. BP estimation performance values for each epoched period were calculated within certain characteristics and recorded in tables (Tables 423). According to these tables, BP estimation can be evaluated at a minimum time. The model proposed in the study differs significantly from the studies in the literature in terms of time epoching. BP calculation was performed for each period. Calculated BP values exhibited a unique structure at each step and increased performance. After first determining the statistical inputs for the ECG signal, it was determined that the model’s performance was low against some features. On the other hand, unnecessary features were removed by optimizing the size. In this way, the performance increase was observed on datasets with high correlation levels. In this system, which tried to be improved gradually, periodic observations were made graphically, and minimum time detection was achieved with a high accuracy rate. The proposed model is among the algorithms with high accuracy obtained in the literature so far (Table 23, mmHg).

The most striking feature of the study is that it detects BP values for each period by epoching the ECG signal. Research on ECG, PPG, and BP properties is carried out extensively in the literature [31, 33, 78]. Among these signals, QRS and other components exhibited in a heartbeat in the ECG signal, which is the research subject of the study, are generally preferred as feature input [27, 33, 34]. Due to changes in the ECG signal, feature extraction can cause workload and computational complexity. On the other hand, statistical parameters were used, and BP estimation performance evaluation was made on epoched ECG signals. The best performance value that can be calculated in minimum time is (Table 24, 16 seconds). It is understood that BP can be detected with high accuracy for two seconds, and applications can be performed within this period. The study used the 16-second epoched signal model for BP estimation. This is because we can reduce the error rate and see the minimum and maximum points in an epoch. ECG and BP are similar signals. The averages of the BP signal’s local minimum points and local maximum points correspond to the DBP and SBP values, respectively. The multiplicity of these points reduces the error rate and increases the execution time. In this condition, the design was realized. In the literature studies, no research has yet been carried out to determine BP in a certain period. Although the article has an original research topic, it has the infrastructure to answer new research questions. Accordingly, with how many seconds of ECG recording can a highly accurate BP prediction model be developed? The study findings show that BP estimation can be made with an ECG signal with at least two-second epochs. However, the highest performance was achieved in the 16-second epoch, but the performance change is not appreciable. For this reason, it is considered that high performance can be obtained from two-second epochs with different signal processing processes.

The datasets are summarized significantly with the help of descriptive statistical parameters. In this study, the ECG signal with 25 feature sets was converted to a clearer dataset (Table 2). Studies in the literature are often based on specific characteristics [31, 33, 34]. In addition, temporal, chaotic, and morphological features are most well-liked in deep learning-based studies [20, 34]. However, studies based on different feature extraction algorithms seem to be insufficient in terms of performance (Table 25) [27, 29]. In this respect, it can be considered that the feature extraction algorithm used in the study has a better structure compared to the literature.

The feature selection process with classical machine learning methods is recommended. However, this process has not been applied in many BP estimation studies [28, 31, 33]. The feature selection algorithm performed better than the literature by optimizing the size of the model used and eliminating unnecessary features.

EBT, GPR, and FT algorithms in the proposed model have been an important element in determining model performance. The performance values of these models on each epoched signal show how compatible that algorithm is with the epoched dataset. This situation indicates that there is algorithm variation between epoched signals in Table 24. The proposed models are one step ahead of the studies in the literature in terms of performance (Table 25).

4.1. Strengths and Limitations

The use of machine learning algorithms in terms of continuous cuffless BP estimation, feature extraction, and evaluation processes that do not require calibration is still a matter of debate for accurate diagnosis and treatment. This study compares their performance using various ML techniques. Statistical feature extraction was applied from each epoched signal by applying the epoching process to the ECG signals. The dataset in which ML techniques were evaluated showed high performance for each epoch. Although different feature sets have been extracted for the model, the morphological features of the ECG signal have not been investigated. In addition, increasing the demographic information of the subjects can improve the BP prediction performance. Even if there are technical failures in the collection of the dataset, the operations applied to the signal in the software environment prevent this data loss, and its effect on performance is minimized.

5. Conclusions

Monitoring of BP is vital for the follow-up and treatment of hypertension. Algorithms and devices that comply with new reliable standards are needed for home BP monitoring. These technologies are expected to offer effortless measurement. This study developed artificial intelligence-based algorithms to monitor BP with ECG. The study’s main question is “What is the minimum time required for BP determination by ECG?” For this purpose, models have been developed for signals with different durations between 2 and 20 seconds, and algorithms have been tested.

It has been determined that BP estimation can be made with high accuracy for any time including 2 seconds. While for 14- and 16-second epochs, it was determined as for 2-second epoch. It varies in the range of 0.98 in other periods. This sitution indicates that BP can be detected with high accuracy for ECG.

The innovations included in the findings of this study are as follows. (1) With ECG, BP can be predicted accurately. (2) Shortening the ECG signal time does not affect the success rate. (3) BP can be estimated with a 2-second epoch. (4) Feature extraction and selection processes improved model performance. (5) Artificial intelligence-based models have increased system reliability. Due to the high performance and reliability of the proposed model, it is considered that it can be used as an auxiliary software for BP monitoring in all systems that can measure ECG signals.

6. Future Work

With these encouraging results, future studies are planned. The scope of the study can be expanded by using a dataset containing many diverse groups of people and more specific ML techniques. Increasing the sampling frequency of the signals will prevent information loss. For this reason, a sampling frequency of at least 2.5 times the signal frequency range is recommended for new studies. In addition, studies can be continued by looking at the relationship between ECG and BP from another perspective by considering the morphological features of the ECG signal and extracting the frequency domain feature groups on the dataset.

Data Availability

We can send the datasets at the request of the authors. The data was obtained from the IEEE Dataport open source site.

Ethical Approval

This article does not contain any studies with human participants. No animal studies were involved in this review.

Conflicts of Interest

The authors declare no conflict of interest.

Authors’ Contributions

All authors contributed equally to this work. In addition, all authors have read and approved the final manuscript and given their consent to publish the article.