Smart Wearables in Healthcare: Signal Processing, Device Development, and Clinical Applications
View this Special IssueResearch Article  Open Access
Xiaoyan Xu, Shoushui Wei, Caiyun Ma, Kan Luo, Li Zhang, Chengyu Liu, "Atrial Fibrillation Beat Identification Using the Combination of Modified Frequency Slice Wavelet Transform and Convolutional Neural Networks", Journal of Healthcare Engineering, vol. 2018, Article ID 2102918, 8 pages, 2018. https://doi.org/10.1155/2018/2102918
Atrial Fibrillation Beat Identification Using the Combination of Modified Frequency Slice Wavelet Transform and Convolutional Neural Networks
Abstract
Atrial fibrillation (AF) is a serious cardiovascular disease with the phenomenon of beating irregularly. It is the major cause of variety of heart diseases, such as myocardial infarction. Automatic AF beat detection is still a challenging task which needs further exploration. A new framework, which combines modified frequency slice wavelet transform (MFSWT) and convolutional neural networks (CNNs), was proposed for automatic AF beat identification. MFSWT was used to transform 1 s electrocardiogram (ECG) segments to timefrequency images, and then, the images were fed into a 12layer CNN for feature extraction and AF/nonAF beat classification. The results on the MITBIH Atrial Fibrillation Database showed that a mean accuracy (Acc) of 81.07% from 5fold cross validation is achieved for the test data. The corresponding sensitivity (Se), specificity (Sp), and the area under the ROC curve (AUC) results are 74.96%, 86.41%, and 0.88, respectively. When excluding an extremely poor signal quality ECG recording in the test data, a mean Acc of 84.85% is achieved, with the corresponding Se, Sp, and AUC values of 79.05%, 89.99%, and 0.92. This study indicates that it is possible to accurately identify AF or nonAF ECGs from a shortterm signal episode.
1. Introduction
Atrial fibrillation (AF) is the most common type of arrhythmia in clinical disease and gradually becomes the world’s rising healthcare burden [1]. According to Framingham heart study, lifetime risk of AF is about 25% [2]. The disease shows that the atrial activity is irregular, and the resulting complications such as stroke and myocardial infarction (MI) [3], endanger the health and lives of humans seriously [4]. Therefore, developing automatic AF detection algorithm is of great clinical and social significance [5, 6].
Generally, AF is significantly different from normal heart rhythm on electrocardiogram (ECG) signals [7]. During AF, RR interval is absolutely irregular and the Pwave is replaced by the continuous irregular Fwave, which is an important feature of AF [8]. Many scholars proposed diverse methods based on RR interval feature, but the accuracy of AF is not sufficient due to the complication of the ECG signals [9], and the pattern recognition ability of the existing statistical and general method is not satisfactory owing to a variety of noise interference [10].
In recent years, AF detection algorithms based on the time domain characteristics have been developed rapidly. Chen et al. [11] developed a multiscale wavelet entropybased method for paroxysmal atrial fibrillation (PAF) recognition. In their work, recognition and prediction used support vector machine (SVM) based method. Fifty recordings from the MITBIH PAF Prediction Database were chosen to test the proposed algorithm, with an average sensitivity of 86.16% and average specificity of 89.68%. Maji et al. [12] used empirical mode decomposition (EMD) to extract Pwave mode components and corresponding parameters to determine the occurrence of AF. This proposed algorithm was tested with a total of 110 cycles of normal rhythm and 68 cycles of AF rhythm from the MITBIH AF Database. An average sensitivity of 92.89% and an average specificity of 90.48% were achieved. Ladavich and Ghoraani [13] constructed the Gaussian mixture model of the Pwave feature space. The model was then used to detect AF, with an average sensitivity of 88.87% and average specificity of 90.67%, while the positive predictive value was only 64.99%. Although relatively fine detection performances were achieved by the aforementioned methods, problems and questions exist. First, different methods used different signal length for AF identification. How will be the accuracy if performed on a very shortterm (such as 1 s) ECG segment? This can show the ability for transient AF detection. In addition, ECG waveforms have various morphology and the abnormal waveforms are different when AF occurs, leading to poor generalization capability of the developed machine learningbased model. Thus, how to improve the model generalization capability is a key issue.
Convolutional neural networks (CNNs) [14] can extract features automatically without manual intervention and expert priori knowledge. Meanwhile, timefrequency (TF) technology as a preprocess operation is to convert 1D ECG signals to 2D TF features which can be used to transfer to a classifier. There are many common TF methods at present, such as the shorttime Fourier transform (STFT), the Wigner–Ville distribution (WVD), and the continuous wavelet transform (CWT) [15]. Luo et al. [16] presented a modified frequency slice wavelet transform (MFSWT) in 2017. MFSWT follows the rules of producing TF representation and contains the information of ECG signals in both time and frequency domains, such as Pwave, QRSwave, and Twave. Additionally, MFSWT can locate the above characters accurately and avoid the complexity of setting parameters. The spectrogram of MFSWT can be expressed as images, while the combination of CNN and images is one of the most excellent choices. For example, Liu et al. [17] proposed the method to learn conditional random fields (CRFs) using structured SVM (SSVM) based on features learned from a pretrained deep CNN for image segmentation. Ravanbakhsh et al. [18] introduced a feature representation for videos that outperforms stateoftheart methods on several datasets for action recognition. Lee and Kwon [19] built a fully connected CNN, which trained on relatively sparse training samples, and a newly introduced learning approach called residual learning for hyperspectral image (HSI) classification.
In this study, MFSWT was adopted to acquire TF images for shortterm AF and nonAF ECG segments from the MITBIH Atrial Fibrillation Database (MITBIH AFDB). A deep CNN with a total of 12 layers was developed to train an AF/nonAF classification model. Indices including accuracy, sensitivity, specificity, and the area under the curve were used for model evaluation based on a 5fold cross validation method [20], to evaluate the stability and generalization ability of the proposed method in comparison with the existing methods. The existing research has achieved very good performance, but there is no validation for large data. In this paper, we used all the data in the database to increase the generalization ability of the model.
2. Methods
2.1. Modified Frequency Slice Wavelet Transform
In our previous work, modified frequency slice wavelet transform (MFSWT) [16] was proposed for heartbeat timefrequency spectrum generation, with following the major principle of frequency slice wavelet transform (FSWT) [21]. The modified transform generates TF representation from the frequency domain, and a bound signaladaptive frequency slice function (FSF) was introduced to serve as a dynamic frequency filter. The window size of FSF smoothly changes with energy frequency distribution of signal in lowfrequency area. The MFSWT has good performance for lowfrequency ECG signals, and its advantages include signaladaptive, accurate timefrequency component locating. The reconstruction is also independent of FSF, and it is readily accepted by clinicians.
Assume is the Fourier transform of . The MFSWT can be defined as follows:where and are observed time and frequency, respectively, “” represents conjugation operator, and is the frequency slice function (FSF): is defined as a scale function of ,
It makes the transform to incorporate signaladaptive property. In (3), corresponds to maximum . is a differential operator, and means signum function, which returns 1 if the input is greater than zero, 0 if it is zero, or −1 if it is less than zero. In (2), , according to the claim in [21], then the original signal can be reconstructed as follows:
Figure 1 shows 4 s normal ECG, atrial fibrillation signals, and their corresponding MFSWT spectra, respectively, from 06426 recording. By the MFSWT, the time domain characteristics in ECG signal wave, such as Pwave, QRSwave, and Twave, have accurately been located in the signal spectrum. At the same time, each component of the spectrum of the TF space distribution is corresponded well with the ECG signal frequency before.
(a)
(b)
(c)
(d)
In this study, the MFSWT is used as a tool to generate spectrograms of an ECG signal for CNNbased classification. The 1 s window, centered at the detected Rpeaks (0.4 s before and 0.6 s after), was used to segment each heartbeat. Subsequently, the TF spectrograms with the size of 250 90 (corresponded 1 s time interval and 0–90 Hz frequency range) were produced by the proposed MFSWT. This is then followed by data reducing. An average 5 2 template operator reduces the size of spectrograms to 50 45.
2.2. Convolutional Neural Networks (CNNs)
Deep CNN was improved by Lecun et al. [22]. CNN had breakthrough performance over the last few decades for solving pattern recognition problems [23], especially in image classification [24]. It has become a popular method for feature extraction and classification without requiring preprocessing and pretraining algorithm [25].
CNN is a composition of sequences of functions or layers that maps an input vector to an output vector. The input is expressed as follows:
Similarly, and are the bias and kernel of the th neuron at layer , respectively. is the output of the th neuron at layer , and means a regular 2D convolution without zero padding on the boundaries. So, the output can be described as follows:
Besides, CNN also involves backpropagation (BP), in order to adjust the delta error of the th neuron at layer . Assuming that the corresponding output vector of the input is , and its ground truth class vector is , we can write the mean absolute error (MSE) as follows:
Thus, the delta error can be concluded as
The implementation of CNN [14] is as shown in Figure 2.
By the MFSWT, we have converted the signals to characteristic waves in a 2D space. Then, we use CNN to learn relevant information from the characteristic waves in a 2D space and achieve classification. The input to the CNN is characteristic waves in a 2D space computed from the exacted signals. The CNN was implemented using the Neural Network Toolbox in Matlab R2017a.
In this paper, we use CNN to automatically extract the features of the labeled image and calculate the scores to classify the predicted image. A 12layer network structure is developed, which contains 3 convolution layers, 3 ReLU layers, a max pooling layer, and 3 fullconnection layers besides the input and output layers. We tested the effect of number of filters in each layer and obtained these values by running a grid search approach. Figure 3 illustrates the architecture of the implemented network and its detailed components for each layer.
3. Experiment Design
3.1. Database
The database was from the MITBIH AFDB [26]. The MITBIH AFDB contains a large number of ECG data that have been annotated by a professional cardiologist, which is the authoritative ECG database in the classification of arrhythmia. This database may be useful for development and evaluation of atrial fibrillation/flutter detectors that rely on timing information only. It consists of 25 recordings (obtained from ambulatory ECG recordings of 25 subjects). The individual recordings are each 10 hours 15 min in duration and contain two ECG signals; each sampled at 250 samples per second with 12bit resolution over a range of 10 millivolts. The reference manual annotation files contain rhythm change annotations (with the suffix.atr) [27] and the rhythm annotations of types: AF, AFL (atrial flutter), J (AV junctional rhythm), and N (used to indicate all other rhythms) [28]. In our experiment, AFL, J, and N are attributed to nonAF category. Figure 4 shows the ECG examples (each 4 s) of the four rhythm types (normal, AF, atrial flutter, and AV junction) from the “06426” recording.
(a)
(b)
(c)
(d)
3.2. Signal Preprocessing
There are a total of 25 recordings in AFDB, and two of them (numbers 00735 and 03665) have no relevant ECG data. Thus, 23 recordings are included in the experiment. In order to split the dataset equally, we divide the 23 recordings into five groups, and the basis for grouping is to reduce the differences of number in the two classes. The recording numbers for 5 groups are 5, 4, 5, 5, 4, subsequently. This experiment uses 5fold cross validation for evaluation. For example, when using the first group to test, it means that the whole data in 04015, 04126, 04936, 07879, and 08405 recordings are used to verify, and the remaining 17 records are all used to train the model. Detailed recordings of grouping conditions are shown in Table 1.

We employ a balanced image dataset to train the model. That is, we choose the same number of AF samples from the nonAF category for training, while all the samples in the remaining fold for test. For example, for testing the fold 1, there are 415,109 normal images and 294,136 AF images as training data from the folds 2–5. We use all the 294,136 AF images and then randomly select 294,136 normal images, resulting in 588,272 images as training CNN model. Then, we test the performance of the developed CNN model using all data in fold 1, that is, 123,083 normal images and 90,403 AF images. Table 2 presents the detailed numbers for each fold testing.

4. Results
4.1. Epoch Number of the CNN
The grid search method [29] is applied to select the optimal epoch number of the CNN. Figure 5 shows the AUC of the test set (AUC is defined as the area under the ROC curve, often used to evaluate the classifier with imbalance data) at varying epoch number, and Figure 6 shows training and test accuracies at varying epoch number. We can see from Figure 5 that the AUC of test set is at a high level while epoch number is 15, and the wave becomes stable in Figure 6. Therefore, we choose the number of epochs as 15.
According to the introduction of Section 2.2 on CNN architecture, the input layer (layer 0) is for images with the size of 50 45 1 and is convolved with a kernel size of 10 9 to produce the layer 1. Layer 2 is the ReLU layer. The output of layer 2 is convolved with a kernel size of 8 7 to develop the layer 3 and layer 4. Similarly, a subsequent feature maps are convolved with a kernel size of 9 to acquire the layer 5 and layer 6. A max pool layer with the stride of 2 (layer 7) is applied to the generated characteristics. Then, the feature has been extracted. Finally, features are transported to layer 8 with 10 neurons and connected to 5 and 2 neurons in layer 9 and layer 10, respectively, to classify. In addition, we select the epoch number as 15, and the learning rate initial value is set to 0.001 while the number of minimal batches is 256. In addition, specific parameter is shown in Table 3.

4.2. Performance Metrics
This research uses the MITBIH AFDB to verify the proposed method to detect AF performance. Four widely used metrics, that is, sensitivity (Se), specificity (Sp), accuracy (Acc), and the area under the curve (AUC), were used (and defined below) for assessment of classification performance. In addition, AUC and Acc refer to the overall system performance, while the remaining indexes are specific to each class, and they measure the generalization ability of the classification algorithm to differentiate events.
Moreover, the Acc includes both test set accuracy and training set accuracy. According to the attribute of the label (positive or negative), the result can generate four basic indexes: true positive (TP), false positive (FP), true negative (TN), and false negative (FN), and in this case, Acc is the radio of the number of correct predicted labels and total number of the labels, thus . Se is the true positive rate and is the probability of incorrectly diagnosing into positive among all positive patients, so . Sp is proportion of incorrectly diagnosing into negative among all negative patients, so . ROC curve is based on a series of different ways of binary classification (boundary value or decision threshold), with true positive rate (Se) as the ordinate, the false positive rate (1−Sp) as the abscissa, and AUC is defined as the area under the ROC curve, often used to evaluate the classifier with imbalance data. Each fold is tested by a specific classifier with the same parameters as shown in Section 4.1; besides, we also selected the average and standard deviation (SD) of the experimental results to be evaluated, and the results are summarized in Table 4.
 
^{#}The results only from the average of the folds 1, 2, 3, and 5. 
From Table 4, a mean Acc of 81.07% from 5fold cross validation is achieved for the test data. The corresponding Se, Sp, and AUC results are 74.96%, 86.41%, and 0.88. It is worth to note that the results from the fourth fold are low. This is because there is an extremely poor signal quality ECG recording in the fourth fold divided as shown in Table 1, which has significantly different timefrequency features compared to the clean ECG signals. So the results from the folds 1, 2, 3, and 5 are recalculated as shown in Table 4 to exclude the lowquality signal effect. Herein, a mean Acc of 84.85% is achieved for the test data, with the corresponding Se, Sp, and AUC values of 79.05%, 89.99%, and 0.92.
5. Discussion and Conclusion
An ECG is widely used in medicine to monitor small electrical changes on the skin of a patient’s body arising from the activities of the human heart. Due to the variability and difficulty of AF, traditional detection algorithm cannot be extracted to distinguish obvious characteristics accurately.
In our work, we present a unique architecture of CNN to distinguish AF beats from all other types of ECG beats. MFSWT is adopted to acquire the TF images of AF and nonAF, respectively, and then, we divide all the data in the MITBIH AFDB into training set and test set different from the existing models and build a deep CNN with a total of 12 layers to extract the characters of training set. Finally, the test set is evaluated by the trained model to obtain the performance indexes (including Acc, Se, Sp, and AUC).
Compared with other studies, the difference is that we use all ECG recordings in the MITBIH AFDB. However, other studies only selected a part of the recordings for training and testing, and Table 5 shows the comparison between our study and other studies. Obviously, the proposed method does not improve the Acc, Se, and Sp significantly, but this paper uses lots of data to train the model in order to improve the generalization ability of the model. Moreover, we use an important indicator called AUC to evaluate the unbalanced data model and obtained a good evaluation standard.

In short, we proposed a protocol for AF beat detection as follows. (1) Use all the recordings of MITBIH Atrial Fibrillation Database for algorithm development and validation. (2) Use 5fold cross validation to examine the algorithm performance. The results have been registered for Acc, Se, Sp, and AUC. Group the folds by recordings rather than heartbeats to prevent heartbeats of the same patient from appearing in both training and test sets. (3) Use a separate database, for instance AF Database, as an independent test to evaluate the generalization ability of the algorithm. We believe that accurate AF beat recognition can facilitate the detection of AF rhythm. As the first step, AF beat is particularly important. Only by improving the accuracy and generalization of AF beat detection can we more effectively implement AF surveillance.
In addition, more data could be used to evaluate the proposed method; for example, we only focus on one ECG lead, and the study can be extended to two ECG leads. We also can try to use more databases for verification. This algorithm can be used for monitoring and prevention of AF, which has great practical meaning.
Data Availability
The data used to support the findings of this study are available from the openaccess MITBIH Atrial Fibrillation Database.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
The project was partly supported by the National Natural Science Foundation of China (61571113 and 61671275), the Key Research and Development Programs of Jiangsu Province (BE2017735), the Natural Science Foundation of Shandong Province in China (2014ZRE2733), and the Fundamental Research Funds for the Central Universities in Southeast University (2242018k1G010). The authors thank the support from the SoutheastLenovo Wearable Intelligent Lab.
References
 J. J. Rieta, F. Ravelli, and L. Sörnmo, “Advances in modeling and characterization of atrial arrhythmias,” Biomedical Signal Processing and Control, vol. 8, no. 6, pp. 956957, 2013. View at: Publisher Site  Google Scholar
 D. M. Lloydjones, “Lifetime risk for development of atrial fibrillation: the Framingham Heart Study,” Circulation, vol. 110, no. 9, pp. 1042–1046, 2004. View at: Google Scholar
 E. Z. Soliman, F. Lopez, W. T. O’Neal et al., “Atrial fibrillation and risk of STsegment elevation versus nonST segment elevation myocardial infarction: the atherosclerosis risk in communities (ARIC) study,” Circulation, vol. 131, no. 21, pp. 1843–1850, 2015. View at: Publisher Site  Google Scholar
 P. Langley, M. Dewhurst, L. Y. Di Marco et al., “Accuracy of algorithms for detection of atrial fibrillation from short duration beat interval recordings,” Medical Engineering and Physics, vol. 34, no. 10, pp. 1441–1447, 2012. View at: Publisher Site  Google Scholar
 C. Bruser, J. Diesel, M. D. H. Zink, S. Winter, P. Schauerte, and S. Leonhardt, “Automatic detection of atrial fibrillation in cardiac vibration signals,” IEEE Journal of Biomedical and Health Informatics, vol. 17, no. 1, pp. 162–171, 2013. View at: Publisher Site  Google Scholar
 G. Y. H. Lip, C. M. Brechin, and D. A. Lane, “The global burden of atrial fibrillation and stroke: a systematic review of the epidemiology of atrial fibrillation in regions outside North America and Europe,” Chest, vol. 142, no. 6, pp. 1489–1498, 2012. View at: Publisher Site  Google Scholar
 D. T. Linker, “Accurate, automated detection of atrial fibrillation in ambulatory recordings,” Cardiovascular Engineering and Technology, vol. 7, no. 2, pp. 182–189, 2016. View at: Publisher Site  Google Scholar
 Y. Li, X. Tang, A. Wang, and H. Tang, “Probability density distribution of delta RR intervals: a novel method for the detection of atrial fibrillation,” Australasian Physical and Engineering Sciences in Medicine, vol. 40, no. 3, pp. 707–716, 2017. View at: Publisher Site  Google Scholar
 A. Petrenas, L. Sornmo, V. Marozas, and A. Lukosevicius, “A noiseadaptive method for detection of brief episodes of paroxysmal atrial fibrillation,” in Proceedings of Computing in Cardiology Conference, pp. 739–742, Zaragoza, Spain, September 2013. View at: Google Scholar
 N. Larburu, T. Lopetegi, and I. Romero, “Comparative study of algorithms for atrial fibrillation detection,” in Proceedings of Computing in Cardiology Conference, pp. 265–268, Hangzhou, China, September 2011. View at: Google Scholar
 Y. Chen, Z. Wang, and Q. Li, “Multiscale wavelet entropy based method for paroxysmal atrial fibrillation recognition,” Space Medicine and Medical Engineering, vol. 26, no. 5, pp. 352–355, 2013. View at: Google Scholar
 U. Maji, M. Mitra, and S. Pal, “Automatic detection of atrial fibrillation using empirical mode decomposition and statistical approach,” Procedia Technology, vol. 10, no. 1, pp. 45–52, 2013. View at: Publisher Site  Google Scholar
 S. Ladavich and B. Ghoraani, “Rateindependent detection of atrial fibrillation by statistical modeling of atrial activity,” Biomedical Signal Processing and Control, vol. 18, pp. 274–281, 2015. View at: Publisher Site  Google Scholar
 S. Kiranyaz, T. Ince, and M. Gabbouj, “Realtime patientspecific ECG classification by 1D convolutional neural networks,” IEEE Transactions on Biomedical Engineering, vol. 63, no. 3, pp. 664–675, 2016. View at: Publisher Site  Google Scholar
 Z. Yan, A. Miyamoto, Z. Jiang, and X. Liu, “An overall theoretical description of frequency slice wavelet transform,” Mechanical Systems and Signal Processing, vol. 24, no. 2, pp. 491–507, 2010. View at: Publisher Site  Google Scholar
 K. Luo, J. Li, Z. Wang, and A. Cuschieri, “Patientspecific deep architectural model for ECG classification,” Journal of Healthcare Engineering, vol. 2017, Article ID 4108720, 13 pages, 2017. View at: Publisher Site  Google Scholar
 F. Liu, G. Lin, and C. Shen, CRF Learning with CNN Features for Image Segmentation, Elsevier Science Inc., Amsterdam, Netherlands, 2015.
 M. Ravanbakhsh, H. Mousavi, M. Rastegari, V. Murino, and L. S. Davis, “Action recognition with image based CNN features,” http://arxiv.org/abs/1512.03980. View at: Google Scholar
 H. Lee and H. Kwon, “Going deeper with contextual CNN for hyperspectral image classification,” IEEE Transactions on Image Processing, vol. 26, no. 10, pp. 4843–4855, 2017. View at: Publisher Site  Google Scholar
 R. Kohavi, “A study of crossvalidation and bootstrap for accuracy estimation and model selection,” in Proceedings of International Joint Conference on Artificial Intelligence, pp. 1137–1143, Montreal, QC, Canada, August 1995. View at: Google Scholar
 Z. Yan, A. Miyamoto, and Z. Jiang, “Frequency slice algorithm for modal signal separation and damping identification,” Computers and Structures, vol. 89, no. 12, pp. 14–26, 2011. View at: Publisher Site  Google Scholar
 Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, “Gradientbased learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, 1998. View at: Publisher Site  Google Scholar
 X. Jiang, Y. Pang, X. Li, and J. Pan, “Speed up deep neural network based pedestrian detection by sharing features across multiscale models,” Neurocomputing, vol. 185, pp. 163–170, 2016. View at: Publisher Site  Google Scholar
 A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” in Proceedings of International Conference on Neural Information Processing Systems, pp. 1097–1105, Lake Tahoe, NV, USA, December 2012. View at: Google Scholar
 S. Chaib, H. Yao, Y. Gu, and M. Amrani, “Deep feature extraction and combination for remote sensing image classification based on pretrained CNN models,” in International Conference on Digital Image Processing, p. 104203D, Prague, Czech, July 2017. View at: Google Scholar
 K. Jiang, C. Huang, S. Ye, and H. Chen, “High accuracy in automatic detection of atrial fibrillation for Holter monitoring,” Journal of Zhejiang University, vol. 13, no. 9, pp. 751–756, 2012. View at: Publisher Site  Google Scholar
 F. D. Murgatroyd, B. Xie, X. Copie, I. Blankoff, A. J. Camm, and M. Malik, “Identification of atrial fibrillation episodes in ambulatory electrocardiographic recordings: validation of a method for obtaining labeled RR interval files,” Pacing and Clinical Electrophysiology Pace, vol. 18, no. 6, p. 1315, 1995. View at: Publisher Site  Google Scholar
 A. Petrėnas, L. Sörnmo, A. Lukoševičius, and V. Marozas, “Detection of occult paroxysmal atrial fibrillation,” Medical and Biological Engineering and Computing, vol. 53, no. 4, pp. 287–297, 2015. View at: Publisher Site  Google Scholar
 J. A. A. Brito, F. E. Mcneill, C. E. Webber, and D. R. Chettle, “Grid search: an innovative method for the estimation of the rates of lead exchange between body compartments,” Journal of Environmental Monitoring, vol. 7, no. 3, p. 241, 2005. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2018 Xiaoyan Xu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.