Abstract

Blood centers are an essential component of the healthcare system, as timely blood collection, processing, and efficient blood dispatch are critical to the treatment of patients and the performance of the entire healthcare system. At the same time, an efficient blood dispatching system through the high-precision predictive capability of artificial intelligence is crucial for the efficiency improvement of the blood centers. However, the current artificial intelligence (AI) models for predicting blood usage do not meet the needs of blood centers. The challenges of AI models mainly include lower generalization ability in different hospitals, limited stability under missing values, and low interpretability. An artificial neural network-based model named the blood usage prediction neural network (BUPNN) has been developed to address these challenges. BUPNN includes a novel similarity-based manifold regularizer that aims to enhance network mapping consistency and, thus, overcome the domain bise of different hospitals. Moreover, BUPNN diminishes the performance degradation caused by missing values through data enhancement. Experimental results on a large amount of accurate data demonstrate that BUPNN outperforms the baseline method in classification and regression tasks and excels in generalization and consistency. Moreover, BUPNN has solid potential to be interpreted. Therefore, the decision-making process of BUPNN is explored to the extent that it acts as an aid to the experts in the blood center.

1. Introduction

Blood products are an essential part of the treatment of bleeding, cancer, AIDS, hepatitis, and other diseases[1]. Blood is also an indispensable resource in treating injured patients; whether promptly transfusing the blood is critical for rehabilitating injured patients. At the same time, early surgical interventions and rapid blood transfusions are the primary measures to reduce mortality. Unfortunately, these measures require large amounts of blood to support them.

Blood, however, is significantly different from other medical products. Currently, blood cannot be manufactured or synthesized artificially and can only be donated by others. In addition, blood has a short shelf life, making emergency blood more specific and irreplaceable than other medical products. For example, patients in mass casualty events caused by earthquakes suffer from fractures, fractures accompanied by multiple organ injuries, and crush injuries. Therefore, the peak period of blood consumption occurs 96 hours after the earthquake [2]. In contrast, patients mainly suffer from burns in mass casualties caused by bombings or fires. The peak blood consumption occurs 24 hours after the event. In addition, ongoing blood transfusions may last for several months [3]. Therefore, modeling the prediction of blood use in patients is a meaningful and challenging topic.

Current blood usage prediction models are not well developed. For example, Wang et al. [4] developed an early transfusion scoring system to predict blood requirements in severe trauma patients in the prehospital or initial phase of emergency resuscitation. However, the abovementioned design is more suitable for triage in a single hospital than in blood centers since the system does not consider how to avoid performance degradation due to domain deviations between different hospitals. Rebecca et al. [5, 6] summarized the currently available studies exploring how to predict the need for massive transfusion in patients with traumatic injuries, listing the blood consumption scoring system (ABC) and the shock index scoring system (SI). These systems use classical or ML-based methods to predict blood consumption during the treatment of patients. Unfortunately, although the progress is remarkable, but it is unsatisfactory in terms of accuracy and generalization, and thus, the blood demand prediction model cannot be widely used. We summarize the problems into the following three points:(i)Data quality is limited. The partnership between blood centers and hospitals makes it very difficult to establish a rigorous feedback system for patient information. In addition, biases and missing data due to differences in hospital equipment create challenges for a constant blood use prediction model.(ii)Model generalizability is unsatisfactory. Blood usage prediction models are often built for specific hospitals or cities without considering extending to a wider range of applications and thus have poor generalization performance.(iii)Model interpretability is inadequate. Most models can only output a category or blood usage forecast but cannot demonstrate the model’s decision process. An interpretable model can better work with experts to help blood centers with blood schedules.

This paper proposes a blood usage prediction neural network, BUPNN, to solve the above problems. First, actual patient clinic data and treatment procedures in 12 hospitals are used as training data. Extensively collected data with biases from various hospitals will provide sufficient information to train a high-performance model. We used multiple data complementation schemes to restore the real problem and overcome missing values. In addition, the BUPNN model MDC is augmented with online data by linear interpolation, which increases the diversity of training data and thus improves the stability of the model under training with missing data. Second, to further improve the generalization performance of BUPNN, a similarity-based loss function is introduced to map data with biases to a stable semantic space by aligning samples from different hospitals in the latent space. Third, we analyze the model based on the deep learning interpretation method to enhance interpretability. The proposed analysis is accompanied by the prediction output of the model in real-time to assist one in understanding the process of the model’s decision-making. The interpretable study of BUPNN provides the conditions for computers and experts to help each other in blood consumption prediction.

The main contributions of this study are as follows:(i)Representative samples. A large amount of data from twelve hospitals is collected for this study to investigate the implied relationship between patients’ chain indicators and blood consumption.(ii)Generalizable model. This study designs MDC for online missing data completion, and thus, data augmentation to enhance the model’s generalization ability. In addition, the various similarity loss function is designed to improve the model’s predictive power across domains.(iii)Excellent Performance. Experiments on six different settings demonstrate that our method outperforms all baseline methods.

The rest of this paper is organized as follows. In Section 2, we provide a literature review of demand forecasting methods for blood products. Section 3 provides an initial exploration of the data. We provide the data description, model background, model development, and evaluation of four different models for blood demand forecasting in Section. In Section 4, a comparison of the models is provided, and finally, in Section 5, concluding remarks are provided, including a discussion of ongoing work for this problem.

2.1. ML Techniques for Medical Problems

The integration of the medical field with ML technology has received much attention in recent years. Two areas that may benefit from the application of ML technology in the medical field are diagnosis and outcome prediction. It includes the possibility of identifying high-risk medical emergencies, such as recurrence or transition to another disease state. Recently, ML algorithms have been successfully used to classify thyroid cancer [7] and predict the progression of COVID-19 [8, 9]. On the other hand, ML-based visualization and dimensionality reduction techniques have the potential to help professionals analyze biological or medical data, guiding them to better understand the data [10, 11]. Furthermore, ML-based feature selection techniques [12, 13] have strong interpretability and the potential to find highly relevant biomarkers for output in a wide range of medical data, leading to new biological or medical discoveries.

2.2. Blood Demand Forecasting

There is limited literature on blood demand forecasting; most investigate univariate time series methods. In these studies, forecasts are based solely on previous demand values without considering other factors affecting the demand. Frankfurter et al. [14] developed transfusion forecasting models using exponential smoothing (ES) methods for a blood collection and distribution center in New York. Critchfield et al. [15] developed models for forecasting blood usage in a blood center using several time series methods, including moving average (MA), winter’s approach, and ES. Filho et al. [16] developed a Box-Jenkins seasonal autoregressive integrated moving average (BJ-SARIMA) model to forecast weekly demand for hospital blood components. Their proposed method, SARIMA, is based on a Box-Jenkins approach that considers seasonal and nonseasonal characteristics of time series data. Later, Filho et al. [17] extended their model by developing an automatic procedure for demand forecasting and changing the model level from hospital level to regional blood center to help managers use the model directly. Kumari and Wijayanayake [18] proposed a blood inventory management model for daily blood supply, focusing on reducing blood shortages. Three time series methods, namely, MA, weighted moving average (WMA), and ES, are used to forecast blood usage and are evaluated based on needs. Fortsch and Khapalova [19] tested various blood demand prediction approaches, such as naive, moving average, exponential smoothing, and multiplicative time series decomposition (TSD). The results show that the Box-Jenkins (ARMA) approach, which uses an autoregressive moving average model, results in the highest prediction accuracy. Lestari et al. [20] applied four models to predict blood component demand, including moving average, weighted moving average, exponential smoothing, exponential smoothing with the trend, and select the best method for their data based on the minimum error between forecasts and the actual values. Volken et al. [21] used generalized additive regression and time-series models with exponential smoothing to predict future whole blood donation and RBC transfusion trends.

Several recent studies consider clinically related indicators. For example, Drackley et al. [22] estimated a long-term blood demand for Ontario, Canada, based on previous transfusions’ age and sex-specific patterns. They forecast blood supply and demand for Ontario by considering demand and supply patterns and demographic forecasts, assuming fixed patterns and rates over time. Khaldi et al. [23] applied artificial neural networks (ANNs) to forecast the monthly demand for three blood components: red blood cells (RBCs), blood, and plasma, for a case study in Morocco. Guan et al. [24] proposed an optimisation ordering strategy in which they forecast the blood demand for several days into the future and build an optimal ordering policy based on the predicted direction, concentrating on minimising the wastage. Their primary focus is on an optimal ordering policy. They integrate their demand model into the inventory management problem, meaning they do not precisely try to forecast blood demand. Li et al. [25] developed a hybrid model consisting of seasonal and trend decomposition using Loess (STL) time series and eXtreme Gradient Boosting (XGBoost) for RBC demand forecasting and incorporated it into an inventory management problem.

Recently, Motamedi et al. [26] presented an efficient forecasting model for platelet demand at Canadian Blood Services (CBS). In addition, C. Twumasi and J. Twumasi [27] compared k-nearest neighbour regression (KNN), multilayer perceptron (MLP), and support vector machine (SVM) via a rolling-origin strategy for forecasting and backcasting blood demand data with missing values and outliers from a government hospital in Ghana. Abolghasemi et al. [28] treat the blood supply problem as an optimisation problem [29] and find that LightGBM provides promising solutions and outperforms other machine learning models.

3. Blood Centers and Datasets

This section describes how blood center works with an example in Zhejiang Province, China. It includes the responsibilities of the blood center and the partnership between the blood center and the hospital, and shows in detail how the proposed model will assist the blood center in accomplishing its mission better.

3.1. Blood Center

Figure 1 shows that, this study considers a provincial centralized blood supply system, including blood centers, blood stations, and hospitals. The entire centralized blood supply system completes the collection, management, and delivery of blood products. The blood center is responsible for collecting blood products, testing for viruses and bacteria, and supplying some hospitals. At the same time, blood centers assume the management, coordination, and operational guidance of blood collection and supply institutions. Blood stations are responsible for collecting, storing, and transporting blood to local hospitals. While receiving blood, hospitals must collect clinical information on current patients for analysis and decision-making at the blood center to make the blood supply more efficient. Our proposed AI blood consumption prediction system (BUPNN) receives clinical information from each hospital and uses it to predict the future blood consumption of each patient. The proposed system helps blood center specialists perform blood collection and transportation better.

3.2. Data Details and Challenges

We collected data in their actual state to build a practically usable model. The data in this study are constructed by processing BS shipping data and the TRUST (Transfusion Research for Utilization, Surveillance, and Tracking) database from Zhejiang Province Blood Center and 12 hospitals in Zhejiang Province. The data include data from 2025 patients, including 1970 data from emergency trauma patients in 10 hospitals in Zhejiang Province from 2018 to 2020 and another 55 from patients in two hospitals in Wenling’s explosion in 2020.

Each dataset mainly included the following parts. (1) General patient information, including case number, consultation time, pretest classification, injury time, gender, age, weight, diagnosis, penetrating injury, heart rate, diastolic blood pressure, systolic blood pressure, body temperature, shock index, and Glasgow coma index. (2) Injury status, including pleural effusion, abdominal effusion, extremity injury status, thoracic and abdominal injury status, and pelvic injury status. (3) Laboratory tests, including hemoglobin, erythrocyte pressure product, albumin, hemoglobin value 24 hours after transfusion therapy, base residual, PH (acidity), base deficiency, oxygen saturation, PT (prothrombin time), and APTT (partial thromboplastin time). (4) Burn situation, including burn area and burn depth. (5) Patient regression, including whether blood use: whether to transfuse suspended red blood cells, the first day of hospital transfusion, and the amount of blood transfusion.

For a more straightforward presentation of the data distribution, preliminary dataset statistics are shown in Table 1, and box plots by age, blood consumption, and missing rate are shown in Figure 2.

After a detailed definition of the problem and a description of the dataset, we summarize the main problems faced in this paper.(i)Data holds missing values. as shown in Table 1 and Figure 2(c), some hospitals have more than 10% missing values. The missing values cause perturbation to the data distribution, severely affecting the model’s training and performance.(ii)Data has domain bias. from Table 1 and Figure 2(c), the missing values rate and blood consumption have obvious differences in different hospitals. Such a data distribution impacts the cross-hospital generalization of the model. In addition, the collection of clinical information from different hospitals may also be biased due to differences in testing devices and habits.

4. Methodology

4.1. Problem Defensions

The following definition is made in this paper to discuss the role of predictors.

Definition 1. (Blood Consumption Prediction Problem, BCPP). Let be a training dataset where clinical information is implicitly linked to the blood usage . The BCPP train a model with , and use the model predict the blood usage of new collected data of clinical information , where is the model parameter.
The BCPP includes a classification subproblem and regression subproblem. For classification subproblem and are one-hot or category values. For regression subproblem and are continuous values.

4.2. Blood Data Complementation

Data complementation is the process of replacing missing data with substituted values. When covering for a data point, it is known as “unit complementation;” when substituting for a component of a data point, it is known as “item complementation.” Missing values introduce substantial noise, making data analysis more complex and less efficient. When one or more patient values are missing, most methods discard data with missing values by default, but data complementation attempts to fill in those values. However, missing data are also a reality, and the model should not require that all data be captured well. Therefore, data complementation is introduced to avoid performance degradation with missing values and improve the actual testing data’s performance.

In this study, a single data is complemented bywhere , , are data components of single data . If any component is missing, the imputed value is used to fill this missing value. The comes from mean value complementation, median value complementation, and KNN complementation:where and are the mean value or median value of training datasets on components . is the KNN neighborhood of data in the sense of European distance, in this paper.

4.3. Cross Hospitals Data Augmentation

Data augmentation is a well-known neural network (NN) training strategy for image classification and signal processing [30]. Data augmentation improves the performance of the methods by precisely fitting the data distribution. First, data augmentation enhances the diversity of data, thereby overcoming overfitting. Second, data augmentation essentially reinforces the fundamental assumption of DR, i.e., the local connectivity of neighborhoods. Finally, it learns refined data distribution by generating more intra-manifold data based on sampled points.

Cross-hospital data augmentation is introduced in a unified framework to generate new data :where the new augmented data is the combination of each feature . is calculated from the linear interpolation of original feature and augmented feature . In addition, the augmented feature is sampled from the neighborhood of . is the k-NN neighborhood of the data on the data of hospitals . is the set of the hospitals. means remove the neighborhood of data ’s hospitals. The combination parameter is sampled from the uniform distribution , and is the hyperparameter.

The cross-hospital augmentation generates new data by combining data with its neighborhoods in different hospitals. It reduces the terrible influence of the missing data and improves the training data’s divergence. Thus, the model learns a precise distribution to enhance the performance of our method. In addition, it works together with the loss function to align data from different hospitals, thus overcoming domain bias.

4.4. Architecture of BUPNN

The proposed BUPNN does not require a unique backbone neural network. The multilayer perceptron network (MLP) is the backbone neural network. In addition to this, a new network architecture is proposed to enhance generalizability. The proposed neural network architecture is shown in Figure 3.

A typical neural network model uses the network output directly to compute a supervised loss function. It may introduce undesirable phenomena such as overfitting. In this paper, similar to the paper [31], a manifold learning regularizer is proposed to suppress problems such as overfitting using the information in the latent space of the network. As shown in Figure 3, a complete neural network is divided into a latent network and an output network . The latent network is a preprocessing network that resists the adverse effects of noise and missing value completion. The output network is a dimensional reduction network that maps the data in high-dimensional latent space to the data in low-dimensional latent space. The latent network maps the data input a network latent space and an output network further map it into the output space:where and are the complementation and augmentation results of origin data .

Neural networks have powerful fitting performance, but at the same time, there is a risk of overfitting. The typical L2 regularization method can reduce overfitting, but it only limits the complexity of the network without constraining the network in terms of the manifold structure of the data. For example, there is no way to guarantee the distance-preserving and consistency of the network mapping. For this reason, we design a manifold regularizer based on manifold learning to solve this problem, and thus improve the generalization performance of the model and its usability for actual data.

4.5. Loss Function of BUPNN

The loss function of BUPNN consists of two parts; one is the cross-entropy loss which uses label information, and the other is the manifold regularizer loss which uses hospital information and latent space information.

4.5.1. Manifold Regularizer Loss

Manifold regularizer loss handles the domain bias in different hospitals during the training phase and provides a manifold constraint to prevent overfitting. Inconsistent medical equipment and subjective physician diagnoses in different hospitals cause domain bias in data between hospitals. The manifold regularizer loss guides the mapping of the neural networks to be insensitive to hospitals, thus overcoming domain bias (shown in Figure 4). Therefore, the pairwise similarity between nodes is defined first. Considering the dimensional inconsistency of the latent space, we use the t-distribution with variable degrees of freedom as a kernel function to measure the point-pair similarity of the data:where is the Gamma function, and the degree of freedom controls the shape of the kernel function. is the Euclidean pairwise distance of node pairs.

Based on the defined pairwise similarity in a single space, we minimize the similarity difference between two spaces by fuzzy set cross-entropy loss (two-way divergence) [32] :where . Notice that is a continuous version of the cross-entropy loss. In BUPNN, equation (6) is used to guide the pairwise similarity of two latent spaces to fit each other.

Therefore, the loss function of the manifold regularizer is defined as follows:where is the batch size and indicates whether is the augmented data of . If is the augmented data of , the loss pulls them together in the latent space; otherwise, the loss keeps a gap between them to preserve the manifold structure. The different degrees of freedom and in t-distribution are basic settings according to the dimensions of the space. Equation (7) describes the behavior of a manifold alignment that pairs data collected by different hospitals together in latent space to avoid the detrimental effects caused by domain bias:where and are the distance between data node and node in spaces and .

4.5.2. Cross-Entropy Loss

The loss function of the manifold regularizer is essentially an unsupervised term that also requires the use of label information while training the network model. The cross-entropy loss function is introduced simultaneously:where is the label of data node , is the output of the network model, and is the natural logarithm. When solving the classification subtask, is the category label, and when solving the regression subtask, is the probability label.

The loss function of BUPNN iswhere is a hyperparameter to balance and .

4.6. Pseudocode and Complexity

BUPNN’s pseudocode is shown in Algorithm 1. The BUPNN includes the initialization and training phases. In the initialization phase, the kNN neighborhood of every single data is discovered. The time complexity of initialization phases is [33], where is the number of data. In the training phase, the complexity of the training phases is the same as that of artificial neural networks (ANN). BUPNN calculates the pairwise distance in a batch, so the complexity of training each batch is , where is the batch size. GPU can well accelerate the pairwise distance, so the training time consumption is the same as that of a typical ANN.

Input: data: , learning rate: , epochs: , batch size: , , , network: ,
Output: graph embedding: .
(1)while; ; ++ do
(2) ⊳ # blood data complementation in equation (1)
(3)  while; ; ++ do
(4)    ⊳ # blood data augmentation in equation (3)
(5)   ; ⊳ # map input data into space in equation (4)
(6)   ; ⊳ # map input data into space in equation (4)
(7)   ; ; ⊳ #calculate distance in and
(8)   ; ; ⊳ #calculate similarity in and in equation (5)
(9)    ; ⊳ # calculate loss function in equation (10).
(10)   , ; ⊳ # update parameters.
(11)  end while
(12)end while
(13); ⊳ # calculate the embedding result.

5. Experiments

5.1. Baseline Methods

In this Section, two subtasks, classification subtask and regression subtask, are defined. Seven state-of-the-art baseline classification and regression methods are chosen to discuss the relative advantages of BUPNN. The baseline approach is as follows.

5.1.1. K-Nearest Neighbor Classification/Regression Method (KNN) [37]

The K-nearest neighbor classification/regression method is a nonparametric statistical method. The KNN classification method outputs the prediction by the “majority vote” of its neighbors. The KNN regression method outputs the prediction by the average of its neighbors.

5.1.2. Decision Tree Classification/Regression Method (DT) [35]

A decision tree builds regression or classification models in the form of a tree structure. It breaks down a dataset into smaller and smaller subsets while, at the same time, an associated decision tree is incrementally developed. The final result is a tree with decision nodes and leaf nodes. A decision node has two or more branches representing values for the attribute tested. The leaf node represents a decision on the numerical target. The root node is the topmost decision node in a tree that corresponds to the best predictor. Decision trees can handle both categorical and numerical data.

5.1.3. Random Forest Classification/Regression Method (RF) [34]

Random forests or random decision forests are ensemble learning methods for classification, regression, and other tasks that operate by constructing many decision trees at training time. For classification tasks, the output of the random forest is the class selected by most trees. For regression tasks, the mean or average prediction of the individual trees is returned. Random decision forests correct decision trees’ habit of overfitting their training set.

5.1.4. Extremely Randomized Trees Classification/Regression Method (ET) [38]

Extremely randomized trees add a further step of randomization to random forest, while similar to ordinary random forests in that they are an ensemble of individual trees, there are two main differences: first, each tree is trained using the whole learning sample (rather than a bootstrap sample), and second, the top-down splitting in the tree learner is randomized. Furthermore, instead of computing the locally optimal cut-point for each feature under consideration (based on, e.g., information gain or the Gini impurity), a random cut-point is selected.

5.1.5. Support Vector Machine Classification/Regression Method (SVM) [39]

Support vector machine is a supervised learning model with associated learning algorithms that analyze data for classification and regression analysis. Given a set of training examples, each marked as belonging to one of two categories, an SVM training algorithm builds a model that assigns new standards to one category or the other, making it a nonprobabilistic binary linear classifier (although methods such as Platt scaling exist to use SVM in a probabilistic classification setting). SVM maps training examples to points in space to maximize the width of the gap between the two categories. New models are then mapped into that space and predicted to belong to a class based on which side of the hole they fall.

5.1.6. Gradient Boost Classification/Regression Method (GB) [40]

Gradient boosting is a machine learning technique used in regression and classification tasks. It gives a prediction model in an ensemble of weak prediction models, typically decision trees. The resulting algorithm is called gradient-boosted trees. When a decision tree is a weak learner, it usually outperforms a random forest.

5.1.7. Adaptive Boost Classification/Regression Method (ADB) [36]

The adaptive boost algorithm obtains a strong learner by combining a series of weak learners and integrating these vulnerable learners’ learning capabilities. Adaptive boost changes the weights of the samples based on the previous learners, increasing the importance of those previously misclassified samples and decreasing the weight of correctly classified samples so that the subsequent learners will focus on those misclassified samples. Finally, these learners are combined into strong learners by weighting.

5.1.8. Light Gradient Boosting Machine Classification/Regression Method (LightGBM) [41, 42]

LightGBM is a distributed gradient boosting framework based on the decision tree algorithm. LightGBM is designed with two main ideas in mind: (1) to reduce the use of data in memory to ensure that a single machine can use as much data as possible without sacrificing speed and (2) to reduce the cost of communication to improve the efficiency when multiple machines are in parallel, and achieve linear acceleration in computation. It can be seen that LightGBM was originally designed to provide a fast and efficient data science tool with a low memory footprint, high accuracy, and support for parallel and large-scale data processing.

5.2. Dataset Partitioning and Grid Search

Table 1 and Figure 2 provide basic information about the data. Five data partitioning schemes are provided in this study to provide a detailed comparison of the performance differences between the different schemes.

Three schemes are with data-complement, including COM-Mea, COM-Med, and COM-KNN as defined in equation (2). First, for COM-Mea, COM-Med, and COM-KNN, the missing values are complemented with a specific method. Following that, the training and testing sets are divided by 90%/10%. Two noncomplement schemes (NC-A and NC-B) are introduced to compare with the complement schemes. NC-A deletes all the data items with missing values, following the typical machine learning scheme. Following the data cleaning, NC-A divides the training and testing sets by 90%/10%. NC-B keeps the same training and testing set division as the data complement schemes and only removes all missing data from the training set and obtains a cleaner training set.

The models are evaluated with 10-fold cross-validation for all the training sets and determine the optimal parameters by grid search. For a fair comparison, we control the search space of all baseline methods to be approximately equal. The search space of the compared process is in Table 2.

5.3. Experimental Setup

We initialize the other NN with the Kaiming initializer. We adopt the AdamW optimizer [43] with learning rate of 0.02 and weight decay of 0.5. All experiments use a fixed MLP network structure, : (−1, 500, 300, 80), : (80, 500, 2), where −1 is the features number of the dataset. The number of epochs is 400. The batch size is 2048. .

5.4. Comparison on Classification Subtask

Although the blood data for each patient is collected for accurate regression subtasks, it is equally important to predict whether a patient needs a blood transfusion. Especially in emergencies, indicating whether a patient needs a blood transfusion in the short term is more important than estimating the amount of blood used throughout the treatment cycle. Therefore, we first evaluate the performance of the proposed BUPNN on the classification subtask. Then, two evaluation metrics, classification accuracy (ACC) and area under the ROC curve (AUC), are used to compare the classifier’s performance from multiple perspectives. The performance comparison of BUPNN and eight baseline methods is shown in Tables 3 and 4. In addition, the ROC curves of all the compared methods on different schemes are shown in Figure 5. The scatter plot the prediction of BUPNN for COM-Mea schemes are shown in Figure 6.

5.4.1. Data Complementation Delivers Performance Improvements

From the performance results of NC-B, COM-Mid, COM-Mea, and COM-KNN in Tables 3 and 4, we observe that the schemes with data complementation yield better performance. We attribute the reasons to the more abundant training data provided by data complementation. Although the complemented data are imperfect, artificial intelligence models can still learn more helpful information.

Although the model trained with only clean data (NC-A) has the highest performance score, scheme NC-A only includes the clean data and cannot be conveniently generated to the actual data with missing values. However, missing values are unavoidable in real-world application scenarios, so data complementation is a better choice to improve the model’s performance.

5.4.2. BUPNN Outperforms Almost All Baseline Methods

From Tables 3 and 4, we observe that the proposed BUPNN has advantages over all the baseline methods. For the AUC metric, BUPNN has the advantage in all five schemes. However, the COM-MID scheme has the most significant benefits, outperforming the second-best method (LGBM) by 3.71%. For the ACC metric, BUPNN has the advantage in all five schemes. However, the COM-KNN scheme has the most significant benefits, outperforming the second-best method (LGBM) by 2.95%. BUPNN has a 1% average advantage over the second-best method in both metrics. The reason is attributed to the fact that BUPNN is a neural network-based model, which performs better with richer data. In addition, the proposed manifold loss function can improve the model’s generalization performance and thus enhance the performance of the testing set.

5.4.3. BUPNN Is Better at Handling Complemented Datasets

From four schemes with the same testing set (NC-B, COM-Mid, COM-Mea, and COM-KNN) in Tables 35, we observe that BUPNN performs better than all the baseline methods when handling complemented datasets. We attribute this to the data augmentation of the proposed BUPNN model. Data augmentation generates new training data from the completed data and attenuates the effect of missing values, thus guiding BUPNN to learn a smooth manifold.

5.4.4. BUPNN Has Better Generalizability between Different Hospitals

To evaluate the generalization performance of our proposed BUPNN and baseline methods, we tested the ACC with the data from various hospitals (Figure 5). The proposed BUPNN has a dominant performance of five hospitals out of a total of 6 hospitals and has a clear advantage over Shaoyifu Hospital. From Table 1 and Figure 1, we observe that the Shaoyifu Hospital has a relatively obvious domain bias. It has a minimum male-to-female rate, the top three average missing rates, and many outliers in its missing values. We argue that the domain bias influences the performance of the baseline methods, and the proposed BUPNN outperforms the baseline methods due to BUPNN overcoming the domain bias in it. Furthermore, the manifold regularizer loss function provides a good manifold constraint to improve the model’s generalizability between different hospitals (as shown in Figure 4).

5.5. Comparison on Regression Subtask

Then, we discuss the performance of BUPNN on the regression subtask. Finally, the regression model is asked to predict the total blood usage in the patient’s treatment for advance scheduling at the blood center. The experimental setting of the regression subtask is the same as that of the classification subtask. The data are collected from the same patients, but the target variable is the total blood usage in the patient’s treatment. The mean square error (MSE) metric is calculated to measure the performance of the seven methods in the five schemes. The performance comparison is shown in Table 6.

5.5.1. BUPNN Shows a Consistency Advantage on Regression Subtasks

From Table 6, we observe that the proposed BUPNN outperforms all the baseline methods on all four schemes. BUPNN has the most significant benefits for the COM-mid scheme, outperforming the second-best method (LGBM) by 0.004 (10.5%). The percentage of improvement shows that the advantage of BUPNN is more evident in the regression subtask. Furthermore, it indicates that the proposed BUPNN is more suitable for handling more difficult regression tasks and that BUPNNs have strong application potential when collecting richer data.

5.6. The Explanatory Analysis of BUPNN

Artificial neural network-based models are considered to have a strong performance. Still, their black-box nature leads to difficulty in interpretability, thus making it difficult for the model to be trusted by domain experts. On the other hand, the proposed BUPNN has stronger interpretability because the smoothness of its mapping leads to models that profound model interpreters can easily interpret. An easy-to-use neural network interpretive tool is introduced to explain the decision process of BUPNN. The visual presentation of the interpretation results is shown in Figure 7.

In Figure 7, the decision processes of BUPNN for three samples are explained by “shap” tool. The vertical coordinates represent the clinical indicators, and the horizontal coordinates represent the predicted values of BUPNN. The images show the decision process of the model from bottom to top. Specifically, the model predicts for an average patient when no information about the patient is observed, i.e., no transfusion is needed. We believe this is reasonable because only patients who are sufficiently harmed need blood transfusions to be treated. Next, using Figure 7(a) as an example, the BUPNN model observed additional patient information, such as “no injury to the lower body,” “no injury to the abdomen,” “no injury to the pelvis,” and the “patient’s heartbeat is not accelerated.” This evidence made the BUPNN firm the original judgment that blood transfusion is unnecessary. Although the penetration injury raised the probability of transfusion, several essential features (e.g., albumin and hemoglobin) ultimately guided the model to predict .

Figure 7(b) provides more examples of how to interpret the decision process of the BUPNN. These two examples show what kind of information needs to be observed by BUPNN that would predict a sample as needing a blood transfusion. Among them, we found some necessary signatures of the need for transfusion, such as the abdomen, pelvis, and pleural.

Next, we calculated the importance of all features in determining whether a blood transfusion is needed or not and displayed them in Figure 8 in the form of a bar chart.

6. Conclusions

In this paper, a neural network model-based blood usage predictor, called deep blood usage neural network (BUPNN), is proposed to serve the scheduling of blood supply in provincial blood centers. The proposed BUPNN receives clinical information from hospital patients and predicts whether a patient needs blood and the amount of blood used. The proposed BUPNN receives clinical data from hospital patients and indicates whether a patient needs blood and the blood consumption amount. The proposed BUPNN mainly solves the problem of predicting blood usage with high availability and generalization in real-life situations. It can accomplish the prediction task well in different hospitals’ missing values and domain bias. BUPNN proposes a manifold learning-based regularizer for the blood prediction problem to improve the model’s generalization performance on data from different hospitals and enhance the model using data augmentation and data complementation.

Data Availability

The blood data used to support the findings this study are included in the article.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This study was supported by the Science Research Foundation of Zhejiang Province (LGF20G030002) “Research on Assisted Decision-making Model of Blood Supply in Public Emergencies ”.