Abstract

Cloud service providers (CSPs) can offer infinite storage space with cheaper maintenance cost compared to the traditional storage mode. Users tend to store their data in geographical and diverse CSPs so as to avoid vendor lock-in. Static data placement has been widely studied in recent works. However, the data access pattern is often time-varying and users may pay more cost if static placement is adopted during the data lifetime. Therefore, it is a pending problem and challenge of how to dynamically store users’ data under time-varying data access pattern. To this end, we propose ADPA, an adaptive data placement architecture that can adjust the data placement scheme based on the time-varying data access pattern and subject for minimizing the total cost and maximizing the data availability. The proposed architecture includes two main components: data retrieval frequency prediction module based on LSTM and data placement optimization module based on Q-learning. The performance of ADPA is evaluated through several experimental scenarios using NASA-HTTP workload and cloud providers information.

1. Introduction

With the development of cloud computing, more and more companies adapt the cloud to store their data for low maintenance costs and reliable SLAs (Service Level Agreements) comparing to the traditional data storage mode. Many mainstream cloud service providers (CSPs) offer a variety of data storage services to satisfy different users’ demands. The pricing of services for the same functionality is diverse among cloud service providers (CSPs). And the pricing policy of the same CSP is various in different regions. The cost of data migration among data centers of the same CSP is cheaper than that in different CSPs.

In addition to the constraints of high migration costs, a single cloud faces the risk of vendor lock-in; i.e., major risks include the price of cloud services and the interruption of SLA. These situations may result in making users pay expensive migration costs. In our previous work, we have proposed an ant colony algorithm-based approach for cost-effective data hosting with high availability in multicloud environments [1]. In order to avoid vendor lock-in, we can divide the original data and store the data chunks into multiple CSPs. Furthermore, there is a more comprehensive solution to this problem in [2]. We aim at providing a cost-effective and high-available data placement in multicloud environments based on users’ demands.

In the previous work [13], the workload of data object affects the choice of data placement scheme. The data workload is related to the data retrieving frequency (DAF) (i.e., Get access rate) during a fixed time period. In the life-cycle of data placement, the workload is time-varying. The data object with high data is read-intensive and in hot-spot status, which is more likely to be stored in CSPs with lower out-bandwidth costs additionally. Conversely, the data object with low DAF is storage-intensive and in cold-spot status, which is more likely to be stored in CSPs with lower storage costs [4]. If a user adapts a data placement scheme with a relatively low-accessible frequency for the entire life-cycle of data storage, it may generate more out-bandwidth costs when DAF increases. In another case, if a user employs the strategies that are more suitable for hot-spot status in the entire life-cycle of data storage, it may produce expensive storage costs when DAF reduces.

In order to reduce the overall cost and enhance availability during the data object lifetime, it is essential to develop a mechanism to dynamically adjust the data placement scheme based on data object workload. Due to the uncertainty of future data workload, the overall cost cannot get the best result. Therefore, predicting the future workload is a key part of the dynamic data placement mechanism. When the future workload is obtained, how to design a dynamic placement scheme based on future access frequency becomes another significant component.

ADPA, which can dynamically host the data object with cost-effective and high availability based on time-varying data workload, is an adaptive data placement architecture proposed in our study. The key contributions of our study are as follows.Firstly, we propose a LSTM [5] algorithm based on workload prediction. It can use historical workload data to predict future workload.Secondly, a dynamic data placement algorithm based on reinforcement learning is proposed. It can migrate data according to the change of workload to ensure cost optimization and availability.Finally, we evaluate ADPA through several experiment scenarios. The results show that ADPA algorithm not only is superior to SOA algorithm, but also can save more time than ACO and DP algorithm to obtain the optimal data placement.

The remainder of this paper is organized as follows. Section 2 discusses the related work. Section 3 presents the motivation of the adaptive data placement. Section 4 describes the architecture of adaptive data placement. Section 5 introduces the proposed algorithm. The performance of our proposed algorithm is shown via extensive experiments by using real-world cloud information in Section 6. Finally, Section 7 summarizes this paper.

There are many studies for data placement optimization in cloud computing. We can divide these studies into two categories based on whether the data storage scheme can be adjusted according to workload: static data placement and dynamic data placement. In static data placement, data placement scheme is determined in advance and does not adapt to variable data object workload. However, in dynamic data placement, data placement scheme can be obtained according to the change of DAF. In the following, we review and discuss them, respectively.

2.1. Static Data Placement

In [6], the authors present the fact that availability of service and data lock-in are two obstacles for cloud computing. In cloud storage, replication and erasure coding are widely used in increasing data availability [7]. To avoid data lock-in, multicloud plays an important role in data storage.

Mansouri et al. [8] propose an approach to replicate data objects in multiple data centers to enhance data availability and avoid data lock-in. In [9], the authors adapt erasure coding to ensure data security and minimize the total cost under many unreasonable assumptions. Hadji [10] focuses on minimizing data storage and presents two efficient algorithms to solve it. But the out-bandwidth cost is ignored. In these studies, each data item is at a fixed number in a static way; Qu et al. [11] present a resilient, fault-tolerance, and high-efficient global replication algorithm (RFH) to determine each data items’ replication, migration, and suicide.

The above studies only optimize to minimize the monetary cost or enhance data availability under other QoS metrics. The trade-off between cost and availability is not considered. Wang et al. [1] propose an approach based on ant colony algorithm that minimizes total cost and enhances data availability through erasure coding. It is still a single objective optimization problem. In [12], Liu et al. use multiobjective particle swarm optimization algorithm to minimize storage space cost, data migration cost, and communication cost as well as enhancing the storage reliability. However, they cannot determine how to choose an optimal solution in the Pareto front.

All the above works only consider the static data placement and ignore the uncertainty of data object workload. Once the workload of data object is changed, the data placement scheme needs to be calculated again.

2.2. Dynamic Data Placement

The dynamic data placement in cloud storage becomes a hot direction in research. Zhang et al. [3] propose a data hosting scheme which contains a transition of storage modes based on changes of data access frequency. However, they only put forward a transfer condition without considering the global data placement sequence optimization problem. In [13], Su et al. present a systematic model, in order to formulate data placement optimization for complex requirements in multicloud environments. They also only discuss the data migration without making the final decision for developers.

Due to the unknown future data access pattern, the above studies do not propose the automatic data migration. Mansouri et al. [4] propose two online algorithms that make a trade-off residential and migration cost. Due to the absence of the future data object workload, the authors use the deterministic online algorithm to solve the problem.

To achieve a dynamic data replication strategy based on real data access pattern, Gill and Singh [14] propose a dynamic cost-aware rereplication and rebalancing strategy. This method firstly determines which file and when to replicate based on files popularity. The more the popular it is, the more possible the file is replicated. Then, it calculates the replica number to meet the availability requirements. Finally, a replica placement mechanism is proposed. Lyapunov optimization is an online control method to solve the problem which closes the optimal solution. In [15], Qiu et al. adapt Lyapunov for dynamic content distribution service deployment without the need for any future request arrival information. In other works of dynamically optimized storage, T. Sreenath Reddy and G. MURALI [16] use GA-based totally records dissemination alternate technique and dynamic request redirection to minimize the cost of grouped gets. The algorithm IMPSO used in [17] also inspires our work.

Except for the above methods, a method based on the historical workload pattern is also effective to solve the dynamic data placement. In [18], Papaioannou et al. propose a cloud storage brokerage solution that can periodically recompute the best provider set using data access statics of the last sampling periods. It can adjust data placement for dynamically changing data access pattern. In recent years, reinforcement learning (RL) has received extensive attention. RL is a powerful approach for a long time decision-making under uncertainties [19]. As far as we know, there are no studies on using RL to optimize the dynamic data placement.

In this paper, we propose an adaptive data placement method that mainly contains data retrieval frequency prediction and data placement optimization. In data retrieval frequency prediction module, we adapt LSTM to predict the future workload of data object. In data placement optimization module, an approach based on Q-learning is used to get a sequential data placement solution according to the prediction data object workload.

3. Motivation

3.1. Dynamic Data Object Workload

The DAF of data object is time-varying. We collect the trace of NASA-HTTP that depicts all HTTP requests to the NASA Kennedy Space Center WWW server in Florida [24]. We calculate the DAF in a cycle of 10 minutes. As shown in Figure 1, it depicts the DAF’s change from 01/Jul/1995 : 0:0 : 0 to 07/Jul/1995 : 23 : 59 : 59. The difference between the maximum DAF value and the minimum DAF value is close to 2. DAF was relatively low at night but increased during the day. The data placement scheme needs to be dynamically adjusted according to the changing DAF, which will be discussed in Section 3.3.

3.2. Heterogeneous Cloud Market

Now there are many CSPs providing storage services, and we collect four most popular CSPs’ pricing strategies: Amazon S3 [20], Microsoft Azure Cloud Storage [21], Alibaba Cloud Object Storage [22], and Google Cloud Storage [23] as shown in Table 1. From the point of users’ cost, the selection of CSP is a research direction in cloud computing. In our previous work, we proposed some approaches for optimizing selection of cloud instance types [2527]. Actually, the price of the same functional storage service provided by the same CSP in different regions is different. And the price of the same functional storage service across CSPs is also different. For example, Amazon S3 in New York, USA, has lower storage price than that in Tokyo. For instance, Amazon S3 in New York has lower storage cost than Microsoft Azure Cloud Storage in New York.

3.3. Discussions

The dynamic data object workload and heterogeneous cloud market inspired us to propose an adaptive data placement scheme. Then, we discuss this in detail through the NASA-HTTP example.

When the access frequency is low, the data is suitable for CSPs with low storage cost. But the bandwidth cost of these cloud service providers may be relatively high. We assume that Amazon S3 in New York, Tokyo, and London is chosen as the data placement scheme because of the lower storage cost. The out-bandwidth price of Amazon S3 in New York, Tokyo, and London is very expensive. As time goes by, DAF gradually becomes larger and the output bandwidth price will be very expensive. If users migrate their data to CSPs with lower out-bandwidth cost, they can save more money even if they need to pay for additional migration cost.

In this study, we propose an adaptive data placement scheme which dynamically adjusts the DAF-based data storage scheme to minimize the total cost.

4. An Adaptive Data Placement Architecture

In this section, we present an adaptive data placement architecture with high cost-effective multicloud availability. Then, we formulate the problem in the architecture.

4.1. Architecture Overview

Figure 2 depicts the architecture of dynamic data placement. There are four components: Cloud Storage Information Collection, Optimization Module, Workload Statistics Module, and Prediction Module.

Cloud Storage Information Collection is used to collect the information of CSPs including storage price, out-bandwidth price, and operation cost. Apart from this, it receives DAF prediction to adjust data placement scheme.

Optimization Module is used to optimize the data placement scheme according to users’ demands which contain the size of data object, required availability, initial DAF, and so on.

Workload Statistics and Prediction Module are responsible for collecting the historical data of workload and were based on these historical data to predict the DAF for the next period of time.

4.2. Problem Definition

To describe a dynamic data placement architecture in detail, we introduce the following definitions. The symbols used in this paper are listed in Table 2.

Definition 1. (data center specification). Assume that there are data centers . Each data center has tuple: , where defines the probability of data center being available.

Definition 2. (data object specification). Assume that there is a data object with the size and the DAF where , the required availability is , and the required data retrieval latency is .
In this work, we also adapt erasure coding to avoid vendor lock-in, enhance data availability, and reduce storage and out-bandwidth costs compared to replication. It is worth noting that the following definitions for availability and cost are similar to those in [1, 3, 13], which is a universal way to define them for data hosting in erasure coding mode. We give the definition of erasure coding at first.

Definition 3. (erasure coding). An -erasure coding denotes the data object which is divided into chunks and encodes chunks into chunks. Users can retrieve their data through any chunks.

Definition 4. (data availability). Since users can tolerate shut-down of data centers at the same time, data availability is the sum probability of all cases that DCs are simultaneously available, . Assume that denotes the data centers chosen for data storage at time slot , where . We use to indicate the number of cases that data centers are available. denotes the th data centers collection in cases at time slot . The data availability at time slot can be defined as follows:where is the data centers that are not available.

Definition 5. (storage cost). the storage cost is the sum of storage cost of data centers. It can be calculated aswhere denotes the size of data stored in each data center.

Definition 6. (network cost). Users can retrieve their data through any data chunks. We choose data centers with the lowest out-bandwidth price to retrieve data. Since the DAF is time-varying, network cost is related to of each time slot. It can be calculated as follows:

Definition 7. (operation cost). The operation cost can be defined as follows:It is worth noting that the value of in (3) and (4) is equal.

Definition 8. (base cost). The base cost is the sum of storage cost, network cost, and operation cost at time slot . It can be defined as

Definition 9. (migration cost). Different DAFs may correspond to different optimal data placement solutions. It generates expensive cost if users adapt the previous solution. Therefore, it can save more cost according to the DAF dynamically adjustment data placement. But the data placement adjustment means data migration which also requires cost. Data migration is not the moment when DAF changes but satisfies the conditions of the cost saved by the new solution (compared to the old one) which can cover the migration cost. We use to indicate the intersection of data placement solutions of time slots and . The data centers that need data migration are . The migration cost can be defined as follows:where the condition is as follows:

Definition 10. (data retrieval latency). The data centers for data retrieval at time slot are data centers with the lowest out-bandwidth cost. The data retrieval latency is the maximum latency of . Since the data retrieval latency is determined by network latency, we use round-trip time to indicate data retrieval latency [4, 28, 29]. We use the calculation method in [30], as follows:where is the distance between users and data centers.

4.3. Optimization Problem

In ADPD, it aims to obtain the data placement solution in each time slot so that the total cost during is minimized. The optimization problem is defined as follows:subject to(1), (2), (3),

Constraint (1) indicates that only data chunks in each time slot. Constraints (2) and (3) ensure that the data placement solutions satisfy the user required availability and data retrieval latency in each time slot. Obviously, this optimization problem is NP-hard.

5. Solution

In this section, we describe the implementation of ADPD. Firstly, we adapt LSTM for data object DAF prediction. In order to make decision-making effectively, the method to solve the problem is mainly based on Q-learning that is an off-policy temporal difference (TD) control algorithm [31, 32].

5.1. DAF Prediction

Recently, there are many machine learning approaches including Support Vector Machines (SVMs), Linear Regression (LR), Random Forest (RF), and k-Nearest Neighbors (kNNs) adopted for prediction. In this study, we use LSTM to predict the future DAF. We can also use other more accurate algorithms to replace LSTM Algorithm 1.

Input: The required length of the series to be forecast, a trained LSTM model .
Output: Predicted future DAF;
(1)  = 0;
(2)Feed the data in the last window into and get a prediction ;
(3)whiledo
(4) Slide the window forward by one step and put the newly predicted value at the end of the slide window;
(5) Feed the data in the new window into and get another prediction
;
(6);
(7)end while

Since our previous work is to predict the price of cloud spot instance whose problem is also for time series data prediction [33]. In this study, we also use sliding window to split the historical DAF data. For DAF prediction, its goal is to find a function that used historical DAF to predict the future DAF.

5.2. Data Placement Optimization

In this section, the data placement optimization aims to obtain the optimal solution according the future DAF. Assuming that denotes the candidate data placement solutions at each time slot, where . Let indicate the optimal data placement solution from to , where . We describe the process of data placement during via Figure 3. For each , it has Markov property. The reason is that the solution choice for the next time slot is only related to the currently selected plan and is independent of the previous choices. It can be expressed as .

In data placement process, the cost directly determines the choice of the plan. And the data placement optimization is a sequential decision-making problem in our paper. So we can attribute it to Markov decision process (MDP) and solve it through reinforcement learning (RL). The MDP of data placement process can be defined as follows [34].

Definition 11. (MDP of data placement process). We can transfer data placement process into MDP, which is a 5-tuple one:(1)A finite set of states is denoted by (2)A finite set of actions are the choice of solution of next time slot(3) is the probability that data placement at time slot t when taking choice results in solution at time t+1(4)A reward function is the negative value of solution transitions total costs from to when taking action between two time slots(5)The discount factor , which is the attention of future rewardsThe goal of RL is to find the optimal policy based on MDP. The basic architecture of RL is shown in Figure 4 [32]. The agent not only interacts with the environment through making action according to the current state, but also gets an immediate reward and the state of observation [32]. RL aims to find an optimal policy to maximize the total rewards that are obtained from every step. In this optimization problem, the reward function is defined as follows:So, the RL objective of this problem can be presented aswhere and is the optimal data placement solution during time slot .
For the above problem, Q-learning is widely used tabular RL algorithm [19, 31] which is an off-policy TD control. It is defined by [31]where that is defined in (10) and is the learning rate.
As stated previously, we transfer the time series data placement optimization into an MDP and adapt the approach based on Q-learning to solve it. For that purpose, we proposed an MDP that comes from data placement optimization being the most important approach. Algorithm 2 represents the transfer methodology. In order to satisfy the constraints (1)–(3) in (9), we filter all candidate data placement solutions according to their availability and latency (Line 5). In the whole life-cycle of data storing, the DAF is time-varying, which results in the fluctuation of cost. Then, we calculate the state matrix that contains basic cost for all satisfied solutions (i.e., ) at each time slot (Lines 6–9). Finally, the migration cost between two solutions at time slots and needs to be calculated as the transfer matrix which is a matrix (Lines 12–16). Since the migration cost between the same two solutions is free, the values of the diagonal on the matrix are 0.
Compared with general MDP, the reward function contains not only the value between states, but also the value of states which is defined in (10). Based on the above MDP, we present an approach based on Q-learning. The core of the method is calculation and update of Q tabular which is the basis for the agent to choose the next time slot’s action. Since this paper solves a time-based sequential decision problem, the Q tabular also has a time dimension. The strategy of Q tabular is based on (12). The value function of next state is adapted to update the current state, which is called bootstrapping [32]. The approach based on Q-learning is an off-policy method; that is, the action strategy and the target strategy are different. In Q-learning, the action chosen strategy is -greedy and the target is greedy. Algorithm 3 represents the data placement methodology based on Q-learning.

Input: Data center specification, , DAF, , the required availability, , the required data retrieval latency, .
Output: The state matrix, , the transfer matrix, .
(1)Calculate all -combinations of data centers .
(2)for alldo
(3) Calculate availability of through (1);
(4) Calculate latency of through (8);
(5)if and then
(6)  for to do
(7)   Calculate the base cost of through (5);
(8)   ;
(9)  end for
(10)end if
(11)end for
(12)for alldo
(13)for alldo
(14)   Calculate the migration cost between and through (6);
(15)end for
(16)end for
(17)return, ;

6. Evaluation

In this section, we firstly describe the dataset used in experiments. Then, we present a basic algorithm to verify the necessity of future DAF prediction. Since ant colony optimization algorithm (ACO) and genetic algorithm (GA) are widely used to solve optimization problem [1], we compare ADPA with them through different experimental scenarios.

6.1. Settings

In this work, the setup for DAF, data centers’ specification, and experiment parameters are as follows.

6.1.1. DAF

We use the real workload trace of NASA-HTTP from [24] which contains two months’ worth of all HTTP requests to the NASA Kennedy Space Center WWW server in Florida from 01/Jul/1995 to 31/Aug/1995. Each piece of data in NASA-HTTP represents a request, including host, timestamp, request, HTTP reply code, and bytes in the reply. In our paper, in order to obtain data access frequency, we count the number of Get operations in a specific time period, which is 600 seconds. The trace is separated as a train and test set for the LSTM at a ratio of , respectively.

Input: The state matrix, , the transfer matrix, .
Output: The optimal data placement solution during ,
(1)Initialize parameters of algorithm including learning rate , discount
(2)Initialize the tabular with zero;
(3)Initialize the start state;
(4)Initialize data placement sequences ;
(5)for to do
(6)for to do
(7)  Choose feasible data placement solutions through equation
(8)  Choose a data placement solution from through -greedy function;
(9)  Append in ;
(10)  Obtain the next state , reward , and the next state’s
(11)  ;
(12)  ;
(13)  Take decay;
(14)end for
(15)end for
(16) Find the sequence with lowest cost in ;
(17)return;
6.1.2. Data Center Specification

The real-world information of data center, including storage price, out-bandwidth price, Get operation price, and latitude and longitude, is collected from CSPs official website [20, 22, 23, 35]. We use 18 data centers from different CSPs, and each CSP’s data center is located in a different city. We also simulate the availability values of each data center in the [95.0%, 99.9%] interval.

6.1.3. Experiment Parameters

In the experiments, we choose -erasure coding as the data splitting method. The size of sliding window for LSTM is 12. The size of data object is initially 200 GB. The required availability and data retrieval latency are initially and 1000 ms, respectively.

6.2. Performance of ADPA

In the time-varying DAF data placement, we cannot obtain the global data placement solutions because of the absence of future DAF. In our method, we firstly predict future DAF for the global optimal data placement solutions during the life-cycle of data storage. In order to verify the performance of ADPA, we propose a step-by-step optimization algorithm called SOA as follows.

6.2.1. SOA Algorithm

This is shown in Algorithm 4. This method minimizes the total cost through Greedy. At the beginning of time slots, the method finds the data placement solution with the lowest basic cost in all candidate solutions. Then, the data objects are allowed to the solution with the lowest total cost at the current time slot (Line 3). In fact, the optimal solution of SOA is locally optimal.

Input: The state matrix, , the transfer matrix, .
Output: The optimal data placement solution during ,
(1) Calculate data placement solution with the lowest
(2)for t = 1 to T-1 do
(3) Find the solution with ;
(4)end for
(5)
(6)return;

Based on the DAF prediction, we can obtain the global optimal data placement solutions, but the solutions of Greedy are locally optimal. In order to verify this situation, we study the total costs of ADPA and SOA by varying data size and data retrieval latency, respectively, as shown in Table 3 and 4. Obviously, the results of ADPA can save , , , , and comparing to Greedy with data size from 100 GB to 500 GB, respectively. Due to the absence of future DAF, SOA only chooses the best solution at current time slot and ignores the importance of future total cost. Conversely, ADPA, based on Q-learning, chooses the current solution which is based not only on current costs, but also on future reward.

As shown in Table 4, we study the results of two methods with the data retrieval latency constraint from 200 ms to 500 ms, the number of time slots is 12 (i.e., 2 hours), and data size is 200 GB. As latency constraint increases from 200 ms to 500 ms, the cost saving of ADPA is . When data retrieval latency is loose, comparing with strict latency, the method explores the solution with lower cost in a larger range. The algorithm explores solutions in the area near users to satisfy the strict latency constraint.

In actual life, there are different levels of data access and these situations have different usage scenarios. In order to verify the universality of the ADPA algorithm, we extend the time range of NASA-HTTP request data to one year (24/Oct/1994–11/Oct/1995) and perform DAF statistics based on one-day cycles. Due to the wide time range of the second data set, it has a lower DAF than our first data set. The experimental environment is the same; we use the second dataset to conduct a comparison experiment as shown in Tables 5 and 6. Similarly, it is clear that ADPA can help users save more money.

6.3. Performance Comparison with Other Methods

We compare our method with ACO and DP through the following experimental scenarios. Assume that ACO and GA can obtain the DAF prediction which is predicted by ADPA.

6.3.1. Cost Saving with Varying Data Size

In this experimental scenario, we evaluate the cost performance of ADPA when data size is increased from 100 GB to 500 GB. From the experimental results, we can know that the total cost of three algorithms increases when the data size increases, and the costs of the three algorithms obtained under different data sizes are the same. Since the DP algorithm can find the optimal solution, it can prove the correctness of ADPA in solving the solution. To demonstrate the advantages of ADPA, the time required to find the optimal solution for the three algorithms is shown in Figure 5. It can be seen that the time required by ACO is greater than that of ADPA and DP algorithm. The algorithm ADPA is higher than DP when the data size is 200 GB, and the others are better than DP.

6.3.2. Cost Saving with Varying Time Slot Count

In this scenario, we explore the impact of time slot count () on total costs when varies from 10 to 18. From these experimental results, we can know that the greater the value, the longer the optimization time and the costs of the three algorithms obtained under different time slot count () are the same. Because of the same reason as above, this proves that ADPA has the ability of solving the solution. In order to reflect the advantages of ADPA, the time required to find the optimal solution for the three algorithms is shown in Figure 6. It can be seen that, for ADPA and DP, except for T = 10, the running time of the algorithm proposed in this paper is only 50% of the DP algorithm; that is, ADPA can solve the optimal solution in the least time.

6.3.3. Cost Saving with Varying Candidate Data Centers

We investigate the impact of the number of candidate data centers () on total cost. Figure 7 gives the results when increases from 12 to 15. The change in candidate data centers count is adding a new data center in the previous set. For example, when , the index of data centers is and we add a new data center with index of 13 to this set when . When increases, the candidate solutions have a clear growth trend. Assuming that data retrieval constraint is 500 ms, , data size is 500 GB and data availability is 0.999; are 792, 1287, 2002, and 3003 when varies from 12 to 15. When N = 12, 13, ADPA and ACO can solve the optimal solution, but when N = 13, 14, the cost of the ADPA and ACO is higher than the optimal scheme by $0.99 and $1.23, respectively. However, from Figure 7, as is manifested, the running time of ADPA is less than that of the other two algorithms in all N values; that is, the efficiency of solving the optimal solution is higher than that of ACO and DP. In this section, we also perform a comparative experiment on the dataset with lower DAF and achieve good results, as shown in Tables 7 and 8.

7. Conclusion

Users should adjust data placement solution based on DAF to minimize the storage, get operation, out-bandwidth, and migration costs in the whole life-cycle of data storage. In order to achieve this goal, we present an adaptive data placement architecture named ADPA. Because of the absence of the future DAF, the DAF prediction module based LSTM of ADPA can predict the future DAF through historical data. And then, the data placement optimization module, which is based on Q-learning of reinforcement learning, solves the optimal data placement solution sequence according to the prediction DAF. The experiments driven by a real workload of NASA-HTTP and cloud providers information indicate that ADPA not only is superior to the algorithm SOA but can save more time than ACO and DP to obtain the optimal data placement. In the future, we intend to present an architecture which can adjust data placement based on the change of cloud market.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Disclosure

Pengwei Wang and Dong Wang are corresponding authors.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was partially supported by the National Natural Science Foundation of China (NSFC) under Grant no. 61602109, DHU Distinguished Young Professor Program under Grant no. LZB2019003, Shanghai Science and Technology Innovation Action Plan under Grant no. 19511101802, Natural Science Foundation of Shanghai under Grant no. 19ZR1401900, Fundamental Research Funds for the Central Universities, and Alliance PlanSpecial Tenders For Difficult Problems under Grant no. LM201819.