Abstract

As an extension of cloud computing, the edge computing has become an important pattern to deal with novel service scenarios of the Internet of Everything (IoE), especially for the rapidly increasing different kinds of service requests from edge equipment. It is still a great challenge to satisfy the demands of delay-sensitive applications, so as to optimize the service provision delay for edge equipment under 5G. In this paper, by introducing virtualized service functions (VSFs) into edge computing pattern, the mechanism of service function detection and placement among multiple Edge Computing Servers (ECSs) is proposed. We firstly improve the Ant colony optimization (ACO) method to adapt to the situation that the service requests may frequently change from different edge network domains. Based on the improved ACO, a scheme of searching for the locations (i.e., ECSs) of the requested service functions is devised, so as to optimize the service searching delay. Then, a service function placement scheme is presented, and it deploys most of appropriate service functions in each ECS by predicting the future requested frequencies of functions, which further reduces the overall service provision delay. In addition, it also improves the ECS computing capacity utilization. The simulation results show that the proposed mechanism is feasible and effective.

1. Introduction

The 5th-generation mobile system (5G) is gradually integrating into people’s daily life. The requirements of the novel service scenarios brought by the 5G have changed significantly [1]. For example, the delay sensitivity has become one of the most important service demands of edge equipment. Many types of research have been done on providing services mainly by the cloud computing center due to its powerful computing capacity, which enables almost all kinds of virtualized service functions (VSFs) [2, 3] to be placed and performed there. However, the cloud computing servers are usually located far away from most of network edges. With the rapid increasing service requests from mobile equipment that mainly locates at the network edges, the service provision based on the conventional cloud computing may lead to problems such as much higher transmission delay and serious congestion [4]. As an extension of cloud computing, the edge computing deploys the service provision capability near to network edges where most of service requests generate [5, 6]. Its distributed edge computing servers (ECSs) locate much closer to edge equipment so as to adapt to the service provision delay demand.

The 5G is able to support high-speed data transmission, which significantly decreases the service delivery delay of ECSs [7]. However, comparing to the service cluster (SC) in the cloud computing center, each ECS’s limited computing capacity means that it can only provide some kinds of services due to the limited number of VSFs placed in it [8]. It is hard for an ECS to satisfy all kinds of service requests for edge equipment, because the corresponding VSFs may have not yet been placed in this ECS when such requests arrive. And the method, by which all such service requests are dealt with by the remote SC or the corresponding VSFs are instantly migrated from the remote SC to this ECS, is obviously unsustainable, especially when the network load is heavy [9]. Considering the fact that other ECSs locating much closer than the SC may have already been placed such VSFs, the corresponding services can be provided by one of these ECSs. How to effectively find an ECS that currently contains a certain VSF becomes a key issue. In practice, a tremendous amount of different kinds of service requests are constantly generated and delivered, which leave the changing traces of recent service provision information in corresponding equipment [10]. In this paper, the VSF concentration on links is presented by leveraging the above changing traces, based on which an improved ACO-inspired service function detection scheme is devised further to effectively search for the placed locations (i.e., nearby ECSs) of the requested VSFs. It optimizes the concentration updating efficiency to overcome the high delay problem caused by several rounds of iteration of the conventional ACO method.

In addition, an ECS should adapt to the frequent changes in service requests by deploying appropriate VSFs in it in time. It is impossible for an ECS to contain all VSFs due to its limited computing capacity [9]. In fact, a lot of service requests are regional and periodic for an ECS [11]. That is, an ECS may often provide some certain kinds of services due to the working features of its nearby equipment [12]. However, the ECS also may be requested the VSFs that have not yet been placed in it because new applications become popular recently. We take the recent frequencies of the VSFs being performed and requested into account, and the VSF placement scheme is devised in this paper. It enables an ECS to retain the already deployed VSFs with higher being performed frequencies in it and migrate the not yet deployed VSFs with higher being requested frequencies to it, respectively. In this approach, the VSF detection delay and the service delivery delay can be further optimized.

Some types of research have been done on service placement and migration over edge computing scenarios (e.g., [1316]). In [13], it introduces a framework for optimal placement of service replicas proactively in the 5G edge network. It deploys the multimedia service instances on the trajectory edge nodes by integrating the user’s path prediction model, so as to provide an optimal deployment technique that traded off between maximizing the QoE and minimizing the deployment cost. In [14], it studies the container-based service migration problem in edge computing and proposes a service migration mechanism based on mobility awareness. Its service migration mechanism firstly triggers the service migration according to the service density of the current node and then selects the service and the corresponding destination node for the optimization object to minimize the service delay. In [15], it proposes a novel service migration policy method based on deep reinforcement learning and dynamic adaptation in multiaccess edge computing. It innovatively analyzes the different states of edge network quantitatively and applies deep Q-learning to migration methods, which can adjust the learning rate adaptively to implement rapid convergence in the learning process. In [16], it combines prediction mechanism and migration research together to optimize the migration of VNF. It built a system cost evaluation model integrating bandwidth utilization and migration time, and devised a heuristic algorithm to obtain the near-optimal solution. However, they mainly focus on instantly deploying functions on real-time service demands or migrating functions based on long-term iterative learning, which cannot well adapt to the service requests with frequent changes and delay sensitivity.

In this paper, the mechanism of ACO-based VSF detection and placement (AVDP) is proposed, so as to minimize the service provision delay to the edge equipment and optimize the ECS computing capacity utilization. The major contribution and innovations can be summarized as follows. The VSF concentration on links is presented to reflect the frequencies of VSFs being detected recently, and the improved ACO-inspired VSF detection scheme based on the VSF concentration is devised to efficiently search for the placed locations of the requested VSFs. It adapts to the novel service scenarios that the service requests for edge equipment and the deployed locations of VSFs may frequently change under 5G. Furthermore, the scheme of dynamically deploying appropriate VSFs in each ECS is devised, and it places VSFs by predicting the future requested frequencies of VSFs according to the variations of VSFs concentration on links and VSFs requested number in ECSs. Thus, the overall service provision delay is significantly optimized and the ECS computing capacity utilization is efficiently improved.

The remainder of this paper is structured as follows. In Section 2, we present the system framework and workflow of the proposed AVDP. In Section 3, we define the VSF detection ant and the VSF concentration, and devise the scheme of searching for VSFs among multiple ECSs based on them. In Section 4, we devise the scheme of placing appropriate VSFs in suitable ECSs with the ECS computing capacity utilization considered. In Section 5, we present simulation experiments and results. Finally, Section 6 concludes the paper.

2. System Framework

The system framework is shown in Figure 1. The conventional cloud computing center is usually located far away from network edges. It can provide almost all kinds of services due to the powerful computing capacity of the service cluster (SC), and VSFs placed in it can be dispatched to ECSs. ECSs are distributed in end network domains (ENDs) where most of service requests generate to efficiently meet the delay sensitivity demand of edge equipment. The service function pool (SFP) in each ECS only contains some VSFs because of the limited computing capacity. The VSFs already deployed in an ECS can be replaced by other VSFs due to the changing service scenarios in practice, and VSFs can be migrated from the SC or other ECSs. In addition, each switching equipment (SE) updates a table named VSFs detection record (VDR). The VDR is to reflect the current probabilities of successfully finding the locations of the requested VSFs by different next hops, so as to efficiently determine the suitable ECSs that can provide corresponding services.

The overall workflow of the system is shown in Figure 2; here, the number is the action order.

3. VSFs Detection

In order to minimize the service provision delay for edge equipment, most of services should be provided by the near ECSs rather than mainly by the SC that is far away from ENDs. It is a key issue to find the ECSs that have already been placed in the requested VSFs as soon as possible. In this section, the VSF detection ant and the VSF concentration are defined in our proposed scheme. The VSF detection ants are sent by the source SE that receives the service request to search for the locations (i.e., ECS) of the requested VSF according to this VSF concentration on links. The information on detecting the VSF is laid over the ants’ trails, and the accumulated amount of information is regarded as the VSF concentration that is recorded and updated in the VDR.

In practice, the VSF concentration on links decreases with time. It also increases according to the ants’ feedback if these ants have found the locations (i.e., ECSs) of the requested VSF. For instance, if the requested VSF is found by any of the VSF detection ants within the ant’s living period, the ant returns back to the source SE, and the VSF concentration on the corresponding links increases. Comparing to the conventional ACO method, our presented VSF concentration updating does not depend on multiple iterations of the same group of ants. In the actual network environment, because of a large number of ongoing service requests for edge equipment, different VSF detection ants can be generated by SEs at any time, which enables different VSF concentrations continuously and efficiently to change in practice combined with the devised detection ant’s feedback method.

3.1. The VSF Detection Ant

In this paper, a VSF detection ant is defined as a four-tuple . Here, is the unique identifier of a detection ant; indicates the VSF that is detected by ; is the set of SEs that has passed; denotes the number of survival hops of .

In detail, is one of the ants generated by the source SE that receives the service request and is forwarded to search for according to the concentration on links. The SEs passed by are orderly recorded in to avoid loopback. Meanwhile, once finds , it returns back to the source SE according to . In addition, an ant is not allowed to detect the ECSs that are far away from the source SE due to the delay limitation. is used to constrain the survival time of , and dies without any feedback if it has been forwarded exceeds .

3.2. The VSF Detection Concentration

We assume two factors will influence the VSF concentration on links, which are time and ants’ feedback. is defined as the concentration on the link from the SE to the SE at time t, and its value changes with time. At time t + 1, its value is shown as follows:

Here, and are the remaining concentration after volatilizing and the additive concentration after ants’ feedback from t to t + 1, respectively. They are defined as follows:

Different from the conventional ACO method, the VSF remaining concentration volatilizes faster and faster with time, because the frequent changing service requests from edge equipment may lead to the frequent migrating of VSFs. is designed to be continuously differentiable, and is a positive constant. The VSF additive concentration can only be brought by the survival ants that have found the requested VSF. The closer the link to the ECS that contains the requested VSF, the more this VSF concentration on the link increases. is designed to promote the searching efficiency for the following ants. m is the number of the detection ants that have returned back between t and t + 1. And is related to the sequential position of in , and it is defined as follows:

Here, is the element number of , and is the position of in .

3.3. The Forwarding Probability

is defined as an adjacent SE of if a detection ant can be forwarded by to by only one hop. Not all adjacent SEs of can receive the detection ants from due to the limited number of the ants generated by the source SE. According to , we define is the set of adjacent SEs of , and is the set of adjacent SEs that can receive the detection ant (e.g., ) at the next hop from , which are shown as follows:

When searching for , stops searching without any feedback if is empty. We assume is the forwarding probability that is forwarded from to , . Here, the adjacent SE with higher only means that it can be forwarded more detection ants rather than be forwarded all detection ants. The is defined as follows:

Especially, .

The working process of searching for the is shown as Algorithm 1. In detail, firstly, when an SE receives the service request, as the source SE, it generates a certain number of detection ants to search for the requested VSF that is here (line 2). Secondly, the same kinds of ants are divided into several groups in each arriving SE, with of each ant being considered to avoid the loopback, the numbers of ants in different groups are determined according to the current concentrations of on different links. Then, different groups of ants are forwarded to adjacent nodes of the current node (lines 3–5). Thirdly, if an detection ant finds , it adds the current node into its and returns back to the source SE. Meanwhile, the concentrations on the corresponding searching links update and the current node (i.e., ECS) is added into (lines 6–11). Else if of the ant is not zero, the ant adds this node into its and joins to the next-hop searching (lines 12–14). Otherwise, the ant dies without any feedback (lines 15–16). Finally, the set of ECSs that are found currently contain (i.e., ) is obtained (line 20).

Input: /the source switching equipment/, /the service function that is requested /
Output: /the set of ECSs that contain , /
(1)Begin
(2)   generates detection ants, each of which is defined as ;
(3)  while the set of detection ants is not empty do
(4)  Divide ants into several groups according to the of each ant and the distributed concentration
(5)  Forward different groups to different adjacent nodes of the current node according to equation (5);
(6)  for each do
(7)   if is found then
(8)    Add this node into ;
(9)     Return back to according to ;
(10)    Update concentrations on links according to equation (2);
(11)    Add this node (i.e., ECS) into ;
(12)   else ifthen
(13)    Add this node into ;
(14)    Join the next-hop searching;
(15)   else ifthen
(16)    Die without any feedback;
(17)   end if
(18)  end for
(19)end while
(20)return
(21)End

4. VSFs Placement

In order to optimize the service delivery delay, the ECSs locating near to the ENDs where the service requests generate should have been deployed in or migrated to the requested VSFs. In this paper, two kinds of VSFs that should have been placed in each ECS are considered, which are deployed VSFs (DVSFs) and migrated VSFs (MVSFs). DVSFs are the VSFs that have already been deployed in the ECS and are still being massively requested recently. MVSFs are the VSFs that should be migrated to the ECS due to the rapidly growing requests for them recently. Therefore, due to the ECS limited computing capacity, the VSFs that have already been deployed but rarely be requested recently in this ECS should be replaced by the MVSFs.

The performed frequencies of the already deployed VSFs in an ECS can be estimated according to these VSFs’ current concentrations on the links around this ECS. Assume that is defined as the set of adjacent SEs of the ECS, . The ratio of performed frequency of to the performed frequencies of all already deployed VSFs in from t to t + 1 is defined as , shown as follows:

Here, is the set of VSFs that have been performed in from t to t + 1.

Let be the concentration stability of on the links around during the recent m time periods, shown as follows:

Assume and are the threshold values of the VSF’s ratio of performed frequency and concentration stability, respectively. The current set of DVSFs that can be retained in is defined as , obviously, . The VSFs in should satisfy the following conditions:or

Let be the set of candidate VSFs that have been requested but not yet been deployed in during the recent m time periods, these VSFs may replace the VSFs in with satisfied. Let be the current set of MVSFs that should be migrated to , obviously, . For the VSF , its total requested number in during the recent m time periods is defined as , and its incremental requested number in in each recent time period is denoted as follows:

Furthermore, the requested number growth rate of in in the day is denoted as follows:

In this approach, the VSFs in can be selected from by comparing the VSF’s requested number growth rate from high to low with the ECS’s available computing capacity considered. And these VSFs can be migrated from the SC in the cloud computing center or other ECSs. In addition, a VSF in may become one element of if it satisfies the conditions of equations (8) and (9), or equations (10) and (11) in the following time periods. Therefore, the VSFs that should be placed in before the next time period are the ones belonging to .

5. Simulation and Results

5.1. The Simulation Setup

In the simulation, the proposed schemes are implemented in Python and all experiments are performed on a computer with one Intel(R) Core(TM) i7-6700 CPU @ 3.40 GHz and 16 GB of RAM. We evaluate the approaches in two typical and real network topologies (e.g., ISP and backbone networks) with different numbers of nodes and links, called Geant and Interroute, which can be obtained from the Internet Topology Zoo [17]. Specifically, Geant is a middle-scale network topology with 41 nodes and 65 links, and Interroute is a large-scale network topology with 110 nodes and 148 links. They are shown in Figure 3.

We assume that the network topology is divided into 6 edge network domains, and one ECS is placed among 10 nodes in the simulation. A variety of VSFs are simulated by ClickOS [18], and it can create small virtual machines, each of which is able to host a VSF. We assume that the computing capacity of each ECS follows a uniform distribution between 100 and 200 units. The number of VSFs requested by each request follows a uniform distribution between 2 and 4, and the type of each VSF is random. The ECS computing capacity needed to support each VSF follows the uniform distribution between 5 and 10 units. The simulation parameters and the corresponding distribution model are motivated by the literature [19, 20] that studies the network function provision problem. We compare the proposed AVDP with the scheme of AI-enabled mobile multimedia service instance placement (AMSP) according to [13]. We use the following performance metrics: the service provision delay (SPD), the computing capacity utilization (CCU), and the service access success ratio (SASR).

5.2. The Simulation Results

We compare the SPDs of the two approaches of AVDP and AMSP under Geant and Interroute. The SPD is defined as the time interval from receiving the service request to successfully providing the service. The results are shown in Figures 4 and 5.

As shown in Figures 4 and 5, the SPD of AVDP is always lower than that of AMSP with the number of service requests increasing (i.e., Figure 4). In addition, we also compare the SPDs of the two approaches with the time period increasing under the network load of 10000 service requests (i.e., Figure 5). In more detail, when the number of service requests is low, the SPDs of AVDP and AMSP are 0.079 s and 0.085 s, respectively. When the number of service requests reaches the maximum, the SPD of AVDP just increases by 0.061 s, while the SPD of AMSP increases by 0.125 s. Moreover, when the network load is the heaviest, the SPD of AVDP reduces to 0.132 s with the time period increasing; that is, the SPD of AVDP is optimized with time, while that of AMSP increases to 0.235 s in the same situation. The reasons are as follows. In AVDP, the improved ACO method enables the VSF concentration on links to be updated in time by efficiently accelerating volatilization and enhancing feedback. Thus, the requested VSFs can be quickly found from the nearest ECSs, which significantly reduces the time overhead of searching for services. Furthermore, AVDP continuously optimizes the deployed locations of VSFs over time, which enables the VSF searching delay of AVDP to be reduced over time. Thus, the SDP of AVDP is further improved. However, AMSP barely changes over time, it mainly provides services on demands by instantly deploying service functions rather than leveraging already deployed functions. Real-timely deploying most of requested service functions leads to large delay for AMSP to provide new services, especially when the network load becomes heavy.

We compare the CCUs of the two approaches of AVDP and AMSP under Geant and Interroute. The CCU is defined as the ratio of the VSFs performed in an ECS to the total already deployed VSFs in this ECS. The results are shown in Figures 6 and 7.

As shown in Figures 6 and 7, the CCU of AVDP increases with the number of service requests increasing, while that of AMSP decreases when the network load becomes heavy (i.e., Figure 6). We also compare the CCUs of the two approaches with the time period increasing under the network load of 10000 service requests (i.e., Figure 7). In more detail, when the number of service requests is low, the CCUs of AVDP and AMSP are about 77% and 84%, respectively. With the network load becoming heavy, the CCU of AVDP increases and it reaches about 94% under the network load of 10000 service requests, while that of AVDP drops to 72% at the same network status. Moreover, when the network load is the heaviest, the CCU of AVDP increases from 71% to 93% with the time period increasing while that of AMSP is always keeping about at 72%. That is, the CCU of AVDP continuously optimizes while that of AMSP barely changes over time. The reasons are as follows. AVDP dynamically deploys VSFs by taking the frequencies of the VSFs being requested and performed in an ECS into account. With the number of service requests increasing, the VSFs rarely performed recently can be replaced by the ones with higher requested frequencies in an ECS; thus, the computing capacity of each ECS can be fully used. In addition, the requested VSFs can be efficiently found by AVDP, which enables the appropriate VSFs to be more frequently performed in each ECS. Thus, the CCU of AVDP is significantly optimized. However, AMSP mainly makes full use of some VSFs that are currently requested frequently in an ECS, and it does not consider replacing already deployed VSFs that are rarely requested by new popular VSFs, which cannot further optimize the ECS computing capacity utilization. In addition, AMSP does not have the ability to achieve the VSF future popularity prediction, and the CCU of AMSP cannot be improved over time.

We compare the SASRs of the two approaches of AVDP and AMSP under Geant and Interroute. The SASR is defined as the ratio of the requests that successfully obtain services to the total requests asking for services. The results are shown in Figures 8 and 9.

As shown in Figures 8 and 9, the SASR of AVDP is much higher than that of AMSP when the number of service requests increases rapidly (Figure 8). We also compare the SASRs of the two approaches with the time period increasing under the network load of 10000 service requests (Figure 9). In more detail, when the number of service requests is low, the SASRs of AVDP and AMSP are almost approaching 1. With the number of service requests increasing, the SASRs of AVDP and AMSP drop to about 84% and 63%, respectively. However, the SASR of AVDP still keeps beyond 84%, especially under the heaviest network load. Moreover, the SASR of AVDP increases from about 72% to about 84% with the time period increasing, while that of AMSP just improves a little and keeps at about 65% over time. Under the heaviest network load, the SASR of AVDP can achieve about 20% higher than that of AMSP. The reasons are as follows. AVDP mainly deploys most of VSFs before these VSFs are massively requested, which does not occupy much network resource to real-timely dispatch the requested VSFs. Thus, the available network resource can support as many new requests as possible, especially under the heavy network load. In addition, in AVDP, multiple ECSs that contain the requested VSFs can be found, which increases the probabilities that successfully providing corresponding services to new receiving requests when most of ECSs are busy. Thus, the SASR of AVDP is significantly optimized. However, AMSP mainly provides services by the server on the mobile path without considering balancing the working load by cooperating with multiple nearby servers. Under the heavy network load, the SASR of AMSP decreases rapidly.

6. Conclusions

In this paper, by introducing VSFs into edge computing pattern, we propose the mechanism of improved ACO-inspired VSFs detection and placement. By efficiently detecting the already deployed locations of VSFs and dynamically placing appropriate VSFs in suitable ECSs, the service provision delay to the edge equipment and the computing capacity utilization of each ECS under the edge computing scenario are significantly optimized. In this mechanism, the approach of searching for the requested VSFs from multiple ECSs is devised, and it improves the ACO method to adapt to the edge computing scenario by defining the VSF detection ant and the VSF concentration on links. Furthermore, the approach of deploying VSFs in each ECS is devised, and it takes the frequency variations of the VSF being performed and being requested in an ECS into account, so as to select the most appropriate VSFs to place in the ECS with this ECS computing capacity utilization considered. Simulation results show that the proposed mechanism has significant improvements in the service provision delay optimization and the computing capacity utilization improvement compared with the current state of the art.

Data Availability

The data used in this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (nos. 62002261 and 62072336) and the Tianjin Municipal Education Commission Scientific Research Project (no. 2018KJ145).