Abstract

The video traffic offloading in edge networks is an effective method for remission of congestion of backward paths in 5G networks by continual optimization of video distribution to promote scale and efficiency of video delivery in edge networks (e.g., D2D-based near-end sharing). Because the video resources are dispersedly cached in local buffer of mobile devices of video users, the management of local video resources of video users in edge networks (e.g., caching and removing of local videos) causes dynamic variation of video distribution in networks. The real-time adjustment of local resources of users in terms of the influence levels (e.g., promotion and recession) of video sharing performance is significant for the continual distribution optimization. In this paper, we propose a novel Social-aware Edge Caching Strategy of Video Resources in 5G Ultra-Dense Network (SECS). SECS designs an estimation method of interest domain of users, which employs the Spectral Clustering to generate initial video clusters and makes use of the Fuzzy C-Means (FCM) to refine the initial video clusters. A user clustering method is proposed, which enables the users with common and similar interests to be clustered into the same groups by estimating similarity levels of interest domain between users. SECS designs a performance-aware video caching strategy, which enables the users intelligently implement management (caching and removing) of local video resources in terms of influence for the intragroup sharing performance. Extensive tests show how SECS achieves much better performance results in comparison with the state-of-the-art solutions.

1. Introduction

The video services (e.g., video-on-demand and living video streaming) provide rich viewing content for video users by making use of mobile smart devices to ubiquitous access to the Internet [15]. The smooth and high-definition watching quality enables the video users obtain great experience, which requires higher bandwidth and lower delay to support video data delivery from video providers to video requesters [68]. The new generation of communication technology 5G relies on bandwidth expansion and transmission acceleration to provide support in capacity and velocity for smooth and high-definition experience [911]. Moreover, the 5G makes use of the ultra-dense deployment to promote network coverage and access capability, which supports more video users to ubiquitously fetch video content via the 5G networks.

The excellent experience and convenient access of video services not only attract the large-scale video users but also speed up the increase of user scale. Obviously, the fast expansion and desired high-quality experience of user scale trigger huge bandwidth consumption, so that the unbalance between supply and demand of bandwidth results in the severe network congestion. In particular, the backhaul paths in the 5G networks inevitably are subjected by the network congestion because of dense access. The high startup delay and unbearable packet loss caused by the network congestion lead to unsmooth playback and video picture distortion, which brings severe negative influence for user quality of experience (QoE). Therefore, offloading video traffic in edge networks without intervention of 5G nodes is significant for relieving congestion levels of backhaul networks.

D2D communications enable mobile devices implement data transmission without intervention of 5G nodes, which become the important method for traffic offloading in underlaying networks [1214]. As Figure 1 shows, the video users not only make use of the 5G networks accessing to the Internet to watch video content but also employ the D2D communications to fetch the video resources from other mobile devices carrying the corresponding videos. The successful traffic offloading based on the D2D communications relies on the two perspectives: (1) search of D2D matching objects which cache video resources requested by the video request nodes; (2) delivery of video data using D2D communications between request and supply nodes of video resources. The successful search of D2D matching objects which cache video resources is the important precondition for the successful delivery of video data using D2D communications. On the other hand, the fast search of D2D matching objects can efficiently reduce startup delay of video request nodes, which becomes the key factor for ensuring high QoE in terms of delay-sensitive property of video services. The successful and fast search of video providers in the range of D2D communications is vital for the offloading video traffic in the 5G networks.

The video resources are carried by the mobile devices and are distributed in networks. The users cache and remove video resources in terms of the interests, so that the dynamic distribution of videos in networks brings severely negative influence for the success ratio and time of searching video providers. On the other hand, the video resources move with the movement of mobile devices, which are dynamically distributed in geographic area of the edge networks. Therefore, the geographical movement and local replacement of video resources are main reasons of dynamic change of video resource distribution: (1) the replacement based on the self-interest change results in the increase in the risk of video search failure (the search failure increases the startup delay of video request users); (2) the geographical movement of video resources results in the decrease in the probability of D2D pairing success with one-hop neighbor relationship (the D2D pairing failure means that the video request users only make use of the multi-hops transmission to fetch video data instead of D2D communications with one-hop transmission). The numerous researchers always focus on making use of resource distribution optimization based on caching management to promote efficiency of resource sharing. For instance, Mehrabi Liu et al. propose a joint QoE-traffic optimization with collaborative edge caching by investigation of impact of collaborative mobile edge caching for both QoE and backhaul data traffic [15], which implements self-tuned bitrate selection to make the decision efficient cache replacement strategy. Thar et al. propose a deep learning-based prediction scheme, which achieves intelligent management of resource leasing and caching [16], which can make the prediction for resource leasing and caching using the Keras and Tensorflow libraries. Jiang et al. propose a cooperative content caching policy based on multiagent reinforcement learning for the architecture without preference and historical content demands of users by designing a cooperative content caching scheme to solve the cooperative content caching problem [17]. However, the most of existing methods neglect the association relationship between interests and behaviors (requesting and caching) of edge users, which results in severe turbulence of resource distribution by the arbitrary replacement of local cached resources. The sharing performance suffers severe negative influence and the traffic load of 5G network cannot be effectively relieved. Therefore, the real-time adjustment of video distribution in terms of variation of video sharing performance is very significant for effect of D2D-based traffic offloading and user QoE.

In this paper, we propose a novel Social-aware Edge Caching Strategy of Video Resources in 5G Ultra-Dense Network (SECS). SECS designs an estimation method of interest domain of users: (1) SECS employs the Spectral Clustering method to generate initial video clusters in terms of similarity between videos by investigating video content and access behaviors of users; (2) SECS makes use of the Fuzzy C-Means (FCM) to refine the video clusters by redefining the distance between videos. The refined video clusters can be considered as the interest domain of users. A group construction method of clustering users is proposed, which relies on estimation results of similarity levels of interest domain to aggregate users, which enables users with common and similar interests to be clustered the same groups. A performance-aware video caching strategy is proposed, which implements intelligent management of local cached videos by estimating influence for the sharing performance in user groups from the three perspectives: promotion of video sharing scale, reduction of response delay, and motivation of near-end sharing. Extensive tests show how SECS achieves better results in comparison with other state-of-the-art solutions in terms of caching hit ratio, caching cost, response delay, and control overhead.

The numerous researchers have focused on content edge caching methods in recent years. Qiao et al. propose a joint optimization method for the content placement and content delivery in the vehicular edge networks by making use of the trilateral cooperation communications among macrocell station, roadside units, and smart vehicles [18]. The joint optimization problem is formulated as a double time-scale Markov decision process. On the large time-scale, the content placement relies on content popularity, vehicle driving paths, and resource availability; on the small time-scale, a scheme based on vehicle scheduling and bandwidth allocation is designed, which decreases the content delivery delay. A deep deterministic policy gradient framework is proposed, which obtains a suboptimal solution corresponding to the long-term mixed integer linear programming problem. However, the formulated optimization problem relies on the precondition that the time-scale of content timeliness changes less frequently during the content delivery process. Kwak et al. propose a hybrid content caching method without the knowledge of content popularity [19]. The content caching location algorithm is designed, which supports average requested content data rates following finite service latency. By employing the Lyapunov optimization approach, a caching control problem with tight coupling between CU caching and BS caching is formulated and solved. By employing the submodularity property of the sum-weight objective function, the practical and heuristic CU/BS caching algorithms are proposed, which deals with a general caching scenario. Liang et al. focus on solving utility-oriented service entity caching problem in edge networks [20]. The positive impact brought by clients from caching a service entity in an edge server is defined as the utility in terms of variation in changed specific scenarios. A utility-based service entity caching problem is formulated, which is extended to other problems via redefinition of utility. The utility problem is proved as a NP-complete problem and is solved by a designed approximation algorithm. Zhang et al. propose an edge caching framework in integrated content-centric mobile 5G networks, which makes use of content-centric networking to achieve efficient management of content-oriented information and promotion of content delivery efficiency [21]. The content caching architecture is consisted of function entities, protocol stack, content retrieval, and edge caching approaches; the edge caching performance also is demonstrated by the authors. Saputra et al. make use of the deep learning to design proactive cooperative caching approaches, which can predict content demand of users in a mobile edge caching network [22]. The content server in edge network is responsible for collecting information of mobile edge nodes and performs the deep learning algorithm to predict content demand of whole network. A framework based on the distributed deep learning is proposed, which allows the mobile edge nodes collaborate and exchange information. The framework can decrease prediction error for content demand and does not need revealing private information of users.

Zhang et al. propose a caching placement method with the multi-winner auction in D2D-enabled caching cellular networks by investigating edge caching incentive and content caching redundancy [23]. A multiwinner edge caching auction is modeled; an optimization problem of content caching revenue maximization is formulated. A semidefinite programming is designed, which obtains an approximate optimal caching placement to reduce the content caching redundancy in a UT movement scenario. A payment strategy with Nash bargaining game based on personal profit fairness is designed and a multiwinner repeated auction based caching placement algorithm is proposed, which reduces the complexity with tiny performance loss. Zhang et al. propose a cooperative caching strategy by construction of a two-tier heterogeneous network consisted of edge servers and caching helpers, which promotes utilization ratio of storage space and caching hit probability [24]. By making use of stochastic geometry and optimizing the caching probabilities of contents, a cooperative caching strategy is designed, which can maximize the hit probability and promotes the utilization ratio of local storage space. Zhang et al. propose a delay-optimal cooperative edge caching in wireless networks, which can optimize content placement and cluster size according to stochastic information of comprehensive measurement of network topology, traffic distribution, channel quality and file popularity [25]. The authors proposes a greedy content placement algorithm using bandwidth allocation optimization and a condition constraining the maximal cluster size based on tradeoff between caching diversity and spectrum efficiency. Liu et al. investigate heterogeneous context without preferences of user to implement management of content caching [26]. An online Bayesian clustering caching algorithm is designed, which maintains sustainable scalability by learning interactive cache hit data between users. The latent number of user groups is explained by a Dirichlet multinomial mixture model based on Bayesian generative framework (users with the same preference). Then, a dynamic clustering policy is proposed, which can obtain preferences of clusters by making use of a collapsed Gibbs sampling algorithm. A cache bandit algorithm is designed, which can support the formulation of cache decisions. Guan et al. propose an edge caching admission strategy of video content based on preference learning [27]. An information collector is designed, which collects the preference-related information without any modification of clients and video providers. A tree-structure model is designed, which learn and compress preferences of users. An explore-and-exploit method is proposed, which makes the decision of video caching. Xu et al. propose a secure edge caching scheme in mobile sensor networks [28]. A secure edge caching framework is designed, which is consisted of content provider, multiple edge caching devices, and some mobile users. An interaction model based on Stackelberg game between the content provider and edge caching devices is designed, where content provider (game-leader) determines payment strategy of secure caching service, and each edge caching device (game-follower) makes the decision for the quality of secure caching service. A zero-payment mechanism is designed, which deals with selfish behaviors of edge caching devices to stimulate content caching.

3. SECS Detailed Design

3.1. Interest Domain of Video Users

The interest for the video content is main reason to drive the video users requesting video resources. The various interests for the different video content make the video users dedicate fetching the desired video resources. However, the user interests for video content are dynamically variational. When the video content is in the range of desired video kind for the video users, the users send the video request to fetch and store the video resources; when the users have watched the video content, the user interest levels for the watched videos fast weaken, so that the watched videos may be removed in the local buffer. The variation of user interests determines the change of requested and cached videos, which brings the severe negative influence for the video distribution. This is as the video system dispatches the video resources to respond the video request of users and efficiently makes use of the buffer space (e.g., removing unpopular videos and caching popular videos) to support the large-scale access. Obviously, the interest analysis is important for understanding and predicting behaviors (e.g. request, caching and replacement of video resources) of video users, which supports beforehand or real-time adjustment of video resource distribution to balance supply and demand and promote video sharing efficiency. Table 1 lists the definition of symbols in the whole paper.

Measurement of user interests relies on the videos which have been watched by the users. However, the single video does not reflect the real interest of users. A watched video only denotes the user has interest for the video, which does not help the video system understand the intention that the user watches the video and predicts the video requested by the user in the future. If the watched videos are aggregated, the clustered video sets not only denote interest kind for the video content but also describe the boundary of interested video resources. Further, the interest domain can be used to estimate the request probability of video users for any video, predict the requested video resources in the future, and cluster the video users with the same or similar interest. The traditional video clustering methods make use of the structured video information to measure the similarity values between videos. For instance, a video can be denoted as , where is any attribute of (e.g., title, actor and abstract of video). If can be considered a vector, the similarity value between and can be defined as the angle cosine between two vectors and . However, the measurement of similarity between videos only investigates the relation levels between video content and does not involve the influence factors from the request behaviors of video users. For instance, the continuous two videos in a TV play series may have the similar actor, but the plot may be different for the two videos. The single measurement based on the content similarity cannot find the two videos belong to the same TV play series. The relation from the request behaviors of video users should be considered as the measurement parameter for estimation of video similarity. Therefore, the merged similarity based on content and request relation can be defined aswhere is the content similarity between and ; is the similarity of user access relation between and . In terms of the abovementioned traditional measurement method of video similarity, and can be the two vectors which include the same attributes, respectively. The angle cosine between the two vectors corresponding to and can be considered as the content similarity between and . The measurement method of value of needs to investigate the association relationship between and in the request behaviors of video users. Let denote the all logs of video request where any item in is the video access log of a user in the video system. If a log includes and and the location of and in is adjacent, and have a request association relation in . The value of can be defined aswhere is the number of request association relation of and and is the total number of all logs in . If the number of request association relation of and is 0, and . The similarity values among and all videos form a vector according to equation (1) where and is the total of all videos. All vectors corresponding to the videos can form an matrix . Because , the values of similarity in the diagonal line of are 0. We employ the Spectral Clustering method to cluster videos in terms of the matrix of video similarity [2931]. Because the Spectral Clustering method is well known, we briefly introduce the process of clustering video. The degree value of any video can be defined as the total sum of similarity values among and other videos, as follows:where is the total number of all videos. In fact, is the sum of all items in line of . The degree matrix of similar matrix can be built and be denoted as . The standardized Laplacian matrix can be calculated according to . The characteristic values of can be further obtained and are sorted according to ascending sequence of values of all items in , namely, . Any item in is equal or greater than 0. The subset in is extracted. The characteristic vectors are calculated corresponding to , where the dimension of any vector is . The characteristic vectors form a matrix , where is the number of videos and is the purposed number of video clusters. The most of Spectral Clustering methods employ the k-means method to divide the videos into clusters based on the matrix . However, the clustering exactitude level of Spectral Clustering method relies on the estimation accuracy of similarity between samples and the clustering performance of k-means method depends on the selection of center samples of clusters [3234]. Therefore, after using the k-means method, the video clusters can be obtained and be defined as . Each item in is denoted as , where is the centric item of .

The accuracy of video clustering is very important for generation of interest domain of users and estimation of video request probability of users. If the accuracy of video clustering is low, the video clusters include videos which are dissimilar with each other, so that the interests of users are not definitely obtained and the probabilities of video request of users are not precisely estimated. This leads to guideless caching and replacement of video resources bring severely negative influence for video sharing performance and user QoE. Therefore, the video clusters based on the Spectral Clustering are considered as the initial clusters and should be refined to promote the accuracy of video clusters (e.g., high cohesion between videos and low coupling between clusters). We employ the Fuzzy C-Means (FCM) method to refine the initial video clusters [35]. The objective function of FCM can be defined aswhere is the distance between two videos and and ; is the total number of video resources and returns the total of all items in video cluster set ; and is the membership of belonging to and can be defined aswhere is the membership between and centric item of and can be defined aswhere is the content similarity between and centric item of , is a weight value, and is the number of ’s items which have the co-occurrence relationship with . For instance, if and any item in jointly are included in any log, and have the co-occurrence relationship. In fact, is the number of items in which jointly are included with in logs. Obviously, and . reflects the continuity between videos for the request behaviors of users; investigates range that a video has the association relationship with the items in a video cluster for the requested content of users. Because the FCM makes use of recalculating membership values of between each item and all clusters to promote intracluster cohesion and reduce intercluster coupling after adjustment of centric item of cluster, the FCM has multiple refinement round and intraround recalculation. In order to clearly describe the refinement process based on the FCM, denotes value of objective function in equation (4), where is the number of refinement round and is the number of recalculation in the current round. The following process shows refinement of clusters in :Step 1: checks whether all items in are marked or not. If all items in have been marked as “refined”, implements Step 7; Otherwise, if all items in are not marked, extract a cluster which is not refined in the current refinement round from .Step 2: the objective function value of calculation of round is defined as . The item which has the minimum distance with centric node of is removed from .Step 3: the centric node of should be reselected because of removing an item in . After reselection of centric item of , the value of objective function needs to be recalculated and is marked as .Step 4: if , the removed item from can be considered as a noise item and is added into a noise set ; is marked and the mark status is “refined”; otherwise, if , the removed item from cannot be a noise item and is re-added into ; The original centric item of still is centric item of ; is marked and the mark status is “unrefined”.Step 5: If still includes “unmarked” items, returns Step 1; otherwise, implements Step 6.Step 6: if all items in are marked as “refined” or “unrefined” and the number of “refined” items is equal or greater than 1, the items in still need to be refined. The number of refinement round is . All items in are redefined as “unmarked” and returns to Step 1; otherwise, if all items in are marked and the number of “refined” items is 0, implements Step 7.Step 7: the current refinement iteration is terminated, and all items in are considered as “refined.”

Because there is only one nested iteration process of membership calculation based on matching video similarity in the above refinement process, the complexity of the above refinement process is . The initial video clusters in are refined in terms of the above refinement process, and the set consisted of refined clusters still is defined as . If the viewing log of a video user is defined as and the items in , respectively, belong to video clusters , the interest domain of is considered as . When a new video is added into , is considered as a member in the cluster where the centric item of has the largest similarity value with among all clusters. In order to avoid the negative influence for the accuracy of clustering videos from the new videos, after is added into , the centric item of should be reselected. and can be calculated, where and are the average membership between all items and centric item before and after joining into , respectively. If , is added into and the reselected centric item becomes the new centric item of ; otherwise, forms a new cluster and the original centric item still act as the centric item of .

3.2. User Group with Common Interest Domain

The interest domain of video users not only denotes range of requested videos but also can be used to measure and predict video sharing among users. For instance, if the two users have the same interest domain, they may supply desired videos with each other by pull and push; otherwise, if the two users have the different interest domain, the video providers cannot supply the requested videos for the video requesters and the video requesters cannot receive the pushed videos from the video providers. Video sharing between users with common interest range not only promotes QoS of video system and QoE of users but also increases utilization of cached resources and energy-efficiency levels of mobile devices. Classifying users with common interests into the same groups is very important for the promotion of video sharing performance.

Let . Each user has the definite interest domain and the interest domain directly shows content and range of representational preference of users. The set consisted of centric items corresponding to interest domain of users can be represented as the interest domain of users. The users which have the same or similar interest domain can be allocated into the same user groups, as follows.

Step 1: if , implements Step 5; otherwise, if includes the items which have the same interest domain, implements Step 2; otherwise, if the items in do not have the same interest domain, implements Step 3.Step 2: extracting an item subset from where the includes the most items among all the item subsets with the same interest domain. can be considered as a user category and an item is selected as centric item of . , where denotes the difference set between and , which means that the classified items are removed from . is added into the set ; return Step 1.Step 3: if (all items in do not have the same interest domain), implement Step 4; otherwise, an item in is selected and the similarity values of interest domain of and all subsets in are calculated according to the equation . and are the interest domain of and , respectively; and return the number of intersection and union of interest domain of and , respectively. joins into the subset which has the largest similarity value with among all subsets in . is removed from and returns Step 1.Step 4: each item in is considered as a subset, is added into , and is removed from ; returns Step 1.Step 5: the current iteration of user classification is terminated.

The above aggregation process can be described in Algorithm 1. Because there are the one nested iteration process of similarity calculation based on matching interest domain in Algorithm 1, the complexity of Algorithm 1 is . The users with common interest domain can be aggregated in terms of the above process and form a user group set . The users in each item of have the same or similar interests for video content. The union set of all users of each item in is considered as the interest domain of current user group.

1: /∗ is user set; is set of user groups;
 2: and are empty sets∗/;
 3: for (; ¡++)
 4:    is added into ;
 5:   for (; ¡++)
 6:    if has same domain with
 7:      is added into ;
 8:    end if
 9:   end for
 10:  if
 11:    is removed from and is added into ;
 12:  else items in are removed from ;
 13:    is added into ;
 14:  end if
 15:   is set to empty set;
 16: end for
 17: all items in are added into ;
 18: if is not
 19:  if is
 20:   for (; ¡++)
 21:     is a group and is added into ;
 22:   end for
 23: else for (; ¡++)
 24:    computes similarity values between and all centric items in ;
 25:    if centric item of subset has the largest similarity with
 26:      is added into ;
 27:    end if
 28:   end for
 29:  end if
 30: end if

4. Video Caching Management Based on Sharing Performance Awareness

The current mobile devices in edge networks have relatively high performance (e.g., fast computation and large storage). However, the development of video quality (e.g., blu-ray video) brings the fast increase of video size, so that the storage capacities of mobile devices become the relatively finitude. When the local videos occupy the large number of storage resources, the users have to remove some local videos in order to store new videos in the future. However, the caching and removing of local videos bring the immeasurable influence for the video resource distribution in the edge networks. For instance, when a user stores a video into local buffer, the supply capacity corresponding to the cached video is enhanced in the edge networks. The user not only makes use of the cached video to promote the supply capacity of upload bandwidth in the same user group, but also provides video data for users in edge networks. The sufficient supply of video resources can reduce the queued response delay of video request and promote the probability of near-end video sharing (e.g. D2D communications). However, when the popularity of a video decreases, the superfluous supply also wastes the local storage resources of mobile devices. On the other hand, when a user removes a video from local buffer, the video supply decreases regardless of user group or edge networks. The lacking supply of videos can increase the queued response delay of request and reduce the probability of near-end video sharing. However, when the popularity of a video decreases, the removing for the superfluous video caching can promote the use efficiency of local buffer. Therefore, the influence for the video sharing performance should be estimated before the caching or removing of local videos. We investigate the video sharing performance in terms of the three perspectives of promotion of video sharing scale, reduction of response delay and motivation of near-end sharing.

4.1. Promotion of Video Sharing Scale

The videos rely on visible content to attract accessing of users. If the videos have irresistible content, the positive request of users promotes popularity of videos and diffusion of video copies in edge networks. After the information of videos (e.g., title and abstract of videos) is obtained by the users, the users which are interested in the videos want to fetch the video resources by sending request messages; instead of the positive request, when the users may receive the pushed messages of video information from other users, they make the decision of receiving the pushed videos.

The interest levels are main driving factors of users requesting video resources. When a video starts to disseminate in networks, the interest level of a user for can be defined aswhere denotes the similarity value between ; and is the centric item of video cluster ; is the number of videos which belongs to and has been watched by ; is the total number of videos which have been watched by ; and denotes the interest level of for the interest subdomain . The higher the value of is, the stronger the intentions of for requesting videos in is. If is high, the membership level between and is strong; If is low, has the weak similarity relationship with the most of items in . In fact, can be considered as a weight of . The larger the value of is, the higher the probability of requesting is. On the other hand, when receives a push message about , not only investigates the interest level for but also considers the social relationship. The acceptance probability of for a pushed can be defined aswhere is the number of videos which are successfully pushed by in the video cluster corresponding to ; is the mean value of similarity between and videos which are successfully pushed by in the video cluster corresponding to ; is the sharing frequency of videos in between and ; and is the frequency of video sharing between and . The larger the value of is, the closer the social relationship between and is.

4.2. Reduction of Response Delay

The users make use of the request and push to disseminate a new video . If has high popularity, the large number of users may request or accept the push of . In the process of users requesting , the supply users which have cached in local buffer receive the request messages and send video data to the request users. The request users receive video data, watch the video content and store the video into local buffer. At the moment, the video copies are generated by the sharing (diffusion) between users. When the number of request users fast increases, the video system needs to make use of the upload bandwidth of users which have cached the copies in networks to provide the video resources for the request users.

Let and be the rate of request and handling, respectively. , where is the number of generated request messages during the time span ; , where is the number of handled request messages during the time span . during the same time span means that the request of users cannot be met due to the deficient supply of requested videos. Because the supply users with limited handling capacities (e.g., low bandwidth, storage, and computation) cannot fast deal with the mass request messages in terms of “early come early service” principle, a large number of request users need to wait for handling request messages. Obviously, the deficient supply leads to the long wait delay. during the same time span means that the request of users can be met due to the sufficient supply of requested videos. The supply users have low handling capacities, but the sufficient number of supply users enables the request messages be uniformly distributed to the supply users. The request messages can be handled in time, and the wait delay of request users can be reduced.

If a user has cached and watched a video , has lost the interest for . In order to save the storage space and promote the utilization of storage resources, replaces videos in local buffer. Before removes videos in local buffer, needs to estimate the influence for intra- supply capacity of . We assume that the request messages generated by users of arriving intra- providers of meet the queuing model. The serving capacity of can be obtained in terms of our previous work [36]. In fact, denotes that the sum of time of handling request message and time of delivering video data. The average serving capacity of users in can be defined aswhere is the number of users which have cached in . Let be the number of request messages generated by intra- users during a future time span . The value of can be calculated according to the following equation:where is the number of all users in ; is the number of users which have watched and lose the interest for ; is the total number of users which are uninterested in and users which have the low interest (), where is the threshold value of interest; denotes that users are interested in ; and denotes that users are uninterested in . Let be the number of users which have cached during . denotes that the supply of in meets the demand for request of , so that the wait delay is 0; if , the average wait delay can be defined as

When the supply of in cannot meet the demand for request of , makes the decision of replacement of according to the relationship between request and supply of after the generation of new copies of .

4.3. Motivation of Near-End Sharing

The delivery quality of video data relies on transmission path of video data. As we know, the transmission delay is the sum of forwarding delay of relay nodes. If the transmission path has the large number of relay nodes, the transmission delay may be high; if the number of relay nodes in transmission path is less or is 0 (the relationship between requesters and providers of videos is one-hop neighbor), the transmission delay is low or is 0. Moreover, if the transmission path has multiple relay nodes which have high mobility, the dynamic change of geographic location of relay nodes leads to the dynamic variation of transmission path. The variation of transmission path further results in the increase of packet loss and rise of forwarding delay.

In order to promote the transmission performance of video data and user quality of experience, the data transmission between one-hop neighbor nodes should be accelerated. Let denote the number of users which have the interest for in () during a future time span . The probability of and becoming one-hop neighbors in users during is defined aswhere is encounter frequency between and during and is the total encounter frequency between and other users in during . If has cached , the probability of requesting to during the encounter period time of and is defined as

The probability of pushing to during the encounter period time of and is defined as

4.4. Video Caching Management Strategy

The video sharing in edge networks can offload the traffic to reduce the load of core networks, which reduces the risk of network congestion. By optimizing distribution of video resources, the video resources can be efficiently allocated in edge networks and promote the utilization efficiency of storage resources. The adjustment of video distribution relies on the demand of video requesters. The performance of video sharing directly reflects the game relationship between supply and demand. The management of video caching is the main tool of adjustment of video distribution. The management of cached videos in local buffer has the important influence for the user quality of experience in terms of the three perspectives of promotion of video sharing scale, reduction of response delay, and motivation of near-end sharing. The management of cached videos based on the performance awareness of video sharing is an efficient method for the effort of distribution optimization and video sharing. We design a video caching management strategy based on the performance awareness of video sharing for a user in .

has watched and has cached in local buffer. Let and be rate of request and handling in , respectively. denotes that the number of users which store is less than the number of requesting in , so that does not remove in local buffer. At the moment, estimates the dissemination scale of in to make the decision of video pushing. can be aware of the users in which are interested in according to equation (10). divides the users into the two subsets: , where each item in has and has more willing to make an active request than accepting pushing of from ; , where each item in has and has higher probabilities of accepting pushing of from than those of active request. preferentially pushes to items in , which can promote supply capacities in by pushing with high acceptation probability. needs to select a user in which has high acceptation probability and low delivery delay according to the following equation:where is the predicted delivery time based on pushing from to . The parameters in (e.g., packet loss rate and transmission delay) can be obtained by sending detection messages to estimate communication quality of transmission path during a small time slot. A user has the largest value of among all items in ; a user has the largest value of among all items in . If , sends data of to in order to generate copies; otherwise, if , sends data of to . After has delivered , makes the decision of caching in terms of the relationship between and .

If , has enough supply capacity relative to request scale. Let denote the number of users which are interested in ; let denote the number of users in which store . If , can remove in local buffer after informs users which store by sending messages; if , estimates the number of users whose state transition from “interested” to “requested” by the following equation:where is a set of time slots; is the number of time slots in ; and are the number of users which request and are interested in during each time slot in , respectively; denotes the transition ratio from “interested” to “requested;” and denotes the average value of transition ratio from “interested” to “requested” during . If , needs to continuously store in local buffer; if , can remove in local buffer.

5. Testing and Test Result Analysis

5.1. Testing Topology and Scenarios

We conduct numerous simulation tests to compare the performance of SECS with current widely used caching policy random caching (with probability equal to 0.5) and state-of-the-art solution OCP [37], based on NS-3. The parameter settings are given as follows: a square scenario with area is implemented, and 500 mobile nodes move in the scenario according to the random way point model. In this model, mobile node first randomly selects a point as the destination by uniform distribution and then determines a velocity and begins to move to the chosen point. After reaching this point, the node repeats the above destination and velocity selection process and begins to move again. The velocity of mobile nodes ranges from 10  to 40 . We reset the physical and MAC layer and modulation schemes of network units according to the 5G industrial standardization.

For evaluating the caching performance, we consider the videos as the transmission data. This is because video service is the driven force behind the current rapid growth of network traffic and the key applications of edge caching. In our simulation, 40 different videos are introduced and shared among the mobile users. The playback bitrate of the video is 2000  and time length of a video is 120 . We further divide each video into 60 small chunks which is 2  long and with the size of 500 . The size of caching space at each mobile node is set 20 to 40 chunks. We further describe the distribution of request behaviors of mobile users at video-level by Zipf distribution [38]. Namely, given a video set of videos, the probability of the -th most popular video being requested by users can be given by [39, 40]where is the Zipf parameter and set to 0.8. For the chunk-level request behaviors, after determining the video to watch, the user sequentially request the chunks of this video for the continuous playback. To implement the OCP caching policy, we uniformly deploy 25 base stations as the coordinators and source video suppliers in OCP.

5.2. Performance Evaluation

We compare the performance of SECS with OCP and random cache in terms of the caching hit ratio, caching cost, response delay, and control overhead, respectively.

Caching hit ratio (CHR): in edge caching, if one node receives a video request whose corresponding chunk is in local cache, then we consider it is a cache hit event; otherwise, it is a cache miss. We define the cache hit ratio as the average ratio between the number of the cache hit events and total number of issued requests (the sum of cache hit and miss events). A higher caching hit ratio indicates that more requests are satisfied by the nearby mobile nodes, namely, a better edge caching utilization and shorter transmission distance. In contrast, a lower caching hit ratio means the edge caching is not fully utilized and mobile users may still have to access the video from far end. Figures 2 and 3 show the CHR of the three solutions with the varying of simulation time, when the size of caching space is 20 and 40 chunks, respectively. We observe that SECS achieves the highest CHR among three solutions. In both figures, we observe that the SECS and OCP first experience an increasing trend and then enter a stable phase when 200 s. The random cache first reaches the highest CHR before 100 s and then slightly decreases. The reason for random cache’s decreasing is large number of requests also results in a frequent cache miss and caching replacement, which significantly decrease the caching utilization. After entering the stable phase, the CHR of SECS is about 8% (5%) and (61%) higher than that of OCP and random cache when caching space is 20 (40), respectively.

OCP and SECS having better performance than random cache is mainly because these two solutions investigate the users’ demand and supplies to achieve higher caching utilization; yet, random cache only simply sets a probability to cache the content which cannot maintain the balance between video demand and supply. Instead of setting a same caching time threshold for all mobile nodes as in OCP, SECS considers the sharing capacity of each mobile users to achieve the flexible management on caching space. With such design, SECS can provide more accurate cache placement and thereby a higher CHR. Besides, SECS also replace the content in cache according to the sharing capacity which further improves the cache utilization.

Caching cost (CC): we define the total number of caching events during the simulation as CC. Figures 4 and 5 show the CC of three solutions with the caching space 20 and 40, respectively. Higher CC means that the caching strategies may consume more storage or energy resource on performing the caching operation, which may harm the system sustainability. In figures, all curves reveal a linear increase trend due to the continuous caching operations during the simulation. Comparing the two figures, we observe that all solutions with caching space 40 has lower CC when comparing with caching space 20. This is because a larger caching space can avoid the unnecessary caching replacement, which in turns reduces the caching cost. In both figures, SECS achieves the lowest CC among three solutions. For example, when caching space is 20, SECS is 8% and 17% lower than OCP and random cache when 400 s, respectively. SECS’s advantage is expended when caching space reaches 40, which is 13% and 19% lower than OCP and SEC when 400 s, respectively.

SECS has the lowest CC mainly because the caching decision making based on sharing capacity provides more accurate caching placement, which avoid frequent caching eviction and replacement. Hence, SECS can significantly reduce the CC. OCP formulates the caching optimization problem by jointly considering the caching cost and system load, which reduces the unnecessary caching operations and thereby achieves lower CC than that of random cache. However, OCP overlooks the caching replacement and simple uses the LRU caching replacement policy, which yields to a higher CC than SECS.

Response delay (RD): we define the response delay as the latency between mobile users sending out request and receiving the first packet of requested content. Low RD indicates that mobile users can access video content from nearby users and also a higher QoE due to the low start up delay on video playback. Figures 6 and 7 show the response delay of three solutions when the caching space is 20 and 40, respectively. In both figures, the curves corresponding to all solutions first reveals a fast growing trend before 300 s and then maintains relatively stable. This is because, with more and more users beginning to request video, one mobile node may need to simultaneously serve several users which increases the response delay. After 300 , due to the number of users that joining equals the number of users quitting the system, the system load and response delay remain stable. Comparing the two figures, we also find that higher caching space can provide better response delay, since larger caching space achieves a higher cache hit ratio, which thereby increases the probability of accessing content from a nearby mobile node. When in stable phase, SECS achieves the lowest RD among three solutions both in the condition of caching spaces 20 and 40. Especially, when caching space is 40, SECS is 11% and 24% lower than OCP and random cache, respectively.

According to Figures 2 and 3, SECS has the highest caching hit ratio, namely, most of video requests can be responded by the mobile nodes nearby the users and hence reduces the response delay. Similarly, OCP achieves the lower RD than random cache because of the similar reason. Random cache policy randomly decide the caching content according to a pregiven fixed probability, whose caching hit ratio can be not guaranteed. Frequent caching miss not only forces the users to access content from distance but also results frequent caching replacement, which further reduce the cache hit ratio. Therefore, random cache performs the worst in terms of the response delay.

Control overhead (CO): in the simulation, we calculate the CO by the averaged bandwidth occupied by delivering the control message of making caching decision. In our SECS, the CO is mainly generated by the information about sharing capacity and interest domain. In OCP, the CO mainly relies on the exchange frequency of mobile node state list and caching decision broadcasting message. In random cache, because each mobile node performs the caching operation based on a fixed probability, there is no control message to exchange. In Figure 8, the curve corresponding to SECS is relatively lower than that of OCP. The main reason is the OCP requires exchange state list which is large in size and more frequent message exchange for accurate predicting the demand variation of the whole system. Although random cache has no CO, this comes at the cost in terms of sacrificing the caching performance including caching hit ratio, caching cost, and response delay.

6. Conclusion

In this paper, we propose a novel Social-aware Edge Caching Strategy of Video Resources in 5G Ultra-Dense Network (SECS). SECS employs the Spectral Clustering to generate initial video clusters and makes use of the Fuzzy C-Means (FCM) to refine the initial video clusters. The refined video clusters are denoted as the interest domain of users. SECS further makes use of estimating similarity levels of interest domain to cluster the users with common and similar interests into the same groups. By estimating influence for the intragroup sharing performance, SECS designs a performance-aware video caching strategy, which enables the users intelligently implements caching and removing of local video resources to continually optimize video distribution and effectively supports video traffic edge offloading. Extensive tests show how SECS achieves better results in comparison with other state-of-the-art solution OCP in terms of caching hit ratio, caching cost, response delay, and control overhead.

Data Availability

The data used to support the findings of the study are available within the article.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported in part by the Science and Technology Key Project of Henan Province under Grant nos. 182102210105 and 202102210172, the Training Plan for Young Backbone Teachers of Colleges and Universities in Henan under Grant nos. 2020GGJS191 and 2017GGJS135, the Program for Science and Technology Innovation Outstanding Talents in the Henan Province under Grant no. 184200510011, the Science and Technology Opening Up Cooperation Project of Henan Province under Grant no. 152106000048, the National Natural Science Foundation of China (NSFC) under Grant no. 61501216, and Open Foundation of the Guangxi Key Laboratory of Trusted Software (Grant no. KX202040).