Abstract

With device-to-device (D2D) communication, user equipment can share data with each other without the involvement of network infrastructures. In order to maintain the Quality of Service (QoS) and Quality of Experience (QoE) for user applications in D2D communications, most existing schemes use proactive content caching that needs to predict content popularity before making caching decisions which may result in privacy leakage, since the information of users is collected to train a deep learning-based model to predict content popularity. Therefore, it is crucial to guarantee secure data collection in machine learning-based framework. In this paper, we propose a privacy-preserving D2D caching scheme with a passive content caching strategy based on node importance, which can deliver more efficient caching and prevent the potential leakage of user privacy. The scheme is based on softwaredefined networking (SDN), in which the controller is responsible for calculating node importance of devices according to the information of requests and encounters collected by SDN switches. Base station will decide which device can establish reliable and secure communication with content requester based on historical information. The simulation results show that the proposed strategy can outperform other D2D caching strategies in terms of cache hit rate and data rate.

1. Introduction

During the recent years, with the rapid development of the mobile Internet, mobile data traffic has increased exponentially. A Cisco VNI report predicted that 79% of global mobile data traffic will be mainly generated from access to video content by 2022 [1]. The current wireless network is facing huge challenges. To reduce the backhaul traffic and base station load, device-to-device (D2D) communication technology has emerged. User devices can directly establish D2D communication links with other devices within the communication range in the D2D communication network without the use of base stations or other access points.

Research results have shown that most mobile traffic is generated from repeated access to popular content [2, 3]. By deploying caches in the core network [4], access network [58], and user device [9, 10] in the 5G mobile communication network architecture, the popular content can be cached at the network edges, which can effectively reduce network congestion and improve network performance. The content is cached to the user devices, and a D2D caching network is constructed. Thus, user devices can share content via the D2D communication link without the use of access points, which can effectively reduce backhaul traffic and base station load. The caching mechanism can determine the caching position and caching contents, is the kernel of the D2D caching network, and determines the caching performance. However, due to the heterogeneity of D2D caching network, it is more vulnerable to security and data privacy threats. It is important to find the trade-off between the performance of security and the cost of protection. Most existing D2D caching mechanisms belong to a proactive caching mechanism that requires content popularity in advance. Before a user sends a request, the content with higher popularity is cached into the D2D caching network in advance. Some owners of mobile devices may be curious about the content cached in their devices and scan the caching content, which may result in privacy leaking [11].

During recent years, short videos distributed by users have attracted massive traffic. The QuestMobile report indicates that the time spent by users in short video applications was 5.5% of the total time spent on mobile applications in 2017. With the quick development of short video applications, such as Bilibili and TikTok, a user can distribute short videos whenever and wherever possible. Short videos can be distributed and requested in a random and burst manner. After popular videos are distributed, they are accessed by massive users in a short period. For example, one user distributed a short video on TikTok, and it was accessed 30,000 times in a couple of minutes. In such cases, a proactive caching strategy cannot predict the popularity of popular video content in time. Thus, the cached contents cannot be updated in real time, which results in massive ineffective caching and wastes network resources.

Motivated by this, we propose a privacy-preserving D2D caching scheme with passive content caching based on node importance to update the cached contents in real time, increase the cache hit rate, and preserve the user privacy. Since the proposed caching scheme adopts a passive caching strategy, the device can only cache the content that it requests, which prevents the leakage of user privacy. By using the network coding technology [12], the diversity of the cached contents can also be improved in limited cache size. It has been proved that using networking coding in content caching can improve the security and performance of the caching system [13]. The SDN switch collects the history request of the terminal devices and meeting information among devices, and the SDN controller can compute the node importance of the terminal devices using the history information. Base station decides which content holder can establish reliable and secure D2D communication link with the content requester. The cached contents can be updated in real time based on user requests and node importance.

The contributions of this paper include the following:(i)Firstly, we introduce a privacy-preserving D2D caching scheme with passive content caching based on node importance to improve the security and performance of D2D caching network. The device with higher node importance and social trust will be selected to establish reliable and secure D2D communication link with the content requester.(ii)Then, we define the node importance as the weighted sum of the physical intimacy and request similarity between devices, which also reflects the social trust of the device.(iii)Finally, we evaluate the performance of the proposed D2D caching strategy and other two well-known caching strategies. The simulation results show that the caching strategy based on node importance proposed by this paper could effectively improve network performance compared to the other two caching strategies.

The remainder of the paper is organized as follows. We introduce related works in Section 2. In Section 3, we define the node importance and propose a privacy-preserving D2D caching strategy based on node importance. Simulation results are presented in Section 4. Finally, we conclude the paper in Section 5.

If the popular content is cached in the user devices, it can effectively reduce the backhaul traffic and the downloading delay for users. Thus, service quality and user experience can be improved. Currently, most D2D caching mechanisms are based on a proactive caching strategy, and the content popularities are assumed to be known. Golrezaei et al. [14, 15] divided the D2D network into multiple D2D clusters, and only the devices in one cluster can establish the D2D communication link. Based on this, the authors proposed two in-cluster D2D caching strategies to improve network performance of cellular networks, including deterministic cache and random cache based on Zipf. In the deterministic caching mechanism [14], devices can cache nonrepeated contents with the top popularities in one virtual cluster and each device only caches one content. In the random caching mechanism based on the Zipf distribution [14], within one virtual cluster, each device can independently and randomly cache content and the popularities of the content cached in the cluster obeying Zipf distribution. Wang et al. proposed a novel D2D caching strategy based on mobile perception, which takes the mobility of users into account. The low-speed and high-speed moving user devices cache content with top popularities, and user devices with middle-speed moving cache content with lower popularities. Thus, the offloading rate can be improved [16]. Chen et al. modelled the offloading benefits and energy consumption of content holders and proposed a proactive caching strategy and user-oriented protocol to obtain higher offloading benefits with lower energy consumption [17]. Malak et al. extended the caching mechanism based on the geographical position and proposed a space-based caching strategy to improve the cache hit rate. To reduce cache redundancy and improve the diversity of cached contents, all the devices in the mutual exclusion area cannot cache the same content [18]. Wu et al. proposed a distributed D2D caching strategy that considers the characteristics of different requests and demands of physical links [19]. Besides a proactive caching strategy that assumes that the content popularity is known, partial D2D caching mechanisms predict content popularity by algorithms such as machine learning. After the future content popularities are predicted, the caching mechanism is determined. Jiang et al. modelled the D2D caching as the multiagent and multiarm gaming machine and determined the cache by using reinforcement learning to reduce downloading delay for users [20]. Li et al. optimized content caching and content distribution jointly to reduce transmission delay and power consumption. They deployed two potential recurrent neural network models, echo state network (ESN) and long short-term memory (LSTM), to predict mobility of users and future popularity of content and then determined the cached contents and cache position. The authors also proposed a content distribution mechanism based on deep reinforcement learning to improve user experience [21]. In another study [22], the authors proposed a proactive caching strategy based on the association between users and content. They predicted the content popularity by using machine learning and collaborative filtering technology. The content with higher popularities is precached to the base stations and user devices in the low peak period to alleviate the backhaul congestion. Chen and Yang proposed a D2D caching strategy based on user preference to improve the offloading rate, which predicts user preferences by using a collaborative filtering algorithm based on the model and then makes caching decisions [23].

In summary, most D2D caching strategies are based on proactive caching mechanisms and the content popularities must be known or be predicted. Then, the caching decision problem is transformed into optimization problems to find a solution. The prediction accuracy of future content popularities decides the caching performance. However, the rise of short video applications makes the proactive caching strategy fail to precisely predict content popularities and update cached content in real time. Thus, caching performance is reduced. Motivated by this, we propose a passive D2D caching strategy based on node importance, and the cached content can be updated in real time according to user requests. Moreover, network coding technology is employed, which can improve the diversity of the cached content and increase the cache hit rate and data rate without increasing the cache size.

3. System Design

3.1. Defining Node Importance

In this paper, we propose a software-defined passive D2D caching strategy (NIC) based on node importance. Different devices have different node importance in the D2D network. In this system, the SDN switch collects meeting information among user devices and information of content requested by user devices to compute the physical intimacy and request similarity among devices. The node importance is defined as the weighted sum of the physical intimacy [24] and request similarity. For the physical network layer, the user devices with higher node importance will have a higher probability to establish stable D2D communication links with other devices in the future. For content requesting, the user devices with higher node importance have a higher probability to request the same content as other devices. In other words, the user devices with higher node importance have a higher probability of providing other user devices with requested content in the future. Thus, the devices with higher importance will cache content with higher popularities, and the original blocks are cached to reduce the network coding and decoding time and computing consumption. The devices with lower importance will cache contents with lower popularities, and the coding blocks are cached to distribute contents in the network more reasonably, improve caching diversity, and increase the caching efficiency without increasing cache size. The physical intimacy between user devices indicates the probability of establishing reliable D2D communication links between two user devices in the future. When device and device are within the D2D communication range, the two devices may have D2D communication potential, which is recorded as one meeting between them. The duration is recorded as the meeting duration. Zhang et al. [24] proved that the user meeting time obeys the gamma distribution , namely, , whereinwhere indicates the meeting duration of device and device and indicates the meeting count between device and device . The physical intimacy can be expressed as [24]where is the minimal meeting duration required by two user devices to successfully transfer one file via the D2D communication link. Thus, the average physical intimacy between device and other devices in the D2D caching network is given by following:where is the quantity of the devices within the D2D communication range of device . The higher the physical intimacy of the user device is, the higher its probability of establishing a reliable D2D communication link with other user devices in the future will be. As shown in Figure 1, the average physical intimacy of user device 1 is

Based on the history requests of the device, the SDN controller can compute the request similarity between user devices by using cosine similarity, which is expressed as follows:where and indicate the interest vector of device and device , respectively. The average request similarity of device and the other devices in the D2D caching network iswhere is the quantity of the devices within the D2D communication range of device . The higher the average request similarity of the user device is, the higher the overlapping degree of requested contents with other devices will be. In this paper, we propose a passive caching mechanism, namely, the user device only caches the ever-requested content to satisfy other users’ requests by providing other users with the desired content. The cached content in the user device with higher average request similarity will have a higher probability to be requested by other users. In this paper, we define the node importance as the weighted sum of the physical intimacy and request similarity. The devices with higher node importance will have a higher probability to provide other user devices with requested content and receive higher caching benefits. The node importance of user device is given by the following equation:where is the normalized average physical intimacy of device , is the normalized average request similarity of device , and and are the design parameters and indicate the importance of physical intimacy and request similarity. In this paper, and are set as 0.5.

3.2. D2D Caching Strategy Based on Node Importance

In this system, to improve the caching efficiency, the base station divides the content into content blocks with the same size. When device requests content , it will first send the request packet to the base station. If the D2D caching network includes this content, the base station will locate a group of user devices with hit caches and determine the caching plan according to the node importance. If of device is higher, then device caches the original blocks. If the node importance of device is lower, then device caches the coding blocks. This is because the device with higher node importance has higher request similarity with other devices, namely, the cached content has higher probabilities to be requested by other devices. The devices with higher node importance have a higher probability to successfully establish D2D communication links with other users and can ensure successful content transfer. Therefore, the devices with higher node importance will cache content with higher popularities and cache original content blocks to improve the cache hit rate and reduce decoding/coding time and computing consumption. The devices with lower node importance will cache contents with lower popularities and cache coding blocks. Each coding block should include all the original block information. This can improve the content diversity of the caching system without increasing the cache size.

In the D2D caching network, the base station maintains the D2D caching network information and locates a group of user device with hit caches, namely, the content holder. indicates the caching information of cached content in the D2D caching network, and , where is the number of user devices with cached content . The content holders will be ranked by the node importance in descending order. indicates the caching information of content cached in the user devices with highest node importance, and the caching information is expressed as follows:where is the device ID, is the node importance of device , indicates the quantity of the cached original blocks or coding blocks, , and indicates if the cached content block in device includes the information of the original content block j. If it is included, then ; otherwise, . When is an all-1 vector and , it indicates that the user device caches the coding blocks of content .

When device requests content , the base station will find the D2D cache information table and check if the D2D caching network can satisfy the user requests. If the quantity of the content blocks cached in the D2D caching network is more than or equal to , then the base station selects a group of content holders with higher node importance, and the selected content holders will establish D2D communication links with the device to transfer corresponding content blocks. If the D2D caching network cannot satisfy the user requests, the base station will send content blocks to respond to device , where is the quantity of the content blocks cached in the D2D caching network, as described in Algorithm 1. The complexity of Algorithm 1 is , where is the number of content holders.

Input:;
Output:, // is the cache identifier, is the code identifier, 0 indicates the original block, and 1 is the coding block;
(1) Initialization: , ;
(2) The SDN switch collects the request records and interactive information of the devices and periodically sends it to the SDN controller;
(3) The SDN controller computes the node importance of the device according to the history information collected by SDN switches, namely, :
(4) While BS receives the request from device for content do
(5) if the node importance of device is higher then
(6)  Make ;
(7) else
(8)  Make ;
(9) end if
(10) if content blocks are cached in the D2D caching network then
(11) BS locates a group of cached content holder with the top node importance to respond to the user request;
(12)  for each cached content holder do
(13)   if the cached content is the original block then
(14)    make ;
(15)   else
(16)    make ;
(17)   end if
(18)   BS sends data packets to the content holder;// is the number of content blocks to be sent by content holder to the requester
(19)   After content holder receives the data packets from BS, it will establish the D2D communication link with device and transfer the data packet ;// is the coding block or content block;
(20)  end for
(21) else
(22)  if then
(23)   BS sends original blocks to the content requester device , namely, ;
(24)  else
(25)   BS sends coding blocks to the content requester device , namely, ;
(26)  end if
(27) end if
(28) end while

In this system, the base station selects a group of content holders with higher node importance and social trust to establish reliable and secure D2D communication links to transmit desired content. Moreover, the base station is also responsible for making caching decisions and instructing how the user devices cache the received content blocks. If the request is from users with higher node importance, then it indicates that caching the content in content requester will bring benefits with a higher probability, i.e., higher cache hit rate. When the base station responds to the device request, it will send the original block and instruct the user devices to cache the original block. Otherwise, when the base station responds to the user request, it sends the coding blocks generated with all the original blocks and instructs the user devices to cache the coding blocks. For details, refer to Algorithm 1. To prevent users from receiving the linearly dependent coding blocks, the user device only caches the coding blocks sent by the base station and does not cache coding blocks obtained by D2D communication links.

When content requester receives the content blocks from the base stations or other devices, it will decide whether to cache the received contents based on cache identifier . If , then the content will be cached locally; otherwise, the content will not be cached. Since the cache size of the device is limited, when the caches are replaced, the user device codes the replaced content blocks into a coding block. In this way, all the original block information of this content will be reserved while the cache size is released to improve the diversity of the caching contents.

4. Experimental Results and Analysis

In the simulation test, the radius of the base station was 500 m and a total of 100 devices were provided. The maximum D2D communication distance was 100 m. The D2D communication belongs to the in-band communication, namely, the D2D communication shares the bandwidth with cellular communication [25]. The cache size of the device included files, and the number of files was 500. The user request obeyed the Poisson distribution, the popularities of the content obeyed the Zipf distribution, and the Zipf parameter was . In the simulation test, the passive caching strategy NIC in this paper was compared with the passive caching strategy, Leave Copy Everywhere (LCE), and the proactive caching strategy, Most Popular Cache (MPC). For the LCE, all the received content will be cached by requester. For the MPC, the most popular content will be cached into devices in advance. The strategies were assessed by the cache hit rate and data rate. The cache hit rate was defined as the ratio of the number of requests responded by user devices to the total requests sent by user devices, which is an important parameter to assess the performance of the caching system. When device requests contents from the base station via cellular communication, the data rate is defined as follows [24]:

When device requests contents from device via D2D communication, the data rate is defined as follows [24]:where , , and indicate the transmission power of the base station, device , and device ; is the Gauss white noise; and are the path loss and are related to communication distance; indicates interference; and indicates no interference.

The influences of the cache size on the three caching mechanisms are shown in Figures 2 and 3. As shown in Figure 2, the cache hit rate of the three caching mechanisms increased with the growth of the cache size. This is because with the growth of the cache size of the user device, more files can be cached in the D2D caching network to make more requests responded to by other terminals. In this case, it can reduce the base station load and backhaul traffic to improve the data rate, as shown in Figure 3. As shown in Figures 2 and 3, the cache hit rate and data rate of the NIC caching strategy proposed in this paper were higher than those of the other two caching strategies. The NIC strategy deploys caches and implements differential caching strategies according to the content popularities and node importance. Thus, the content can be distributed more reasonably in the D2D caching network and more content can be obtained via D2D communication. It can reduce the base station load and improve the caching hit rate. Compared to the other two caching strategies, the NIC mechanism can consider the physical intimacy between user devices and make the cache hit node closer to the request devices and obtain a higher data rate.

The influences of the Zipf parameters on the three caching mechanisms are shown in Figures 4 and 5. The bigger the Zipf parameter is, the more similar the user requests will be, and there will be a greater repeated request time for a small part of the files. As time elapses, the cached contents in the D2D network increase, and the cached contents are centralized in a small number of files. Then, users have a higher probability to receive files via the D2D communication link. Thus, the cache hit rate and data rate of the three caching mechanisms grow as Zipf parameters enlarge. The caching performance of NIC was always superior to the other two caching strategies, which was more significant when the Zipf parameter was smaller. The reason for this is that NIC can improve the diversity of the cached contents and improve caching performance by using network coding technology without increasing the cache size.

5. Conclusion

In this paper, we propose a privacy-preserving D2D caching strategy based on the node importance, which can improve the diversity of the cached content by using the network coding technology and preserve the privacy of users and data. The SDN switch collects the information of requests and meeting information of the devices, and the SDN controller can compute the physical intimacy and request similarity between user devices and other devices by using the history information collected by SDN switches to obtain the node importance of the device. The node importance is defined as the weighted sum of the physical intimacy and request similarity. The nodes with higher importance have higher cache benefits and will cache the original blocks; the nodes with lower importance have lower cache benefits and will cache the coding blocks to make the cached contents be distributed more reasonably in the network. Base station will decide which device can establish reliable and secure communication with the requester based on historical information, which reflects the importance and social trust of devices. The simulation results show that the caching strategy based on node importance in this paper could improve the cache hit rate and data rate and effectively improve the performance and security of the caching network compared to the other two proactive and passive caches.

Data Availability

No data were used to support this study.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Acknowledgments

This research was funded by the National Natural Science Foundation of China (62002072) and the Guangzhou Science and Technology Project (GZHKJXM20170117).