Abstract

Caching popular contents at base stations (BSs) has been regarded as an effective approach to alleviate the backhaul load and to improve the quality of service. To meet the explosive data traffic demand and to save energy consumption, energy efficiency (EE) has become an extremely important performance index for the 5th generation (5G) cellular networks. In general, there are two ways for improving the EE for caching, that is, improving the cache-hit rate and optimizing the cache size. In this work, we investigate the energy efficient caching problem in backhaul-aware cellular networks jointly considering these two approaches. Note that most existing works are based on the assumption that the content catalog and popularity are static. However, in practice, content popularity is dynamic. To timely estimate the dynamic content popularity, we propose a method based on shot noise model (SNM). Then we propose a distributed caching policy to improve the cache-hit rate in such a dynamic environment. Furthermore, we analyze the tradeoff between energy efficiency and cache capacity for which an optimization is formulated. We prove its convexity and derive a closed-form optimal cache capacity for maximizing the EE. Simulation results validate the proposed scheme and show that EE can be improved with appropriate choice of cache capacity.

1. Introduction

With the proliferation of new wireless devices and smart phones, as well as bandwidth-intensive applications, user demand for wireless data traffic has been growing tremendously, and the corresponding network load increases exponentially [1, 2]. For more effectively handling the explosive growth of mobile data services, caching at the edge of wireless network has been proposed and has become a hot research issue [3, 4]. Recent studies have shown that caching popular contents at the edge of wireless networks can not only reduce the backhaul cost, access latency, and energy consumption but also boost the throughput [5, 6].

Meanwhile, energy efficiency (EE) has become a major performance metric for future 5G cellular networks [7]. However, only few works in the literature consider the energy efficiency analysis for cache-enabled cellular networks. In general, the EE of a cellular network with caching can be improved from two aspects: improving the cache-hit rate and optimizing the cache size. The work in [8] takes content popularity and request arrival density into account for deriving the EE. By analyzing the EE for the caching-enabled network, [9] elaborates whether and how caching at BSs can improve the EE. The authors in [10] show that introducing cache can improve the EE if the content catalog size is not infinitely large and placing cached contents at small cell base stations (SBSs) is more energy efficient. Furthermore, a user-centric cooperative model is proposed in [11], and the relationship between EE and the cache-hit rate is analyzed. In [12], the authors analyzed the impact of various factors on EE such as interference level, backhaul and cache capacity, user and BS density, power consumption parameters, BS work state, and content popularity. Also, the condition when the EE can benefit from caching is derived. This work also points out that the backhaul power consumption and caching power consumption are not negligible for analyzing EE. It is worth noting that there are tradeoffs between backhaul power consumption and caching power consumption. Specifically, caching more contents could reduce the backhaul load and correspondingly reduce the backhaul power consumption, at the cost of increasing the caching power consumption. However, it is still not clear what is the optimal amount of cached popular contents at the BS which could maximize the EE of the downlink system.

Furthermore, all the above works are based on an ideal assumption that the content popularity is known and fixed. Popularity refers to the cumulative number of requests for each content, or the requested probability, or the requested arrival rate over a specific area and period of time [13]. In reality, content popularity is dynamic which needs to track an ever changing popularity profile of contents. Substantial number of experimental results have shown that requests for contents frequently appear during a certain period of time at a specific location, while the number of requests outside this time period is small [14]. Apparently, considering dynamic content popularity will pose some challenges, such as how to predict content popularity and how to analyze cache performance in such a dynamic environment [15].

In this work, we investigate the energy efficient caching problem in backhaul-aware cellular networks by jointly improving cache-hit rate and deriving optimal cache size. And dynamic content popularity is considered. To timely estimate the dynamic content popularity, we propose a method based on shot noise model (SNM), which can predict the request hit rate of contents. Then we propose a distributed caching policy to improve the cache-hit rate in such a dynamic environment. Also, the content redundancy can be reduced by the proposed policy. Finally, we analyze the tradeoff between energy efficiency and cache capacity for which an optimization is formulated. Analytically closed-form solutions are obtained. And the simulation results show effectiveness of the proposed scheme under limited backhaul and cache capacity.

The remainder of this paper is organized as follows: in Section 2, we introduce the system model. The dynamic caching policy and distributed cache policy are proposed in Section 3. The EE optimization problem of a cache-enabled cellular network is formulated and analyzed in Section 4. Numerical results and analysis are provided in Section 5, and the conclusions are summarized in Section 6.

2. System Model

In this section, we describe our network, transmission, and BS sleep pattern models. For ease of reading, we summarize the notation used throughout in Notations.

2.1. Network Model

We consider a backhaul-aware cellular network consisting of base stations. These BSs are randomly distributed on a two-dimensional Euclidean plane by a homogeneous spatial Poisson point process (HPPP) with density [16]. Each BS is equipped with transmit antennas and serves multiple users with single antenna. Each BS has a cache device which is able to cache contents and is connected to the core network through a backhaul link with limited backhaul capacity . An illustration of the network model is depicted in Figure 1.

In the network model, the users are modeled by an independent homogeneous Poisson Point Process with density , where . In the paper, we assume that user associating with BS is based on both content and signal-to-interference-plus-noise ratio (SINR). Specifically, if there are multiple BSs which have cached the requested content, a user will select the nearest one. Otherwise, a user will be associated with an BS with maximal receiving SINR. For mathematical tractability, we use a circular area with a radius of to approximate the hexagonal shaped area, where a number of single antenna users are evenly distributed. The average user density of each cell is ; then the probability density of users in each cell can be expressed as

Denoting as the set of contents cached at the th BS, . Each user requests different content, such as video, music, and news. We build the contents of the user request into a content catalog, that is, . We assume that each content has equal size bits in the content catalog, and the cache capacity of each BS is bits. If the content requested by the user is cached in the local BS, the base station will directly deliver the content to the user from the caching device. And this user is called the cache-hit user, where the transmission rate is determined by the SINR of the downlink channel [17]. On the contrary, when the user requested content is not cached in the local BS, the BS firstly obtains the content from the core network through the backhaul link and then transmits it to the user through the radio link. Similarly, this user is called the cache-miss user.

2.2. Transmission Model

Spatial multiplexing is taken into consideration. At the same time, the number of simultaneously served users can not exceed the number of antennas . And a round-robin scheduling policy is applied [18]. To capture the essence of the problem and simplify the analysis, we assume that each BS serves users with zero-forcing beam-forming (ZFBF) [19], which is a spatial signal processing technique to eliminate multiuser interference. And the downlink transmission power for all users is the same. We introduce a parameter to reflect the percentage of intercell interference (ICI) that can be eliminated by interference management methods (i.e., coordinated beam-forming and serial interference suppression) [20].

The signal transmitted from a given BS is affected by two propagation losses before reaching the user:(1)Path-loss: a distance dependent path-loss , where is the distance from the BS to user and is the path-loss exponent.(2)Rayleigh fading: is the small-scale Rayleigh fading channel vector from the BS to user.

Particularly, we describe the channel gain between the th user and the th BS as , which can be expressed as where denotes the integrated channel power gain between the th user and the th BS due to Rayleigh fading. With the above expressions, we can get the SINR of a typical user as where is the transmit power of each base station and the random variable denotes the small-scale Rayleigh fading of the wireless channel and follows the exponential distribution. denotes the cumulative interference experienced from all BSs except the th BS, and is the distance between the user and all base stations except the th BS. is the variance of the white Gaussian noise.

2.3. BS Sleep Pattern of the Network

The energy consumption of a BS is determined by its working state [21]. When a BS is in active state, the components (e.g., power amplifier, signal processing unit, and antennas) will generate energy consumption. On the other hand, if the BS switches to sleep state, some units will be closed, and parts of the energy consumption can be saved. Once a BS has no active user, it will be switched into the sleeping state. According to the system model, the probability of a BS being in sleep state can be expressed as

3. Distributed Caching with Dynamic Content Popularity

Appropriate modeling has important impacts on the performance analysis of caching policies. However, most existing works do not take the dynamic changes of popular content distribution into consideration. In this work, we consider dynamic content popularity and propose a method to handle the time-varying popularity based on Poisson Shot Noise Model (SNM) [22].

Shot Noise Model is a “content-level” business model, which can reveal the dynamic changing of content catalogs and the arrival rates of different contents in the same area, and can be used to evaluate the performance of a cached network. SNM describes the process of each content generation (i.e., content from active to inactive) and each content request process [23]. In [24], the author assumes that the process of generating contents is a Poisson process, and each content is taken out after being active for certain period. The request process of each content is an independent nonhomogeneous Poisson process. In this paper, we assume that the process of generating contents and the request process follows the process of [24]. Figure 2 shows the Poisson Shot Noise Model, which includes two different request arrival rates and time during the lifetime. We consider the active lifetime and shape of all contents to be fixed in this model.

The lifetime of each content is associated with a time-varying request, which can be described by following characteristics:(1)The start time of request content : the request start time of contents follows a Poisson process with average arrival rate (i.e., for a unit of time there is an average of contents start to be active). For rectangular pulses, the average total number of contents generated in time is and the average number of active content is , where is the file life-cycle.(2)Shape : the shape describes the regularity of the request arrival rate of file changing with time (i.e., rectangular pulse). It is reported that the this feature has a smaller impact on the hit probability under the Least Recently Used (LRU) cache management policy.(3)The total number of requests : this feature represents the total number of times that content is requested by a user during the active time.(4)The life-cycle of content : this feature describes the length of the request for the content ’s arrival time, which can be defined by rectangular pulses, so the file life-cycle can be expressed as

We assume that all contents have the same life-cycle. Denote with the arrival time of each request content . At time the lifetime content catalog is expressed as

We can describe the alive content by the active “age” of each content and the total number of requests . The “age” of active content at time can be expressed by , where the “age” refers to the active time of content at time since the content appears in the system. According to the above characteristics, for the shot noise model with file life-cycle and rectangular pulse , the request arrival rate of a content at time can be expressed as .

3.1. Request Hit Rate with Estimated Popularity

In this section, we assume that we do not know the popularity of each content; then we estimate the dynamic content popularity by observations of the past user’s requests to the file in the caching controller and try to maximize the request hit rate in lifetime . However, considering that there is no traffic power for transmitting a content from the core network to the BSs, we can achieve the time horizon request hit rate optimization via solving the individual problem of maximizing the request hit rate in each time instance.

In this model, we choose the pointwise traversal method to study the request hit rate in each time instance. For ease of analysis, we choose to study the case where and observe the random events in the time interval . We use binary caching vector to determine whether the content is cached at time , where represents the fact that content is cached in the BS at time , and otherwise. Due to the limited cache capacity, the constraint must be satisfied in each time interval.

In the following, we consider using binary caching vector to maximize the instantaneous request hit probability at the origin. Taking the fact into account that contents popularity is unknown, the instantaneous request hit probability at the origin can be expressed as . Next we would like to optimize the expected request hit rate which can be expressed as follows: With SNM, predictive popularity can be expressed as where is the power-law density and is the “age” of active content at time . Accordingly, the request hit rate maximization problem based on predicted popularity can be formulated as It can be seen from the above expression that we can calculate the instantaneous request hit rate by the value of and .

3.2. Age-Based Threshold Caching Policy

To solve the optimal request hit probability at every time instant, we design a caching policy with age-based threshold, and this policy does not need to calculate the estimation . Intuitively, the “age” affects the estimation . Due to the uncertainty in the estimation is low; then if we want to maximize the request hit rate we must cache more contents with larger ages. Accordingly, we propose an age-based threshold which depends on the content age . Under this policy, if the content satisfies , it will be cached. When the content is fixed and the age is known, the cache decision depends only on the total number of requests , which makes caching decision more simpler.

Note that an important progress of the age-based threshold estimation is to compute . We can compute the age threshold function by iteratively filling cache capacity in the algorithms, which make the marginal request hit rate improvements for each request age approximately equal. The details can be found in Algorithm 1.

Step 1: Selecting Parameter. (i) is the cache capacity, (ii) is the average request rate, (iii) is the average
total number of request content at any time instant, (iv) is the fraction of caching content catalog,
(v) is the function of distribution. (vi) Define is the th tipper-percentile of , where
.
Step 2: Setting the Age-Based Threshold. Denote is the age-based threshold.
Step 3: Caching Content Vector. Using binary caching vector to determine whether the content is cached
at time .
Step 4: Caching Capacity. We define a constraint condition for the capacity of the cache, if , we will
select with , otherwise set for content.
3.3. The Correlation of Content Request Popularity

Due to differences in social and cultural backgrounds among users, restrictions on specific office areas, or the impact of certain group activities, content popularity among BSs may vary. However, there are still possible correlations. We assume that the contents exhibit geographical correlations and hence there exist groups of contents which are very popular in subset of local caches [2527]. For local caches with the same popularity, we can more effectively constrain the aggregation to a subset of caches that witness similar traffic patterns by aggregating all cached request contents.

Request Model for Correlated Popularity. We design a model for the correlation of content request popularity . The aggregate total popularity is expressed as , and the global average request hit rate is . We obtain the correlation between the intercell request arrival rate through community detection [28] and inhomogeneous random graphs [29] as follows:(i)Each content is associated with a property vector , where is the property of the content (e.g., the probability of the content as comedy). In order to simplify the model we consider that is an independent uniformly distributed random variable taking values in .(ii)Each BS is also associated with a property vector , where is the property of the BSs (e.g., the probability of the user request comedy in intercell ). The vector is selected uniformly and randomly in .(iii)We assume that and are uniformly distributed in . We define , where is a symmetric and continuous function with period of , strictly decreasing on , and for all .

When the content popularity of each intercell is different, according to [30], the request arrival rate of content in BS can be expressed as where are the contents in the core network. As the number of caching contents increases in the core network and the kernel function satisfies some base conditions, the normalization constant almost surely becomes deterministic:

In the section, the request hit rate of the future contents is predicted according to the request information of the historical contents of each BS, and then each BS caches the most popular different contents.

3.4. BS Cache Policy of the Network

Research shows that multiple users could access the same content, like some videos, news, music, and so on. In this case, the network will be flooded with requests of same contents, which can largely increase the latency, eventually leading to network congestion. Therefore, under the limited cache capacity, assuming that the user’s choice distribution matches with the dynamic content popularity, then each BS can cache most popular content to reduce service latency and backhaul congestion in such a dynamic environment.

A content catalog is defined as , and the size of each content in the catalog has a length of Mbytes. According to the description of Section 3, we can predict the content popularity, and the instantaneous popularity follows the Zipf distribution [31]. Each user requests a content from the content catalog , and the probability of requesting the th content is where characterizes the steepness of the distribution. The value of depends on the average number of users per BS, and higher values of correspond to steeper distributions which means that the higher the number of users attached to a BS is, the more accurately the trend is sampled.

As mentioned above, each BS caches most popular content according to the dynamic popularity distribution. Hence, the cache-hit rate of each BS can be obtained by the Zipf distribution:

It can be proved from (13) that converges to 1 when the content catalog in the BS goes to infinity. Consider the relation between maximal EE of the network and the cache size. To this end, we provide the cache-hit rate for large values of and . In order to show the impact of the cache capacity and content library on cache-hit rate, we analyze a normalized cache size . Thus (13) is normalized as where the approximation in (14) is accurate when and .

Cache-hit rate affects the performance of wireless networks, like throughput, energy efficiency, user quality of service (QoS), and so on. In order to increase cache-hit rate, we propose a distributed content caching strategy in backhaul-aware cellular networks, where four adjacent BSs cache different popular contents. According to the system model, we can find that each user is associated with a BS based on both content and SINR. Meanwhile, if there are several BSs caching the requested content, the user selects one whose distance to him is the shortest and delivers the content via radio link directly. Otherwise, user associates with a BS whose SINR is maximal from its neighbouring BSs and then gets the requested content through the backhaul link from the core network. As illustrated in Figure 3, each hexagon represents a BS with cached popular content. This distributed content caching strategy can reduce content redundancy by caching different contents in different BS. When each BS caches contents with the distributed cache strategy, each user can requests popular contents.

4. Energy Efficiency Optimization

In recent years, the research on green communications has made great progress. It shows that the energy efficiency of the cellular network has become an essential factor in the study of its performance, which is the key to reduce operating costs and is aiming to provide wireless access in the way of sustainable development and rational utilization of resources [32]. We define the EE as the ratio of the average throughput of the network to the average total power consumption at the BS, which can be expressed as where is the average throughput and is the average power consumption.

4.1. Average Throughput of the Network

If user request contents are cached at local BS, the user can immediately obtain the requested content from the local cache via the radio links, where the data rate is limited only by the capacity of the wireless channel. On the contrary, if the contents requested by users are not cached at local BSs, the BSs need to obtain the contents from the core network via the backhaul links and then transmit the contents to users through radio links, so the data rate is subject to backhaul link capacity and radio channel capacity. Considering the above two cases, the downlink average throughput of the th cell can be expressed as where is the average sum rate of the cache-hit users and is the average sum rate of the cache-miss users.

4.1.1. The Average Sum Rate of Cache-Hit Users

Considering that the SINR for every user shown in (3) are identically distributed, the average sum rate of the cache-hit users can be approximated as where is a constant only depending on the path-loss exponent when , is the distance from th BS to the th user.

4.1.2. The Average Sum Rate of Cache-Miss Users

Considering the transmit power of base stations to be the same, the cumulative interference from all BS approaches to its expectation when according to the law of the ultradense BSs deployment. This suggests that the comparison of and depends on the distance from each user to the local base station when the number of BSs is very large. Therefore, the average sum of the cache-miss users can be approximated as Given that the user’s distributions are independent and random, we can apply the probability density function of to proving that are random variables which are independently and exponentially distributed. Thus, we can get that is a random variable following Gamma distribution . In this case, when the probability density function , we can derive . Therefore, when , (18) becomes . When , (18) can be approximated as where , , , , , and .

In summary, the average sum rate of the cache-miss users can be approximated as (i)When is large and , by simplifying (17), the average number of bits totally transmitted can be approximated as follows: (ii)When and , we can get , , and , , . So the formula can be approximated as

4.2. Average Total Power Consumption of the Network

In this work, we investigate the performance of EE in cellular networks with caching at BSs. To this end, we need to calculate the overall energy consumption caused by the transmission and storage of contents.

We have known that when the BS is in sleeping state, some units are turned off, such that the BS reduces part of the power consumption. As a result, we can divide the circuit power consumption and transmit power consumption into two states according to the working state of BS, respectively.

Finally, the overall power consumption of the network can be divided into four parts:(i)The average total transmit power of the base station in the entire downlink can be expressed as (ii)The average total power consumption of the power supply circuit can be expressed as where is the active state of the base station, reflects the impact of power amplifier, is the transmit power of the BS in active mode, is the supply power of the BS in active mode, and is the circuit power of the BS in sleeping mode.(iii)The average caching power consumption: in this paper, we consider DRAM as cache hardware. According to the power model in [33], if the power coefficient of the cache hardware is fixed, the average caching power consumption is related to the size and the number of content cached in the local caching device. The average caching power consumption of BS can be expressed as where is the power coefficient of caching hardware and is the normalized cache capacity.(iv)The average backhaul power consumption: we use fiber as the backhaul link (with capacity 1 Gbps). When we download contents, the backhaul link will generate power consumption, and the average backhaul power consumption of BS can be expressed as where is the power coefficient of backhaul equipment and denotes the power consumed by the backhaul for supporting the maximum data rate . From formula (26), we can find that the energy consumption for backhauling will reduce sharply with a little decrease of the cache size, but decreasing more and more slowly with the growth of cache size. Therefore, we need to find the best cache capacity to minimize power consumption.

Then, the average total power consumption at all BSs can be expressed as

4.3. Energy Efficiency of the Network

Substituting (16) and (27) into (15), we can approximate the network EE as

The goal of this paper is to maximize the energy efficiency of the downlink wireless network by finding out the optimal cache content in the local base station. Mathematically, we have the following optimization problem: where represents the probability of the cache-hit rate, is the limit of backhaul capacity, and is the condition that the cache can increase energy efficiency. The following result can be obtained for the maximization of EE and the analysis is carried out later.

Proposition 1. Consider that the BS has limited cache capacity and let denote the optimal cache size that maximizes EE. Then where is a Lambert-W function satisfying .

Proof. (i)We first derive the derivative of in (28); we obtain where and are considered as constants.(ii)Let , and the derivative of in the equation is (iii)So when ; then decreases with the increase of the value . When , the equation ; then we can get (iv)Since decreases with , we can obtain that EE monotonically increases when and EE monotonically decreases when . Through the above descriptions, we find that the EE can be maximized when .

5. Numerical and Simulation Results

In this section, we evaluate the energy efficiency of downlink networks with caching at BS. Meanwhile, we investigate the impacts of some system parameters on throughput and energy efficiency, for example, interference level , the ratio of user density to BS density , and backhaul capacity .

In the work, the distribution of BS and user depends on HPPP in the whole downlink network, and the user density is set to be . In order to facilitate calculation, we record the parameters in Table 1 that will be used throughout the numerical analysis.

5.1. Caching and Backhaul Power Consumption

In this section, we validate whether both the energy consumption for content update and the power consumption for transmit content influence the total power consumption, where the transmission power involves the power consumed by the core network transmitting the content to the BS through backhaul link.(i)Caching power consumption: in interval , we assume that percent of the cached dynamic contents are updated in BSs. According to (25), the percentage of caching energy consumption with respect to the average total energy consumption at all BSs during can be expressed as where is the caching energy consumed for updating contents. In Section 3, we consider that many contents are dynamic changes at lifetime , and the popularity of many contents changes slowly. We set up several sets of data in Table 2 to validate the assumption that the cache energy consumption for content updates can not be ignored when the content popularity changes slowly. According to the result of in the table, we find that . Then, the caching energy consumption for content update can not be ignored when content popularity is dynamic changes in lifetime .(ii)Backhaul power consumption: we assume that percent of the requested contents are transmitted over the backhaul link; the percentage of backhaul power consumption to the total power consumption in interval is According to the above parameters, we can obtain ; then we must consider the backhaul power consumption for the total power consumption in all BS.

Based on the above analysis, we have a knowledge that the power consumption of cache and backhaul can not be ignored in the total power consumption when analyzing EE.

5.2. Impact of Parameters on Throughput

We compare the simulation results of the average throughput per cell with the numerical results obtained from (16) versus in Figure 4. We can see that the throughput increases with increasing backhaul capacity. Moreover, the throughput increases with more sharply when is small.

To show the impact of different cache size on the average throughput per cell, we provide the numerical results derived from (16) versus in Figure 5. It can be seen when the interference level is the same and the throughput of the downlink increases with the cache capacity. Moreover, we can also find when is small, the average throughput increases with more sharply. This suggests that the throughput can be boosted more efficiently by caching at the BS if the ICI level can be reduced.

5.3. Impact of Parameters on EE

According to the derivation of the optimal cache capacity for EE maximization in Section 3, we derive numerical results of EE with both backhaul capacity and cache capacity. As shown in Figure 6, when only a small portion of contents are cached at each BS, the EE firstly increases rapidly with the increasing backhaul capacity (i.e., ); then the performance gain approaches a constant. This is due to the fact that though the throughput is increased and the backhaul power consumption is reduced by caching, ICI limits the final throughput.

In Figure 7, we show the numerical results of EE obtained from (28) versus the normalized cache capacity with different ratios of BS density to user density. It can be seen that the EE firstly increase with quickly and then decreases gradually. This suggests that there is an optimal normalized cache capacity that maximizes EE. Moreover, with the same cache capacity, EE increases with .

In Section 3, we propose a distributed caching strategy to cache different contents at four adjacent BSs. Figure 8 shows the simulation results of EE for nondistributed caching strategy, distributed caching strategy which caches different contents at two adjacent base stations, and distributed caching strategy which caches different contents at four neighbouring BSs. We can see that distributed caching can achieve higher EE as shown in the figure. This is because when each BS uses a distributed cache strategy to cache contents, each user can access cache contents; that is, the cache-hit rate is increased. No matter what kind of cache policy is adopted, energy efficiency firstly increases and then decreases, which indicates the existence of an optimal cache capacity to maximize energy efficiency.

6. Conclusion

In this paper, we have investigated the energy efficient caching problem in backhaul-aware cellular networks jointly considering the cache-hit rate and the optimal cache. Considering that the content popularity is dynamic, we have proposed a policy with “age” threshold based on the shot noise model to timely estimate content popularity. Then, we have further proposed a distributed caching policy to improve the hit rate. Finally, based on the analysis of the tradeoff between energy efficiency and cache capacity, we have formulated an optimization problem and have derived a closed-form solution of optimal cache capacity for maximizing the EE. The numerical results show that the EE can be improved with appropriate choice of cache capacity. Meanwhile, compared with the nondistributed cache scheme, the distributed cache scheme can further improve the energy efficiency by increasing the cache-hit rate.

Notations

, :Distribution density of BSs and users
:Number of BSs
:Number of antennas
:Content catalog
:Number of content cached at the th BS
:Each content size (in bits)
, :Limited cache capacity and backhaul capacity
:Round radius of cells
:Distance from BS to user
:Rayleigh fading channel vector
:Transmit power of each BS
:The supply power of BS in active mode
:Circuit power of the BS in sleeping mode
:White Gaussian noise
:The percentage of intercell interference
:Path-loss exponent
:Normalized cache size
, :Power consumption per bit (in Watt/bit)
:Average arrival rate
:The request arrival rate of a content at time
:Represent that content is cached in BS at time
:The “age” of active content at time .

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Acknowledgments

This work is supported by Natural Science Foundation of Jiangsu Province (BK20170805, BK20160874, and BK20160812), Natural Science Foundation of China (61701230), and by research funding (2016-PYS/K-KY-J061).