Abstract

Smart device users spend most of the fragmentation time in the entertainment applications such as videos and films. The migration and reconstruction of video copies can improve the storage efficiency in distributed mobile edge computing, and the prediction of video hits is the premise for migrating video copies. This paper proposes a new prediction approach for video hits based on the combination of correlation analysis and wavelet neural network (WNN). This is achieved by establishing a video index quantification system and analyzing the correlation between the video to be predicted and already online videos. Then, the similar videos are selected as the influencing factors of video hits. Compared with the autoregressive integrated moving average (ARIMA) and gray prediction, the proposed approach has a higher prediction accuracy and a broader application scope.

1. Introduction

At present, smart device users spend more than 70% of the fragmentation time in the entertainment applications such as videos and films. The video content providers (e.g., Netflix) desire to know the future video view counts of all their videos, especially the new ones, to provide a better experience for consumers. In the era of the 5G-based mobile edge computing, the explosive increase of video resources requires placing multiple copies to the edge of the network for better performance [13]. The distributed model brings many problems for video service providers, such as how to keep the storage efficiency and improve the energy efficiency of storage [4, 5]. The key to solve these problems is to effectively manage data copies and data nodes [68]. The current popular cloud storage platforms generally use a static storage mechanism, that is, setting the number of copies before placing them, such as Google File System [9], Hadoop Distributed File System (HDFS) [10], and Amazon Dynamo [11]. The static placement of copies is easy to implement, but it may lead to unbalanced access. For example, it is found that 90.26% of the data in Yahoo’s Hadoop Cluster can only be accessed within two days after constructing, 89.61% of the data from the last access to deletion do not exceed 10 days, and 40% of the data have a dormant period (not accessed) for more than 20 days [12].

The current research [13] shows that the migration and reconstruction of video copies is an effective means to improve the storage efficiency, and the prediction of the number of video hits is the prerequisite for migrating video copies. There are many data prediction and recommendation approaches [14, 15]; however, they do not consider the fine-grained granularity of every video copy and cannot provide the required information for migration and reconstruction of video copies. Additionally, the videos occupy a high proportion of storage space and have more rich attributes, and the trend of video hits is influenced by various factors, which are difficult to predict accurately. Based on the combination of correlation analysis and wavelet neural network (WNN), this paper proposes a new prediction approach for video hits by analyzing the correlation between the video to be predicted and already online videos, and selecting the similar videos as the influencing factors.

2.1. Feature Analysis of Video Copies

A video copy covers a variety of attributes, such as news, teaching, and viewing, and it is the most representative type of data copies which are stored on the cloud. The existing hits prediction are coarse-grained (see Figure 1(a)), and they do not consider the fine-grained hits of every copy and cannot provide information for migration and reconstruction of video copies. The prediction approach that provides the hits of video copies is an urgent problem.

Figure 1(a) shows the trend of a long-timescale (in months) from March 2014 to February 2015. The reason for the smooth rising during the first few months is that “Transformers 4” released on June 2014 stimulates the previous series and even related science fiction series. From the short-timescale (in weeks) counterpart in Figure 1(b), it is found that the video hits have a certain periodicity and autocorrelation. This means that the video hits are affected by the hits of other similar videos. Therefore, the following are the general directions to follow to predict video hits: the short learning time of time series data and the small number of available parameters.

2.2. Existing Prediction Approaches

With more requirement of new entertainment, various prediction approaches for guaranteeing quality of service (QoS) have attracted increasing attention in recent years [16, 17]. Zhang et al. [18] proposed a covering-based quality prediction method for web services via neighborhood-aware matrix factorization. Qi et al. [15] proposed a novel privacy-aware data fusion and prediction approach for smart city industrial environment, which is based on the classic locality-sensitive hashing technique. Zhang et al. [19] proposed a distributed edge QoS prediction model with privacy-preserving for the edge computing networks. However, these works are from macroscopic point and cannot apply to the hits prediction of video copies which requires a micro perspective.

The autoregressive integrated moving average (ARIMA) model is based on the autocorrelation of time series, which is characterized by the fact that the first and the time series are broad and stable. Additionally, if the value of the individual data does not fluctuate up and down in the sequence mean value, ARIMA can do smooth processing of the original data through differential way. In this way, even if the data have a certain degree of fluctuation, the prediction accuracy can be achieved through the smooth processing. Therefore, ARIMA is suitable for prediction of the data with the flat trend and feature of linear wide stationary processes. It is often used in network traffic prediction [20, 21], while the traffic of each data copy has a more fine-grained granularity.

Gray prediction is a knowledge acquisition method with incomplete or uncertain data [22]. It processes the original data through the analysis of the difference degree of the changes in the system factors and predicts the future data by establishing the gray differential prediction model based on a small amount of information. Gray prediction model applies to the situation that the video copy is just launched into the market, which has little original data. These two methods seem to be feasible, but because of the video copy of the fluctuations in many factors, we analyze this video from the factors.

This paper considers the fact that the number of hits for each video is closely related to its similar videos. This is achieved by analyzing the principal components of the sources of the video viewing modes. For example, a systematic quality correlation model was proposed that described three different types of quality correlations between services [23]. Logically, there are some relevant video recommendations after watching a certain video on the web, and the recommendation and impact of the associated videos are the principal components of video-on-demand. By investigating and analyzing the Storm website, after a video is watched, the average viewing rates of the top three similar videos recommended by Storm are 47%, 30%, and 25%, respectively. The average value of the recommendation results comes from 200 videos which are randomly sampled from the Storm website. This means that when the total number of viewers remains stable, it is possible to infer the number of potential viewers of a certain video in the future, according to the current number of viewers of its similar videos. There exists delay during the process, and it is essential for intelligent analysis. This paper applies the wavelet neural network to predict future video hits. The wavelet-based neural network replaces the activation function of hidden nodes with wavelet function. In recent years, several researches applied the wavelet neural network model to predict network traffic [24], but the hits prediction of video copies is a more micro perspective. The key to solve this problem is to inversely deduce the videos that are used as prediction-relevant parameters, according to the video to be predicted.

2.3. Selection of Associated Videos

The choice of related video in this paper is divided into two steps: the first step is to select 12 videos that are similar to the video to be predicted; the second step further reduces the parameter dimension, from the similarity of the video selecting the highest degree of relevance of the four.

2.3.1. Similar Video Selection

Similar video selection is in the same type of videos through the establishment of vector space model (VSM) to calculate and select the level of the video, the degree of the audience, the type of video, the age of the audience, the influence of the producers, etc., and then refine and quantify the score between 0 and 100, as shown in Table 1.

Let m1 and m2 be the video objects to be compared, and f1, f2,……, fn be the attributes of every object. The similarity between m1 and m2 can be calculated by using the attributes of video objects. The values of the video attribute fi corresponding to m1 and m2 are denoted by ai and bi, respectively. By using support vector machine (SVM), the similarity between m1 and m2, denoted by , is calculated as follows:

From equation (1), we can calculate the most similar 12 videos, in order to further reduce the dimension, and from the 12 videos to select the most relevant video as a prediction parameter.

2.3.2. Correlation Degree Analysis

Table 2 shows the hits of 6 sample videos within 10 days, where x1 is from 2014.10.26 to 2014.11.24, and the data of x2, …, x5 cover the time span from 2014.10.20 to 2014.11.18. x1 is the video to be predicted, which can be obtained by the hits of other videos in the last week.

The calculation of the association degree can be divided into the following steps.Step 1: standardize the data in Table 2, where x, y represent longitudinal data and horizontal data, respectively. The sample mean of -th video is (j = 1, 2, ……, n); through it, we can calculate the sample variance of -th video: (j = 1, 2, ……, n); standardized form is . After standardized calculation, we haveStep 2: calculate the correlation matrix, , ; then we can analyze the correlation between the videos after getting the correlation matrix, as shown in Table 3, in which the x1 is the video that needs to be predicted; x2, x3, x4, x5, x6 is the similar video that needs to be analyzed for the degree of change.

From 12 similar videos, select the correlation matrix and x1 closest to the four videos used as predictive video; the above matrix is the first batch to be compared with video x1; it is clear that x4, x5, and x1 are highly correlated with clicking playback. In this way, select the videos x3–x5 and use the current number of hits to predict the hits of video x1.

2.4. Wavelet Neural Network

Neural networks are often used to predict and analyze non-linear time series. In theory, neural network prediction accuracy can be achieved arbitrarily. But in practical application, it may meet many difficulties, such as the delineation of the structure of the neural network, training and learning process being too slow, and being trapped in local second-best in the optimization process. WNN combines the characteristics of wavelet analysis and neural network. It replaces the neurons in the neural network with wavelet neurons, replaces the Sigmoid in the neural network with wavelet functions (as activation function), and establishes the relationship between wavelet transform and network coefficients through affine transformation. The WNN has the following characteristics:(1)High prediction accuracy due to strong learning ability and excellent function approximation ability.(2)The fact that local optimum can be avoided as the shift factor and scaling factor of the network are determined in advance.(3)High learning and training speed as wavelet function as an activation function is easy to implement.

The WNN has three layers: input layer, hidden layer, and output layer, as shown in Figure 2. The output layer adopts the linear output. The neurons of the input layer are (x1, x2, …, xm), the hidden layer has neurons, and the output layer has neurons. In this case, since the number of associated videos is 4, there are 4 related parameters as the basis for prediction and the input neuron selection is 4. In the set of daily visits, it is found that the visits are recorded in the unit of day; the visit cycle is 7 days. Then, the number of neurons used in the middle of the hidden layer is set to 7, and the output layer is 7. That is why this case uses 4-7-7 structure neural network. is the wavelet basis function, instead of the previous activation function Sigmoid.

Let denote the weights of neurons from layer to layer , denote the -th input of neuron in layer , denote the transfer function of layer , and denote the corresponding output of the layer.

Since the case is designed in a three-tier network, , are the transfer functions in the hidden layer. This case is called “Morlet wavelet,” which is given by

Substituting (5) into (3) and (4), we get

According to (6) and (7), we can conclude that the video prediction on time series is

Given the input and output samples of group, the error function can be expressed as where is the expected output of the -th node and is the actual output of the -th node. , , and are constantly adjusted to minimize the error. When the error is less than the given value, the program will end, and an appropriate prediction value can be calculated.

For a series of videos that have similar and stable attributes, the proposed approach can achieve reasonable prediction results. However, for certain videos that have a worse or better reputation than other videos in the series, the proposed approach may produce a biased prediction result. This is because a potentially unbalanced reputation has affected the final results.

2.5. Performance Evaluation

As the actual video copy, if the prediction granularity is too low, such as one day, it does not make much sense, because it is impossible for the video copy to undergo frequent evaluation migration with a 24-hour time unit. If the prediction granularity is too long, such as in the unit of month, the weekly trends of a video copy cannot be timely captured. So, this experiment is expected to have a length of weeks and data collection is based on days, the estimated collection of 84 days of data. The number of targets to be tested is from 2014.10.26 to 2015.1.17, and the video data as the influencing factor are from 2014.10.20 to 2015.1.11, the first 56 days of the two sides are used as training data, and the last 28 days of data are used as the test data. The experiment runs on 64-bit Windows 7 Professional with 10 GB of RAM and 2.1 GHz Intel Core i3 processor and uses MATLAB R2014a to conduct the simulation and numerical analyses. The neural network is designed to be 4-7-7 structure, and the output is 7-node data.

From Figure 3, it can be seen that ARIMA through differential algorithm fits the time series data well before prediction, but it is not accurate in grasping the trend of the data. Figure 4 shows that the gray prediction can predict the general trend of the data but has a large error range. In terms of the change trend and prediction accuracy, the performance of the proposed approach exhibits the optimal results, as shown in Figure 5.

If it is predicted after a week, it will become a training set and then forecast the way of the next week. The time granularity of each prediction is not the same, which is 7 days, 14 days, and 28 days, respectively. As can be seen from Table 4, if the time granularity for prediction is 1 week or 4 weeks, the proposed approach is the most accurate and far higher than those of ARIMA and gray. If the time granularity for prediction is two weeks, the accuracies of the three prediction methods are comparable. In the proposed approach, the histories of hits of the similar video are used as the influencing factors. Therefore, it can be inferred that the prediction approach based on the combination of the correlation analysis and WNN has better practical effect than the alternatives.

When applying this method to several other types of videos, the result is similar to the above case. The average prediction accuracy of the proposed approach is 10% higher than that of ARIMA and 5–7% higher than that of gray prediction.

From the above analysis, each prediction approach has its own advantages and disadvantages. ARIMA has certain advantages if the video enters a certain life cycle, which have the relatively flat change trend and the feature of linear wide stationary processes. But if the change trend of hits is gentle, it has little significance for migrating video copies.

3. Conclusions

This paper proposes a new prediction approach for video hits based on the combination of correlation analysis and WNN. This is achieved by establishing the video index quantification system and analyzing the correlation between the video to be predicted and already online videos. Then, the similar videos are selected as the influencing factors of video hits. Compared with the ARIMA and gray prediction, the proposed approach has a higher prediction accuracy and a broader application scope.

Data Availability

No data were used to support this study.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Authors’ Contributions

Xiulei Liu and Junyang Yu proposed the overall idea of this manuscript. Shoulu Hou and Xuhong Liu designed the experimental plan. Qiang Tong and Zhihui Qin wrote and revised the manuscript. All authors have contributed to this research work. All authors have read and approved the final manuscript.

Acknowledgments

This work was supported in part by the National Key R&D Program of China (No. 2018YFC0830202); supported in part by Qin Xin Talents Cultivation Program, Beijing Information Science and Technology University (2020); supported in part by Construction Project of Innovative Scientific Research Platform for Edge Computing (No. 2020KYNH105); and supported in part by the Opening Foundation of State Key Laboratory of Digital Publishing Technology.