Journal of Advanced Transportation

Journal of Advanced Transportation / 2020 / Article
Special Issue

Data Analysis and Optimization for Intelligent Transportation in Internet of Things

View this Special Issue

Research Article | Open Access

Volume 2020 |Article ID 8859538 | https://doi.org/10.1155/2020/8859538

Ya Zhang, Mingming Lu, Haifeng Li, "Urban Traffic Flow Forecast Based on FastGCRNN", Journal of Advanced Transportation, vol. 2020, Article ID 8859538, 9 pages, 2020. https://doi.org/10.1155/2020/8859538

Urban Traffic Flow Forecast Based on FastGCRNN

Academic Editor: Naixue Xiong
Received23 Jun 2020
Revised09 Aug 2020
Accepted15 Sep 2020
Published27 Sep 2020

Abstract

Traffic forecasting is an important prerequisite for the application of intelligent transportation systems in urban traffic networks. The existing works adopted RNN and CNN/GCN, among which GCRN is the state-of-the-art work, to characterize the temporal and spatial correlation of traffic flows. However, it is hard to apply GCRN to the large-scale road networks due to high computational complexity. To address this problem, we propose abstracting the road network into a geometric graph and building a Fast Graph Convolution Recurrent Neural Network (FastGCRNN) to model the spatial-temporal dependencies of traffic flow. Specifically, we use FastGCN unit to efficiently capture the topological relationship between the roads and the surrounding roads in the graph with reducing the computational complexity through importance sampling, combine GRU unit to capture the temporal dependency of traffic flow, and embed the spatiotemporal features into Seq2Seq based on the Encoder-Decoder framework. Experiments on large-scale traffic data sets illustrate that the proposed method can greatly reduce computational complexity and memory consumption while maintaining relatively high accuracy.

1. Introduction

Traffic forecasting using timely information provided by Internet of Things technology (IoT) is an important prerequisite for the application of intelligent transportation system (ITS) [1] in urban traffic networks, because an accurate and efficient prediction model can be used for travellers to select high-quality reference routes, maximize the utilization of road networks, and provide a basis for the reasonable planning of urban construction departments. However, along with worldwide urbanization, urban road networks have been expanded significantly [2], which brings challenges for traffic forecasting because the corresponding computation complexity will greatly increase due to the expanded road networks [3].

This paper mainly studies the problem of urban traffic forecasting based on the Internet of Things technology (IoT) in large urban road traffic networks. This problem is how to use historical traffic flow data to predict traffic flow data in future timestamps in large urban road traffic networks. In the literature, there has been plenty of studies in traffic forecasting, including traffic volume, taxi pick-ups, and traffic in/out flow volume. Initially, numerous statistical based methods, such as Historical Average (HA) [4], Time Series [5], K Nearest Neighbors Algorithm (KNN) [6], and Kalman Filter [7], have been proposed to predict road traffic. However, these models are generally suitable for relatively stable traffic flow, which cannot well reflect the temporal correlation of traffic flow data, nor can they reflect the real-time nature of traffic flow. In order to solve the unstable characteristics of traffic flow data, ARIMA [8] and its variants [9, 10] are used in this field [11]. Although these studies show that the prediction can be improved by considering various other factors, they are still unable to capture the complex nonlinear spatiotemporal correlation. The latest advances in deep learning enable researchers to model complex nonlinear relationships and show promising results in multiple fields. This success has inspired many attempts to use deep learning technology in traffic flow prediction. Recent studies have proposed the use of improved LSTM [12] and GRU [13] to predict traffic flow. Furthermore, considering the influence of spatial structure on the traffic flow of different roads, Li et al. [14, 15] proposed modeling the traffic volume of the city as an image and partitioning the city map (the image) into a large number grid. Within each grid cell, the traffic volume within a period of time can be regarded as a pixel value. Based on that, Li adopted ConvLSTM [16] to model the spatial-temporal correlation among traffic flows, where the convolution operation and the LSTM unit are utilized to model spatial and temporal correlation, respectively. However, the conversion of traffic flow into images loses the spatial topology of urban roads. Li et al. [17] modeled the traffic flow as a diffusion process on a directed graph and captured the spatial dependency using bidirectional random walks on the graph and the temporal dependency using the Encoder-Decoder architecture with scheduled sampling. Seo et al. [18] used GCN [1921] to extract the spatial topology of the traffic network and RNN to find dynamic patterns to optimize traffic forecasting. However, GCN suffers from the scalability issue, because it requires a lot of space to maintain the entire graph and embed each node in memory [2226], and it has a very high computational complexity [27].

In order to solve the above problem, we propose forming the road network into a geometric graph and constructing a spatiotemporal graph convolution network based on the abstract graph to capture the spatiotemporal features of traffic flow for prediction. We propose using GCN as the spatial topology extractor of the model and applying the sampling method [2830] to GCN. The method can put the nodes in the graph into the model in batches, sample the neighbors of the nodes in each batch, extract the nodes that have a greater impact on the nodes in this batch, and perform convolution operations, which greatly reduces the calculation complexity and memory. FastGCN can effectively process the large graph by importance sampling so that memory overflow is not easy to occur. Then, we further combine GRU to extract temporal features to achieve the extraction of spatiotemporal features of traffic flow. Finally, we embed spatiotemporal features into Seq2Seq [31] based on Encoder-Decoder framework for prediction.

2. Problem Analysis

Urban traffic flow prediction is based on historical traffic flow sequences, which are highly time-varying, nonlinear, and uncertain. The traffic flow in the road network usually has the following temporal characteristics [32]:(a)Periodicity. Traffic flows change periodically. The time series of traffic flow usually presents a wavy or oscillatory fluctuation around the long-term trend.(b)Trend and Trend Variability [33]. The time series of traffic flow shows a regular change trend. It will not change randomly, but it will continuously change with time. For example, from spring to summer, the traffic volume of the morning peak will gradually advance. Present a trending change.(c)Continuity. Traffic flow has continuity in time; that is, there is a correlation between the value of traffic flow at different times, especially in adjacent time periods.

At a certain time, traffic flow also has some spatial characteristics, such as the impact of traffic flow upstream and downstream of the road on the current road, and the rules of speed limit and traffic flow limit of the same level of road.

In view of these two main influence factors, especially considering the large scale of the road network [3439], which requires a lot of time for spatial calculation, this paper proposes the Fast Graph Convolution Recurrent Neural Network (FastGCRNN).

It uses recurrent neural network to capture the long-term temporal dependency of traffic flow and the graph convolution neural network (GCN) to capture the spatial correlation among roads in different geographical locations. At the same time, importance sampling is applied to GCN to reduce the computational complexity of large road networks.

3. Preliminaries

3.1. Notations

Given an undirected graph , where is a set of nodes with , is a set of edges that can be represented as an adjacency matrix , and is a feature matrix with denoting a feature vector of node . is the length of the historical time series, and each feature in corresponds to the traffic flow at a certain time. Our target is to obtain the traffic information ( is the length of traffic flow time series to be predicted) of a certain period of time in the future according to the historical traffic information X.

3.2. Graph Convolution Networks

As a semisupervised model, GCN can learn the hidden representation of each node. The hidden vectors of all nodes in layer can be represented recursively by the hidden vectors of layer as follows:where , denotes the learnable weight matrix at layer , , and is an activation function, such as ReLu. Initially, .

4. Fast Graph Convolution Recurrent Neural Network

The traffic flow of a road is affected by the traffic flow of the surrounding roads and the historical traffic flow of the road itself, so the prediction model should consider these two factors. To model the temporal dependency of historical traffic on the road, GRU unit is embedded in the Seq2Seq model based on Encoder-Decoder framework to complete sequence prediction, and, to model the spatial correlation among neighbor roads, FastGCN is used in the traffic map of the road network to reduce the computational complexity and improve the efficiency. We integrate a model for quickly extracting spatiotemporal features, so we propose the FastGCRNN (Fast Graph Revolution Recurrent Neural Network) model. The overall architecture of the model is shown in Figure 1.

This model mainly includes six parts, namely:(a)Input Sequence X. It is the input data of the whole prediction model, which is fed into the encoder part. In the road network traffic graph, it is the traffic flow of each node in a continuous period of time.(b)Output Sequence Y. It is the output of the whole prediction model (the output of decoder part). In the road network traffic graph, it is the traffic flow of each node road in the future.(c)FastGCN Unit. It can extract the spatial structure information of the road network through graph convolution. Based on that, it further uses sampling to reduce computational complexity.(d)GRU Unit. Traffic flows are time series signals, so we use GRU units to capture the long-term or short-term temporal dependence between the input traffic flow time series and embed two FastGCN units in its internal.(e)Encoder Unit. It is composed of GRU unit, and the output state of hidden layer is obtained by encoding the time series of the input traffic flow network graph.(f)Decoder Unit. It is also composed of GRU units. When it receives the encoder output, the decoder will continuously predict the traffic flow of each node.

The whole FastGCRNN model adopts the Seq2Seq model based on the Encoder-Decoder framework, which can use traffic flow of each road within the road network to predict the future traffic flow. Firstly, the continuous traffic flow data on the road network is fed into the encoder part, and the data instance at each timestamp needs to go through FastGCN units to extract the spatial structure information between the road nodes, and then it needs to be processed by the GRU units in the encoder to get the temporal features of the traffic flow. After encoding, the hidden state output by the GRU units in the encoder is fed to the GRU units in decoder, and spatial features are further extracted by FastGCN. The final GRU units continuously predict the traffic flow .

4.1. Fast Spatial Feature Extractor: FastGCN

Each road in the urban road network does not exist in isolation but connects with the surrounding roads to form a whole. The traffic flow between roads is interactive; particularly, on the two-way road, there are vehicles flowing in and out. To model spatial correlation of traffic flows among road networks, we abstract the roads in road networks as nodes and their intersections as edges, as shown in Figure 2, where blue lines and dots represent road and intersections in road networks, respectively. Because we intend to predict traffic flows of the roads, while GCN can only make prediction on nodes, we model roads as nodes and their intersections as edges, as illustrated through the red triangles and yellow lines in Figure 2, respectively.

In order to consider the influence of multihop in GCN, the number of layers of GCN will be increased recursively to realize the information exchange between multiple upstream and downstream roads. However, the recursive neighborhood expansion across layers poses time and memory challenges for training with large, dense graphs. To solve this problem, the FastGCN method is used, which interprets GCN as the integral transformation of the embedded function under the probability measure. The integration at this time can use the Monte Carlo method for consistency estimation, and the node training in the graph can also be performed in batches. Because the node training is carried out in batches, the structure of the graph is not limited; that is, when performing test prediction, the number of nodes and the connection relationship in the graph can change, and it does not have to be the same as the graph structure during training. This increases the generalization ability and scalability of the model to a certain extent.

The nodes in the graph of FastGCN can be regarded as independent and identically distributed sampling points that satisfy a certain probability distribution, and the calculated loss and convolution results are expressed as the integral form of the embedding function of each node. The estimation of integration can be expressed by Monte Carlo approximation which defines the sampling loss and sampling gradient. In order to reduce the variance of estimation, the sampling distribution can be further changed to make it more consistent with the real distribution. For example, the simplest way is to use uniform distribution for sampling convolution. The improved method is to use importance sampling to make it continuously approach the real distribution and reduce the error caused by sampling.

If a node in the graph is taken as the observation object, its convolution can be considered as the information embedding expression of node and all nodes in the graph in the upper layer through the addition of other forms of adjacency matrix and then the transformation of feature dimension through the trainable parameter matrix, which is equivalent to a discrete integral, and the adjacency matrix is equivalent to a weight given to each node. Therefore, the convolution process of node in the graph is expressed in integral form as

GCN in the form of integration is integrated by Monte Carlo method, and then it is transformed into the discrete form of sampling. At layer , points are sampled independently and identically with probability , and the approximate estimation is

If each layer of convolution uses this method for sampling and information transfer, after layer , the embedded expression of node is

In the above integral form of GCN, the embedded information expression of node V needs to be obtained from all nodes in the graph. However, after sampling, only nodes in the graph need to exchange and fuse information in FastGCN, so the calculation complexity of the whole graph changes from to , and the efficiency is greatly improved.

Here is an example to illustrate the advantages of FastGCN compared with GCN, if the abstract road network graph has 5 nodes and 6 edges, as shown in Figures 3 and 4.

In GCN, each epoch must be put into a complete graph, instead of using only a few nodes in the graph; that is, each node in the graph needs to convolute and exchange information with all other nodes in the graph. In FastGCN, we decompose the large graph into several small graphs by batch operation and put them into memory, as well as the method of sampling to remove the information exchange with some low correlation nodes. Each node only interacts with the sampled nodes in the graph. As shown in Figure 4, each node only interacts with node A and node E. In this way, the computing efficiency is greatly improved, especially when it can be calculated on a large graph without memory overflow.

For the sampling method, in order to make the sampling closer to the real connected nodes, FastGCN does not use uniform sampling [40], but importance sampling. That is, each node is not sampled according to the same probability, but using probability distribution . No matter what probability distribution sampling is used, the mean value of the sample is constant, but it will affect the variance of the sample. In order to minimize the error, the distribution which can minimize the sample variance is selected here. At this time, the calculation output of node passing through FastGCN layer is

In the experiment, only two FastGCN units were used to extract spatial features. This is because we need to avoid the problem of oversmoothing [41]. The specific calculation process is as follows:

4.2. Fast Temporal Feature Extractor: GRU

This is a key issue to effectively capture the long-term temporal dependence of traffic flow. The observed value of each timestamp is shown in Figure 5. The flow value of each node will change with time. The prediction is a typical time series prediction problem; that is, given the observed value of each road at timestamps in history, the traffic flow value of timestamps in the future will be predicted.

LSTM and GRU are commonly used in time series prediction. Both models use gating mechanisms to remember as much long-term information as possible and are equally effective for various tasks. To maximize efficiency, we chose GRU with relatively simple structure, fewer parameters, and faster training ability. GRU unit has update gate, reset gate, and memory unit, which can make it have a process of screening memory for historical data, so it can retain long-term memory. In GRU, time sequence information is saved by memory unit, which can capture long- and short-term memory in time and improve the accuracy of prediction.

In order to complete the sequence prediction, the Seq2Seq model based on the Encoder-Decoder structure is used. Seq2Seq puts the input history sequence into GRU, extracts the timing features, and obtains the hidden state vector of the input sequence as the coding result of the encoder. This state vector contains the feature information of all previous moments, which is a centralized embodiment of their temporal features. In the decoder, is used as the initial input of decoder to generate the predicted time series. In this way, Seq2Seq can extract the temporal characteristics of the traffic volume in the previous period, such as the proximity, trend, and periodicity of the traffic flow in the time dimension. When predicting the traffic volume, the model can obtain a smoothly changing traffic volume according to the proximity, and the characteristics of the proximity can be adjusted according to the trend and periodicity.

5. Experiment

In order to illustrate the role of the model in the large graph, 1865 roads in Luohu District of Shenzhen city are selected for the experiment, and the specific roads and areas are shown in Figure 6.

To calculate the traffic flows in each road, we map the GPS coordinates to the corresponding roads through the Frechet method [42]. The format of the mapped data is shown in Table 1. The core fields are road number (road_id), license plate number (car_id), and upload time (time). Each data record represents the information; the taxis with the car_id are on the road with road_id at the specific time.


Road_idCar_idTime

92230023412015-01-01 00 : 03 : 46
92230039822015-01-02 06 : 23 : 12

5.1. Data Preprocessing

In data preprocessing, the taxi data in Shenzhen is transformed into the form of continuous timestamps on the road network, i.e., the traffic data shown in Figure 5. Specifically, we map the original GPS upload data to the road and count the traffic flow on each road in each time period. The data preprocessing algorithm is shown in Algorithm 1.

(1)Initialize: time_interval = 5 min (or 30 min), begin_time=2015-01-01 00:00:00, roadflow[roadid][time_num] = 0
(2)For Every data record do
(3)time_num ←(timebegin_time)/time_interval
(4)End for
(5)All data records are grouped by car_id, sorted by time_num within the group
(6)For each group records do
(7) Remove duplicates records based on road_id and time_num
(8) Count roadflow
(9)End for
(10)Output: roadflow
5.2. Comparative Experiment

The biggest advantage of FastGCRNN model is that it can be applied to large graphs, and it can reduce the computational complexity without losing the accuracy of the model. On the road network data of Shenzhen, the experiment is conducted with the traffic flow series of different time intervals to compare with some classic traffic flow prediction models: (1) HA, (2) ARIMA, (3) SVR, (4) LSTM, (5) ConvLSTM, (6) GCRN [18], and (7) GCRNN-nosample. The evaluation standard used in the experiment is Root Mean Squared Error (RMSE) [43]. The specific experimental results are shown in Table 2.


ModelTime
RMSE
5 min30 min

HA19.50223.158
ARIMA17.54119.097
SVR17.89519.005
LSTM13.10216.930
ConvLSTM19.48121.038
GCRN11.89216.265
GCRNN-nosample9.95016.231
FastGCRNN9.86716.2734

From the table results, we can find that FastGCRNN model has reached the best prediction performance in terms of RMSE. In these comparison models, HA, ARMIMA, SVR, and LSTM only consider the temporal correlation without considering the spatial correlation, which is also one of the reasons for their poor accuracy. ConvLSTM divides the urban area into a grid and maps the traffic volume in each time period to the grid, and the traffic volume is regarded as the pixel value of the grid. Although this method considers the spatial correlation of vehicle flow, it also loses the topological structure relationship of the road network graph.

For verification, the proposed GCRNN can reduce the computational complexity, compared with the GCRN model, which also captures the topology information of the road network; the result is shown in Figure 7.

In Figure 7, we only compare the baselines with higher prediction accuracy, namely, GCRN and GCRNN-nosample. From Figure 7, it can be observed that the computational complexity of FastGCRNN is the lowest. The training time of FastGCRNN is about 0.03 times that of GCRN. Moreover, FastGCRNN reduces the training time to 1/3 times that of GCRNN-nosample, i.e., the GCRNN model without sampling. From the experiment results, it can be concluded that both the GCRNN model and the sampling method can reduce the training time.

5.3. Model Parameter Analysis

In FastGCRNN, each sampling point has a certain effect on the accuracy and training time of the model. When using 1685 roads in Shenzhen for experiments, different sampling sizes were set to compare the accuracy and time changes. The experimental results are shown in Figure 8. The abscissa in the figure shows the sampling size of FastGCN unit in the first and second layers, respectively. The blue column represents the RMSE of the prediction results. The red line indicates the time consumption in each epoch, and the upper and lower ends are the maximum and minimum values of time consumption in the training process.

From the experimental results, it can be seen that choosing different sampling sizes has little effect on accuracy, and it does not necessarily mean that the more the samples, the more the information obtained, and the better the prediction effect. For example, the accuracy of sampling 50 nodes for each layer in the figure is not the best, because there are “bridge” type (other nodes affecting the central node will spread to other unrelated distant areas) and “tree” type (other nodes affecting the central node will be limited to the small area to which the node belongs) of connection relationship between nodes [44]. If more nodes are sampled, the influence relationship of the nodes will spread to unrelated areas, resulting in information redundancy, misleading the update of node features, and reducing the prediction accuracy. In addition, in the road network graph, intersections generally connect four roads; that is to say, selecting four nodes in one hop can complete the extraction of feature information. Here is the statistics of 1865 selected roads' degrees, as shown in Figure 9. Among them, the nodes with degree 4 are the most, and the degrees of 70% of the nodes are less than 5, and the degrees of nearly 99% of the nodes are less than 7. Therefore, the case of sampling size 5 can already include the neighbors in one hop around it. In this case, not only the training time is reduced, but also the accuracy is not reduced.

And we compared the time consumption of FastGCN and standard GCN in different sizes of graphs. The experimental results are shown in Figure 10.

From the experimental results, it can be seen that FastGCRNN has obvious advantages in dealing with large graph problems. Particularly, when the size of graph reaches a certain degree, FastGCRNN is still running normally when GCRNN-nosample model has overflowed memory and cannot be trained.

6. Conclusions

This paper mainly deals with the problem of large graphs with spatiotemporal properties by constructing the FastGCRNN model and applies them to road network traffic graphs. The model predicts the traffic flow by extracting the temporal and spatial attributes of the traffic flow on the large-scale road networks. Among them, FastGCN is used to extract the topological structure in the space and accelerate training and reduce complexity. GRU is used to extract time series features, and the Seq2Seq model based on the Encoder-Decoder framework can complete sequence prediction tasks of unequal length. The most prominent advantage of this model is the FastGCN embedded in it, which uses the sampling method to accelerate the extraction of spatial features, reduce computational complexity, and improve efficiency. Moreover, the model is not prone to memory overflow in processing large-scale graph-structured data.

It is worth mentioning that this model is not only applicable to traffic flow data, but also applicable to all graph structure data with spatiotemporal characteristics, especially the largerscale data.

Data Availability

The data used to support the findings of this study are available upon request to Ya Zhang, zndxxxxyzy@csu.edu.cn.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. U. Mori, A. Mendiburu, M. Álvarez, and J. A. Lozano, “A review of travel time estimation and forecasting for advanced traveller information systems,” Transportmetrica A: Transport Science, vol. 11, no. 2, pp. 119–157, 2015. View at: Publisher Site | Google Scholar
  2. Y. Zhang, T. Cheng, and Y. Ren, “A graph deep learning method for short-term traffic forecasting on large road networks,” Computer-Aided Civil and Infrastructure Engineering, vol. 34, no. 10, pp. 877–896, 2019. View at: Publisher Site | Google Scholar
  3. P. Wang, J. Lai, Z. Huang, Q. Tan, and T. Lin, “Estimating traffic flow in large road networks based on multi-source traffic data,” IEEE Transactions on Intelligent Transportation Systems, pp. 1–12, 2020. View at: Publisher Site | Google Scholar
  4. A. Reggiani and L. A. Schintler, Introduction: Cross Atlantic Perspectives in Methods and Models Analysing Transport and Telecommunications, Springer Science & Business Media, Berlin, Germany, 2005.
  5. M. S. Ahmed and A. R. Cook, Analysis of Freeway Traffic Time-Series Data by Using Box-Jenkins Techniques, Springer, Berlin, Germany, 1979.
  6. G. A. Davis and N. L. Nihan, “Nonparametric regression and short-term freeway traffic forecasting,” Journal of Transportation Engineering, vol. 117, no. 2, pp. 178–188, 1991. View at: Publisher Site | Google Scholar
  7. I. Okutani and Y. J. Stephanedes, “Dynamic prediction of traffic volume through Kalman filtering theory,” Transportation Research Part B: Methodological, vol. 18, no. 1, pp. 1–11, 1984. View at: Publisher Site | Google Scholar
  8. C. Kim and A. G. Hobeika, “Short-term demand forecasting model from real-time traffic data,” in Proceedings of the Infrastructure Planning and Management, pp. 540–550, Denver, CO, USA, June 1993. View at: Google Scholar
  9. X. Luo, L. Niu, and S. Zhang, “An algorithm for traffic flow prediction based on improved SARIMA and GA,” KSCE Journal of Civil Engineering, vol. 22, no. 10, pp. 4107–4115, 2018. View at: Publisher Site | Google Scholar
  10. N. K. Chikkakrishna, C. Hardik, K. Deepika, and N. Sparsha, “Short-term traffic prediction using sarima and FbPROPHET,” in Proceedings of the 2019 IEEE 16th India Council International Conference, INDICON 2019-Symposium Proceedings, pp. 1–4, Rajkot, Gujarat, India, December 2019. View at: Google Scholar
  11. B. Liu, X. Tang, J. Cheng, and P. Shi, “Traffic flow combination forecasting method based on improved LSTM and ARIMA,” International Journal of Embedded Systems, vol. 12, no. 1, pp. 22–30, 2020. View at: Publisher Site | Google Scholar
  12. B. Yang, S. Sun, J. Li, X. Lin, and Y. Tian, “Traffic flow prediction using LSTM with feature enhancement,” Neurocomputing, vol. 332, pp. 320–327, 2019. View at: Publisher Site | Google Scholar
  13. G. Dai, C. Ma, and X. Xu, “Short-term traffic flow prediction method for urban road sections based on space-time analysis and GRU,” IEEE Access, vol. 7, pp. 143025–143035, 2019. View at: Publisher Site | Google Scholar
  14. P. Li, M. Sun, and M. Pang, “Prediction of taxi demand based on convLSTM neural network,” in Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 11305, pp. 15–25, Springer, Berlin, Germany, 2018. View at: Google Scholar
  15. R. He, N. Xiong, L. T. Yang, and J. H. Park, “Using multi-modal semantic association rules to fuse keywords and visual features automatically for web image retrieval,” Information Fusion, vol. 12, no. 3, pp. 223–230, 2011. View at: Publisher Site | Google Scholar
  16. X. Shi, Z. Chen, H. Wang, D. Y. Yeung, W. K. Wong, and W. C. Woo, “Convolutional LSTM network: a machine learning approach for precipitation nowcasting,” in Advances in Neural Information Processing Systems, pp. 802–810, Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 2015. View at: Google Scholar
  17. Y. Li, R. Yu, C. Shahabi, and Y. Liu, “Diffusion convolutional recurrent neural network: data-driven traffic forecasting,” 2017, http://arxiv.org/abs/1707.01926. View at: Google Scholar
  18. Y. Seo, M. Defferrard, P. Vandergheynst, and X. Bresson, “Structured sequence modeling with graph convolutional recurrent networks,” in Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 11301, pp. 362–373, Springer, Berlin, Germany, 2018. View at: Google Scholar
  19. J. Bruna, W. Zaremba, A. Szlam, and Y. LeCun, “Spectral networks and locally connected networks on graphs,” 2013, http://arxiv.org/abs/1312.6203. View at: Google Scholar
  20. C. Lin, Y.-X. He, and N. Xiong, “An energy-efficient dynamic power management in wireless sensor networks,” in Proceedings of the 2006 Fifth International Symposium on Parallel and Distributed Computing, pp. 148–154, Rhodes Island, Greece, April 2006. View at: Google Scholar
  21. Y. Liu, M. Ma, X. Liu, N. Xiong, A. Liu, and Y. Zhu, “Design and analysis of probing route to defense sink-hole attacks for internet of things security,” IEEE Transactions on Network Science and Engineering, vol. 7, no. 1, 2018. View at: Publisher Site | Google Scholar
  22. L. Shu, Y. Zhang, Z. Yu, L. T. Yang, M. Hauswirth, and N. Xiong, “Context-aware cross-layer optimized video streaming in wireless multimedia sensor networks,” The Journal of Supercomputing, vol. 54, no. 1, pp. 94–121, 2010. View at: Publisher Site | Google Scholar
  23. Y. Wang, A. V Vasilakos, J. Ma, and N. Xiong, “On studying the impact of uncertainty on behavior diffusion in social networks,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 45, no. 2, pp. 185–197, 2014. View at: Publisher Site | Google Scholar
  24. H. Zheng, W. Guo, and N. Xiong, “A kernel-based compressive sensing approach for mobile data gathering in wireless sensor network systems,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 48, no. 12, pp. 2315–2327, 2017. View at: Publisher Site | Google Scholar
  25. Z. Wan, N. Xiong, N. Ghani, A. V. Vasilakos, and L. Zhou, “Adaptive unequal protection for wireless video transmission over IEEE 802.11e networks,” Multimedia Tools and Applications, vol. 72, no. 1, pp. 541–571, 2014. View at: Publisher Site | Google Scholar
  26. J. Li, N. Xiong, J. H. Park, C. Liu, S. Ma, and S. Cho, “Intelligent model design of cluster supply chain with horizontal cooperation,” Journal of Intelligent Manufacturing, vol. 23, no. 4, pp. 917–931, 2012. View at: Publisher Site | Google Scholar
  27. W.-L. Chiang, X. Liu, S. Si, Y. Li, S. Bengio, and C.-J. Hsieh, “Cluster-gcn: an efficient algorithm for training deep and large graph convolutional networks,” in Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 257–266, Anchorage, AK, USA, June 2020. View at: Google Scholar
  28. J. Chen, T. Ma, and C. Xiao, “FastGCN: Fast learning with graph convolu-tional networks via importance sampling,” in Proceedings of the 6th International Conference on Learning Representations, ICLR 2018-Conference Track Proceedings, Vancouver, Canada, May 2018. View at: Google Scholar
  29. Z. Wang, T. Li, N. Xiong, and Y. Pan, “A novel dynamic network data replication scheme based on historical access record and proactive deletion,” The Journal of Supercomputing, vol. 62, no. 1, pp. 227–250, 2012. View at: Publisher Site | Google Scholar
  30. Y. Yang, N. Xiong, N. Y. Chong, and X. Défago, “A decentralized and adaptive flocking algorithm for autonomous mobile robots,” in Proceedings of the 2008 The 3rd International Conference on Grid and Pervasive Computing-Workshops, pp. 262–268, Kunming, China, May 2008. View at: Publisher Site | Google Scholar
  31. I. Sutskever, O. Vinyals, and Q. V Le, “Sequence to sequence learning with neural networks,” in Advances in Neural Information Processing Systems, pp. 3104–3112, Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 2015. View at: Google Scholar
  32. Z. Pan, Y. Liang, W. Wang, Y. Yu, Y. Zheng, and J. Zhang, “Urban traffic prediction from spatio-temporal data using deep meta learning,” in Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 1720–1730, Anchorage, AK, USA, July 2019. View at: Publisher Site | Google Scholar
  33. Q. Hou, J. Leng, G. Ma, W. Liu, and Y. Cheng, “An adaptive hybrid model for short-term urban traffic flow prediction,” Physica A: Statistical Mechanics and Its Applications, vol. 527, Article ID 121065, 2019. View at: Publisher Site | Google Scholar
  34. Y. Zeng, C. J. Sreenan, N. Xiong, L. T. Yang, and J. H. Park, “Connectivity and coverage maintenance in wireless sensor networks,” The Journal of Supercomputing, vol. 52, no. 1, pp. 23–46, 2010. View at: Publisher Site | Google Scholar
  35. C. Lin, N. Xiong, J. H. Park, and T.-h. Kim, “Dynamic power management in new architecture of wireless sensor networks,” International Journal of Communication Systems, vol. 22, no. 6, pp. 671–693, 2009. View at: Publisher Site | Google Scholar
  36. Y. Sang, H. Shen, Y. Tan, and N. Xiong, “Efficient protocols for privacy preserving matching against distributed datasets, information and communications security,” in Proceedings of the International Conference on Information and Communications Security, pp. 210–227, Raleigh, NC, USA, December 2006. View at: Publisher Site | Google Scholar
  37. F. Long, N. Xiong, A. V. Vasilakos, L. T. Yang, and F. Sun, “A sustainable heuristic QoS routing algorithm for pervasive multi-layered satellite wireless networks,” Wireless Networks, vol. 16, no. 6, pp. 1657–1673, 2010. View at: Publisher Site | Google Scholar
  38. W. Guo, N. Xiong, A. V. Vasilakos, G. Chen, and C. Yu, “Distributed k-connected fault-tolerant topology control algorithms with PSO in future autonomic sensor systems,” International Journal of Sensor Networks, vol. 12, no. 1, pp. 53–62, 2012. View at: Publisher Site | Google Scholar
  39. N. Xiong et al., “A self-tuning failure detection scheme for cloud computing service,” in Proceedings of the 2012 IEEE 26th International Parallel and Distributed Processing Symposium, pp. 668–679, Shanghai, China, May 2012. View at: Publisher Site | Google Scholar
  40. M. Setia, “Methodology series module 5: sampling strategies,” Indian Journal of Dermatology, vol. 61, no. 5, p. 505, 2016. View at: Publisher Site | Google Scholar
  41. Q. Li, Z. Han, and X. M. Wu, “Deeper insights into graph convolutional networks for semi-supervised learning,” in Proceedings of the 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, pp. 3538–3545, New Orleans, LA, USA, Feburary 2018. View at: Google Scholar
  42. X. Song, V. Raghavan, and D. Yoshida, “Matching of vehicle GPS traces with urban road networks,” Current Science, vol. 98, no. 12, pp. 1592–1598, 2010. View at: Google Scholar
  43. T. Chai and R. R. Draxler, “Root mean square error (RMSE) or mean absolute error (MAE)?-arguments against avoiding RMSE in the literature,” Geoscientific Model Development, vol. 7, no. 3, pp. 1247–1250, 2014. View at: Publisher Site | Google Scholar
  44. K. Xu, C. Li, Y. Tian, T. Sonobe, K. I. Kawarabayashi, and S. Jegelka, “Representation learning on graphs with jumping knowledge networks,” in Proceedings of the 35th International Conference Machine Learning ICML 2018, pp. 8676–8685, Stockholm Sweden, July 2018. View at: Google Scholar

Copyright © 2020 Ya Zhang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views428
Downloads275
Citations

Related articles

Article of the Year Award: Outstanding research contributions of 2020, as selected by our Chief Editors. Read the winning articles.