<<

arXiv:1810.10161v1 [cs.NE] 24 Oct 2018 nvriy aghu302,Cia (email: China, 310027, Hangzhou to University, designed been already intelli- have telecommunications, of mechanisms of domain gent the array In an so on. and management, in traffic forecasting, planning, technique weather transportation including key problems, the practical been always input of length the or samples change training we sequences. to of how comparable matter number no is the LSTM conventional RCLSTM that the demonstrate of of that further accuracy we prediction benefit as directly the improvement could traces this prediction mobility from the and series networks, traffic telecommunication I of of scenarios. field application the latency-stringent more become in model RCLSTM applicable sparsity, the computational makes the of in and level decrease complexity appealing certain an way, to a leads this exhibits which In model networks. RCLSTM neural the of the in formation between breakthrough architecture significant connectivity a LSTM, achieves stochastic which to via neurons, mobility formed Compared user is and networks. RCLSTM traffic predicting telecommunication by in it test we (RCLSTM) and LSTM, LSTM model Connectivity of Random at the cost aiming forward a computing put Then, propa- considerable model. give forward the LSTM first reducing the and we of structure mechanism article, the effectively gation this to to introduction In specific promise brief learning, problem. a deep the as in overcome Long scheme solutions, whereas series of , (LSTM) kind time most for Memory in obstacle Short-Term embedded an are often is that dependencies long- Learning range values. historical future from determines then information and useful records extracts that process a 003 hn eal [email protected]). (email: China 100053, xianfu.chen@vtt.fi). (email: Finland FI-90571, Oulu iogeg honggangzhang lirongpeng, epLann ihLn hr-emMmr for Memory Short-Term Long with Learning Deep h nlssadpeito ftm eishas series time of prediction and analysis The Abstract .Ha .Za,R iadH hn r ihZhejiang with are Zhang H. and Li R. Zhao, Z. Hua, Y. .Lui ihCiaMbl eerhIsiue Beijing Institute, Research Mobile China with is Finland, Liu of Centre Z. Research Technical VTT with is Chen X. Tm eispeito a egnrlzdas generalized be can prediction series —Time ui u,ZiegZa,Rnpn i inuCe,ZiigL Zhiming Chen, Xianfu Li, Rongpeng Zhao, Zhifeng Hua, Yuxiu .I I. NTRODUCTION } @zju.edu.cn). ieSre Prediction Series Time { 1307 zhaozf, 21631087, n ogagZhang Honggang and n x nispeiu data previous its on isa nigafunction a finding at aims nesvl tde o 0yas[ years 40 for studied intensively sa ls otegon truth ground the to close as is aa hrfr,i ean hlegn ocapture to series challenging past to remains the tendency it of values natural Therefore, mean data. its one the is on However, concentrate ARIMA processes. of used time-varying limitation been study has Integrated model to (ARIMA) Autoregressive Average the Moving processing, signal vital of likely is network, also [ complexity overall importance and but the accuracy QoS prediction of the desired performance the the only degrade support not will to information fail resources predicted network outdated of and reservation misplaced since sbudu ihpol’ ie n aiu ra of areas various and [ lives society people’s that with up infrastructure bound cloud social is indispensable and an (IoT) has become network Things communication mobile of the computing, Internet mobile the of expansion of the Internet, proliferation as well explosive as the terminals mobile with hand, other the od n evc euss omninafw[ few a channel mention location, to requests, user service traffic, and load, data as time-dependent such of events, number large a analyze and track { which service at network demand location likely the [ will varying and user the mobile traffic a predict data to of the is movements. pattern of solution user regardless promising and traffic One experience network of of (QoS) service dynamics quality of the quality the and guarantee to how that ewr hne nna eltm [ time real to near effectively in react changes and network resources, network some 1 y = .Acrigy ewr prtr a reserve can operators network Accordingly, ]. 1 iesre nlssadpeito aebeen have prediction and analysis series Time aial,tepeito olo ieseries time a of goal prediction the Basically, y , { 2 y ... , i − 1 k .Hwvr trmisacalnigissue challenging a remains it However, ]. } y , st siaetevlea time at value the estimate to is i − 3 ]. k +1 .,y ..., , i y − i − 1 } 1 y , i , f ( i − = x 2 ) y ... , { i othat so ,.. n ..., k, spossible. as 4 fw denote we If . iu, .I statistical In ]. 2 .However, ]. } y ˆ h goal the , i = i 1 based .On ]. f ( x ) 1 2 a rapidly changing process [5]. Support Vector yield significant performance changes on different Regression (SVR) has been successfully applied subsets taken from the same database. Therefore, it for time series prediction, but it also has dis- would be hard to draw conclusions from one single advantages like the lack of structured means to dataset for comparing different algorithms. Second, determine some key parameters of the model [5]. the prediction of both network traffic and user In recent years, owing to the flexible structure, mobility is of significant importance in the design models are increasingly used in time and optimization of telecommunication networks. series prediction [6]. Specifically, Recurrent Neural For these reasons, we take advantage of the two Networks (RNNs), one of deep learning models, different datasets. Our simulation results show that establish the reputation to cope with time series when compared to LSTM, the RCLSTM model by recurrent neural connections. However, for any is highly capable of traffic prediction and user- standard architecture, the influence of a given location forecasting with less than half the neural input on the hidden layers and eventually on the connections. Particularly, in the traffic prediction neural network output would either decay or blow task, RCLSTM with even 1% neural connections up exponentially when cycling around recurrent performs better than ARIMA, SVR, and FFNN, connections. To tackle this problem, Long Short- while reducing the computing time by around 30% Term Memory (LSTM) has been revolutionarily compared with the conventional LSTM. More inter- designed by changing the structure of the hidden estingly, when the length of input traffic sequences neurons in traditional RNN [7]. Today, research increases from 50 to 500, the prediction error of and applications of LSTM for time series predic- RCLSTM remains basically unchanged, whereas tion are proliferating. For example, Wang et al. that of LSTM increases by about 10%. [2] used LSTM-based model to predict the next- moment traffic load in a specific geometric area II. AN OVERVIEW OF ARTIFICIAL NEURAL and Alahi et al. [3] predicted the motion dynamics NETWORKSAND LSTM in crowded scenes based on LSTM. Generally, without customized hardware and Artificial Neural Networks (ANNs) are con- software acceleration, the LSTM’s computing time structed as a class of models is proportional to the number of parameters. Given that can eliminate the drawbacks of the tradi- this disappointing characteristic, in this article, we tional learning algorithms with rule-based program- present an approach to decrease the number of ming [11]. ANNs can be classified into two main involved parameters, and thus put forward a new categories—FFNNs and RNNs. FFNNs usually model that reduces the computational cost. Inspired consist of an input , an output layer and hidden by the interesting finding that Feed Forward Neural layers (if necessary). Each layer is composed of Networks (FFNNs) with sparse neural connections a number of neurons and an activation . have a similar or even superior performance in A simple diagram of FFNNs is illustrated in Fig. many experiments compared to the conventional 1(a). In FFNNs, there is no connection between FFNNs [8], we introduce random connectivity to the neurons within the same layer, and all neurons the conventional LSTM model, thus forming a cannot be connected across layers, which means new architecture with sparse neural connections, the information flows in one direction, from the called the Random Connectivity Long Short-Term input layer, through the hidden layers (if any), to Memory (RCLSTM) model. the output layer. FFNNs are widely used in various Our simulation model is a three-layer stack fields like data classification, object recognition, RCLSTM neural network with a memory cell size and image processing. However, constrained by of 300 per layer. The simulation data comprise traf- their internal structure, FFNNs are unsuitable for fic data from GEANT´ networks—a pan-European handling historical dependencies. research network [9], and realistic user-trajectory RNNs, as another type of ANNs, are similar to data [10]. The reasons to use the two different FFNNs in the structure of neural layers, but allow datasets are twofold. First, ANN-based models the connections between the neurons within the like LSTM are sensitive to the dataset and may same hidden layer. An illustration of RNNs can be observed in the left-hand side of Fig. 1(b). In addi- 3

(a) (b) (c) Output Output cell state layer layer

hidden state j Hidden Hidden layer layer forget input input output Unfold s s j s gate gate update gate

Input Input input layer layer

Fig. 1: An illustration of FFNN, RNN and LSTM memory block: (a) FFNN; (b) RNN; (c) LSTM memory block. tion, the right-hand side of Fig. 1(b) is the expanded useful activation functions need to be reviewed. 1 form of the RNN model, indicating that RNNs The σ(x) = 1+e−x and the tanh calculate the output of the current moment from the function ϕ(x)=2σ(2x) − 1 are commonly used input of the current moment xt and the hidden state as the in ANNs. The domain of the previous moment ht−1. Therefore, RNNs of both functions is the field, but allow historical input information to be stored in the return value for the sigmoid function ranges the network’s internal state, and are thereby capable from 0 to 1, while the tanh function ranges from of mapping all of the historical input data to the -1 to 1. Fig. 1(c) explains in detail how LSTM final output. Theoretically, RNNs are competent works. The first step is to decide what kind of to handle such long-range dependencies. However, information will be removed from the memory cell in practice, RNNs seem unable to accomplish the state, which is implemented by a sigmoid layer task. This phenomenon has been explored in depth (i.e., the forget gate). The next step is to decide by Hochreiter and Schmidhuber [7], who explained what new information will be stored in the memory some pretty fundamental reasons why such learning cell state. This operation can be divided into two might be difficult. steps. First, a sigmoid layer (i.e., the input gate) Long Short-Term Memory networks, usually just determines what will be updated, and a tanh layer called “LSTMs”, are a special RNNs that are creates a vector of new candidate values zt that suitable for learning long-term dependencies [7]. can be added to the memory cell state, where the The key part that enhances LSTMs’ capability subscript t denotes the current moment. Next, these to model long-term dependencies is a component two parts are combined to trigger an update to the called memory block [7]. As illustrated in Fig. 1(c), memory cell state. To update the old memory cell the memory block is a recurrently connected subnet state ct−1 into the new memory cell state ct, we that contains functional modules called the memory can first multiply the corresponding elements of cell and gates. The memory cell is in charge ct−1 and the output of forget gate layer (i.e. ft), of remembering the temporal state of the neural which is just like the oblivion mechanism in the network and the gates formed by multiplicative human , and then add it ∗ zt, where it denotes units are responsible for controlling the pattern of the output of input gate and ∗ denotes element- information flow. According to the corresponding wise multiplication. The last step is to decide practical functionalities, these gates are classified what to output, which is realized by element-wise as input gates, output gates and forget gates. Input multiplication between the value obtained from a gates control how much new information flows tanh function of ct and the output of a sigmoid layer into the memory cell, while forget gates control (i.e., the output gate), ot. Through the cooperation how much information of the memory cell still between the memory cell and the gates, LSTM is remains in the current memory cell through re- endowed with a powerful ability to predict time current connection, and output gates control how series with long-term dependences. much information is used to compute the output Since the invention of LSTM, a number of activation of the memory block and further flows scholars have proposed several improvements with into the rest of the neural network. Before going respect to its original architecture. Greff et al. [12] through the details of LSTM, some simple yet evaluated the aforementioned conventional LSTM 4

j Inputs forget s input s input j output s gate gate update gate Hidden state Memory cell Data flow Operations Parameters

j j

forget s input s input j output s forget s input s input j output s gate gate update gate gate gate update gate

LSTM block RCLSTM block

Fig. 2: The comparison between LSTM and RCLSTM in the view of generating process of neural connections. and eight different variants thereof (e.g., Gated whether a strategy of forming more random neu- Recurrent Unit (GRU) [13]) on three ral connectivity, like in the human brain, might problems—TIMIT, IAM Online and JSB Chorales. yield potential benefits to LSTM’s performance and Each variant differs from the conventional LSTM efficiency. With this conjecture, we built up the by a single and simple change. They found that RCLSTM model. the conventional LSTM architecture performs well In RCLSTM, neurons are randomly connected on the three datasets, and none of the eight in- rather than being fully connected as in LSTM. vestigated modifications significantly improve the Actually, the trainable parameters in LSTM only performance. exist between the input part—the combination of the input of the current moment (i.e. xt) and the output of the previous moment (i.e. ht−1), and III. RANDOM CONNECTIVITY FOR LSTM the functional part—the combination of the gate The conventional LSTM (including its variants) layers and the input update layer. Therefore, the follows the classical pattern that the neurons in LSTM architecture can be further depicted in Fig. each memory block are fully connected and this 2. In our approach, whether the LSTM neurons connectivity cannot be changed manually. How- are connected or not can be determined by certain ever, it has been found that for certain functional randomness. Therefore, we use dashed lines to connectivity in neural microcircuits, random topol- denote that the neural connections can be added ogy formation of synapses plays a key role and or omitted, as depicted in the upper part of Fig. 2. can provide a sufficient foundation for specific If the neurons are fully connected, then it becomes functional connectivity to emerge in local neural a standard LSTM model. On the other hand, if the microcircuits [14]. This discovery is different from neurons are randomly connected according to some the conventional cases where neural connectivity rules (which are covered in detail below), then an is considered to be more heuristic so that neurons RCLSTM model is created. The lower right part need to be connected in a more fully organized of Fig. 2 shows an example RCLSTM structure in manner. It raises a fundamental question as to which the neural connections are randomly sparse, 5 unlike the LSTM model. The fundamental differ- ence between RCLSTM and LSTM is illustrated in Fig. 2, so let us move to the implementation RCLSTM RCLSTM RCLSTM RCLSTM strategy of randomly connecting neurons. block block block block First, we attach a value to each pair RCLSTM RCLSTM RCLSTM RCLSTM of neurons that are connected by a dashed line block block block block Unfold in the upper part of Fig. 2. The probability val- ues can obey arbitrary statistical distributions, and RCLSTM RCLSTM RCLSTM RCLSTM block block block we choose uniform distribution in our simulations block given its computational efficiency. The probability value indicates the tendency that the corresponding pair of neurons will be connected. Then we assume Fig. 3: The designed RCLSTM network for simulations. all neurons are connected with the same probability and carefully set a threshold to determine the percentage of connected neurons. If the probability these factors on the prediction accuracy of the values are greater than the threshold, the corre- RCLSTM network. sponding pairs of neurons are connected. Other- wise, they are prohibited from being connected. A. Data Description and Processing This process can be visualized as turning dashed We evaluate the model’s performance on traffic lines into solid lines, as shown in the right-hand prediction depending upon real traffic data from transformation of Fig. 2. Therefore, the RCLSTM a link in the GEANT´ backbone networks [9]. structure can create some sparsity, considerably GEANT´ is a pan-European data source for the decreasing the total number of involved parameters research and education communities. These traf- to be trained and reducing the computational loads fic data are sampled every 15 minute during a of the whole RCLSTM network. 4-month period and the unit for data points is Kbps. In this study, we selected 10772 traffic data points from 2005-01-01 00:00AM to 2005- IV. NUMERICAL SIMULATIONS FOR TRAFFIC 04-30 00:00AM. The raw data are so uneven in AND MOBILITY PREDICTION numerical size that there are about three orders of In this section, we focus on verifying the perfor- magnitude difference between the maximum and mance of the proposed RCLSTM model on traffic minimum. To reduce the numerical unevenness, we prediction and user-location forecasting. In partic- first take the logarithm to base 10 of the raw data, ular, we construct a three-layer RNN whose recur- then carry out a normalization process according rent neurons are all the newly designed RCLSTM x − min(x) to , where x is the vector after memory blocks (for the sake of simplicity, this max(x) − min(x) model is called the RCLSTM network in the taking the logarithm of the raw data, min(x) and following statement). The corresponding expanded max(x) denote the minimum and maximum value form is described in Fig. 3. Because the purpose of of x, respectively. Through this process, the raw our simulations is to predict the value at the next data are limited to a range between 0 and 1, which moment given some historical data, the input data makes the training phase of ANN-based models are from y1 to yT where T denotes the length of the converge faster and effectively avoid the bad local input sequences, and the output of the RCLSTM optimal solution [11]. Real-time prediction of data network is a prediction of the actual value at the traffic requires continuous data input and learning. next moment denoted as yˆT +1. First of all, we Hence, we introduce a notion of sliding window, take advantage of the RCLSTM network to predict which indicates a fixed number of previous times- traffic and user mobility, particularly compare the lots to learn and then predict the current data traffic. prediction accuracy of the RCLSTM network with Finally, we split the processed data into two sets that of other algorithms or models. Then, we adjust (i.e., a training set and a test set). The training set the number of training data samples and the length is used to train the RCLSTM network, and the test of input sequences to investigate the influence of set is used to evaluate its prediction accuracy. 6

The other dataset to evaluate the capability of the is the indicator function that equals to 1 if yi =y ˆi RCLSTM model comes from reference [10], which or 0 if yi =6 y ˆi. contains the location history of several mobile users together with manually defined important places for C. Testing Results and Analyses every person, and is thus used for user-mobility prediction. The data for every person are taken 1) Traffic Prediction from the Android location history service, and Fig. 4(a) reveals the RMSE and the computing mobile users are asked to mark the important places time under different percentages of neural connec- that are further constructed as polygons on the tivity in the RCLSTM model (note that 100% con- map. The dataset contains four attributes, i.e., date- nectivity means the fully-connected LSTM model). time, latitude, longitude, and assigned location ID, Notably, the probability of neural connections within which the last item is used to map the user’s obeys a uniform distribution between 0 and 1. In location to one of the geographical polygons. These addition, the size of the RCLSTM’s memory cell data are collected in 1-hour intervals from 2015-08- is set at 300, the ratio between the number of 06 to 2015-11-04. In this article, we use location training samples and the number of test samples is data of five users. First, we encode the location set at 9:1, and the length of input traffic sequences ID to one–hot code. For example, suppose that in is 100. Fig. 4(a) shows that the RMSE of the User A’s trajectory, the number of different location RCLSTM model is slightly larger than that of the IDs marked by User A is m. Then we map these LSTM model, but the RCLSTM with very sparse different location IDs to the Numbers 1 to m, and neural connections (i.e. 1%) reduces the computing use them as new location IDs. If the newly current time by around 30% compared with the baseline location ID is i, we construct an m-dimensional LSTM. In addition, the computing time almost vector where only the element with index i is one stops increasing when the percentage of neural and the others are all zero. After transforming all connections is larger than 20%, which is because the raw data to one-hot vectors, we use the sliding the method we use for accelerating calculation only window to slice the processed data and finally split works efficiently on extremely sparse matrices. Fig. them into training set and test set, which follows 4(b) intuitively illustrates the actual and predicted the same procedure as mentioned earlier in the traffic values. It can be observed from the subfigure traffic data processing. that the predicted values can match the variation trend and features of the actual values very well. Therefore, the simulation results indicate that the B. Evaluation Metrics RCLSTM model can yield acceptable prediction capability, and effectively decrease the computa- To evaluate the performance of our proposed tional loads and complexity. RCLSTM model on traffic prediction, Root Mean Then we investigate how the predictive capa- Square Error (RMSE) is applied to estimate the bility of the RCLSTM model is influenced by prediction accuracy. RMSE measures the squared the number of training samples and the length of root of the mean of the deviation squares, which input traffic sequences. First we train the models quantifies the difference between the predicted with 90%, 80%, 70% and 60% of the processed values and the actual ones. The RMSE can be data, respectively, while fixing both the size of the 1 N 2 expressed as RMSE= (yi − yˆi) , where yi q N Pi=1 memory cell (set at 300) and the length of the input is the actual value, yˆi is the predicted value and N sequences (set at 100), and test the trained models represents the total number of predicted traffic data. with the remaining data. Then we vary the length of On the other hand, the evaluation metric for the input sequences from 50 to 500 while keeping human mobility prediction is the accuracy level, the values of the other hyperparameters fixed. The which indicates the percentage of the correct loca- simulation results are shown in Fig. 4(c) and (d). tion predictions. In our study, the accuracy level is N It can be observed from Fig. 4(c) that RCLSTM Pi=1 1(yi=y ˆi) defined as Acc= N , where yi is the actual models are more sensitive to the number of training value, yˆi is the predicted value, N represents the samples than LSTM, because when the number total number of predicted locations and 1(yi =y ˆi) of training samples increases, the RMSEs of the 7

Fig. 4: Results of traffic prediction using RCLSTM network: (a) RMSE and computing time of RCLSTMs ; (b) predicted traffic data and actual traffic data; (c) RMSE of RCLSTMs over different number of training samples; (d) RMSE of RCLSTMs over different length of input sequences; (e) comparing RCLSTM with SVR, ARIMA, FFNN and LSTM.

RCLSTM models vary more significantly than that shown in Fig. 4(e), which reveals that LSTM with of the LSTM model. Fig. 4(d) gives the related a memory cell size of 300 performs much better results and shows that RCLSTM models are less than the others, followed by the RCLSTM with the susceptive to the length of the input sequences than memory cell size of 300 and 1% neural connec- LSTM model. tions. Interestingly, the RCLSTM model performs Finally, we compare the RCLSTM with three better than the LSTM with the memory cell size of well-known prediction techniques—SVR, ARIMA, 30, which is probably due to a degree of overfitting and FFNN. The hyper-parameters of these algo- that exists in the latter [11]. rithms are as follows: 2) Human Mobility Prediction • SVR: The number of input features is 100, The results of user-mobility prediction with the the is a Radial Basis Function (RBF) RCLSTM model are shown in Fig. 5. Fig. 5(a) and the tolerance for the stopping criterion is shows the respective prediction accuracy for the 0.001. five users with the RCLSTM model, where the • ARIMA(p, d, q): The number of autoregres- probability of neural connections obeys a uniform sive terms (i.e. p) is 5, the number of nonsea- distribution between 0 and 1. In addition, the size sonal differences needed for stationarity (i.e. of the memory cell is 150, the length of input d) is 1, and the number of lagged forecast sequences is 12, and the training samples account errors in the prediction equation (i.e. q) is 0. for 90% of the processed data. There is a slight • FFNN: The number of input features is 100 difference in the prediction results for different and the numbers of neurons in both the first users. For example, Users A and D are both hidden layer and the second hidden layer are university students with a part-time job, and thus 50. they almost follow the same behavioral pattern in In addition, since the LSTM with a memory cell school or work, which results in highly expected size of 30 has almost as many trainable parameters predictability. On the other hand, User E is running as the RCLSTM with a memory cell size of 300 his own business and is more likely to travel a and 1% neural connections, we put it into the lot with unfixed schedule, consequently having low comparison list as well. The simulation results are expected predictability [10]. Although the predic- 8

Fig. 5: Results of human mobility prediction using RCLSTM: a) prediction accuracy of RCLSTMs for different users; b) prediction accuracy of RCLSTMs over different length of input sequences for User A; c) performance of RCLSTMs over different length of input sequences for User E. tion accuracy of the RCLSTM model is not as of traffic prediction and user-mobility forecasting good as that of the LSTM model, the RCLSTM with deep learning and proposed a new model model with high sparsity of neural connections can named RCLSTM by revolutionarily redesigning the compute faster than the LSTM, similar to the traffic conventional LSTM model. The basic idea behind prediction scenario. the RCLSTM model is to construct neural networks In order to further investigate the capability by forming and realizing a random sparse graph. of the RCLSTM model to characterize long-term We have checked the effectiveness of the RCLSTM dependencies of user mobility, we carry out simu- model by predicting the dynamics of traffic and lations with the different length of input sequence user locations through various temporal scales. In data for Users A and E, consistent with the scenario traffic prediction, we have demonstrated that the of traffic prediction. The results are shown in Fig. RCLSTM model with 1% neural connections re- 5(b) and (c), respectively. It can be observed from duces the computing complexity by 30% compared the subfigures that the RCLSTM models can catch with the standard LSTM model. Although the char- up with LSTM in terms of prediction accuracy, acteristic of sparse neural connections may cause regardless of the length of the input sequences. a degradation of approximately 25% in prediction Remarkably, for User A, increasing the length of accuracy, the RCLSTM model still outperforms input sequences scarcely affects the prediction ac- SVR, ARIMA, FFNN, even the LSTM model curacy, whereas for User E, the prediction accuracy with the same number of parameters. In addition, improves when the length of the input sequences along with the adjustment of the length of input increases. From the perspective of data analysis, sequences and the number of training samples, the the mobility data of User E is more irregular than RCLSTM models fluctuate in line with the LSTM that of User A, so both the RCLSTM model and model in terms of prediction accuracy. In summary, the baseline LSTM model need more input data to it can be expected the RCLSTM model with lower find the regular movement patterns of User E. computing costs and satisfactory performance will play an essential role in time series prediction in Remark. The RCLSTM model shows promise for the future intelligent telecommunication networks. manifesting strong traffic and user-mobility pre- diction capabilities while reduces the number of REFERENCE parameters to be trained, which in effect decreases [1] R. Li, Z. Zhao, X. Zhou, G. Ding, Y. Chen, Z. Wang, the computational load and complexity. and H. Zhang, “Intelligent 5G: When Cellular Networks Meet Artificial Intelligence,” IEEE Wireless Communi- V. CONCLUSION cations, vol. 24, no. 5, Oct 2017, pp. 175–183. [2] J. Wang, J. Tang, Z. Xu, Y. Wang, G. Xue, X. Zhang, In this article, we have addressed the importance and D. Yang, “Spatiotemporal Modeling and Prediction of leveraging deep learning for time series predic- in Cellular Networks: A Big Data Enabled Deep Learn- tion. In particular, we have reinvestigated the issues ing Approach,” Proc. 36th International Conference on 9

Computer Communications (INFOCOM), Atlanta, GA, USA, May 2017, pp. 1–9. [3] A. Alahi, K. Goel, V. Ramanathan, A. Robicquet, L. Fei-Fei, and S. Savarese, “Social LSTM: Human Trajectory Prediction in Crowded Spaces,” Proc. 29th IEEE Conference on and (CVPR), Las Vegas, NV, USA, June 2016, pp. 961–971. [4] N. I. Sapankevych and R. Sankar, “Time Series Predic- tion Using Support Vector Machines: A Survey,” IEEE Computational Intelligence Magazine, vol. 4, no. 2, May 2009, pp. 24–38. [5] W.-C. Hong, “Application of Seasonal SVR with Chaotic Immune in Traffic Flow Forecast- ing,” Neural Computing and Applications,vol. 21, no. 3, Apr 2012, pp. 583–593. [6] I. H. Witten, E. Frank, M. A. Hall, and C. J. Pal, : Practical Machine Learning Tools and Techniques, Morgan Kaufmann, Burlington, MA, USA, 4 edition, 2016, https://www.amaxon.com/exec/obidos/ASIN/0128042915/departmofcompute, accessed 10/11/2018. [7] S. Hochreiter and J. Schmidhuber, “Long Short-Term Memory,” Neural Computation, vol. 9, no. 8, 1997, pp. 1735–1780. [8] M. J. Shafiee, P. Siva, and A. Wong, “Stochasticnet: Forming Deep Neural Networks via Stochastic Connec- tivity,” IEEE Access, vol. 4, 2016, pp. 1915–1924. [9] S. Uhlig, B. Quoitin, J. Lepropre, and S. Balon, “Providing Public Intradomain Traffic Matrices to the Research Community,” SIGCOMM Comput. Commun. Rev., vol. 36, no. 1, Jan 2006, pp. 83–86. [10] J. Tkaˇc´ık and P. Kord´ık, “ for Sequential Learning of Human Mobility Patterns,” Proc. 2016 International Joint Conference on Neural Net- works (IJCNN), Vancouver, BC, Canada, July 2016, pp. 2790–2797. [11] Y. LeCun, Y. Bengio, and G. Hinton, “Deep Learning,” , vol. 521, no. 7553, 2015, p. 436. [12] K. Greff, R. K. Srivastava, J. Koutnk, B. R. Steunebrink, and J. Schmidhuber, “LSTM: A Search Space Odyssey,” IEEE Transactions on Neural Networks and Learning Systems, vol. 28, no. 10, Oct 2017, pp. 2222–2232. [13] J. Chung, C¸. G¨ulc¸ehre, K. Cho, and Y. Bengio, “Empir- ical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling,” CoRR, vol. abs/1412.3555, 2014. [14] S. L. Hill, Y. Wang, I. Riachi, F. Sch¨urmann, and H. Markram, “Statistical Connectivity Provides a Suf- ficient Foundation for Specific Functional Connectivity in Neocortical Neural Microcircuits,” Proceedings of the National Academy of Sciences, vol. 109, no. 42, 2012, pp. E2885–E2894.