<<

entropy

Article Optimal Randomness for Stochastic Configuration Network (SCN) with Heavy-Tailed Distributions

Haoyu Niu 1,† , Jiamin Wei 2 and YangQuan Chen 1,*,†

1 Electrical Engineering and Computer Science Department, University of California, Merced, CA 95340, USA; [email protected] 2 School of Telecommunications Engineering, Xidian University, No.2, Taibai Road, Xi’an 710071, Shaanxi, China; [email protected] * Correspondence: [email protected]; Tel.: +1-209-2284672 † MESA Lab address: Room 22, 4225 Hospital Road, Atwater, CA 95301, USA.

Abstract: Stochastic Configuration Network (SCN) has a powerful capability for regression and classification analysis. Traditionally, it is quite challenging to correctly determine an appropriate architecture for a neural network so that the trained model can achieve excellent performance for both learning and generalization. Compared with the known randomized learning algorithms for single hidden layer feed-forward neural networks, such as Randomized Radial Basis Function (RBF) Networks and Random Vector Functional-link (RVFL), the SCN randomly assigns the input weights and biases of the hidden nodes in a supervisory mechanism. Since the parameters in the hidden layers are randomly generated in uniform distribution, hypothetically, there is optimal randomness. Heavy-tailed distribution has shown optimal randomness in an unknown environment for finding some targets. Therefore, in this research, the authors used heavy-tailed distributions to randomly initialize weights and biases to see if the new SCN models can achieve better performance than the original SCN. Heavy-tailed distributions, such as Lévy distribution, , and , have been used. Since some mixed distributions show heavy-tailed properties,  the mixed Gaussian and Laplace distributions were also studied in this research work. Experimental  results showed improved performance for SCN with heavy-tailed distributions. For the regression Citation: Niu, H.; Wei, J.; Chen, Y. model, SCN-Lévy, SCN-Mixture, SCN-Cauchy, and SCN-Weibull used less hidden nodes to achieve Optimal Randomness for Stochastic similar performance with SCN. For the classification model, SCN-Mixture, SCN-Lévy, and SCN- Configuration Network (SCN) with Cauchy have higher test accuracy of 91.5%, 91.7% and 92.4%, respectively. Both are higher than the Heavy-Tailed Distributions. Entropy test accuracy of the original SCN. 2021, 23, 56. https://doi.org/ 10.3390/e23010056 Keywords: SCN; optimal randomness; heavy-tailed distribution; Lévy; Weibull; Cauchy

Received: 27 October 2020 Accepted: 28 December 2020 Published: 31 December 2020 1. Introduction Publisher’s Note: MDPI stays neu- The Stochastic Configuration Network (SCN) model is generated incrementally by tral with regard to jurisdictional clai- using stochastic configuration (SC) algorithms [1]. Compared with the existing randomized ms in published maps and institutio- learning algorithms for single-layer feed-forward neural networks (SLFNNs), the SCN can nal affiliations. randomly assign the input weights (w) and biases (b) of the hidden nodes in a supervisory mechanism, which is selecting the random parameters with an inequality constraint and assigning the scope of the random parameters adaptively. It can ensure that the built randomized learner models have universal approximation property. Then, the output Copyright: © 2020 by the authors. Li- censee MDPI, Basel, Switzerland. weights are analytically evaluated in either a constructive or selective manner [1]. In This article is an open access article contrast with the known randomized learning algorithms, such as the Randomized Radial distributed under the terms and con- Basis Function (RBF) Networks [2] and the Random Vector Functional-link (RVFL) [3], ditions of the Creative Commons At- SCN can provide good generalization performance at a faster speed. Concretely, there tribution (CC BY) license (https:// are three types of SCN algorithms, which are SC-I, SC-II, and SC-III. SC-I algorithm uses creativecommons.org/licenses/by/ a constructive scheme to evaluate the output weights only for the newly added hidden 4.0/). node [4]. All of the previously obtained output weights are kept the same. The SC-II

Entropy 2021, 23, 56. https://doi.org/10.3390/e23010056 https://www.mdpi.com/journal/entropy Entropy 2021, 23, 56 2 of 13

algorithm recalculates part of the current output weights by analyzing a local problem with user-defined shifting window size. The SC-III algorithm finds all the output weights together by solving a global least-squares problem. SCN algorithms have been commonly studied and used in many areas, such as image data analytics [5,6], prediction of component concentrations in sodium aluminate liquor [7], and etc. [8,9]. For example, in [5], Li et al. developed a two-dimensional SCNs (2DSCNs) for image data modelling tasks. Experimental results on hand written digit classification and face recognition showed that the 2DSCNs have great potential for image data analytics. In [7], Wang et al. proposed a SCN-based model for measuring component concentrations in sodium aluminate liquor, which are usually acquired by titration analysis and suffered from larger time delays. From the results, the mechanism model showed the internal relationship. The improved performance can be achieved by using the SCN-based compensation model. In [10], Lu et al. proposed a novel robust SCN model based on a mixture of the Gaussian and Laplace distributions (MoGL-SCN) in the Bayesian framework. To improve the robustness of the SCN model, the random noise of the SCN model is assumed to follow a mixture of Gaussian distribution and Laplace distributions. Based on the research results, the proposed MOGL-SCN could construct prediction intervals with higher reliability and prediction accuracy. Neural Networks (NNs) can learn from data to train feature-based predictive models. However, the learning process can be time-consuming and infeasible for applications with data streams. An optimal method is to randomly assign the weights of the NNs so that the task can become a linear least-squares problem. In [11], Wang et al. classified the NN models into three types. First, the feed-forward networks with random weights (RW-FNN) [12]. Second, recurrent NNs with random weights [13]. Third, randomized kernel approximations [14]. According to [11], there are three benefits of the randomness: (1) simplicity of implementation, (2) faster learning and less human intervention, (3) possibility of leveraging and classification algorithms. Randomness is used to define a feature map, which converts the data input into a high dimensional space where learning is more simpler. The resulting optimization problem becomes a standard linear least-squares, which is a simpler and scalable learning procedure. For the original SCN algorithms, weights and biases are randomly generated in uni- form distribution. Randomness plays a significant role in both exploration and exploitation. A good NNs architecture with randomly assigned weights can easily outperform a more deficient architecture with finely tuned weights [11]. Therefore, it is critical to discuss the optimal randomness for the weights and biases in SCN algorithms. In this study, the authors mainly discussed the impact of three different heavy-tailed distributions on the performance of the SCN algorithms, Lévy distribution, Cauchy distribution, and Weibull distribution [15]. Heavy-tailed distribution has shown optimal randomness for finding targets [16], which plays a significant role in exploration and exploitation [17]. It is impor- tant to point out that the proposed SCN models are very different from Lu et al. [10]. As mentioned earlier, Lu at al. assumed that the random noise of the SCN model following a mixture of Gaussian distribution and Laplace distributions. In this research study, the authors randomly initialize the weights and biases with heavy-tailed distributions instead of uniform distribution. To compare with the mixture distributions, the authors also used the mixture distributions for weight and bias generation. A more detailed comparison of the two heavy-tailed methods is shown in Results and Discussion section. There are two objectives for this research, (1) compare the performance of SCN al- gorithms with heavy-tailed distributions on a linear regression model [18]; (2) evaluate the SCN algorithms performance on MNIST handwritten digit classification problem with heavy-tailed distributions. The rest of the paper is organized as follows: Section2 in- troduces fundamental definitions of the heavy-tailed distributions and how to generate the random numbers according to the given distribution. Results and discussions are presented in Section3. A simple regression model and MNIST handwritten digit classifi- cation problem are used to demonstrate the heavy-tailed SCNs. In Section4, the authors Entropy 2021, 23, 56 3 of 13

draw conclusive remarks and share views in SCN with heavy-tailed distributions in future research.

2. Materials and Methods 2.1. Heavy-Tailed Distributions Heavy-tailed distributions are widely used for modeling in different scenarios, such as finance [19], insurance [20], and medicine [21]. The distribution of a real-valued X is said to have a heavy right tail if the tail probabilities P(X > x) decay more slowly than those of any if

P(X > x) lim ( ) = ∞, (1) x→∞ e−λx for every λ > 0. For the heavy left tail, it has similar definition [22].

2.1.1. Lévy Distribution Lévy distribution, named after a French mathematician Paul Lévy, is a random walk that has a heavy-tailed distribution [23]. As a fractional-order stochastic process with heavy-tailed distributions, Lévy distribution has better computational effects [24]. Lévy distribution is stable and has probability density functions that can be expressed analytically. The probability density function of Lévy flight [25] is √  γ 2π  γ , x > µ, (x−µ)3/2 p(x, µ, γ) = e 2(x−µ) (2) 0, x ≤ µ. where µ is the and γ is the . In practice, the Lévy distribution is updated by u Lévy(β) = , (3) |ν|1/β where u and ν are random numbers generated from a with mean of 0 and standard deviation of 1. The stability index β ranges from 0 to 2. Moreover, it is interesting to point out that the well-known Gaussian and Cauchy distribution are its special cases when its stability index is set to 2 and 1, respectively.

2.1.2. Weibull Distribution A random variable is subject to a Weibull distribution if it has a tail function F as

λw F(x) = e−(x/k) , (4)

where k > 0 is the scale parameter, λw > 0 is the shape parameter. If the λw < 1, the Weibull distribution will be a heavy-tailed distribution.

2.1.3. Cauchy Distribution A random variable is subject to the Cauchy distribution if its cumulative distribution is:

1 2(x − µ ) 1 F(x) = arctan( c ) + , (5) π σ 2

where µc is the location parameter, σ is the scale parameter.

2.2. Mixture Distributions A is derived from a collection of other random variables. First, a random variable is selected by chance from the collection according to given probabilities of selection. Then, the value of the selected random variable is realized. The mixture distribution densities are complicated in terms of simpler densities, which provide a Entropy 2021, 23, 56 4 of 13

good model for certain data sets. The different subsets of the data can exhibit different characteristics. Therefore, the mixed distributions can effectively characterize the complex distributions of certain real-world datasets. In [10], a robust SCN based on the mixture of the Gaussian and Laplace distributions were proposed. Thus, Gaussian and Laplace distributions are mentioned in this section for comparison purpose.

2.2.1. Gaussian Distribution 2 A random variable X has a Gaussian distribution with mean µG and σG (−∞ < µG < ∞ and σG > 0) if X has a continuous distribution for which the probability density function is as follows [26]:

1 x−µG 2 2 1 − ( ) f (x|µ σ ) = e 2 σG f or − < x < G, G 1/2 , ∞ ∞. (6) (2π) σG

2.2.2. Laplace Distribution The probability density function of the Laplace distribution can be written as fol- lows [10]: √ | − | 1 (− 2 x µl ) F(x|µ , η) = e η , (7) l (2η2)1/2

where µl and η represent the location and scale parameters, respectively.

2.3. SCN with Heavy-Tailed Distribution SCN was first proposed by Wang et al. in 2017 [1]. Compared with the known randomized learning algorithms for single hidden layer feed-forward neural networks, the SCN randomly assigns the input weights and biases of the hidden nodes in the light of a supervisory mechanism. The output weights are analytically evaluated in a constructive or selective method. The SCN has better performance than other randomized neural networks in terms of the fast learning, scope of the random parameters, and the required human intervention. Therefore, it has already been used in many data processing projects, such as [5,9,27]. Since the parameters of hidden layers are randomly generated in uniform distribution, it might exist optimal randomness in the SCN model. Heavy-tailed distribution has shown optimal randomness in an unknown environment for finding some targets [17], which plays a significant role in both exploration and exploitation. Therefore, in this research, the authors used heavy-tail distributions to randomly update the weights and biases in the hidden layers to see if the SCN models have improved performance with heavy-tailed distributions. Some of the key parameters of the SCN are listed in Table1. For example, the maximum times of random configuration Tmax is set as 200. The scale factor Lambdas in the activation function, which directly determines the range for the random parameters, is examined by performing different settings (0.5–200). The tolerance is set as 0.05. Most of the parameters for SCN with heavy-tailed distributions are kept the same with the original SCN algorithms for comparison purpose. For more details, please refer to [1] and AppendixA. Entropy 2021, 23, 56 5 of 13

Table 1. SCNs with key parameters.

Properties Values Name: ‘Stochastic Configuration Networks’ version: ‘1.0 beta’ L: hidden node number W: input weight matrix b: hidden layer bias vector Beta: output weight vector r: regularization parameter tol: tolerance Lambdas: random weights range Lmax: maximum number of hidden neurons Tmax: maximum times of random configurations nB: number of node being added in one loop

3. Results and Discussion In this section, some simulation results are discussed over one regression problem and the MNIST handwritten digit classification problem. For the regression model, performance evaluation goes into the eventual number of hidden nodes required for achieving the expected training error tolerance. For the classification problem, performance evaluation goes into the eventual training and testing accuracy. In this study, the sigmoid activation function g(x) = 1/(1 + exp(-x)) is being used, which is the default in the original SCN algorithms [1].

3.1. A Regression Model The dataset of the regression model was generated by a real-valued function [18]:

2 2 2 f (x) = 0.2e−(10x−4) + 0.5e−(80x−40) + 0.3e−(80x−20) , (8) where x ∈ [0, 1]. There are 1000 points randomly generated from the uniform distribution [0, 1] in the training dataset. The test set has 300 points generated from a regularly spaced grid on [0, 1]. The input and output attributes are normalized into [0, 1], and all results reported in this research took averages over 1000 independent trials. The settings of parameters are similar to the SCN in [1].

3.1.1. Parameter Tuning for Regression Model The Materials and Methods section shows that the three heavy-tailed distribution algorithms have user-defined parameters, for example, the power-law index for SCN-Lévy, and location and scale parameters for SCN-Cauchy and SCN-Weibull, respectively. Thus, to illustrate the effect of parameters on the optimization results and to offer reference values to the proposed SCN algorithms, parameter analysis is conducted, and correspond- ing experiments are performed. Based on the experimental results, for the SCN-Lévy algorithm, the most optimal power law index is 1.1 to achieve the minimum number of hidden nodes (Figure1). For the SCN-Weibull algorithm, the optimal location parame- ter α and scale parameter β for the minimum number of hidden nodes are 1.9 and 0.2, respectively (Figure2). For the SCN-Cauchy algorithm, the optimal location parameter α and scale parameter β for the minimum number of hidden nodes are 0.9 and 0.1, respec- tively (Figure3). The values of colorbar in Figures2 and3 represent for the mean hidden node number. Lower values mean that the SCN algorithms use less hidden node numbers to reach the training tolerance. Entropy 2021, 23, 56 6 of 13

Figure 1. SCN-Lévy parameters tuning for regression model.

Figure 2. SCN-Weibull parameters tuning for regression model.

Figure 3. SCN-Cauchy parameters tuning for regression model. Entropy 2021, 23, 56 7 of 13

3.1.2. Performance Comparison among SCN, SCN-Lévy, SCN-Cauchy, SCN-Weibull, and SCN-Mixture on the Regression Model In Table2, the performance of SCN, SCN-Lévy, SCN-Cauchy, SCN-Weibull, and SCN- Mixture are shown, in which mean values are reported based on 1000 independent trials. In [1], Wang et al. used time cost to evaluate the SCN algorithms performance. In this study, the authors used the mean hidden node numbers to evaluate the performance. The number of hidden nodes associates with the modeling accuracy. Therefore, the authors are interested to analyze if SCN with heavy-tailed distributions can use less hidden nodes to generate high performance, which makes the NNs less complex. According to the experimental results, the SCN-Cauchy used the least number of mean hidden nodes, 59 with an RMSE of 0.0057. The SCN-Weibull has a mean hidden nodes number of 63, with an RMSE of 0.0037. The SCN-Mixture has a mean hidden nodes number of 70, with an RMSE of 0.0020. The mean hidden nodes number for SCN-Lévy is 70. The original SCN model has a mean hidden nodes number of 75. A more detailed training process is shown in Figure4. With fewer hidden node numbers, the SCN models with heavy-tailed distribution can be faster than the original SCN model. The neural network structure is also less complicated than SCN. Our experimental results on the regression task demonstrate remarkable improvements in modeling performance compared with the current SCN model results. For example, the approximation performance of SCN-Cauchy is shown in Figure5. The other SCN models with heavy-tailed distributions also have similar performance.

Table 2. Performance comparison of SCN models on regression problem.

Models Mean Hidden Node Number RMSE SCN 75 ± 5 0.0025, SCN-Lévy 70 ± 6 0.0010, SCN-Cauchy 59 ± 3 0.0057, SCN-Weibull 63 ± 4 0.0037, SCN-Mixture 70 ± 5 0.0020.

-1 SCN Training RMSE SCN Cauchy Training RMSE SCN Lévy Training RMSE -2 SCN Mixture Training RMSE SCN Weibull Training RMSE

-3

-4 log(RMSE)

-5

-6

-7 0 10 20 30 40 50 60 70 80 L

Figure 4. Performances of SCN, SCN-Lévy, SCN-Weibull, SCN-Cauchy, and SCN-Mixture. Entropy 2021, 23, 56 8 of 13

(a) (b) Figure 5. (a) SCN-Cauchy training performance for regression model; (b) SCN-Cauchy test performance for regres- sion model.

3.2. MNIST Handwritten Digit Classification The handwritten digit dataset contains 4000 training examples and 1000 testing ex- amples, a subset of the MNIST handwritten digit dataset. Each image is a 20 by 20 pixels grayscale image of the digit (Figure6). Each pixel is represented by a number indicating the grayscale intensity at that location. The 20 by 20 grid of pixels is “unrolled” into a 400-dimensional vector.

Figure 6. The handwritten digit dataset example.

3.2.1. Parameter Tuning for Classification Model Similar to the parameter tuning for the regression model, parameter analysis was conducted to illustrate the impact of parameters on the optimization results and to offer reference values to the MNIST handwritten digit classification SCN algorithms. Cor- responding experiments are performed. According to the experimental results, for the SCN-Lévy algorithm, the most optimal power law index is 1.6 to achieve the best RMSE performance (Figure7). For the SCN-Cauchy algorithm, the optimal location parameter α and scale parameter β for the least RMSE are 0.2 and 0.3, respectively (Figure8). The values of colorbar in Figure8 represent for the RMSE. Entropy 2021, 23, 56 9 of 13

Figure 7. SCN-Lévy parameters tuning for classification model.

Figure 8. SCN-Cauchy parameters tuning for classification model.

3.2.2. Performance Comparison among SCN, SCN-Lévy, SCN-Cauchy, and SCN-Mixture on the Classification Model The performance of the SCN, SCN-Lévy, SCN-Cauchy, and SCN-Mixture are shown in Table3. Based on the experimental results, both of the SCN-Cauchy, SCN-Lévy and SCN-Mixture have better performance on training and test accuracy, compared with the original SCN model. A detailed training process is shown in (Figure9). Within around 100 hidden nodes, the SCN models with heavy-tailed distribution perform similarly to the original SCN model. When the number of the hidden node is greater than 100, the SCN models with heavy-tailed distribution have lower RMSE. Since more parameters for weights and biases are initialized in heavy-tailed distribution, this may cause that the SCN with heavy-tailed distributions converges to the optimal values at a faster speed. The experimental results on the MNIST handwritten classification problem demonstrate Entropy 2021, 23, 56 10 of 13

improvements in modeling performance. It also shows that SCN models with heavy-tail distributions have a better search ability to achieve lower RMSE. The training and test performance of the SCN, SCN-Cauchy, and SCN-Lévy are shown in Figure 10.

Table 3. Performance comparison between SCN, SCN-Lévy and SCN-Cauchy.

Models Training Accuracy Test Accuracy SCN 94.0 ± 1.9% 91.2 ± 6.2%, SCN-Lévy 94.9 ± 0.8% 91.7 ± 4.5%, SCN-Cauchy 95.4 ± 1.3% 92.4 ± 5.5%, SCN-Mixture 94.7 ± 1.1% 91.5 ± 5.3%.

1 1

0.9 Training ACC for SCN-Lévy 0.9 0.8 Training ACC for SCN Training ACC for SCN-Cauchy Training ACC for SCN-Mixture 0.7 Training RMSE for SCN-Lévy 0.8 Training RMSE for SCN Training RMSE for SCN-Cauchy 0.6 Training RMSE for SCN-Mixture 0.7

0.5 RMSE Accuracy

0.4 0.6

0.3 0.5 0.2

0.1 0.4 0 50 100 150 200 250 L

Figure 9. Classification performance of training accuracy between SCN, SCN-Lévy and SCN-Cauchy.

(a) (b)

Figure 10. Cont. Entropy 2021, 23, 56 11 of 13

(c) (d)

(e) (f)

Figure 10. (a) SCN training performance for classification model; (b) SCN test performance for classification model; (c) SCN-Cauchy training performance for classification model; (d) SCN-Cauchy test performance for classification model; (e) SCN-Lévy training performance for classification model; (f) SCN-Lévy test performance for classification model.

4. Conclusions The SCN models have been widely used for regression and classification analysis. Compared with the known randomized learning algorithms for single hidden layer feed- forward neural networks, the SCN randomly assigns the input weights and biases of the hidden nodes in a supervisory mechanism. Since the parameters of the hidden layers are randomly generated in uniform distribution, it might exist optimal randomness in the SCN models. Heavy-tailed distribution has shown optimal randomness in an unknown environment for finding some targets, which plays a significant role in both exploration and exploitation. Therefore, in this research, the authors mainly discussed the effect of different heavy-tailed distributions on the SCN algorithms performance. In this study, heavy-tailed distributions, such as Lévy distribution, Cauchy distri- bution, and Weibull distribution, are being used for generating optimal randomness of weights and biases initialization. Experiment results showed improved performance for SCN with heavy-tailed distributions. The experimental results of regression tasks demon- strate remarkable improvements in modeling performance compared with the existing SCN model results. SCN-Mixture, SCN-Lévy, SCN-Cauchy, and SCN-Weibull used less Entropy 2021, 23, 56 12 of 13

hidden nodes to achieve similar performance with SCN. For the classification model, SCN- Mixture, SCN-Lévy and SCN-Cauchy have higher test accuracy of 91.5%, 91.7% and 92.4%, respectively. Both are higher than the test accuracy of the original SCN. The experimental results showed that SCN models with heavy-tail distributions have a better search ability to achieve better performance. For future research, the authors will try the heavy-tailed distributions for the DeepSCN models. Since heavy-tailed distributions, such as SCN-Lévy, SCN-Cauchy, and SCN- Weibull, showed improved performance for SCN models. Other heavy-tailed distributions, such as Mittag-Leffler distribution, Pareto-distribution, and more, will be applied for SCN and other algorithms. Randomly assigning a subset of parameters in a learner model is an efficient method for training NNs, which provides an effective solution for machine learning problems. The proposed SCN algorithms with heavy-tailed distributions will be applied for more regression and classification problems in the future.

Author Contributions: H.N. wrote the original manuscript draft. J.W. and Y.C. contributed to the results interpretation and discussions and editing of the manuscript. All authors have read and agreed to the published version of the manuscript. Funding: This research received no external funding. Institutional Review Board Statement: Not applicable. Informed Consent Statement: Not applicable. Data Availability Statement: Not applicable. Acknowledgments: We gratefully acknowledge the of NVIDIA Corporation with the dona- tion of the Titan X Pascal GPU used for this research. Conflicts of Interest: The authors declare no conflict of interest.

Abbreviations The following abbreviations are used in this manuscript:

SCN Stochastic Configuration Network SLFNNs Single-layer Feed-forward Neural Networks RBF Randomized Radial Basis Function (RBF) Networks RVFL Random Vector Functional-link NN Neural Networks RW-FNN Feed-forward Networks with Random Weights RMSE Root Mean Square Error MNIST Modified National Institute of Standards and Technology database

Appendix A. SCN Codes The Matlab and Python codes are attached here, which can be found at DeepSCN homepage: http://deepscn.com/software.php.

References 1. Wang, D.; Li, M. Stochastic Configuration Networks: Fundamentals and Algorithms. IEEE Trans. Cybern. 2017, 47, 3466–3479. [CrossRef][PubMed] 2. Broomhead, D.; Lowe, D. Multivariable Functional Interpolation and Adaptive Networks. Complex Syst. 1988, 2, 321–355. 3. Pao, Y.H.; Takefuji, Y. Functional-link Net Computing: Theory, System Architecture, and Functionalities. Computer 1992, 25, 76–79. [CrossRef] 4. Wang, D.; Li, M. Deep stochastic configuration networks with universal approximation property. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. 5. Li, M.; Wang, D. 2-D Stochastic Configuration Networks for Image Data Analytics. IEEE Trans. Cybern. 2019, 51, 359–372. [CrossRef][PubMed] 6. Niu, H.; Wang, D.; Chen, Y. Estimating crop coefficients using linear and deep stochastic configuration networks models and UAV-based Normalized Difference Vegetation Index (NDVI). In Proceedings of the International Conference on Unmanned Aircraft Systems (ICUAS), Athens, Greece, 1–4 September 2020. Entropy 2021, 23, 56 13 of 13

7. Wang, W.; Wang, D. Prediction of Component Concentrations in Sodium Aluminate Liquor Using Stochastic Configuration Networks. Neural Comput. Appl. 2020, 32, 13625–13638. [CrossRef] 8. Lu, J.; Ding, J.; Dai, X.; Chai, T. Ensemble Stochastic Configuration Networks for Estimating Prediction Intervals: A Simultaneous Robust Training Algorithm and Its Application. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 5426–5440. [CrossRef][PubMed] 9. Huang, C.; Huang, Q.; Wang, D. Stochastic Configuration Networks Based Adaptive Storage Replica Management for Power Big Data Processing. IEEE Trans. Ind. Inform. 2019, 16, 373–383. [CrossRef] 10. Lu, J.; Ding, J. Mixed-Distribution-Based Robust Stochastic Configuration Networks for Prediction Interval Construction. IEEE Trans. Ind. Inform. 2019, 16, 5099–5109. [CrossRef] 11. Scardapane, S.; Wang, D. Randomness in neural networks: An overview. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2017, 7, e1200. [CrossRef] 12. Pao, Y.H.; Park, G.H.; Sobajic, D.J. Learning and Generalization Characteristics of the Random Vector Functional-link Net. Neurocomputing 1994, 6, 163–180. [CrossRef] 13. Lukoševiˇcius,M.; Jaeger, H. Reservoir Computing Approaches to Recurrent Neural Network Training. Comput. Sci. Rev. 2009, 3, 127–149. [CrossRef] 14. Lee, D.D.; Pham, P.; Largman, Y.; Ng, A. Advances in Neural Information Processing Systems. Technical Report. 2009. Available online: http://toc.proceedings.com/34099webtoc.pdf (accessed on 27 December 2020). 15. Rinne, H. The Weibull Distribution: A Handbook; CRC Press: Boca Raton, FL, USA, 2008. 16. Wei, J. Research on Swarm Intelligence Optimization Algorithms and Their Applications to Parameter Identification of Fractional- Order Systems. Ph.D. Thesis, Beijing Jiaotong University, Beijing, China, 2020. 17. Wei, J.; Chen, Y.; Yu, Y.; Chen, Y. Optimal Randomness in Swarm-Based Search. Mathematics 2019, 7, 828. [CrossRef] 18. Tyukin, I.Y.; Prokhorov, D.V. Feasibility of random basis function approximators for modeling and control. In Proceedings of the IEEE Control Applications, (CCA) & Intelligent Control, (ISIC), St. Petersburg, Russia, 8–10 July 2009. 19. Bernardi, M.; Petrella, L. Interconnected Risk Contributions: A Heavy-tail Approach to Analyze US Financial Sectors. J. Risk Financ. Manag. 2015, 8, 198–226. [CrossRef] 20. Ahn, S.; Kim, J.H.; Ramaswami, V. A New Class of Models for Heavy Tailed Distributions in Finance and Insurance Risk. Insur. Math. Econ. 2012, 51, 43–52. [CrossRef] 21. Resnick, S.I. Heavy-Tail Phenomena: Probabilistic and Statistical Modeling; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2007. 22. Foss, S.; Korshunov, D.; Zachary, S. An Introduction to Heavy-Tailed and Subexponential Distributions; Springer: Berlin/Heidelberg, Germany, 2011; Volume 6. 23. Niu, H.; Chen, Y.; Chen, Y. Fractional-order extreme learning machine with Mittag-Leffler distribution. In Proceedings of the International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, Anaheim, CA, USA, 18–21 August 2019; American Society of Mechanical Engineers: New York, NY, USA, 2019. 24. Hariya, Y.; Kurihara, T.; Shindo, T.; Jin’no, K. Lévy flight PSO. In Proceedings of the IEEE Congress on Evolutionary Computation (CEC), Sendai, Japan, 25–28 May 2015. 25. Yang, X.S. Nature-inspired Metaheuristic Algorithms; Luniver Press: Beckington, UK, 2010. 26. Spiegel, M.R.; Schiller, J.J.; Srinivasan, R. Probability and ; McGraw-Hill: New York, NY, USA, 2013. 27. Niu, H.; Wang, D.; Chen, Y. Estimating actual crop evapotranspiration using deep stochastic configuration networks model and UAV-based crop coefficients in a pomegranate orchard. In Proceedings of the Autonomous Air and Ground Sensing Systems for Agricultural Optimization and Phenotyping V, San Diego, CA, USA, 27 April–8 May 2020; International Society for Optics and Photonics: San Diego, CA, USA, 2020.