International Journal of Environmental Research and Public Health
Article Exploring Spatial Influence of Remotely Sensed PM2.5 Concentration Using a Developed Deep Convolutional Neural Network Model
Junming Li 1, Meijun Jin 2,* and Honglin Li 3 1 School of Statistics, Shanxi University of Finance and Economics, Wucheng Road 696, Taiyuan 030006, China; [email protected] or [email protected] 2 College of Architecture and Civil Engineering, Taiyuan University of Technology, Yingze Street 79, Taiyuan 030024, China 3 Shanxi Centre of Remote Sensing, 136 Street Yingze, Taiyuan 030001, China; [email protected] * Correspondence: [email protected]
Received: 2 January 2019; Accepted: 31 January 2019; Published: 4 February 2019
Abstract: Currently, more and more remotely sensed data are being accumulated, and the spatial analysis methods for remotely sensed data, especially big data, are desiderating innovation. A deep convolutional network (CNN) model is proposed in this paper for exploiting the spatial influence feature in remotely sensed data. The method was applied in investigating the magnitude of the spatial influence of four factors—population, gross domestic product (GDP), terrain, land-use and land-cover (LULC)—on remotely sensed PM2.5 concentration over China. Satisfactory results were produced by the method. It demonstrates that the deep CNN model can be well applied in the field of spatial analysing remotely sensed big data. And the accuracy of the deep CNN is much higher than of geographically weighted regression (GWR) based on comparation. The results showed that population spatial density, GDP spatial density, terrain, and LULC could together determine the spatial distribution of PM2.5 annual concentrations with an overall spatial influencing magnitude of 97.85%. Population, GDP, terrain, and LULC have individual spatial influencing magnitudes of 47.12% and 36.13%, 50.07% and 40.91% on PM2.5 annual concentrations respectively. Terrain and LULC are the dominating spatial influencing factors, and only these two factors together may approximately determine the spatial pattern of PM2.5 annual concentration over China with a high spatial influencing magnitude of 96.65%.
Keywords: spatial influence; PM2.5 pollution; deep convolutional network; remote sensing
1. Introduction Remote sensing technology has developed rapidly since the 1960s [1], and an abundance of remote sensing data has been accumulated in this 50-year period. Although abundant remotely sensed data have been applied to many fields, such as ecology, environment, geography, etc., the spatial analysis method for remotely sensed lattice data desiderates innovation. A single spatial variable generally has autocorrelation [2] (i.e., spatial dependence [3,4]), and various spatial variables have correlation. Spatial autocorrelations can be analysed with local indicators of the spatial association (LISA) index [5] (e.g., local Moran I [6], local Geary c index [7]). The main objective of spatial analysis is to identify the natural relationships that exist between variables [8,9]. The mainstream classical spatial analysis models, e.g., spatial lag model [10,11], spatial error model [10,11], and Bayesian spatial regression model [12], can only evaluate the overall or average linear correlation feature over a whole study region, neglecting the details of local area. These methods ignore the consequences of spatial heterogeneity [13]. The majority of spatial analysing methods assume stationary space.
Int. J. Environ. Res. Public Health 2019, 16, 454; doi:10.3390/ijerph16030454 www.mdpi.com/journal/ijerph Int. J. Environ. Res. Public Health 2019, 16, 454 2 of 11
However, assuming spatial convariance structure to be stationary is not so reasonable [14]. The spatial influencing relationship can better be explored when the analysis is local and more detailed results can be yielded [15]. The inclusion of a spatial heterogeneity resulting from differences in environmental conditions, socioeconomic dynamics, and other factors reinforces the need for more regionalized spatial analyses in exposure assessment and public health [16]. Although, the geographically weighted regression (GWR) [17] method considers local details; however, it can only describe a linear or simple non-linear spatial influencing relationship. In the era of big data, the need for developing advanced spatial analysis methods (e.g., machine learning methods) for remotely sensed data is urgent. Previously, there are several studies that have applied machine learning methods to address the influencing factors on PM2.5 concentrations. Zheng et al. [18] used traditional artificial neural networks to model spatial correlation between Beijing’s air qualities and influencing factors, e.g., meteorology, traffic flow, human mobility. Yan et al. [19] predicted the daily average PM2.5 concentration in Nanjing, Beijing, and Sanya, combining meteorological and contaminant factors based on the Long Short-Term Memory (LSTM) model. Suleiman et al. [20] presented a machine learning model to predict the traffic-related PM10 and PM2.5 concentrations from various variables (e.g., traffic variables). Hsieh et al. [21] proposed a semi-supervised learning algorithm to optimize the monitoring locations of air quality in Beijing based on spatial correlation. Certainly, there are some studies which utilized typical methods to investigate the influence of satellite-based PM2.5. He et al. [22] used empirical orthogonal function (EOF) to analyse the relationship between remotely sensed PM2.5 and climate circulation transformation in East China. Hajiloo et al. [23] employed geographical weight regression (GWR) to investigate impact of meteorological and environmental parameters on PM2.5 concentrations in Tehran, Iran. Yang et al. [24] quantified the influence of natural and socioeconomic factors on PM2.5 pollution using the GeoDetector model [25,26]. This study proposed a spatial analysis method that exploits the spatial influencing feature of remotely sensed data based on the deep CNN. CNN is a mainstream deep learning method and can effectively extract the feature representations from a large number of images [27] and object detection [28]. Some researchers have applied deep CNN in remote sensing classification. Q. Zou et al. [29] and Zhao et al. [30] proposed a DBN method for high-solution satellite imagery classification. H. Liang and Q. Li, C. Tao et al., and F.P.S. Luus et al. [31], Nogueira et al. [32], Volpi et al. [33], Chen et al. [34], have employed deep CNN in hyperspectral imagery classification or feature extraction. Some researchers have also employed deep CNN in synthetic aperture radar (SAR) image classification, e.g., Du et al. [35] and Geng et al. [36]. To our knowledge, the research applying deep CNN into spatial influencing of remotely sensed lattice data is very rare. This study aimed to present a deep CNN model exploiting the magnitude of spatial influence of four factors—population, gross domestic product (GDP), terrain, and land-use and land-cover (LULC)—to remotely sense the annual mean concentration of PM2.5 over China. This model not only considers local spatial heterogeneity but also has super nonlinear fitting ability. Therefore, the presented model is rooted in a deep learning framework and may reduce uncertainty of the results obtained from a simplistic correlation analysis or simple regression model, therefore giving better information to decision makers of public health.
2. Materials and Methodology
2.1. Materials
The materials used in this research contain five types of data: remotely sensed PM2.5 concentration, population spatial distribution density, GDP spatial distribution density, terrain data, and LULC in China. The remotely sensed PM2.5 annual concentration dataset in 2010 was produced by the Atmospheric Physics Institute of Dalhousie University in Canada [37] with a resolution of 0.1◦ × 0.1◦. The population density data in this paper are cited in the Gridded Population of the World (GPW), data of the UN-Adjust Population Density-v4 [38], published by a data centre in NASA’s Earth Observing Int. J. Environ. Res. Public Health 2019, 16, 454 3 of 11
System Data and Information System (EOSDIS), with a resolution of 300 × 300. GDP spatial distribution density, terrain data, and LULC datasets were drawn from the Resources and Environmental Science Centre of the Chinese Academy of Sciences (http://www.resdc.cn). All above-mentioned data were projected by Albers Conic Equal Area with WGS-84 datum, and the resolution was unified to 10 km × 10 km.
Int. J. Environ.2.2. Methodology Res. Public Health 2018, 15, x 3 of 11 The methodology in this paper consists of two modules: processing geospatial data and 2.2 Methodologystructuring the deep CNN model. The purpose of the former is to establish the dataset for the Thedeep methodology CNN model. in The this deep paper CNN consists model of undertakestwo module thes: processing mission of fittinggeospatial the complexdata and function structuring of the deep spatialCNN model. correlation The relationship.purpose of the former is to establish the dataset for the deep CNN model. The deep CNN model undertakes the mission of fitting the complex function of spatial correlation relationship. 2.2.1. Processing Geospatial Data 2.2.1. ProcessingThe deep Geospatial CNN method Data is usually applied in image identification or classification, not directly Thetransplanted deep CNN into method analysing is usually geospatial applied data. in Inimage the geospatial identification issue, or spatial classification, correlation not and directly geographical transplanted into analysingattribute needgeospatial to be data. considered. In the geospatial Hence, geospatial issue, spatial data correlation require technical and geographical processing attribute to match need the to be considered.deep CNN Hence, model geospatial structure. data The require four influencing technical pr factorsocessing generate to match inputs. the deep Each CNN pixel locationmodel structure. contains The PM2.5 concentrations as output and four influencing factors. In view of spatial correlation, the pixel four influencing factors generate inputs. Each pixel location contains PM . concentrations as output and four influencinglocation factors. and the In surroundingview of spatial locations correlation, should the be pixel considered. location Theand deepthe surrounding CNN model locations has the ability should be considered.of processing The deep big CNN data; model therefore has the the ability order of processi spatial correlationng big data; can therefore be amplified. the order In of this spatial paper, correlation the can beorder amplified. of spatial In this correlation paper, the adopts order of n-order spatial shape, correlation(2n + adopts1) × ( 2nn-order+ 1) pixelsshape, (n(2n= +1, 1)2, ... ×). (2n Figure + 1) 1pixels (n = 1,2,shows …). Figure an illustration 1 shows ofan 5-orderillustration shape of 5-order of the spatialshape of correlation the spatial extent,correlation including extent,11 including× 11 pixels. 11 × 11 pixels.Subsequently, Subsequently, it canit can extract extract the the corresponding corresponding four four sets sets of influencing of influencing factor factor attribute attribute data data for a for pixel a pixel location.location. Each dataset Each dataset of influenc of influencinging factors factors comprises comprises the correspon the correspondingding values valuesof the surrounding of the surrounding (2n + 1) × (2n + 1) × (2n + 1) pixels. In short, the PM annual concentration of a pixel location is affected (2n + 1) pixels. In short, the PM . annual concentration2.5 of a pixel location is affected by the four influencing factorsby of the its four own influencing and the surrounding factors of its n-order own and spatial the surroundingcorrelation extent, n-order (2n spatial + 1) correlation× (2n + 1) extent,pixels. The mathematic(2n + 1 form) × (2n can+ be1) expressedpixels. The as mathematic follows: form can be expressed as follows:
(PM ) ( ) C 2.5 = F Pop , GDP . ,=Ter , LULC + ξ (1) i i|(2n+1)×(2n+1) i|(2n+1)× (2n+1) i|(2n+1)×(2n+1) i|(2n+1)×(2n+1) i (1) |( )×( ), |( )×( ), |( )×( ), |( )×( ) + (PM ) ( ) 2.5 where C . i is the PM2.5 annual concentration of the i-th pixel, Popi|(2n+1)×(2n+1), where is the PM . annual concentration of the i-th pixel, |(2 +1)×(2n+1) , GDPi|(2n+1)×(2n+1), Teri|(2n+1)×(2n+1), LULCi|(2n+1)×(2n+1) represent the four influencing factor |(2 +1)×(2n+1), |(2 +1)×(2n+1), |(2 +1)×(2n+1) represent the four influencing factor attribute values of attribute values of the i-th pixel and its surrounding (2n + 1) × (2n + 1) pixels, and ξ represents the i-th pixel and its surrounding (2n + 1) × (2n + 1) pixels, and represents the error. The spatiali influencing ( ) functionthe F(. error. ) can The be spatial learned influencing by the deep CNN function model. F . can be learned by the deep CNN model.
Figure 1. Illustrating 5-order shape of the extent of spatial correlation. Figure 1. Illustrating 5-order shape of the extent of spatial correlation
2.2.2. A Developed Deep Convolutional Neural Network Model CNN contains two categories of cells in the visual cortex, simple cells which exploit local features and complex cells which “pool” (e.g., maximizing, averaging) the outputs of simple cells within a neighbourhood. The structure of CNN model which has two special aspects of local connections and sharing weights is different from general deep learning models. A complete deep CNN stack three types of layers, convolutional layers, pooling layers, and full connected layers. Int. J. Environ. Res. Public Health 2019, 16, 454 4 of 11
2.2.2. A Developed Deep Convolutional Neural Network Model CNN contains two categories of cells in the visual cortex, simple cells which exploit local features and complex cells which “pool” (e.g., maximizing, averaging) the outputs of simple cells within a neighbourhood. The structure of CNN model which has two special aspects of local connections and sharing weights is different from general deep learning models. A complete deep CNN stack three types of layers, convolutional layers, pooling layers, and full connected layers. The commonly used CNNs are 2-Dimensional CNN and 3- Dimensional (3-D) CNN. Figure2 showsInt. J. Environ. a 3-D Res. CNN Public illustration Health 2018, 15, withx m (m = 1, 2, ... ) filters and k (k = 1, 2, ... ) convolution kernels.4 of 11 xyz The value of a neuron Sij at position (x, y, z) of the j-th convolutional feature in the i-th layer can be expressedThe commonly as follows used [34]: CNNs are 2-Dimensional CNN and 3- Dimensional (3-D) CNN. Figure 2 shows a 3- D CNN illustration with m (m=1,2,…) filters and k (k=1,2,…) convolution kernels. The value of a neuron S Pi−1 Qi−1 Ki−1 at position ( , , ) of the j-xyzth convolutional feature in thepqk i-th(x +layerp)(y+ canq)(z be+k )expressed as follows [34]: S = C( w S + bij) (2) ij ∑ ∑ ∑ ∑ ijm (i−1)m m p=0 q=0 k= 0 ( )( )( ) = ( ( ) + ) (2) where m indexes the convolutional feature in the (i − 1)th layer connected to the j-th convolutional feature,where m and indexesPi and theQ convolutionali are the height feature and in the the width ( − 1) ofth the layer convolutional connected to kernel.the j-th convolutionalKi is the size feature, of the pqk spatialand influencing and are the factors, heightw andijm isthe the width value of the of convolutional position (p, q kernel., k) connected is the tosize the of m-theth spatial convolutional influencing feature,factors, and b ij isis the the value bias of of position the j-th ( , , ) convolutional connected feature to the in m- theth convolutionali-th layer. feature, and is the bias of the j-th convolutional feature in the i-th layer.
Factor-1
Convolutional Feature-1
Factor-2 ……
……
Convolutional Feature-m
Factor-k
FigureFigure 2.2. TheThe illustration illustration of 3-D of 3-Dconvolution convolution with m with (m=1,2 m,…) (m filters = 1, 2,and... k (k=1,2,…)) filters and convolution k (k = 1,kernels, 2, ... the) convolutionweights are color-coded. kernels, the weights are color-coded.
ThisThis paperpaper designs a a deep deep 3-D 3-D CNN CNN model model for for exploiting exploiting spatial spatial influencing influencing feature feature of remotely of remotely sensed senseddata. Figure data. Figure 3 illustrates3 illustrates the presented the presented deep deep CNN CNN model model architecture architecture which which contains contains including including four fourconvolutional convolutional layers, layers, four polling four polling layers, layers,and three and hidden three layers. hidden And layers. the activation And the function activation for function hidden layer for hiddenadopted layer the adoptedRectified theLinear Rectified Unit (ReLU) Linear function. Unit (ReLU) The pooling function. mode The employed pooling mode average employed mode. The average batch mode.normalization The batch was normalization set in each layer was except set infor eachthe output layer layer. except The for dropout the output ratio layer.and learning The dropout ratio were ratio set as 25% and 1%, respectively. The dimension of the pre-processed input neural layer is (2n + 1) × (2n + 1) × 4, and learning ratio were set as 25% and 1%, respectively. The dimension of the pre-processed input including four sets of influencing factors with the i-th pixel and its surrounding (2n + 1) × (2n + 1) pixels. neural layer is (2n + 1) × (2n + 1) × 4, including four sets of influencing factors with the i-th pixel Table 1 lists the experimental results when the spatial correlation parameter n was assigned various values. It andshows its surroundingthat the validation(2n + accuracy1) × (2n reaches+ 1) pixels. the highest Table wh1 listsen the the spatial experimental correlation results parameter, when n, the is spatialtaken 9, correlationalthough the parameter training accuracy n was assigned is improved various along values.with theIt increase shows of that the the parameter, validation n. Considering accuracy reaches that the thevalidation highest accuracy when the is spatialbetter indica correlationtor representing parameter, the n,accuracy is taken of 9,a althoughmodel. Hence, the trainingthe spatial accuracy correlation is improvedparameter, along n, is withassigned the increasewith nine. of Then the parameter, the input laye n. Consideringr contains 19 that× 19 the× 4 validationneurons with accuracy the four is factors’ better indicatorattribute representingvalue. The number the accuracy of the ofoutput a model. neuron Hence, is 11, the labelled spatial by correlation PM . annual parameter, concentration n, is assigned with 11 categories: <10μg/m ,10~20 μg/m ,20~30 μg/m , 30~40 μg/m , 40~50 μg/m , 50~60 μg/m , 60~70 μg/ m , 70~80 μg/m , 80~90 μg/m , 90~100 μg/m , > 100 μg/m .
Int. J. Environ. Res. Public Health 2019, 16, 454 5 of 11
with nine. Then the input layer contains 19 × 19 × 4 neurons with the four factors’ attribute value. The number of the output neuron is 11, labelled by PM2.5 annual concentration with 11 categories: <10 µg/m3, 10 ∼ 20 µg/m3, 20 ∼ 30 µg/m3, 30 ∼ 40 µg/m3, 40 ∼ 50 µg/m3, 50 ∼ 60 µg/m3, 3 3 3 3 3 60Int.∼ J. Environ.70 µg/m Res. Public, 70 Health∼ 201880 ,µ 15g/m, x , 80 ∼ 90 µg/m , 90 ∼ 100 µg/m , >100 µg/m . 5 of 11
Influencing factors In p ut
Depth=8 Depth=8 Depth=16 Depth=16 Depth=32 Depth=32
Pop GDP Ter LULC
8 filters Average 16 filters Average 32 filters Average 64 filters 3D Conv. Pooling 3D Conv. Pooling 3D Conv. Pooling 3D Conv.
FC FC FC ReLU
ReLU ReLU Softmax Depth=64 Depth=64
Average Pooling
Labeled . concentration Pop: population; GDP: Gross Domestic Product; Ter: Terrain; LULC: Land use and land cover; FC: Full connected; Conv.: Convolutional; ReLU: Rectified Linear Unit
Figure 3. Illustration of the presented deep CNN model architecture exploiting spatial influencing feature of
Figureremotely 3. sensedIllustration PM . concentration, of the presented including deep four convolutional CNN model layers, architecture four polling exploiting layers, and three spatial hidden influencing featurelayers, the of remotelyfirst layer sensedis input containingPM2.5 concentration, four influencin includingg factors’ values four convolutionalon a pixel, the output layers, layer four with polling 11 layers, andneurons three consisting hidden of layers, 11 categories the first of PM layer . annual is input concentrations containing on th foure pixel influencing location in the factors’ middle.values on a pixel, the output layer with 11 neurons consisting of 11 categories of PM annual concentrations on the 3. Results 2.5 pixel location in the middle. The remotely sensed PM . annual concentration and influencing factors possessed 96,337 pixels, among 3.which, Results 86,903 pixels (accounting for the ratio of 90%) were used for deep learning, and the remaining 9434 pixels (accounting for 10%) were reserved for validation. Training accuracy is defined as the accuracy applied to theThe training remotely data (i.e., sensed 86,903PM pixels),2.5 annual while validation concentration accuracy and is the influencing accuracy for factors the remaining possessed data (i.e., 96,337 pixels, among9434 pixels), which, and 86,903 estimated pixels accuracy (accounting is the accuracy for thefor the ratio total of data 90%) (i.e., were 96,337 used pixels). for To deep investigate learning, the and the integrated and respective spatial influence of the four various factors, we exploited the congregate magnitude remaining 9434 pixels (accounting for 10%) were reserved for validation. Training accuracy is defined of spatial influence from the four factors and the separate influencing magnitude from one or two factors. as the accuracy applied to the training data (i.e., 86,903 pixels), while validation accuracy is the accuracy for the remaining data (i.e., 9434 pixels), and estimated accuracy is the accuracy for the total data (i.e., 96,337 pixels). To investigate the integrated and respective spatial influence of the four various factors, we exploited the congregate magnitude of spatial influence from the four factors and the separate influencing magnitude from one or two factors. Int. J. Environ. Res. Public Health 2019, 16, 454 6 of 11
Int. J. Environ. Res. Public Health 2019, 16, 454 6 of 11 Table 1. Experimental results of the training and validation accuracy of the deep 3-D CNN model when the spatial correlation parameter, n, is assigned various values Table 1. Experimental results of the training and validation accuracy of the deep 3-D CNN model when the spatialSpatial correlation correlation parameter, n, is assigned various values. Training accuracy Validation accuracy parameter, n Spatial Correlation Parameter, n Training Accuracy Validation Accuracy 1 67.94% 80.17% 12 77.71% 67.94% 82.37% 80.17% 2 77.71% 82.37% 33 88.08% 88.08% 86.11% 86.11% 44 92.01% 92.01% 90.50% 90.50% 55 94.51% 94.51% 91.83% 91.83% 66 96.53% 96.53% 92.14% 92.14% 7 97.35% 92.90% 87 97.35% 98.30% 92.90% 92.46% 98 98.30% 98.71% 92.46% 93.29% 109 98.71% 98.87% 93.29% 92.40% 1110 98.87% 99.29% 92.40% 93.28% 12 99.53% 93.25% 11 99.29% 93.28% 12 99.53% 93.25% 3.1. Integrated Spatial Influencing Feature 3.1 IntegratedIf the spatial four impact influencing factors feature were all fed into the input layer, after 1000 epochs of learning, the training accuracy of 86,903 pixels were 98.71%, and the validation accuracy of the remaining 9434 pixels If the four impact factors were all fed into the input layer, after 1,000 epochs of learning, the training reached 93.29%. Figure4 illustrates the spatial distribution of the original and estimated PM annual accuracy of 86,903 pixels were 98.71%, and the validation accuracy of the remaining 9,434 pixels reached2.5 93.29%. concentration of the total 96,337 pixels using the trained deep learning model fed with the four Figure 4 illustrates the spatial distribution of the original and estimated PM . annual concentration of the total influencing factors. 96,337 pixels using the trained deep learning model fed with the four influencing factors.
(A) (B)
(C)