APP下载

Accurate Multi-Site Daily-Ahead Multi-Step PM2.5 Concentrations Forecasting Using Space-Shared CNN-LSTM

2022-03-14XiaoruiShaoandChangSooKim

Computers Materials&Continua 2022年3期

Xiaorui Shao and Chang Soo Kim

1Department of Information Systems,Pukyong National University,Busan,608737,Korea

Abstract: Accurate multi-step PM2.5 (particulate matter with diameters≤2.5 um)concentration prediction is critical for humankinds’health and air population management because it could provide strong evidence for decisionmaking.However,it is very challenging due to its randomness and variability.This paper proposed a novel method based on convolutional neural network(CNN) and long-short-term memory (LSTM) with a space-shared mechanism,named space-shared CNN-LSTM(SCNN-LSTM)for multi-site dailyahead multi-step PM2.5 forecasting with self-historical series.The proposed SCNN-LSTM contains multi-channel inputs,each channel corresponding to one-site historical PM2.5 concentration series.In which, CNN and LSTM are used to extract each site’s rich hidden feature representations in a stack mode.Especially, CNN is to extract the hidden short-time gap PM2.5 concentration patterns; LSTM is to mine the hidden features with long-time dependency.Each channel extracted features are merged as the comprehensive features for future multi-step PM2.5 concentration forecasting.Besides, the space-shared mechanism is implemented by multi-loss functions to achieve space information sharing.Therefore, the final features are the fusion of short-time gap,long-time dependency,and space information,which enables forecasting more accurately.To validate the proposed method’s effectiveness,the authors designed,trained,and compared it with various leading methods in terms of RMSE,MAE,MAPE,and R2 on four real-word PM2.5 data sets in Seoul, South Korea.The massive experiments proved that the proposed method could accurately forecast multi-site multi-step PM2.5 concentration only using self-historical PM2.5 concentration time series and running once.Specifically,the proposed method obtained averaged RMSE of 8.05,MAE of 5.04,MAPE of 23.96%,and R2 of 0.7 for four-site daily ahead 10-hour PM2.5 concentration forecasting.

Keywords: PM2.5 forecasting; CNN-LSTM; air quality management;multi-site multi-step forecasting

1 Introduction

With the rapid development of industrialization and economics, air pollution is becoming a serious environmental issue, which threatens humankinds’health significantly.The PM2.5concentration could reflect the air quality and has been widely applied for air quality management and control [1,2].Thus, accurate PM2.5forecasting has been a hot topic and attracted massive attention as it could provide in-time and robust evidence to help decision-makers make appropriate policies to manage and improve air quality.There are two kinds of PM2.5forecasting tasks: onestep (also called single step) and multi-step.One-step PM2.5forecasting provides one-step ahead information, while multi-step forecasting provides multi-step ahead information.Citizens could benefit from them by taking peculiar actions in advance.

The current methods for PM2.5forecasting could be divided into regression-based, time series-based, and learning-based methods (also called data-driven methods) [3].Regression-based methods aim to find the linear patterns among multi-variables to build the regression expression.e.g., Zhao et al.[4] applied multi-linear regression (MLR) with meteorological factors including wind velocity, temperature, humidity, and other gaseous pollutants (SO2, NO2, CO, and O3) for one-step PM2.5forecasting.Ul-saufie et al.[5] applied principal component analysis (PCA) to select the most correlated variables to forecast one-step PM10with MLR model.Time series-based methods aim at mining the PM2.5series’hidden patterns between past historical and future values.The most popular time series-based method is auto-regressive integrated move average (ARIMA),which models the relationship between historical and future values by calculating three parameters:(p,d,q).It has been widely used for one-step PM2.5and air quality index (AQI) forecasting [6,7].However, both regression-based and time-series methods only consider the linear correlations in those meteorological factors while it is usually nonlinear [8].Thereby, the forecasting accuracy is still not satisfactory.

Learning-based methods, including shallow learning and deep learning, could extract the nonlinear relationships between meteorological variables and future PM2.5concentrations that have been applied for PM2.5forecasting.Support vector machine (SVM), one of the most attractive shallow learning-based methods, uses various nonlinear kernels to map the original meteorological factors into a higher-dimension panel to improve forecasting accuracy.e.g., Deters et al.[9]applied SVM for daily PM2.5analysis and one-step forecasting with meteorological parameters.Sun et al.[2] applied PCA to select the most correlated variables as the input of SVM for one-step PM2.5concentration forecasting in China.ANN, another shallow learning-based method, utilizes two or three hidden layers to extract hidden patterns for PM2.5concentration forecasting [10,11].However, SVM requires massive memory to search the high-dimension panel and is easy to fall into overfitting [12,13].ANN cannot extract the full hidden patterns due to it is not “deep”enough.Therefore, the forecasting accuracy is still not satisfactory and can be improved.In addition, some methods combined serval models to achieve better performance for AQI forecasting.e.g., Ausati et al.[1] combined ensemble empirical mode decomposition and general neural network (EEMD-GRNN), adaptive neuro-fuzzy inference system (ANFIS), principal component regression (PCR), and LR models for one-step PM2.5concentration forecasting using meteorological data and corresponding air factors.Cheng et al.[7] combined ARIMA, SVM, and ANN in a linear model to predict daily PM2.5concentrations in five of China’s cities.However, those combined methods still cannot address each model’s shortcoming, and they require handcrafted feature selection operations, which is time-consumption and increases the development’s cost.

Deep learning technologies, including deep brief network (DBN), convolutional neural network (CNN) [14], and recurrent neural network (RNN) [15], provide a new view for PM2.5forecasting due to their excellent feature extraction capacity.Xie et al.[16] applied a manifold learning-locally linear embedding method to reconstruct low-dimensional meteorological factors as the DBN’s input for daily single-step PM2.5forecasting in Chongqing, China.Kow et al.[17]utilized CNN and backpropagation (CNN-BP) to extract the hidden features of multi-sites in Korea with multivariate factors, including temperature, humidity, CO, and PM10, for multi-step and multi-sites hourly PM2.5forecasting.Park et al.[18] applied CNN with nearby locations’meteorological data for daily single-step PM2.5forecasting.Ayturan et al.[19] utilized a combination of gated recurrent unit (GRU) and RNN to forecast hourly ahead single-step PM2.5concentrations with meteorological and air pollution parameters.Zhang et al.[20] used VMD to obtain frequency-domain features from the historical PM2.5series as the input of the bidirectional LSTM (Bi-LSTM) for hourly single-step PM2.5forecasting.Moreover, some hybrid models combined CNN and LSTM have been developed for PM2.5concentration forecasting.For instance, Qin et al.[21] used one classical CNN-LSTM to make hourly PM2.5predictions.They collected wind speed, wind direction, temperature, historical PM2.5series, and pollutant concentration parameters as CNN’s input.CNN extracted features are fed into LSTM to mine the features consider the time dependence of pollutants for PM2.5forecasting.Qi et al.[22] developed a novel graph convolutional network and long short-term memory networks (GC-LSTM) for single-step hourly PM2.5forecasting.Pak et al.[23] utilized mutual information (MI) to select the most correlated factors to generate a spatiotemporal feature vector as CNN-LSTM’s input to forecast daily single-step PM2.5concentration of Beijing, China.Tab.1 gives a summary of recent important references using deep learning for PM2.5forecasting.

Although the above deep learning-based methods achieved good performance, Tab.1 showed that most of them require collecting meteorological and air quality data except for [20].Collecting those kinds of data is time-consumption and even is not available for most cases [24,25].Besides, Zhang et al.[20] proposed method requires the PM2.5series is long enough to do VMD decomposition while day-ahead forecasting cannot satisfy.Another observation showed that only CNN-BP [17] focused on multi-step hourly PM2.5concentration forecasting while others are single-step, which cannot satisfy human beings’needs.Motivated by those, this manuscript proposed a novel deep model to extract PM2.5concentration’s full hidden patterns for multisite daily-ahead multi-step PM2.5concentration forecasting only using self-historical series.In the proposed method, multi-channels corresponding to multi-site PM2.5concentration series is fed into CNN-LSTM to extract rich hidden features individually.Especially, CNN is to extract short-time gap features; LSTM is to mine the features with long-time dependency from CNN extracted feature representations.Moreover, the space-shared mechanism is developed to enable space information sharing during the training process.Consequently, it could extract rich and robust features to enhance forecasting accuracy.The main contributions of this manuscript are summarized as follows:

• To our best of understanding, we are the first to make multi-site multi-step PM2.5forecasting with the space-sharing mechanism only using the self-historical PM2.5series.

• A novel framework named SCNN-LSTM with a multi-output strategy is proposed to forecast daily-ahead multi-site and multi-step PM2.5concentrations only using self-historical PM2.5series and running once.Sufficient comparative analysis has confirmed its effectiveness and robustness in multiple evaluation metrics, including RMSE, MAE, MAPE,and R2.

• The effectiveness of each part in the proposed method has been analyzed.

Table 1: The recent references related to PM2.5 forecasting using deep learning technologies

The rest of the paper is arranged as follows.Section 2 gives a detailed description of the proposed SCNN-LSTM.The experimental verification is carried out in Section 3.Section 4 discusses the effectiveness of the proposed SCNN-LSTM for daily-ahead multi-step PM2.5concentrations forecasting.The conclusion is conducted in Section 5.

2 The Proposed SCNN-LSTM for Multi-Step PM2.5 Forecasting

2.1 Multi-Step PM2.5 Forecasting

Assume that we collected a long PM2.5concentration series, denoted as Eq.(1).Where the PM2.5series consists ofNvalues,Ttis the PM2.5concentration value at timet.

The current methods for multi-step forecasting consist of direct and recursive strategies [26].The direct strategy useshdifferent modelsfhfor each step forecasting, and each step forecasting results are independent, as described in Eq.(2).The futureh-step PM2.5concentrations from timet+1 tot+hare obtained using differenthmodels such as ARIMA, MLR, ANN, CNN, etc., with historical PM2.5concentrations from(t-m)thtotth.However, it did not consider the influence of each step, which leads to low-precision forecasting.The recursive strategy trains one modelfathtimes forh-step forecasting with dynamic inputs to overcome this shortcoming.The previous-step forecasting resultTt+1is used as the next step’s input to forecastTt+2, as described in Eq.(3).However, the accumulated error of each-step forecasting will decrease the forecasting accuracy significantly.Moreover, training one model many times is costly.

To avoid the above shortcomings, the proposed SCNN-LSTM method adopted a multi-output strategy for multi-step PM2.5forecasting, as shown in Eq.(4).Theh-step future PM2.5concentrationsT{t+1,t+2,...,t+h}are calculating by trained modelfonce with historical PM2.5concentrations(Tt-n,Tt-n-1,...,Tt-1,Tt).Our purpose is to build modelfto extract the full hidden feature representations that existed in the PM2.5series to forecast accurately.

2.2 Proposed SCNN-LSTM

The proposed SCNN-LSTM for multi-site daily-ahead multi-step PM2.5concentration forecasting consists of four steps: input construction, feature extraction, multi-space feature sharing,output and update the network, as shown in Fig.1.More details of each part are introduced in the following sections.

2.2.1 Input Construction

The proposed method adopted nearssites’self-historical PM2.5concentration series as input to mine its hidden patterns considering the influence on space.Before modeling, we utilized a non-overlapped algorithm [8] to generate each site’s corresponding input matrixSiteiandh-step label matrixLabelSiteisince CNN-LSTM is a supervised algorithm, as shown in Eqs.(5) and(6).Where a long time-series defined in Eq.(1) with the length ofNcould obtainMsamples using a non-overlapped algorithm in a sample rate ofmand stridek, each sampleSitei,j=[Tt-m+1,Tt-m+2,...,Tt] containsmPM2.5values from time step(t-m+1)th to tth.Moreover,m,k,M,N,hshould satisfy the constraint of Eq.(7).Theis a round-down operation.Therefore,the proposed SCNN-LSTM input matrix containsssites’PM2.5concentration series, as written in Eq.(8), the corresponding output matrix is defined in Eq.(9).

Figure 1: The proposed SCNN-LSTM for multi-site daily-ahead multi-step PM2.5 concentration forecasting

2.2.2 Feature Extraction

Due to the excellent feature extraction ability of CNN and the ability of LSTM to process time series with long-time dependency [27].The proposed method adopts one-dimensional (1-D)CNN to extract short-time gap features, the extracted features are fed into LSTM to extract the features with long-time dependency.There are two sub-steps in the feature extraction part, as described following.

Short-time gap feature extraction:Two 1-D non-pooling CNN layers [28] are utilized to extract hidden features in the short-time gap as the PM2.5series is relatively less-dimension.The process of 1-D convolution operation, as described in Eq.(10).Where the convoluted outputis calculated using(l-1)th’s outputxl-1andZfilters with a biasbz, each filter is denoted asAfter the convolution operation, the convoluted values are processed by one activation functionf(·)to activate the feature maps.Significantly, Rectified Linear Unit (ReLU) could enhance feature expression ability by improving the nonlinear expression of the feature maps, which was applied in the proposed SCNN-LSTM, as shown in Eq.(11).By utilizing two 1-D CNN layers,sitei’s short-time gap featuressfsiteicould be obtained, as defined in Eq.(12).WhereConv1Dis a 1-D convolution operation in the format of Conv1D(filters, kernel_size, stride).Especially, the proposed method utilized 32, 64 as two layers’filters and 5, 3 as the kernel size to extract the daily PM2.5short-time gap features.

LSTM feature extraction:Although CNN extracted short-time gap features, it loses some critical hidden patterns with long-time dependency.LSTM [29], a special RNN, could extract this kind of feature in a chain-like structure was utilized.The structure of LSTM, as shown in Fig.2.Three cells existed in Fig.2 over the timet-1,t,t+1.Three gate-like components in each LSTM cell, including input gateit, output gateot,and forgot gateftand control stateCtcontrolled the whole information flow.The calculations of LSTM’s components at timet, as described in Eqs.(13)-(18).

wheresfsitei,tis CNN extracted short-time gap features;wi,wf,wo,wcare connection weights of above three gates and cell state, andbi,bf,bo,bcare corresponding offset vectors;σis activation function sigmoid and * represents an element-wise calculation.The forgot gate decides what information from perceived hidden outputht-1should be deleted according to current inputsfsitei,t.In contrast, the dynamic states of the current cell are remembered by calculatinMoreover, the valuable partft*Ct-1at the previous cell and new informationit*will be added into theCt.The hidden informationhtis worked out according to the current statsCtand output informationot, it reduces some meaningless information meanwhile adding some new useful knowledge over time as each site’s final features for future PM2.5forecasting.Thus, some unnormal information such as bad weather, holiday, big events could be remembered to enhance forecasting accuracy.Throughsparallel LSTM layers, the short-time gap features with long-time dependency for each site could be obtained, as shown in Eq.(19).In which, each LSTM layer has 100 hidden nodes.

Figure 2: LSTM structure in the proposed SCNN-LSTM

2.2.3 Multi-space Feature Sharing

The space information is vital for PM2.5forecasting as one space PM2.5concentration is affected by adjacent spaces such as weather, environmental statues.To make full use of space information, the proposed method merged extracteds-site features from Eq.(19) as the fusion features to implement space sharing, as shown in Eq.(20).The fusion features contain shorttime gap, long-time dependency, and space information are utilized for multi-site multi-step PM2.5concentrations forecasting.

2.2.4 Multi-Site h-Step Output and Update the Network

The fusion features are utilized to forecast futures-siteh-step PM2.5concentrations in a linear mode.Each site’s features contribute the same, implemented by one fully connected Dense layer withhnodes, as shown in Eq.(21).Moreover, multi-site outputs are implemented by the multiple loss functions, which is the key to achieve space-sharing through concatenate layer.It calculates the mean square error (MSE) between each site’s ground truth and forecasting values to update the hidden layers and implement weights-sharing with a gradient descent algorithm.The total loss is the sum ofssites, as defined in Eq.(22).Each site contributes the same weight to the total loss.

3 Experimental Verification

To validate the proposed method’s effectiveness, the authors implement the proposed SCNNLSTM based on the operating system of ubuntu 16.04.03, TensorFlow backend Keras.Moreover,the proposed method adopted “Adam” as the optimizer to find the best convergence path and“ReLu” as the activation function except the output layer is “linear.”

3.1 Data

The authors adopted four real-word PM2.5concentration data sets to validate the proposed method’s effectiveness, including Jongno-gu, Jung-gu, Yongsan-gu, and Guangdong-gu from Seoul,South Korea, which is available on the website of http://data.seoul.go.kr/dataList/OA-15526/S/1/datasetView.do.Each data set is collected from 2017-01-01 00:00:00 to 2019-12-31 23:00:00.Noticed that some missing values caused by sensor failures or unnormal operations existed in each subset.The authors replaced them with the mean value of the nearest two values to reduce the influence of missing values.Then, we adopted Eqs.(5) and (6) to generate the input samples and the corresponding 10-step (h) labels in a sample ratem=24 and stridesk=24.It means that we utilize 24-hour historical PM2.5concentrations to forecast future 4-site daily-ahead 10-hour PM2.5concentrations.Finally, we got four subsets with a length of 1,095=samples.The more detailed information about the data, as described in Tab.2.

Table 2: Data description

3.2 Evaluation Metrics

We utilized multiple metrics including root mean square error (RMSE), mean absolute error(MAE), mean absolute percentage error (MAPE), and R square (R2) to evaluate the proposed method from multi-views.The calculation of each metric is given in Eqs.(23)-(26).Whereyiis the ground truth of PM2.5concentration at the timei,is the predicted value,is the meaning value of totaly, andnis the number of samples.

3.3 Workflow

The workflow for multi-site and multi-step PM2.5concentration forecasting using the proposed SCNN-LSTM, as shown in Fig.3.Firstly, four-site historical series are normalized with Eq.(27)to reduce the influence of different units.WhereTis PM2.5series,tiis PM2.5concentration at the timei,is normalized value.Then, utilizing Eqs.(5) and (6) to generate the corresponding input and output matrix.Thirdly, two-year (2017 and 2018) data are utilized for training the model,while the data in 2019 is for testing.In the training part, 20% of them are used to find the best model within minimum time with the early-stop strategy.Especially, we set patience as ten and epoch as 100.If the ‘val_loss’does not decrease in ten epochs, the training process will be ended.The model with the lowest ‘val_loss’will be saved as the best model.Otherwise, it will stop until 100 epochs.Lastly, the forecasting results are used to evaluate the model in terms of RMSE,MAE, MAPE, and R2.

3.4 Comparative Experiments

3.4.1 Multi-Step Forecasting

To validate the effectiveness and priority of the proposed method for daily-ahead multi-step PM2.5concentration forecasting, we compared the proposed SCNN-LSTM with some leading deep learning methods, including CNN [17], LSTM [19], CNN-LSTM [30].It is worth noticing that previous CNN and LSTM required additional meteorological or air pollutant factors; we use their structure for forecasting.The configurations of those comparative methods, as given in Tab.3.Each comparative model runs four times for four-site forecasting, while the proposed method only needs to run once.The comparison results using averaged RMSE, MAE, MAPE,and R2for four-site daily ahead 10-step PM2.5concentration forecasting, as shown in Tab.4.The findings indicated that the proposed method outperforms others, which won 13 times of 16 metrics on four subsets.Especially, the proposed method has an absolute priority at all evaluation metrics for all subsets compared to CNN.Although LSTM performs a little better than the proposed method at R2on ‘Jongno-gu’and MAE on ‘Jung-gu’.CNN-LSTM performs a little better than the proposed method at MAPE on ‘YongSan-gu.’They require to run various times to get the forecasting results for each site while the proposed method only needs to run once.Moreover, the standard error proved the proposed method has good robustness.The performance of each method is ranked as: Proposed >CNN-LSTM >LSTM >CNN by comparing the averaged evaluation metrics on four sites.Especially, the proposed method has an averaged MAPE of 23.96% with a standard error of 1.94%, while others are greater than 25%, and only the proposed method’s R2is more significant than 0.7.Besides, we found CNN could not forecast multi-step PM2.5concentration well due to all MAPE are greater than 100% (that is not caused by a division by zero error).The forecasting results for different methods, as shown in Fig.4.The findings indicated that only the proposed method could accurately forecast PM2.5concentration’s trend and value while others cannot.Also, Fig.4 shows long-step forecasting is more complicated than short-step.In summary, the proposed SCNN-LSTM could accurately, effectively, and expediently forecast multi-site daily-ahead multi-step PM2.5concentrations.

Figure 3: The workflow for multi-site and multi-step PM2.5 concentration forecasting

Table 3: The configuration information for three comparative methods

3.4.2 Each-step Forecasting

To explore and verify the effectiveness of the proposed method for each-step PM2.5forecasting, we calculated averaged RMSE, MAE, MAPE, and R2for each step on four sites using the above deep models except CNN as its MAPE does not make sense.The comparison results showed that the proposed method has an absolute advantage for each step forecasting on all evaluation metrics, which could be conducted from Fig.5.Especially, the proposed method has the lowest RMSE and R2for each-step forecasting.For MAE, the proposed method has the lowest values except for the third step is a little greater than LSTM.The MAPE indicated that the proposed method performs very well on the first five-step forecasting.Especially, the first three steps’MAPE is lower than 20%, which improves a lot compared to the other two methods.R2indicated that the proposed method could explain more than 97% for the first step, 92% for the second step.Moreover, the results indicated that the forecasting performance decreases with the steps.Significantly, the relationship between each evaluation metric and the forecasting step for the proposed method is denoted as Eqs.(28)-(31).It shows that if increasing one step, the RMSE will increase by 0.6588, MAE will increase by 0.4890, MAPE will increase by 2.2174, while R2will decrease by 0.0595, respectively.

Table 4: The comparative results for four-site daily ahead 10-hour PM2.5 concentration forecasting using averaged RMSE, MAE, MAPE, and R2 (“-” means greater than 100%)

Figure 4: The forecasting results using different methods

3.5 Ablation Study

To explore each part’s effectiveness in the proposed SCNN-LSTM, we designed four subexperiments.Specially, designed space-shared CNN (SCNN) to verify the effectiveness of LSTM;Designed space-shared LSTM to verify the effectiveness of CNN; and designed CNN-LSTM without a space-shared mechanism (CNN-LSTM alone) to verify its effectiveness; designed SCNN-LSTM with a recursive strategy to validate the effectiveness of the multi-output strategy.All configurations of those methods are the same as the proposed method.The results based on subset 2 (Jung-gu), as shown in Tab.5.

The results indicated that the utilization of LSTM had improved RMSE by 5.97%, MAE by 6.74%, MAPE by 32.57%, and R2by 1.49%, which conducts by comparing SCNN and the proposed method.The application of CNN has improved 5.14% of RMSE, 8.51% of MAE,7.37% of MAPE, and 10.29% of R2.By comparing the CNN-LSTM alone with the proposed method, we can conduct that the space-shared mechanism has improved 2.45% of RMSE, 2.02%of MAE, 3.18% of MAPE, and 1.49% of R2.Moreover, by comparing the SCNN-LSTM with a recursive strategy to the proposed method, the findings derived that the multi-output strategy has absolute priorities as it performs much better on RMSE, MAE, and MAPE except for R2is a little worse.In summary, the above evidence proved that CNN could extract the shorttime gap feature; LSTM could mine hidden features which have a long-time dependency; The space-shared mechanism ensures full utilization of space information; The multi-output strategy could save training cost simultaneously keeping high forecasting accuracy.Combining those parts properly could accurately forecast the multi-site and multi-step PM2.5concentrations only using self-historical series and running once.

Figure 5: Averaged metrics for each step forecasting on four sites: (a) RMSE, (b) MAE, (c)MAPE, and (d) R2

4 Discussion

We have proposed a novel SCNN-LSTM deep model to extract the rich hidden features from multi-site self-historical PM2.5concentration series for multi-site daily-ahead multi-step PM2.5concentration forecasting, as shown in Fig.1.It contains multi-channel inputs and outputs corresponding to multi-site inputs and future outputs.Each site’s self-historical series is fed into CNN-LSTM to extract short-time gap and long-time dependency individually first, then extracted features are merged as the final features to forecast multi-site daily-ahead multi-step PM2.5concentrations.

Table 5: The ablation study of the proposed SCNN-LSTM based on subset 2 (Jung-gu)

To validate the proposed method’s effectiveness, we compared it with three leading deep learning methods, including CNN, LSTM, and CNN-LSTM, on four real-word PM2.5data sets from Seoul, South Korea.The comparative results indicated that the proposed SCNN-LSTM outperforms others in terms of averaged RMSE, MAE, MAPE, and R2, which could be conducted from Tab.4.Especially, the proposed method got averaged RMSE, MAE, MAPE, and R2are 8.05%, 5.04%, 23.96%, and 0.70 on four data sets.Fig.4 confirmed its excellent forecasting performance again and showed the long-step forecasting is more changeling and difficult than short-step.Also, the proposed method has good robustness, which could be conducted from Tab.4 by using standard error.

Moreover, the authors have explored the proposed method’s effectiveness for each-step forecasting, as shown in Fig.5.The comparison results showed that the proposed method has an absolute advantage for each step forecasting on all evaluation metrics.Also, the relationship between each evaluation metric and the forecasting step has been conducted at Eqs.(28)-(31).It shows that if the forecasting step increases one, the RMSE will increase 0.6588, MAE will increase 0.4890, MAPE will increase 2.2174, while R2will decrease 0.0595, respectively.

To validate each component’s effectiveness, an ablation study is done, as described in Tab.5.The results indicate that CNN could extract the short-time gap feature; LSTM could mine hidden features which have a long-time dependency; The space-shared mechanism ensures full utilization of space information; The multi-output strategy could save training cost simultaneously keeping high forecasting accuracy, respectively.

By setting the parameters of the proposed SCNN-LSTM, the forecasting accuracy could improve.Future studies will focus on using deep reinforcement learning technology to find the best parameter under the structure of the proposed SCNN-LSTM.

5 Conclusions

This manuscript has developed an accurate, convenient framework based on CNN and LSTM for multi-site daily-ahead multi-step PM2.5concentration forecasting.In which, CNN is used to extract the short-time gap features; CNN extracted hidden features are fed into LSTM to mine hidden patterns with a long-time dependency; Each site’s hidden features extracted from CNNLSTM are merged as the final features for future multi-step PM2.5concentration forecasting.Moreover, the space-shared mechanism is implemented by multi-loss functions to achieve space information sharing.Thus, the final features are the fusion of short-time gap, long-time dependency, and space information, which is the key to ensure accurate forecasting.Besides, the usage of the multi-output strategy could save training costs simultaneously keep high forecasting accuracy.The sufficient experiments have confirmed its state-of-the-art performance.In summary, the proposed SCNN-LSTM could forecast multi-site daily-ahead multi-step PM2.5concentrations only by using self-historical series and running once.

Funding Statement:This work was supported by a Research Grant from Pukyong National University (2021).

Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.