Climate change affects soil development, plant growth, food security, water safety, air quality, traffic safety, tourism, agricultural production, forest fire, flood, ecosystems, renewable energy sources, energy management, sustainable socio-economic development, and human health1,2,3,4,5,6,7,8,9,10,11. Especially, extreme rainfall and temperature events often lead to serious disasters such as drought, mudslides and floods12. Climate change is having adverse effects on crop productivity13,14. Climate change is also the main driving factor for the annual growth of trees. The increment of tree growth decreases with the increase of daily maximum temperature, but increases with the increase of relative humidity 15. The global average annual air temperature in 2023 is 1.45 ± 0.12 °C higher than pre-industrial levels (1850–1900). Moreover, the average air temperature from 2014 to 2023 is 1.20 ± 0.12 °C higher than the temperature (1850–1900) (World Meteorological Organization). The Paris Agreement aims to control the global average temperature rise lower than pre-industrial levels by 2 °C, and strives to limit the temperature rise to within 1.5 °C. If humans significantly reduce greenhouse gas emissions and accelerate carbon neutrality, we still have the potential to avoid the worst climate disasters. Meanwhile, robust long-term climate predictions contribute to the formulation of climate management policies. Compared to seasonal forecasting, skillfully predicting monthly averages has significant social importance, and predictive information can be incorporated into time sensitive decisions, such as agricultural, urban, energy and reservoir management practices 16. However, current climate model forecast is not satisfying yet17. Therefore, accurate climate prediction is an urgent scientific issue that needs to be addressed. The sub seasonal time scale, ranging from approximately two weeks to three months, is often described as a “predictable desert” due to a lack of predictive skills.

Artificial intelligence (AI) models have been widely applied in multidisciplinary fields18,19,20,21,22,23,24,25,26,27,28,29,30,31. With the deepening of research, AI models are able to explore the context of time series and have been introduced into climate modeling32,33,34,35,36. Machine Learning (ML) models provide a new perspective to climate forecasting37,38,39,40. Artificial Neural Network (ANN) and Recurrent NN (RNN) are adapted to predict of outdoor air temperature in four European cities. ANNs are used to estimate precipitation in the Brazilian Legal Amazon41. The comparison of measured and predicted air temperature confirms the accurate training of the neural networks42. ANNs have nonlinear generalizability and predictive ability43. A deep learning algorithm is employed to predict the El Niño–Southern Oscillation (ENSO). At the six-month lead, the performance of the DL is significantly better than the multi-models44. Deep learning (DL) NN models are used for statistical downscaling45. Long short-term memory (LSTM) can learn nonlinear relationships with long-term dependencies, which may be difficult for traditional ML and NN models to capture46,47,48,49. A convolutional LSTM is developed for precipitation nowcasting50. The correlation skill of the Nino3.4 index of the convolutional neural networks (CNN) model is higher than those of current dynamical forecast systems51. The CNN could more accurately forecast future storm surges52. The CNN has been found to have the potential to extract spatial information from high-resolution predictions. It has been successfully applied to predict precipitation53 and tropical cyclone wind radius54. The MTL-NET model is utilized to forecast the Indian Ocean Dipole (IOD). The MTL-NET can forecast the IOD seven month ahead, outperforming most of climate dynamical models. The multi-task learning model (MTL-NET) contains four convolutional layers, two maximum pooling layers, and a LSTM layer55. The ConvLSTM and U-Net models are utilized to forecast Marine heatwaves (MHWs) in the the South China Sea (SCS) region. The innovative forecasting system gives better predictions of MHWs56. The Precipitation Nowcasting Network Via Hypergraph Neural Networks (PN-HGNN) is first adopted for precipitation forecasts57. The probabilistic method (DEUCE) integrates Bayesian neural networks and the U-Net to predicting precipitation58. Climate-invariant machine learning is proposed to predict climate59. The latent deep operator network (L-DeepONet) is used to forecast weather and climate60. The GCN-GRU (Graph Convolutional Network and Gated Recurrent Unit) is used to forecast visibility weather61. Nevertheless, Establishing robust artificial intelligence models for climate predicting remains a challenging task as there is still no clear understanding of climate data.

In summary, artificial intelligence can be used to predict climate variables and its predictive performance has been evaluated. The LSTM model, as an artificial intelligence method, has also made certain progress in predicting climate factors. CNN is one of the most representative network structures in DL. Compared with other DL approaches, CNN has stronger feature extraction and information mining capabilities. Combining CNN and LSTM can more effectively extract spatiotemporal features from input data. The experimental results show that the combined CNN-LSTM model has better prediction accuracy than independent LSTM and CNN models62,63,64. Various ANN, RNN, LSTM and CNN methods are employed to predict climate factors. However, the comparison between the multiple climatic elements prediction results of artificial intelligence models (ANN, RNN, LSTM, CNN, and CNN-LSTM) in the jinan city is not clear. This study conducted in-depth exploration and experimentations to find the best climate simulation model. In this article, we use a set of climate datasets to demonstrate that the CNN-LSTM hybrid model is an effective climate prediction system and can be applied to various climate prediction problems in principle.

In this study, CNN-LSTM was applied to predict climate time series. The proposed CNN-LSTM hybrid scheme was mainly tested in Jinan City to predict monthly climate factors. The following research issues are addressed: (1) What are the skills of the CNN-LSTM hybrid method compared to traditional ANN methods? (2) What are the hyperparameters of the CNN-LSTM hybrid model? (3) Is the CNN-LSTM hybrid solution of practical significance in East China? By answering these questions, the author aims to develop more proficient climate forecasting and introduce artificial intelligence methods into climate forecasting on a broader scale, inspiring future creative work.

Study area and data

Study area

Jinan is located in East China, central Shandong, and the southeast edge of North China Plain (Fig. 1). Jinan belongs to the continental monsoon climate zone. Jinan has excellent agricultural resources and a long tradition of vegetable cultivation, earning the reputation of “China’s high-quality vegetable basket”. In 2022, the total agricultural output value was 52.6 billion yuan, with a total grain output of 2.962 million tons. Jinan has completed afforestation of 773.3 hectares and forest nurturing of 5333.3 hectares, with over 60 families and more than 300 types of tree and shrub resources.

Figure 1
figure 1

Location Map of the research area. The map is generated using ArcGIS Pro 2.5 (ArcGIS Pro), URL: https://www.esriuk.com.

Data and data standardization

The monthly climatological data during 1951–2022 in Jinan city are obtained from the meteorological platform (http://www.nmic.cn/). The monthly climatological data include average atmospheric temperature (AT), extreme minimum atmospheric temperature, extreme maximum atmospheric temperature, precipitation (P), average relative humidity (RH), and sunlight hours (SH).

The climatological datasets are partitioned into three divisions for calibration (training) (80%) (from January 1951 to October 2008), verification (10%) (from November 2008 to November 2015), and testing (predicting) (10%) (from December 2015 to December 2022). Then, three divisions sets are standardized to ensure uniformity among the climatological data. This helps to improve the prediction accuracy of the artificial intelligence models. At the training phase, the optimal parameter values for the artificial intelligence models are assigned. At the prediction period, the standardized data are offered to the networks to climate forecast. At the predicting phase, the climatic elements are accurately predicted by the models.

In order to improve the training speed of the artificial intelligence models and the prediction accuracy, it needs standardization of the climate data and unify its value range between 0 and 165. The formula is as follows:

$$l_{m} = \frac{{L_{m} - L_{\min } }}{{L_{max} - L_{MIN} }},$$
(1)

where, lm is the normalized climate data, Lm represents the m-th value of the original climate data, Lmax expresses the maximum value of the climate sequence, Lmin expresses the minimum value of the climate sequence.

Methodology

Artificial neural network (ANN)

A three-layer back propagation (BP) ANN consists of a single input layer, a single hidden layer and a single output layer (Fig. 2). The artificial neurons (nodes) in each layer of the BPANN are closely connected with nodes in other layers66.

Figure 2
figure 2

Structure of ANN model.

The output value of each layer is estimated by the following equation:

$$Y_{i} = f\left( {\sum\limits_{j = 1}^{a} {\left( {W_{ij} X_{j} } \right)} + b_{t} } \right),$$
(2)

where Yi is the output of the ith node; f means activation function (transfer function); a is the number of nodes; Wij represents weight; Xj and bt is input of the jth neuron and threshold of nodes.

The error is calculated based on the forecasted and expected outputs. The mean square error (MSE) of the ANN is calculated using the equation:

$$MSE = \frac{1}{m}\sum\nolimits_{g = 1}^{m} {(D_{g} - K_{g} )}^{2} ,$$
(3)

where m, Dg and Kg are respectively the number of training climate data, the expected and actual output of the gth sample, and the actual output of the gth sample.

Four kinds of activation functions (transfer functions) are typically logsig (Sigmoid) in Eq. (4), purelin in Eq. (5), tansig (tanh) in Eq. (6) and Rectified Linear Unit (ReLU, or positive linear transfer function(poslin)) in Eq. (7), respectively.

$$logsig({\text{z}}) = \frac{1}{{1 + e^{ - z} }},$$
(4)
$$purelin({\text{z}}) = {\text{z,}}$$
(5)
$$tansig(z) = \frac{{e^{{\text{z}}} - e^{ - z} }}{{e^{z} + e^{ - z} }},$$
(6)
$$relu({\text{z}}) = \max (0,{\text{z}}),$$
(7)

After model training and validation, the settings of the model have been determined. The input variables are 12, the output variable is 1, and the neurons of hidden layer are 5. The activation functions are logsig and purelin, and training algorithm is trainbr. The learning rate is 0.001, epochs are 100 and the goal is 10–5.

Recurrent neural network (RNN)

Memory loss is the main drawback of feedforward artificial neural networks. To overcome this limitation, Recurrent Neural Networks (RNNs) integrate the feedback connections into their structure67. The RNNs employ an internal state to solve sequential climate time series forecasting.

Transferring information from one neuron to the next. For example, the last hidden state sn−1 and the recent input xn give the recent hidden state sn in Eq. (8). The recent output yn extracts features from Sn in Eq. (9)68.

$${\text{s}}_{n} = f\left( {Ax_{n} + B{\text{s}}_{n - 1} + b_{s} } \right),$$
(8)
$$y_{{\text{n}}} = W{\text{s}}{}_{n} + b_{0} ,$$
(9)

where A, B, W are the weights, and bs and b0 are the biases.

Similarly, the parameters of the model are set. The input variables are 12, the output variable is 1, and the neurons of hidden layer are 8. The activation functions are tansig for hidden layer and purelin for output layer. training function is trainlm. learning function is learngdm. Performance function mse is 10–5.

Long short-term memory neural network (LSTM)

LSTM is the most popular method in climate prediction, and used to address long-term dependency issues in RNN. Moreover, LSTM can more widely capture nonlinear features in climate data through three gates. The LSTM can be trained utilizing climate data to forecast climate change. The input gate determines how much input can be stored in the memory unit. The forgetting gate controls how much climate data from the last memory unit should be trashed. The output gate determines how much output from the memory unit. These gates operate together to regulate the climate data and assist LSTM in capturing and processing long-term dependencies in climate data. By utilizing multiple gates to filter the climate data, short-term memory (sn−1) and long-term memory (Dn−1) are processed. Therefore, LSTM can recall data on diverse timescales, making it suitable for climate forecast69.

Figure 3 illustrates the structural framework of the LSTM cell (unit). Inputs of the cell comprise the climate information kn, the hidden state of the previous cell sn−1 and the cell state of the previous cell Dn−1. Outputs of the unit include the hidden state st and the unit state Dn, which are regulated by the forget gate fn, the input gate jn, and the output gate on70. The regulating course and the outputs of the cell are calculated as follows:

$$f_{t} = \, Sigmoid(W_{f} \times \left[ {{\text{s}}_{n - 1} ,{\text{k}}_{n} } \right] + b_{f} ),$$
(10)
$$j_{t} = Sigmoid(W_{j} \times \left[ {s_{n - 1} ,k_{n} } \right]) + b_{j} ),$$
(11)
$$D_{n}^{\prime } = tanh(W_{d} \times \left[ {s_{n - 1} ,k_{n} } \right]) + b_{d} ),$$
(12)
$$D_{n} = f_{n} \times D_{n - 1} + j_{n} \times D_{n}^{\prime } ,$$
(13)
$$O_{n} = \, Sigmoid\left( {W_{o} \times \left[ {s_{n - 1} ,k_{n} } \right]} \right) + b_{o} ),$$
(14)
$$s_{n} = o_{n} \times tanh\left( {D_{n} } \right),$$
(15)

where b and W are the coefficient (Weight) and bias vector.

Figure 3
figure 3

The cell logic structure of the LSTM network.

Similarly, the hyperparameters of the model are set. The input size is 12, the number of output is 1, and the number of hidden units is 100. The activation functions are tanh and sigmoid. The Adam Optimizer is used and the learning rate is set to 0.001. The goal is 10–5.

Convolutional neural network (CNN)

CNN includes convolutional computation and has a deep structure (Fig. 4). Compared with ANNs, the advantage of the CNN is that it can improve the ability to extract inputting features by increasing depth. The operation steps for the CNN include: (a) after inputting the data into CNN, a large number of features are extracted and weights were reduced through several convolutions and pooling; (b) Transforming the features into regression results of nodes through a fully connected layer; (c) the CNN uses a resharing mechanism to assign the same weights to different neurons, thereby using the sum of the same weights, which reduces the number of parameters62. To process the input through layers, neurons, and activation functions, the ReLU is utilized. Normalization techniques are used to avoid overfitting. The Adam Optimizer is used and the learning rate is set to 0.001. The goal is 10–5. Epochs of trainning parameter are 200. Batch sizes are 20.

Figure 4
figure 4

Basic structure diagram of CNN.

CNN-LSTM

Figure 5 describes the climate simulation process using CNN-LSTM. Weight sharing and local receptive fields are the ways in which CNN learns the features of climate data. The output characteristics of each time step in the LSTM models are extracted through a gate mechanism. The fully connected layer of time distribution does not only obtain the output of the last time step, but also obtains the output of all time steps. On this foundation, this article puts forward a regression system for climate modeling using CNN and LSTM networks.

Figure 5
figure 5

The CNN-LSTM architecture.

This proposed CNN-LSTM model is directly fed the 3D climate data, and its shape is 12 × 1 × 1. In the first convolutional layer, shapes of convolutional kernels are 3 × 1 and stride is 1, which allows to extract abstract features of the climate data. In the second layer (max pooling layer), their shapes are 2 × 1 and their stride is 1. The third and fourth layers are respectively batch normalization and ReLU. In convolutional Layer 2, 32 kernels in the size of 3 × 1. To reduce the size of the feature matrix, followed a max-pooling layer, which has a size of 2 × 1 with 1 stride. One LSTM layer is utilized in the model. the LSTM contain 100 neurons. LSTM layers are translated into the features using a fully connected layer. Finally, a regression layer is employed to predict climatic element. Epochs of trainning parameter are 200. Batch sizes are 20. The learning rate is 0.005, and the Adam Optimizer is used.

Experiment

Cross-validation

Fivefold cross validation is utilized to avoid overfitting. The validation climate data are utilized for part-to-part training and validation. The dats are divided into five parts. The folding process is repeated five times. In the 1st fold, 20% is utilized for validation and 80% for training. Similarly, in the 2nd fold, 20% is utilized for validation. Therefore, by folding the climate data, overfitting of the artificial intelligence models is prevented. Because it uses different training climate data each time, it does not lean towards the training climate data. Therefore, overfitting issues can be prevented71.

Evaluation criteria

Performance evaluation of the different models is performed, by comparing the differences between the observed and simulated climate values. The various parameter metrics namely Pearson correlation coefficient (R), root mean square error (RMSE), and mean absolute error (MAE) are utilized. The different methods such as ANN, RNN, LSTM, CNN, and the CNN-LSTM are applied to get high performance72. Calculations of R, RMSE and MAE are in following equations.

$$R = \frac{{\sum {(L_{m} - \overline{L})(O_{m} - \overline{O})} }}{{\sqrt {\sum {(L_{m} - \overline{L})^{2} (O_{m} - \overline{O})^{2} } } }},$$
(15)
$$RMSE = \sqrt {\frac{{\sum {(L_{m} - O_{m} )^{2} } }}{I}} ,$$
(16)
$$MAE = \frac{1}{I}\sum {\left| {L_{m} - O_{m} } \right|} ,$$
(17)

where Lm is the value of observed climate data, Om is the predicted climate value, and I is the length of climate data. \(\overline{L}\) is the average of the observed climate data, \(\overline{O}\) represents the average of the forecasted climate values.

Simulating monthly average atmospheric temperature

The models have a “input–output” structure, where output is climatic element for t + 1 month and inputs are climatic element in t and t-s previous months. Through periodic analysis, the cycle of climate factors is 11. Meanwhile, there are 12 months in a year. Therefore, the input variables are 12. The models are used to simulate and forecast the climate changes of 1-time-step ahead for monthly time scale. The data from the past 12 months are used as input for the five models, and the next 1 month data is used as output for the five models.

The simulation results of the different algorithms in the ANN are shown in Table 1. Comparing the results, it can be seen that their training performances have significant differences. Trainbr has the highest R (0.9894), the lowest RMSE (1.8199), and the lowest MAE (1.4787). The trainbr algorithm also performs well in the validation stage, with R (0.9870), RMSE (1.9923) and MAE (1.6352), respectively. In the prediction phase, R, RMSE, and MAE of trainbr are respectively 0.9907, 2.0669, and 1.8042, indicating its good generalization ability.

Table 1 Comparison between various training algorithms in the ANN model.

The simulation results of different combinations of training functions (tansig (T), purelin (PU), logsig (L), poslin (PO)) are shown in Table 2. The combination (logsig-purelin) training results have the lowest RMSE and MAE values, and the highest R value, indicating a successful combination. In addition, this combination also performs well in cross validation stage and has good generalization ability in the prediction stage.

Table 2 Comparison between various transfer functions in the ANN model.

Table 3 illustrates the statistical results achieved using various artificial intelligence models (ANN, RNN, LSTM, CNN) and the proposed CNN-LSTM model in terms of different evaluation indicators namely R, RMSE, and MAE in training stage, verification stage and predicting stage. The proposed CNN-LSTM approach attains monthly average atmospheric temperature prediction accuracy with R (0.9981) and a minimal error in RMSE (0.6292) and MAE (0.5048). This indicates the CNN-LSTM model is well effective than other in monthly average atmospheric temperature prediction. The minimum prediction error is achieved by the CNN-LSTM model compared to other artificial intelligence models. The lower the predicting error, the greater will be predicting accuracy and performance. The hybrid CNN-LSTM model performed better than the four models in simulating the average atmospheric temperature for one month ahead.

Table 3 Comparison analysis between various models for simulated monthly average atmospheric temperature.

Figures 6 represent the predicted monthly average atmospheric temperature with respect to different models. The predicted value of the hybrid CNN-LSTM model is very close to the actual value. The five models can efficiently capture the monthly average atmospheric temperature trends as well as peaks. However, the performance of the CNN-LSTM model at simulating the monthly average atmospheric temperature is optimal.

Figure 6
figure 6

Comparison of observed and predicted values of monthly average atmospheric temperature in Jinan from 2015 to 2022. (a) Plot of the prediction results, (b) Scatterplot of the prediction results.

Comparison of observed and simulated monthly extreme minimum atmospheric temperature

Table 4 compares observed and simulated monthly extreme minimum atmospheric temperature using the best ANN, RNN, LSTM, CNN and CNN-LSTM models for one month ahead simulated during the three periods. It is worth noting that the five models have better performance in the calibration (training) period, the verification period and predicting period according to R, RMSE and MAE values. The CNN-LSTM model is the most accurate monthly extreme minimum atmospheric temperature forecast that offers one step ahead based on R, RMSE and MAE performance criteria.The value of R, RMSE and MAE are respectively 0.9971, 0.8350 °C and 0.6583 °C while training the CNN-LSTM model with 80% of basic data. The CNN-LSTM has nicely forecasted the monthly extreme minimum atmospheric temperature with the R of 0.9970, RMSE of 0.8287 °C, MAE of 0.5979 °C, respectively.

Table 4 Comparison analysis between various models for simulated monthly extreme minimum atmospheric temperature.

Figure 7 represents simulated results of the five artificial intelligence models for one-time-step-ahead extreme minimum atmospheric temperature forecasted in the prediction period. The five models capture the temporal variations and the peaks of the monthly extreme minimum atmospheric temperature. However, the performance of the CNN-LSTM model at simulating the monthly extreme minimum atmospheric temperature is optimal.

Figure 7
figure 7

Comparison of observed and predicted values of monthly extreme minimum atmospheric temperature in Jinan from 2015 to 2022. (a) Plot of the prediction results, (b) Scatterplot of the prediction results.

Comparison of observed and simulated monthly extreme maximum atmospheric temperature

Table 5 shows the R, MSE and RMSE of the five models for forecasting one time step in the training, verification and testing periods. The CNN-LSTM model has the slightest error and the most correlation in simulating monthly extreme maximum atmospheric temperature in the calibration, verification and predicting periods. Simulated results in the predicting stage for the R, RMSE and MAE criterion values for the CNN-LSTM model in the one step ahead are 0.9987, 0.4932 °C and 0.3931 °C, respectively. However, the forecasting results at one month ahead for ANN, RNN, and LSTM and CNN models are acceptable. The CNN-LSTM model performs better than the ANN, RNN, LSTM, and CNN models. For instance, the RMSE values in the predicting period for the best ANN, RNN, LSTM, and CNN models, to forecast monthly extreme maximum atmospheric temperature one time-steps-ahead 2.2345 °C, 2.2979 °C, 2.1666 °C, and 0.8212 °C, respectively, are obtained.

Table 5 Comparison analysis between various models for simulated monthly extreme maximum atmospheric temperature.

Figure 8 compares observed and forecasted extreme maximum atmospheric temperature using the ANN, RNN, LSTM, CNN and CNN-LSTM models for one-time-step-ahead forecasted in the predicting stage in the monthly time scale. The results indicate that CNN-LSTM model with good accuracy forecasted extreme maximum atmospheric temperature peaks. In general, the CNN-LSTM model forecasts in one time step are better than the ANN, RNN, LSTM, and CNN models.

Figure 8
figure 8

Comparison of observed and predicted values of monthly extreme maximum atmospheric temperature in Jinan from 2015 to 2022. (a) Plot of the prediction results, (b) Scatterplot of the prediction results.

Comparison of observed and simulated monthly precipitation

Table 6 represent the R analysis, MAE and RMSE analysis respectively with respect to the five models. The error values (MAE and RMSE) achieved by the CNN-LSTM model are less when compared to other four models such as ANN, RNN, LSTM, and CNN. The MAE and RMSE values achieved by the CNN-LSTM model in predicting period are 8.1762 mm and 6.7051 mm, respectively. The R values of the proposed CNN-LSTM model is 0.9962 which is larger compared to other existing models. From ANN to CNN-LSTM model, the RMSE and MAE decrease, and R increases.

Table 6 Comparison analysis between various models for simulated monthly precipitation.

Figure 9 shows the predicted results of five models and actual precipitation. The figure shows clearly that the overall performance of CNN-LSTM model is superior to ANN, RNN, LSTM, and CNN models. The trends of CNN and CNN-LSTM have better agreement with observations than ANN, RNN, and LSTM models. Moreover, Fig. 4 indicates that ANN model performs less than the RNN and LSTM models. The ANN model cannot efficiently capture the monthly precipitation trends as well as the peaks and it has a poor correlation with the observed precipitation. The ANN simulation cannot meet climate simulation objective, while the CNN and CNN-LSTM simulated results can.

Figure 9
figure 9

Comparison of observed and predicted values of monthly precipitation in Jinan from 2015 to 2022. (a) Plot of the prediction results, (b) Scatterplot of the prediction results.

Comparison of observed and simulated monthly average relative humidity

Table 7 represents the R, RMSE, MAE for the five models. The best CNN-LSTM with the slightest value RMSE and MAE and the most value R is selected in both the verification and predicting stages. In Table 5, the RMSE, MAE, and R criteria values for the CNN-LSTM model in predicting stages 0.6615% are 0.5093% and 0.9984, respectively.

Table 7 Comparison analysis between various models for simulated monthly average relative humidity.

Figure 10 shows the superior performance of the CNN-LSTM over the ANN, RNN, LSTM, and CNN models in forecasting monthly average relative humidity. The CNN-LSTM hybrid model has been able to simulate and predict relative humidity peaks well.

Figure 10
figure 10

Comparison of observed and predicted values of monthly average relative humidity in Jinan from 2015 to 2022. (a) Plot of the prediction results, (b) Scatterplot of the prediction results.

Comparison of observed and simulated monthly sunlight hours

Table 8 shows the R, RMSE and MAE values of the ANN, RNN, LSTM, and CNN, CNN-LSTM models in training, verification and predicting stages. The optimized CNN-LSTM model in predicting stage has better assessment criteria, with R = 0.9984, RMSE = 2.9058 h, MAE = 2.3794 h, respectively.

Table 8 Comparison analysis between various models for simulated monthly sunlight hours.

Figure 11 shows results comparing the observed and predicted monthly sunlight (sunshine) hours using the ANN, RNN, LSTM, and CNN, CNN-LSTM models for one month ahead predicting during the predicting period. Also, the forecasted values of the CNN-LSTM model are more compact in the proximity of the observed values, compared to the ANN, RNN, LSTM, and CNN models. this means better performance of the CNN-LSTM model.

Figure 11
figure 11

Comparison of observed and predicted values of monthly sunlight hours in Jinan from 2015 to 2022. (a) Plot of the prediction results, (b) Scatterplot of the prediction results.

Comparison with other literatures

It can be inferred from Tables 1, 2, 3, 4, 5 and 6 that the CNN-LSTM model improves the CNN and LSTM models, for its RMSE and MAE are smaller and R is larger than CNN and LSTM models in training, verification and testing stages. Therefore, model performance can be ranked from high to low as: CNN-LSTM > CNN > LSTM > RNN > ANN. The generalization ability of the five models is consistent with their fitting performance rank. The CNN-LSTM model provides strong guarantee for the flexibility in capturing inherit features of climate data.

In addition to the intuitive error metrics described above, we also calculated the average running time by recording 100 random runs of each model, and the running results are recorded in Table 9. Among all the models, CNN-LSTM is the most computationally expensive model, and ANN is the least computationally expensive model. However, CNN-LSTM increases in PM2.5 prediction accuracy.

Table 9 Average time for each model run.

Different artificial intelligence models have been applied to climate prediction in different regions. The RMSE and R2 values of the Feed-forward neural network (FNN) model for the monthly minimum atmospheric temperature (Tmin) prediction in Türkiye were calculated as about 1.09 °C and 0.98 for the testing process, respectively. The R2 value of of the FNN model for the monthly maximum air temperature (Tmax) prediction was calculated as 0.9928, whereas the RMSE value was obtained as 0.9136 °C. The FNN is more successful than the Elman neural network (ENN) in modeling Tmin, Tmax and monthly mean air temperature (T)73. A diurnal regressive NN algorithm is presented to predict urban air temperature in the U.S. Statistics generate a bias of 0.8 K, an RMSE of 2.6 K, and R2 of 0.874. The atmospheric temperature forecasting GATG model integrates the graph attention network (GAT) module and the gated recurrent unit (GRU) cell. When the predicting length is 96 h in china, the RMSE, MAE, and Correlation Coefficient (CORR) results of the GATG model are about 3.161, 2.849, and 0.831, respectively75. The CEEMD-BiLSTM is builded using CEEMD (complementary ensemble empirical mode decomposition) and BiLSTM (directional LSTM), and it is used to predict monthly temperature in Zhengzhou City. The average relative error of the CEEMD-BiLSTM is about 0.22%. The performance indicates that the model has a good prediction quality76. The CEEMDAN–BO–BiLSTM model has a MAE of 1.17, and a RMSE of 1.43 for monthly mean air temperature in Jinan City77. Our model's prediction accuracy is similar to those of these models.

Precipitation (P) series have unpredictable randomness and nonlinear. Consequently, the accurate prediction of P is very laborious. For improving P series forecast accuracy, P data is decomposed into some subcomponents through different decomposition approaches (wavelet transform (WT), time varying filtering (TVF-EMD), and CEEMD). Then, the ENN is used to construct, WT-ENN, TVF-EMD-ENN, and CEEMD-ENN models to forecast the subcomponents of P. Finally, the models are applied to P series in twelve different regions in China. The predictions by TVF-EMD-ENN are better than those of WT-ENN and CEEMD-ENN78. A hybrid monthly P prediction model (ABR-LSO) is built by integrating the AdaBoostRegressor (ABR) with the Lion Swarm Optimization (LSO) algorithm. RMSE and MAE values of the ABR-LSO are 52.39 and 39.89 in in Idukki, India79. A multiscale LSTM model with attention mechanism (MLSTM-AM) is employed to improve the accuracy of monthly precipitation prediction. The average Nash–Sutcliffe efficiency (NSE) and relative absolute error (RAE) in the YRB for the model are 0.66 and 0.35. The performance of the MLSTM-AM is better than the LSTM, MLSTM, and multiple linear regression (MLR)80. Our model has higher prediction accuracy than these models.

Two ANN models namely radial base function (RBF) and multiple layer perceptron (MLP) are employed for the forecasts of hourly and monthly meteorological variables in Ongole and Hyderabad, India. The RBF and MLP have given about 91–96% accuracy for the forecasts of air temperature (AT), and soil temperature (ST) and relative humidity (RH). In addition, the predictions demonstrated a strong correlation with the recorded data in between about 0.61 and 0.9481.

The different machine learning algorithms are used to predict AT and RH at Terengganu state in Malaysia. The multi-layered perceptron (MLP-NN) performs well in forecasting daily RH and AT with R of about 0.633 and 0.713, respectively. However, in monthly forecast AT, MLP-NN can be utilized to forecast monthly AT with R of about 0.846. Whereas in prediction monthly RH, the efficiency of radial basis function (RBF-NN) is higher than other models with R of about 0.711. The results indicate that there is great potential for predicting AT and RH values with an acceptable range of accuracy on either architecture of the ANN82. The ANNs are used to predict monthly relative humidity (RH) in Sivas Province, Turkey. The R, MAE, and RMSE values ranged between about 0.97–0.98, 1.92–2.59, and 2.47–3.39, respectively. The results show that the ANN provides satisfactory predictions for monthly RH83. The seasonal autoregressive moving average (SARIMA) and ANN models are utilized to forecast the monthly RH in Delhi, India. The SARIMA model provides the predicted RH with RMSE of about 6.0 and MAE of about 4.6. However, ANN model reported the predicted RH with RMSE of about 4.7 and MAE of about 3.4. The ANN is more reliable for forecasting RH than SARIMA84. The LSTM and adaptive neuro-fuzzy inference system (ANFIS) are used to forecast one day ahead of RH in Turkey. In the testing period, the MAE, R and RMSE values of the LSTM are 5.76%, 0.892 and 7.51%, respectively, in Erzurum province. Moreover, their values are 5.95%, 0.887 and 7.67%, respectively, in Erzurum province using the ANFIS. The results indicate that the ANFIS and LSTM perform satisfactory performance in RH prediction85. Our model has lower prediction errors than these models.

ANNs are used to estimate sunshine duration (SD) in Iran. Average determination coefficient (R2), mean bias error (MBE), RMSE for the comparison between measured and estimated SD are calculated resulting 0.81, 0.1% and 6.3%, respectively86. Three different ANN models, namely, general regression NN (GRNN), MLP, and RBF, are used in the estimation of sunshine duration (SD) in Turkey. The average of estimated SD is about 6.795 h 6.772 h, and 0.672 h by MLP, GRNN and RBF with mean bias error (MBE) 0.072 h, 0.049 h and − 0.054 h, percentage mean absolute error (MAPE) 9.36%, 9.88% and 10.75%, RMSE 0.626, 0.797, and 1.128, percentage root mean square error (%RMSE) 9.30%, 11.75% and 16.78% and R is found to be 0.956, 0.953 and 0.890, respectively. Statistical indicators show that MLP and GRNN models produce better results than the RBF87. Our model has higher prediction accuracy than these models.

In this article, we propose a more effective method to predict the monthly climate of Jinan city. The proposed method is based on the fusion of CNN and LSTM models. In order to further evaluate the predictive effectiveness of the CNN-LSTM model, the prediction results of the CNN-LSTM model are compared with the other three models (ANN, RNN, LSTM, and CNN). In summary, in the performance comparison of the five models, it can be seen that CNN-LSTM can better handle nonlinear and non-stationary climate signals. For the processing of climate time series, the LSTM is suitable for the forecast of nonlinear climate data, and it can further improve the forecast accuracy of the CNN. In addition, the LSTM model can capture a series of climate information before and after, which is more conducive to the prediction of non smooth and nonlinear climate time series. Considering the advantages of these two models, the constructed CNN-LSTM model is feasible for predicting nonlinear climate sequences with better predictive performance and higher accuracy. Consequently, the CNN-LSTM runs best in climate forecast, further indicating that the CNN-LSTM is effective and feasible for forecasting non-stationary climate data. When proposing, testing, and evaluating this new method, we intend to provide decision support for future rational planning, especially flood control and disaster reduction.

Conclusion

  1. (1)

    The potential of artificial intelligence models for monthly climate factors prediction during 1951–2022 in Jinan city is investigated in this study. The novel climate prediction model “CNN-LSTM” is presented to accurately predict climatic elements based on historical data. The effectiveness of the CNN-LSTM is validated by climate data in Jinan city. The performance of the CNN-LSTM is investigated using three evaluation indicators namely MAE, R and RMSE.

  2. (2)

    The CNN-LSTM is compared with ANN, RNN, LSTM, and CNN utilizing the same monthly climate data. The input variables lag times are derived from the cycle of monthly climate data. The performance achieved by the CNN-LSTM model is greater than other compared artificial intelligence models (ANN, RNN, LSTM, CNN). The prediction results of CNN-LSTM model is superior to other models in generalization ability and precision. This indicates the effectiveness of the proposed CNN-LSTM model over other existing artificial intelligence models. This means that compared to other models, the CNN-LSTM has the lowest RMSE and MAE values and the highest R value.

  3. (3)

    In climate change monitoring, extreme climate is of extraordinary importance, and the CNN-LSTM model also has higher efficiency than other four models in extreme climate prediction. The surveys show that predicted extreme air temperature is more accurate than predicted precipitation. The CNN model is close to CNN-LSTM model in some single coefficients but the errors demonstrate that the CNN-LSTM has more stable performance. The CNN-LSTM can reasonably predict the magnitude and monthly variations in climate elements, capture the peaks of the climate data, and meet the evaluation criteria of model performance. The CNN-LSTM is a promising approach for climate simulation.

Limitations and suggestions

  1. (1)

    The simulated monthly climate values of the hybrid CNN-LSTM model are substantially similar to the observed climate values, but there are some differences which require more accurate multi-time scale climate data. This study only considered historical climate data of the month as the input variable for the model. Therefore, building hourly model requires a higher resolution data. The application of accurate multi-time scale climate data can provide improvements in the climate prediction.

  2. (2)

    In future, the prediction accuracy of CNN-LSTM will also compare with current physics-based dynamical models. The models can be further extended by applying the decomposition techniques (WT, TVF-EMD, and CEEMD), bidirectional LSTM (BiLSTM), transfer learning, bidirectional GRU (BiGRU), and considering the residual components of climate data as input for the models development. The performance of climate modeling will be further improved by exploring multi-time scales, spatial features and other factors such as solar radiation, greenhouse temperature, terrain, vegetation, El Niño-Southern Oscillation, etc. This strategy can be utilized for any prediction application to make short-term, medium-term, and long-term operational plans. Therefore, the robustness of the models can be utilized to forecast multi-time scale climatic elements. In the simulation and prediction of different extreme climates, we will integrate multiple data such as remote sensing and ground observations for spatiotemporal multi-scale climate modeling. In the process of climate modeling, we will integrate physical mechanisms into machine learning models to improve the interpretability and generalization ability of the models. It provides a new technological approach for the combination of traditional physical models and advanced artificial intelligence algorithms.