Research article Special Issues

A time series image prediction method combining a CNN and LSTM and its application in typhoon track prediction


  • Typhoon forecasting has always been a vital function of the meteorological department. Accurate typhoon forecasts can provide a priori information for the relevant meteorological departments and help make more scientific decisions to reduce the losses caused by typhoons. However, current mainstream typhoon forecast methods are very challenging and expensive due to the complexity of typhoon motion and the scarcity of ocean observation stations. In this paper, we propose a typhoon track prediction model, DeepTyphoon, which integrates convolutional neural networks and long short-term memory (LSTM). To establish the relationship between the satellite image and the typhoon center, we mark the typhoon center on the satellite image. Then, we use hybrid dilated convolution to extract the cloud features of the typhoon from satellite images and use LSTM to predict these features. Finally, we detect the location of the typhoon according to the predictive markers in the output image. Experiments are conducted using 13, 400 satellite images of time series of the Northwest Pacific from 1980 to 2020 and 8420 satellite images of time series of the Southwest Pacific released by the Japan Meteorological Agency. From the experimentation, the mean average error of the 6-hour typhoon prediction result is 64.17 km, which shows that the DeepTyphoon prediction model significantly outperforms existing deep learning approaches. It achieves successful typhoon track prediction based on satellite images.

    Citation: Peng Lu, Ao Sun, Mingyu Xu, Zhenhua Wang, Zongsheng Zheng, Yating Xie, Wenjuan Wang. A time series image prediction method combining a CNN and LSTM and its application in typhoon track prediction[J]. Mathematical Biosciences and Engineering, 2022, 19(12): 12260-12278. doi: 10.3934/mbe.2022571

    Related Papers:

    [1] Faisal Mehmood Butt, Lal Hussain, Anzar Mahmood, Kashif Javed Lone . Artificial Intelligence based accurately load forecasting system to forecast short and medium-term load demands. Mathematical Biosciences and Engineering, 2021, 18(1): 400-425. doi: 10.3934/mbe.2021022
    [2] Xingyu Tang, Peijie Zheng, Yuewu Liu, Yuhua Yao, Guohua Huang . LangMoDHS: A deep learning language model for predicting DNase I hypersensitive sites in mouse genome. Mathematical Biosciences and Engineering, 2023, 20(1): 1037-1057. doi: 10.3934/mbe.2023048
    [3] Wen Li, Xuekun Yang, Guowu Yuan, Dan Xu . ABCNet: A comprehensive highway visibility prediction model based on attention, Bi-LSTM and CNN. Mathematical Biosciences and Engineering, 2024, 21(3): 4397-4420. doi: 10.3934/mbe.2024194
    [4] Xin Jing, Jungang Luo, Shangyao Zhang, Na Wei . Runoff forecasting model based on variational mode decomposition and artificial neural networks. Mathematical Biosciences and Engineering, 2022, 19(2): 1633-1648. doi: 10.3934/mbe.2022076
    [5] Wenbo Yang, Wei Liu, Qun Gao . Prediction of dissolved oxygen concentration in aquaculture based on attention mechanism and combined neural network. Mathematical Biosciences and Engineering, 2023, 20(1): 998-1017. doi: 10.3934/mbe.2023046
    [6] Guanghua Fu, Qingjuan Wei, Yongsheng Yang . Bearing fault diagnosis with parallel CNN and LSTM. Mathematical Biosciences and Engineering, 2024, 21(2): 2385-2406. doi: 10.3934/mbe.2024105
    [7] Ying Chang, Lan Wang, Yunmin Zhao, Ming Liu, Jing Zhang . Research on two-class and four-class action recognition based on EEG signals. Mathematical Biosciences and Engineering, 2023, 20(6): 10376-10391. doi: 10.3934/mbe.2023455
    [8] Xintong Wu, Yingyi Geng, Xinhong Wang, Jucheng Zhang, Ling Xia . Continuous extraction of coronary artery centerline from cardiac CTA images using a regression-based method. Mathematical Biosciences and Engineering, 2023, 20(3): 4988-5003. doi: 10.3934/mbe.2023231
    [9] Vasileios E. Papageorgiou, Georgios Petmezas, Pantelis Dogoulis, Maxime Cordy, Nicos Maglaveras . Uncertainty CNNs: A path to enhanced medical image classification performance. Mathematical Biosciences and Engineering, 2025, 22(3): 528-553. doi: 10.3934/mbe.2025020
    [10] Keruo Jiang, Zhen Huang, Xinyan Zhou, Chudong Tong, Minjie Zhu, Heshan Wang . Deep belief improved bidirectional LSTM for multivariate time series forecasting. Mathematical Biosciences and Engineering, 2023, 20(9): 16596-16627. doi: 10.3934/mbe.2023739
  • Typhoon forecasting has always been a vital function of the meteorological department. Accurate typhoon forecasts can provide a priori information for the relevant meteorological departments and help make more scientific decisions to reduce the losses caused by typhoons. However, current mainstream typhoon forecast methods are very challenging and expensive due to the complexity of typhoon motion and the scarcity of ocean observation stations. In this paper, we propose a typhoon track prediction model, DeepTyphoon, which integrates convolutional neural networks and long short-term memory (LSTM). To establish the relationship between the satellite image and the typhoon center, we mark the typhoon center on the satellite image. Then, we use hybrid dilated convolution to extract the cloud features of the typhoon from satellite images and use LSTM to predict these features. Finally, we detect the location of the typhoon according to the predictive markers in the output image. Experiments are conducted using 13, 400 satellite images of time series of the Northwest Pacific from 1980 to 2020 and 8420 satellite images of time series of the Southwest Pacific released by the Japan Meteorological Agency. From the experimentation, the mean average error of the 6-hour typhoon prediction result is 64.17 km, which shows that the DeepTyphoon prediction model significantly outperforms existing deep learning approaches. It achieves successful typhoon track prediction based on satellite images.



    A typhoon is a kind of tropical cyclone. When the sustained wind speed in the center of a tropical cyclone reaches category 12 to 13 (32.7–41.4 meters per second), it is called a typhoon [1]. Typhoons usually form on the sea surface in tropical areas 3–5 latitudes away from the equator, such as the North and South Pacific, the North Atlantic, and the Indian Ocean. Its movement is mainly affected by large-scale weather systems, and eventually dissipates at sea, becomes an extratropical cyclone, or dissipates after arriving on land. A typhoon is sudden and destructive; it will bring strong winds and rainstorms wherever it goes. The storm surge disaster caused by typhoons is one of the most serious natural disasters in the world. Sun et al. [2] found that typhoons will become stronger, larger, and more destructive in the context of global warming. Kossin [3] mentioned in his study that the translation speeds of tropical-cyclones have decreased globally by 10% and decreased by 30% in Asia during 1949–2016. The lower the velocity of a cyclone, the more rain falls on the location, and the higher the risk of flooding and other disasters. Therefore, slower translation speeds will lead to more serious disasters. In order to reduce the economic losses and casualties caused by typhoons, it is urgent to establish accurate and timely prediction methods.

    The motion of a typhoon is highly complex. Unlike many time series problems, this cannot be solved by using the characteristics of periodicity [4]. A meteorological department mainly uses numerical simulation to achieve typhoon track prediction. This requires a lot of typhoon observational data. In order to establish complex atmospheric equations to simulate the motion and variation of the atmosphere, researchers need to fully understand the internal structural changes, key physical processes, and the mutation mechanism of the atmosphere [5]. South Korea, for example, conducts numerical simulations on a Cray XC40 supercomputer with 139, 329 CPUs. Due to its difficulty, numerical simulation requires a lot of calculation time. Even so, numerical simulation cannot obtain the analytical solution, only the numerical solution. Also, expensive hardware means high maintenance costs.

    A deep learning method can use a single graphics processing unit (GPU). The prediction results can be generated within seconds by updating the input data of the prediction model. Compared with using hundreds of thousands of CPUs in a simulation, this method is obviously quicker and cheaper. There have been many investigations into the use of deep learning methods for typhoon track prediction. Song et al. [6] used the spatial location information and meteorological factors of typhoons to train a neural network which is constructed based on convolutional neural networks (CNNs) and the gated recurrent neural network (GRU). They state that the proposed model is better than the traditional prediction methods for typhoon track prediction. Song et al. [7] used BiGRU with an attention mechanism to predict typhoon tracks. Gao et al. [8] trained a LSTM with typhoon track data. In addition to describing the relevant physical quantities of the typhoon process, satellite images can also effectively reflect the spatial information.

    Satellite images can provide the distribution of clouds, and the changes between satellite images can reflect the evolution of weather systems. With the help of satellite images, more accurate weather forecasts can be made. Lee et al. [9] were probably the first to consider using satellite images of tropical cyclones for typhoon track prediction with the help of neural networks. However, the images did not function as input data for the LSTM. Therefore, they extract information such as Dvorak number, maximum wind speed, or spatial location of the cyclone from satellite images and input these into the network for time series prediction. Kovordanyi et al. [10] used satellite images as neural network input data for the first time. They predicted the shape of cloud clusters and the direction of movement of the typhoon via a neural network. Rüttgers et al. [11] marked the typhoon center on the satellite images and used generative adversarial networks (GANs) [12] to generate the satellite images to predict the shape of the cloud cluster and the typhoon track.

    The prediction of typhoon tracks is affected by many factors such as the typhoon model structure, terrain, initial field, and boundary conditions of the model [13]. For example, the movement of a typhoon is affected by the steering flow, which is determined by the environment around the storm. When a typhoon is affected by multiple weather systems at the same time, it is difficult to predict the steering flow [14]. Therefore, a single neural network cannot accurately predict the typhoon track. Therefore, the fusion of multiple neural networks is more suitable for the task of typhoon track prediction rather than a single neural network [7]. Xu et al. [15] modeled the spatiotemporal relationship. They attached the trained generator to the output of the LSTM to generate the cloud images. Hong et al. [16] applied a convolutional LSTM network (ConvLSTM) to typhoon track prediction.

    At present, research based on deep learning methods combined with satellite images mostly focuses on typhoon eye tracking [17] or typhoon intensity prediction [18], and less on typhoon track prediction. We believe that it is necessary to extract the temporal and spatial features of time series satellite images for typhoon path prediction. Therefore, we propose a new typhoon track prediction model based on a CNN and LSTM: DeepTyphoon. The specific contributions of this paper are as follows: We mark the satellite images obtained from the Japan Meteorological Agency to create the data set. According to the characteristics of the images in the data set, we establish a prediction network. The features of the clouds and marked squares in the image are extracted by dilated convolution. In addition, a convolutional block attention module (CBAM) is used to enhance the feature extraction ability of the network from two dimensions: Channel and spatial.

    The attention module can focus on the local information of the network and suppress invalid information, so as to enhance the expression of effective features. The attention mechanism was initially applied to natural language processing and achieved good results. However, there are few studies on the application of attention mechanisms to time series prediction in meteorology. The convolutional block attention module is a new convolutional attention module proposed by Sanghyun et al. in 2018 [19]. The convolution operation extracts information features by mixing cross channel information and spatial information. Therefore, they innovatively proposed the attention mechanism of integrating channel attention and spatial attention and applied a channel attention module and a spatial attention module for the two respective dimensions. The principle of CBAM is shown in Figure 1.

    Figure 1.  Principle of CBAM.

    In this paper, CBAM is applied to the encoder of the neural network to enhance the feature extraction ability of the convolution network, and extracts the features of marked squares and clouds from the two dimensions.

    The long short-term memory network is a special recurrent neural network. It was first proposed by Hochreiter and Schmidhuber [20] and later improved by Grave [21], which can address the problems of gradient disappearance and gradient explosion in the process of long sequence training. Like RNNs, LSTMs can also connect previous information to the current task, such as using the previous video frame information to understand the current frame. Compared with RNNs, LSTMs perform better in learning long-term dependencies.

    The operation of LSTM is mainly realized by the input gate it, forget gate ft, and output gate ot. First, the LSTM uses the forget gate ft to decide which information in ct-1 to delete, and the activation state of the forget gate is determined by the activation function σ(·):

    ft=σ(Wf[ht1,xt]+bf)σ (1)

    where Wf is the weight matrix of the forget gate; bf is the offset term of the forget gate; ft represents the forget gate at time t; xt is the input of network at time t. Second, the LSTM uses the input gate it to determine the information stored in the new memory unit ct:

    it=σ(Wi[ht1,xt]+bi)σ (2)
    ~ct=tanh(Wc[ht1,xt]+bc) (3)

    where Wi is the weight matrix of the input gate; Wc is the weight moment of the cell state; bi is the offset term of the input gate; bc is the bias term of the cell state; it is the input gate at time t; tanh is the hyperbolic tangent function; ~ct indicates the candidate value ct added to the new memory unit status. Then, the LSTM updates the status of the previous memory unit ct-1 to the status ct:

    ct=ft×ct1+it×~ct (4)

    where ct is the input status of the memory unit. Finally, the output gate ot is used to calculate the hidden state ht. The specific formula is as follows:

    ot=σ(Wo[ht1,xt]+bo) (5)
    ht=ottanh(ct) (6)

    where Wo is the weight matrix of the output gate; bo is the output gate offset term; σ is the sigmoid function; represents the multiplication of matrix elements.

    The dilated residual network [22] was proposed to improve the receptive field of convolutions. It maintains the width and height of input features by setting dilation rates. However, the setting of dilation rates leads to the "gridding issue"; when the dilation rate increases, the input sampling will become very sparse, which will be detrimental to the learning of features. Therefore, Wang et al. [23] proposed hybrid dilated convolution (HDC) which alleviates the gridding issue by setting the dilation rates for a series of convolution layers and connecting them. In this paper, the size of the marked square on the satellite images is 3 × 3 pixels. As shown in Figure 2, the three-layer HDC layer with a dilation rate of 1/2/3 can ensure that the input feature of the square is not lost.

    Figure 2.  An example of Hybrid Dilated Convolution. (a)–(c) are continuous convolution layers, and the dilation rates are 1/2/3, respectively.

    The self-built typhoon image data set used in this paper is provided by the National Institute of Information (NII). The satellite image data are taken from Himawari-1~8 series and GOES meteorological satellites. Of these, Himawari-8 can achieve a 10-minute time resolution and 500 m spatial resolution. It has an improved imager observing five bands (one for visible (0.64 μm) and four for infrared (10.4, 12.4, 6.2 and 3.9 μm)). GOES is the geostationary orbit operational satellite series of NOAA in the United States. It also has an improved imager observing five bands (one for visible (0.55–0.75 μm) and four for infrared (10.20–11.20, 11.50–12.50, 6.50–7.00 and 3.80–400 μm)). Satellite images in the recent 40 years are selected as data samples, which are all infrared images (10.4 μm) with a size of 128 × 128 pixels, representing an actual range of 2600 × 2600 km. The middle of the image is the typhoon center, and NII provides the longitude and latitude of the typhoon position corresponding to each image.

    Due to the different life cycles of typhoons, the typhoon processes with typhoon cycle less than 60 hours are screened out through data cleaning in order to ensure that the neural network can obtain sufficient information in the training process. We selected satellite images located in the Northwest Pacific and Southwest Pacific to demonstrate the effectiveness of the model. Of these, 513 typhoon processes located in the Northwest Pacific from 1980 to 2020, that is, 513 typhoon sequences, and a total of 13400 satellite images are selected to construct the training set, verification set, and test set, which are 338, 103 and 72 typhoon sequences, respectively, corresponding to 9110, 2680 and 1610 satellite images. Also, 198 typhoon processes located in the Southwest Pacific from 2000 to 2020, a total of 8420 satellite images, are selected to construct the training set, verification set, and test set, which are 108, 60 and 30 typhoon sequences, respectively, corresponding to 5470, 2110 and 840 satellite images.

    The original infrared image is a single channel gray image, and the range of digital number of satellite image is 0–255. Typhoons are sequential from start to finish. Moreover, the same typhoon has different forms in different lifecycle stages, and the spiral cloud system shown on the satellite cloud map is also different [24]. Figure 3 shows some images in the lifecycle of typhoon Damrey, which lasted for 11 days and 18 hours.

    Figure 3.  Some images in the lifecycle of typhoon Damrey. (a)–(d) respectively represent the ground truth of Damrey at 3:00 (131.4° E, 13.4° N), 9:00 (131.1° E, 13.5° N), 15:00 (131.3° E, 13.6° N), and 21:00 (131.4° E, 13.8° N) on May 7, 2000.

    As described in Section 2.2.1, the typhoon center in the satellite image provided by NII is located in the middle of the image. However, the typhoon center on the satellite images is an absolute position. We need the relative positional information between satellite image sequences to judge the movement of clouds. In order to determine the direction of movement and distance of the typhoon through the satellite image, for each satellite image, we mark the typhoon position at the previous time (for example, six hours ago) on the satellite image at the current time. The marking method is as follows: The spatial coordinate of the typhoon at the current time is (φ, λ). The horizontal offset Dφ and vertical offset Dλ (in km) between the coordinates of the previous time and the coordinates of the current time is calculated according to the Havers formula [25]; the earth radius R is taken from the real coordinate system:

    Dφ=2Rarcsinsin2(φφ2) (7)
    Dλ=2Rarcsincosφcosφsin2(λλ2) (8)

    Then, according to Dφ and Dλ, we calculate the horizontal and vertical offset Pφ and Pλ at the pixel level on the image using the reference system. This determines the position of the typhoon at the previous time on the satellite image at the current time.

    We use black squares to mark satellite images (gray image) for the experiments. Compared with using RGB image (image data marked with red squares) as the input of neural network, the use of gray images as the input of neural network means a shorter training time which is convenient for the experiments to adjust the parameters of the neural network. In the experimentation, it is found that using a red square marked image can eliminate the interference of the gray background, so as to determine the position of the typhoon more accurately. As shown in Figure 4: Figure 4(a) is the original satellite image, Figure 4(b) is the satellite image marked with a black square, and Figure 4(c) is the satellite image marked with a red square. In Figure 4(b), (c), the center of the image is the typhoon center at the current time, and the marked square represents the typhoon center at the previous time. The size of the marked square is 3 × 3 pixels. There are 13400 single channel gray images in the typhoon image data set with black square marks, and 13, 400 RGB three channel images in the typhoon image data set with red square marks. The range of digital number of image data marked with square (black or red) are both 0–255.The size of the dataset does not change as a result of marking. The data set size of the Northwest Pacific is 9110 images for the training set, 2680 images for the verification set, and 1610 images for the test set. For the Southwest Pacific data set, this is 5470, 2110 and 840, respectively.

    Figure 4.  (a) Ground truth. (b) Satellite image marked with a black square. (c) Satellite image marked with a red square.

    In the research of using deep learning to predict typhoon tracks, most researchers use reanalysis data as the input of neural network. For example, the work in [26] converts the reanalysis data into a density map and uses it as the input of a neural network to determine the spatial position of the typhoon according to the predicted density map. However, compared with the thermodynamic diagram transformed according to physical quantity data, satellite images contain more intuitive information. In order to consider the temporal and spatial characteristics of the typhoon process at the same time, we propose the DeepTyphoon prediction model for satellite image training and prediction. DeepTyphoon combines the ability of a CNN to extract advanced semantic features and the utility of LSTM for time series prediction. It takes the marked satellite image as the input, learns the relative position between the square mark and the cloud cluster and the direction of movement of the cloud cluster in the training process, and finally predicts the position of the typhoon according to the square mark on the predicted image.

    The prediction process of DeepTyphoon is shown in Figure 5. The whole neural network model includes a convolution neural network (as encoder), an LSTM, and a deconvolution neural network (as decoder). Taking the satellite images marked with black squares as an example, the prediction process is as follows: First, the marked satellite images (denoted by m) are input into the encoder. We use the hybrid dilated convolution layer to extract the features of the satellite image m. At the same time, we add the channel attention module and the spatial attention module to further improve the feature extraction ability of the coding network. After each image of a satellite image sequence is input into the encoder, the encoder outputs the feature vector of m (denoted by t) as the input of LSTM. The output of the LSTM network t' is the predicted value of t. Finally, we deconvolute t' into an image m' in order to determine the typhoon center according to the prediction result t' of LSTM. Due to the black square mark in the input image m, the final output of the model, the predicted image m', will also have a predicted black square mark.

    Figure 5.  Structure of DeepTyphoon.

    The black square in the predicted image is detected by applying a filter on the predicted image and a convolution with the size of the black square on the filtered image. The pixel with the highest value in the convolved image gives the location of the predicted typhoon center. After identifying the pixel coordinates of the predicted black square, they are converted back to coordinates of latitude (φ) and longitude (λ) through the geographic reference system.

    The experimental environment of this paper is mainly based on the PyTorch framework and built on the Ubuntu operating system. The specific configuration is shown in Table 1 below. In order to fully verify the performance, the experiments in this paper are completed in the same experimental environment. The experimental hyperparameters are set as follows: The batch size is 100, the learning rate is 0.0001, and the number of epochs is 100. The Adam algorithm is used to optimize the network parameters, and the weight attenuation coefficient is set to 0.00002.

    Table 1.  Configuration of training platform.
    Name Configuration
    Operating System Ubuntu 18.04.5
    Framework PyTorch 1.8.0
    Memory 60 Gb
    GPU GeForce RTX 3080Ti
    CUDA CUDA 10.1

     | Show Table
    DownLoad: CSV

    In order to test the influence of the number of convolution layers on the experimental results, we set different convolution layers for testing. First, we set the dilation rates of the three continuous convolution layers to 1/2/3, respectively, and take the three convolution layers as one HDC layer. We use the satellite image marked with the red square as the input of the model, and calculate the mean average error (MAE) of the predicted results when one, two, and three HDC layers are used.

    As shown in Table 2, we achieved the minimum MAE when using two HDC layers. That is to say, when we use six convolution layers and set the dilation rates to 1/2/3/1/2/3 respectively, the prediction result of the neural network is the best.

    Table 2.  MAE of prediction results corresponding to different HDC layers.
    Number of HDC layers MAE (km)
    1 78.22
    2 66.38
    3 69.43

     | Show Table
    DownLoad: CSV

    An important hyperparameter of the model is the number of input image sequences. If the number of input image sequences (denoted by n) is too large, the memory usage will be increased, and the training time will be increased. Moreover, the lifecycle of a typhoon is not too long, and too large a value of n will make most typhoon processes unable to be used as training data. However, too small a value of n will lead to insufficient input information and affect the prediction performance. We use the satellite images marked with red squares as the input. The time interval of the satellite image is 6 hours. After training and testing, for the three cases where n is 5, 10 and 20, the MAE after 6 hours is calculated.

    As shown in Table 3: In the case of n = 5, the change of the typhoon seems to have not been fully captured. In the case of n = 20, due to the reduction of the number of training data sets, the results are not good compared with n = 10. For example, considering typhoon Mitag with 29 sequences, using n = 10 allows to start training from sequence 11, and provides a number of 19 training sequences. However, when n = 20, this only allows to start training from sequence 21, that is, 9 training sequences.

    Table 3.  MAE of prediction results corresponding to different values of n.
    n MAE (km)
    5 96.05
    10 68.49
    20 79.43

     | Show Table
    DownLoad: CSV

    During the experiment, it is found that different convolution kernel sizes will affect the shape of the final generated square. In order to accurately predict the square in the predicted images, the convolution kernel size is tested. We select the satellite image marked with a black square as the input, which can save calculation time.

    Figure 6 shows the prediction results corresponding to different convolution kernels. Figure 6(a) shows the prediction results when the convolution kernel size is 3. It is obvious from the prediction image that a 3 × 3 pixel black square is generated. Figure 6(b), (c) correspond to the shape of the predicted square in the generated image when the convolution kernel size is 2 and 4, respectively. Although it can be seen from the predicted image that the marks are generated, they are blurred to varying degrees and the central position cannot be accurately determined by convolution. Finally, the size of the convolution kernel is set to 3 so that the exact position of the square can be obtained.

    Figure 6.  Prediction results when the convolution kernel is 2, 3 and 4. (a) Prediction results when convolution kernel is 3. (b) Prediction results when convolution kernel is 2. (c) Prediction results when convolution kernel is 4.

    For the typhoon satellite image at the later stage of the typhoon lifecycle, as shown in Figure 7(a), if the black square is used to mark the satellite image, the square in the predicted image almost coincides with the gray background of the image. It is impossible to accurately detect the mark position in the prediction image through convolution. Therefore, we choose the satellite image marked by the red square as the model input so that we can accurately detect the location of the marked pixels. The channel attention module is added after the convolution layer to enhance the weight of the pixel value of the image R channel (red). In the predicted image, only the pixel value on the R channel in the generated image is detected, and the position of the marked square can be obtained accurately. As shown in Figure 7(b), the red square in the figure is located and redrawn after convolution on the R channel.

    Figure 7.  Comparison of satellite image prediction results. (a) Prediction results of satellite images marked with a black square. (b) Prediction results of satellite images marked with a red square.

    The parameters of the neural network are finally determined through experiments, as shown in Table 4. Of these, the number of HDC layers is obtained from Table 2, which is six convolution layers with different dilation rates. In order to decode the prediction vector into an image, the number of deconvolution layers is the same as that of convolution layers. The kernel size is determined according to Figure 6. Based on previous experience, we set the number of hidden layers of LSTM to two.

    Table 4.  Parameters of the neural network.
    Name Parameter
    HDC layers 2
    Deconvolution layers 6
    LSTM layers 2
    Kernel 3 × 3

     | Show Table
    DownLoad: CSV

    In order to evaluate the performance of the prediction model, the root mean square error (RMSE) and the mean absolute error are selected to measure the prediction accuracy of the model.

    The absolute error (E) describes the distance between the predicted coordinates (φpred, λpred) and the real coordinates (φreal, λreal). It is calculated in kilometers (km) by applying the haversine formula, and the earth radius R is taken from the real coordinate position:

    E=2Rarcsinsin2(φpredφreal2)+cosφpredcosφpredsin2(λpredλreal2) (9)

    The mean absolute error formula is as follows, in km, where E represents the absolute error of the typhoon:

    MAE=1mi=1m|E| (10)

    The root mean square error formula is as follows, in km, where E represents the absolute error of the typhoon:

    RMSE=1mmi=1(Ei)2 (11)

    As an example of a predicted image, Figure 8 shows the prediction results of typhoon Gray sequence 11, in which Figure 8(a) is the ground truth and Figure 8(b) is a predicted image.

    Figure 8.  Comparison between the satellite image and predicted image. (a) Ground truth of sequence 11. (b) Predicted image of sequence 11.

    It is obvious that the predicted image does not have the same definition as the real image; in particular, the clouds are blurred compared with the ground truth. However, we focus on accurately predicting the position of the square in the predicted images, so as to accurately predict the spatial position of the typhoon rather than improve the quality of the generated images.

    By calculating the relative position between the square and the center of the predicted image, the spatial position of the predicted typhoon is determined. We compare the position of the predicted typhoon with the position of the real typhoon, and finally we can obtain the prediction accuracy. We select an example in the Northwest Pacific and the Southwest Pacific to show the prediction results of DeepTyphoon. Each typhoon is predicted in chronological order, and the sequence is named by the date and time of the corresponding real image UTC. For example, "1993072718" represents 6 p.m. on July 27, 1993 (UTC).

    Table 5 shows the coordinate prediction results and prediction errors of Typhoon Mitag. As can be seen from the table, this typhoon has taken 29 steps from occurrence to extinction; each time step is 6 hours, giving a total of 174 hours. Take ten points as a sequence as the input of the prediction model, the predicted values of Mitag sequence 11 to sequence 29 can be obtained in order. After calculation, the average error of Typhoon Mitag from sequence 11 to sequence 29 is 55.037 km.

    Table 5.  Absolute error for Typhoon Mitag.
    Sequence Real Predicted E (km)
    11(2007112300) 13.7, 127.2 13.6, 126.8 44.629
    12(2007112306) 13.7, 127.0 14, 127.1 35.062
    13(2007112312) 14.0, 126.7 14.5, 126.6 56.632
    14(2007112318) 14.1, 126.5 14.3, 126.1 48.516
    15(2007112400) 14.3, 126.2 14.8, 126.4 59.619
    16(2007112406) 14.5, 125.9 15, 126.2 64.278
    17(2007112412) 14.8, 125.5 15, 124.9 68.201
    18(2007112418) 15.1, 125.2 15.4, 125.5 46.353
    19(2007112500) 15.5, 124.5 16.1, 124.7 70.065
    20(2007112506) 16.4, 123.9 16.1, 124.1 39.606
    21(2007112512) 16.9, 122.9 16.8, 122.5 43.997
    22(2007112518) 17.3, 121.5 17, 122.1 71.951
    23(2007112600) 18.3, 121.0 18.7, 121.4 61.298
    24(2007112606) 19.0, 120.7 19.4, 120.5 49.187
    25(2007112612) 19.5, 120.7 19.2, 121.1 53.609
    26(2007112618) 20.0, 121.1 20.6, 121.4 73.688
    27(2007112700) 20.1, 121.5 20.4, 122.6 56.719
    28(2007112706) 20.5, 122.8 20.6, 123.6 23.619
    29(2007112712) 20.6, 124.0 20, 124.4 78.685
    MAE (km) 55.037

     | Show Table
    DownLoad: CSV

    As shown in Figure 9, in order to display the forecast results more intuitively, they are marked in the map according to the real track (yellow) and predicted track (blue) of Typhoon Mitag.

    Figure 9.  Prediction results for Typhoon Mitag.

    Table 6 shows the coordinate prediction results and prediction errors of Typhoon Marcia. Typhoon Marcia has taken 22 steps from occurrence to extinction; each time step is 6 hours, giving a total of 132 hours. Take ten points as a sequence as the input of the prediction model, the predicted values of Marcia sequence 11 to sequence 22 can be obtained in order.

    Table 6.  Absolute error for Typhoon Marcia.
    Sequence Real Predicted E (km)
    11(2015021909) –20.5, 150.6 –20.4, 150.3 33.175
    12(2015021912) –20.8, 150.5 –20.6, 150.9 47.177
    13(2015021915) –21.1, 150.5 –21.5, 150.8 54.261
    14(2015021918) –21.5, 150.5 –21.5, 151.1 70.413
    15(2015021921) –22.1, 150.5 –21.9, 150.9 46.853
    16(2015022000) –22.7, 150.5 –22.5, 150.6 24.494
    17(2015022003) –23.1, 150.5 –22.7, 150.2 54.061
    18(2015022006) –23.7, 150.6 –23.3, 150.9 53.983
    19(2015022009) –24.2, 150.9 –24.7, 150.2 90.065
    20(2015022012) –24.5, 151.3 –24.9, 150.8 67.302
    21(2015022015) –24.9, 151.6 –25.1, 151 64.426
    22(2015022018) –25.3, 151.8 –25.4, 151.4 41.705
    MAE (km) 53.992

     | Show Table
    DownLoad: CSV

    As shown in Figure 10, we also display the forecast results for Typhoon Marcia in the map according to Table 6. The real track is represented by yellow points, and the predicted path is represented by blue points.

    Figure 10.  Prediction results for Typhoon Marcia.

    We tested two test sets located in the Northwest Pacific and Southwest Pacific. Table 7 shows the prediction results of DeepTyphoon for the two test sets. The predicted time range is 6 hours. The average absolute error of the prediction results of DeepTyphoon for the two test sets is less than 70 km. In addition, we recorded the calculation time of the DeepTyphoon prediction model. After 14 hours of training, the DeepTyphoon prediction model can process 1000 satellite images, that is, 100 typhoon points, in one minute.

    Table 7.  Comparison of prediction results on data sets in the Northwest Pacific and Southwest Pacific.
    Data set Test set size RMSE (km) MAE (km) Computational time (s)
    Northwest Pacific 1610 72.72 65.39 87.42
    Southwest Pacific 840 69.58 62.41 49.80

     | Show Table
    DownLoad: CSV

    Table 8 shows the prediction results for RNN, GAN, AE–GRU, TraijGRU, and DeepTyphoon. The performance indicators include RMSE and MAE.

    Table 8.  Evaluation results of different methods.
    Model Predicted Time Range RMSE (km) MAE (km)
    RNN[27] 12 h 164.22
    AE-GRU[28] 12 h 138.67
    TraijGRU[29] 6 h 66.6
    DeepTyphoon 6 h 73.96 64.17
    12 h 134.70 129.82

     | Show Table
    DownLoad: CSV

    As shown in Table 8, the evaluation indexes for DeepTyphoon are the best after 6 and 12 hours. Through comparison, it is shown that DeepTyphoon can more accurately predict the location of typhoons using satellite images.

    In order to further verify the validity of the proposed model, we calculated the MAEs in longitude and latitude and compared it with SOTA models. As shown in Table 9, our model has the best prediction result at 6 hours and 12 hours. This verifies the validity of the proposed model.

    Table 9.  MAEs in longitude and latitude at times 6 and 12 hours.
    Model Latitude Longitude
    6 h 1 h 6 h 12 h
    DeepFR[30] 0.9100 1.1765 0.8910 1.2430
    Trj-DMFMG[31] 0.6602 0.9275 0.8747 1.1329
    DeepTyphoon 0.4876 0.8812 0.6645 1.0874

     | Show Table
    DownLoad: CSV

    The experimental results show that marking the typhoon center at the previous time on the satellite images and using CNN and LSTM for feature extraction can effectively learn cloud motions in the satellite image. Also, we can use the deconvolution network to detect the typhoon center at a future time so as to accurately predict the typhoon path in the future.

    However, we expect to get more information from the prediction image obtained by the deconvolution network. Therefore, the focus of future work should be to further optimize the neural network and better extract cloud features, so as to obtain more realistic satellite images through prediction. In this way, we can not only detect the typhoon center in the prediction image, but also determine the intensity of the typhoon. In addition, the prediction accuracy of the typhoon path should be further improved. Then, how to ensure the accuracy of typhoon track prediction and the authenticity of the prediction image at the same time is the difficulty in the future.

    In this paper, we use a deep learning method combined with satellite images to predict typhoon tracks, and propose the DeepTyphoon prediction model. The typhoon center is marked on the satellite image. Combined with the characteristics of CNN and LSTM, we extract the features of the satellite image by HDC and CBAM, and use LSTM to predict the features. After the predicted feature vector is transformed into an image, the position of the typhoon is detected on the predicted image. More than 20000 time series satellite images in two data sets released by the Japan Meteorological Agency in recent years are used for training and prediction. Experimental results on the Northwest Pacific and Southwest Pacific data sets show that our prediction method is better than existing deep learning methods in typhoon track prediction.

    This work was supported by "Shanghai science and technology innovation plan" (Project number: 20dz1203800) and "the Capacity Development for Local College Project" (Project number: 19050502100).

    The authors declare there is no conflict of interests.



    [1] L. S. Chen, Y. H. Ding, Introduction to Typhoons in The Western Pacific, Science Press, 1979.
    [2] Y. Sun, Z. Zhong, T. Li, L. Yi, Y. J. Hu, H. C. Wan, et al., Impact of ocean warming on tropical cyclone size and its destructiveness, Sci. Rep., 7 (2017), 1–10. https://doi.org/10.1038/s41598-017-08533-6 doi: 10.1038/s41598-016-0028-x
    [3] J. P. Kossin, A global slowdown of tropical–cyclone translation speed, Nature, 558 (2018), 104–107. https://doi.org/10.1038/s41586-018-0158-3 doi: 10.1038/s41586-018-0158-3
    [4] H. Kim, U. Yun, B. Vo, J. C. W. Lin, W. Pedrycz, Periodicity-oriented data analytics on time-series data for intelligence system, IEEE Syst. J., 15 (2021), 4958–4969. https://doi.org/10.1109/JSYST.2020.3022640 doi: 10.1109/JSYST.2020.3022640
    [5] T. T. Tang, Q. L. Li, G. X. Li, Y. L. Peng, Research on statistical model of typhoon intensity prediction based on meteorological big data, Integr. Technol., 2 (2016), 73–84. https://doi.org/10.12146/j.issn.2095-3135.201602006 doi: 10.12146/j.issn.2095-3135.201602006
    [6] T. Song, Y. Li, F. Meng, P. F. Xie, D. Y. Xu, A novel deep learning model by BiGRU with attention mechanism for tropical cyclone track prediction in the northwest pacific, J. Appl. Meteorol. Climatol., 61 (2022), 3–12. https://doi.org/10.1175/JAMC-D-20-0291.1 doi: 10.1175/JAMC-D-20-0291.1
    [7] J. Lian, P. P. Dong, Y. P. Zhang, J. G. Pan, K. H. Liu, A novel data–driven tropical cyclone track prediction model based on CNN and GRU with multi-dimensional feature selection, IEEE Access, 8 (2020), 97114–97128. https://doi.org/10.1109/ACCESS.2020.2992083 doi: 10.1109/ACCESS.2020.2992083
    [8] S. Gao, P. Zhao, B. Pan, Y. R. Li, M. Zhou, J. L. Xu, et al., A nowcasting model for the prediction of typhoon tracks based on a long short term memory neural network, Acta Oceanol. Sin., 37 (2018), 8–12. https://doi.org/10.1007/s13131-018-1219-z doi: 10.1007/s13131-018-1219-z
    [9] R. S. T. Lee, J. N. K. Liu, Tropical cyclone identification and tracking system using integrated neural oscillatory elastic graph matching and hybrid RBF network track mining techniques, IEEE Trans. Neural Network, 11 (2000), 680–689. https://doi.org/10.1109/72.846739 doi: 10.1109/72.846739
    [10] R. Kovordányi, C. Roy, Cyclone track forecasting based on satellite images using artificial neural networks, ISPRS J. Photogramm., 64 (2009), 513–521. https://doi.org/10.1016/j.isprsjprs.2009.03.002 doi: 10.1016/j.isprsjprs.2009.03.002
    [11] M. Rüttgers, S. Lee, S. Jeon, D. You, Prediction of a typhoon track using a generative adversarial network and satellite images, Sci. Rep., 9 (2019), 1–15. https://doi.org/10.1038/s41598-019-42339-y doi: 10.1038/s41598-018-37186-2
    [12] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, et al., Generative adversarial networks, Commun. ACM, 63 (2020), 139–144. https://doi.org/10.1145/3422622 doi: 10.1145/3422622
    [13] K. Q. Liu, J. Y. Yuan, W. Z. Chen, Y. P. Chen, Verification of typhoon storm surge forecast model based on adcirc hydrodynamic model, Water Resour. Hydropower Express, 40 (2019), 27–34. https://doi.org/10.15974/j.cnki.slsdkb.2019.04.005 doi: 10.15974/j.cnki.slsdkb.2019.04.005
    [14] L. G. Wu, Z. P. Ni, J. J. Duan, H. J. Zong, Sudden tropical cyclone track changes over the western north pacific: A composite study, Mon. Weather Rev., 141 (2013), 2597–2610. https://doi.org/10.1175/MWR-D-12-00224.1 doi: 10.1175/MWR-D-12-00224.1
    [15] Z. Xu, J. Du, J. J. Wang, C. X. Jiang, Y. Ren, Satellite image prediction relying on GAN and LSTM neural networks, in IEEE International Conference on Communications, (2019), 1–6. https://doi.org/10.1109/ICC.2019.8761462
    [16] S. Kim, J. S. Kang, M. H. Lee, S. Song, Deeptc: Convlstm network for trajectory prediction of tropical cyclone using spatiotemporal atmospheric simulation data, in Conference and Workshop on Neural Information Processing Systems (NIPS), (2018). https://openreview.net/pdf?id=HJlCVoPAF7
    [17] C. Tan, Tclnet: Learning to locate typhoon center using deep neural network, in International Geoscience and Remote Sensing Symposium (IGARSS), (2021), 4600–4603. https://doi.org/10.1109/IGARSS47720.2021.9554524
    [18] R. Pradhan, R. S. Aygun, M. Maskey, R. Ramachandran, D. J. Cecil, Tropical cyclone intensity estimation using a deep convolutional neural network, IEEE Trans. Image Process., 27 (2018), 692–702. https://doi.org/10.1109/TIP.2017.2766358 doi: 10.1109/TIP.2017.2766358
    [19] S. Woo, J. Park, J. Y. Lee, I. S. Kweon, CBAM: Convolutional block attention module, in Proceedings of the European conference on computer vision (ECCV), (2018), 3–19. https://doi.org/10.1007/978-3-030-01234-2-1
    [20] S. Hochreiter, J. Schmidhuber, Long short-term memory, Neural Comput., 9 (1997), 1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735 doi: 10.1162/neco.1997.9.8.1735
    [21] A. Graves, Supervised sequence labelling with recurrent neural networks, in Studies in Computational Intelligence, 385 (2008). https://doi.org/10.1007/978-3-642-24797-2
    [22] F. Yu, V. Koltun, T. Funkhouser, Dilated residual networks, in Proceedings-30th IEEE Conference on Computer Vision and Pattern Recognition, (2017), 636–644. https://doi.org/10.1109/CVPR.2017.75
    [23] P. Q. Wang, P. F. Chen, Y. Yuan, D. Liu, Z. H. Huang, X. D. Hou, et al., Understanding Convolution for Semantic Segmentation, in Proceedings-2018 IEEE Winter Conference on Applications of Computer Vision, WACV 2018, (2018), 1451–1460. https://doi.org/10.1109/WACV.2018.00163
    [24] J. B. Yu, The Research on the Recognition of the Classification of Cloud, Typhoon Segmentation and Locating of the Typhoon's Centre Based on the Meteorological Satellite, Master's thesis, Wuhan University of Technology, 2008. https://doi.org/10.7666/d.y1365665
    [25] H. Mahmoud, N. Akkari, Shortest path calculation: A comparative study for location-based recommender system, in Proceedings-2016 World Symposium on Computer Applications and Research, (2016), 1–5. https://doi.org/10.1109/WSCAR.2016.16
    [26] S. Giffard-Roisin, M. Yang, G. Charpiat, C. K. Bonfanti, B. Kégland, C. Monteleoni, Tropical cyclone track forecasting using fused deep learning from aligned reanalysis data, Front. Big Data, 3 (2020), 1. https://doi.org/10.3389/fdata.2020.00001 doi: 10.3389/fdata.2020.00001
    [27] S. Alemany, J. Beltran, A. Perez, S. Ganzfried, Predicting hurricane trajectories using a recurrent neural network, in Proceedings of the AAAI Conference on Artificial Intelligence, (2019), 468–475. https://doi.org/10.1609/aaai.v33i01.3301468
    [28] J. Lian, P. P. Dong, Y. P. Zhang, J. G. Pan, A novel deep learning approach for tropical cyclone track prediction based on auto–encoder and gated recurrent unit networks, Appl. Sci., 10 (2020), 3965. https://doi.org/10.3390/app10113965 doi: 10.3390/app10113965
    [29] S. F. Tekin, F. Ilhan, Cs559 project report forecasting of tropical cyclone trajectories with deep learning, preprint, arXiv: 1910.10566v2. https://doi.org/10.48550/arXiv.1910.10566
    [30] W. Qin, J. Tang, S. Y. Lao, Deepfr: A trajectory prediction model based on deep feature representation, Inf. Sci., 604 (2022), 226–248. https://doi.org/10.1016/j.ins.2022.05.019 doi: 10.1016/j.ins.2022.05.019
    [31] W. Qin, J. Tang, C. Lu, S. Lao, A typhoon trajectory prediction model based on multimodal and multitask learning, Appl. Soft Comput., 122 (2022), 108804. https://doi.org/10.1016/j.asoc.2022.108804 doi: 10.1016/j.asoc.2022.108804
  • This article has been cited by:

    1. Hui Wen Loh, Chui Ping Ooi, Shu Lih Oh, Prabal Datta Barua, Yi Ren Tan, Filippo Molinari, Sonja March, U. Rajendra Acharya, Daniel Shuen Sheng Fung, Deep neural network technique for automated detection of ADHD and CD using ECG signal, 2023, 241, 01692607, 107775, 10.1016/j.cmpb.2023.107775
    2. Bihao You, Jiahao Qin, Yitao Xu, Yize Liu, Yunfeng Wu, Sijia Pan, 2023, Multi-Modal Lightweight Deep Learning Model for Typhoon Prediction, 9798400716300, 112, 10.1145/3637494.3637514
    3. Zuoming Zou, Shuming Yang, Liang Zhao, Dual-loop control and state prediction analysis of QUAV trajectory tracking based on biological swarm intelligent optimization algorithm, 2024, 14, 2045-2322, 10.1038/s41598-024-69911-5
    4. Bihao You, Jiahao Qin, Yize Liu, Yitao Xu, Sijia Pan, 2023, Attention ResNet-GRU model for Typhoon Prediction, 979-8-3503-0887-7, 135, 10.1109/ICCEIC60201.2023.10426677
    5. Yang Zhang, Liang Zhao, Jyotindra Narayan, Adaptive control and state error prediction of flexible manipulators using radial basis function neural network and dynamic surface control method, 2025, 20, 1932-6203, e0318601, 10.1371/journal.pone.0318601
    6. Qurat ul Ain, Soyiba Jawed, Wolfgang Osten, 2025, Diagnosing ADHD with 1D-CNN: efficient analysis of fMRI data, 9781510688278, 59, 10.1117/12.3056040
  • Reader Comments
  • © 2022 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(4395) PDF downloads(524) Cited by(6)

Figures and Tables

Figures(10)  /  Tables(9)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog