Communication Special Issues

Biomedical approach in autism spectrum disorders—the importance of assessing inflammation

  • Autism spectrum disorders (ASD) are severe heterogeneous neurodevelopmental disorders characterized by dysfunctions in social interaction and communication skills, repetitive and stereotypic verbal and nonverbal behaviors. Published findings have identified widespread changes in the immune systems of children with autism, at the systemic and cellular levels, suggesting that autism may, in fact, be a systemic disorder with connections to abnormal immune responses. Evaluating autism is hindered by a lack of specific biomarkers, making these pathologies difficult to diagnose. A critical priority for the future of ASD management is the identification of potential targets for the development of diagnostic and therapeutic strategies. The purpose of this brief report is to raise awareness regarding the involvement of different inflammatory processes in ASD and the need to assess them as a part of a biomedical evaluation. An extensive analysis of biomarkers relating to inflammation, immune dysfunctions, intestinal dysfunctions and infections will assist in the management of the autistic patient through a more personalized therapy.

    Citation: Tatjana Mijatovic, Dario Siniscalco, Krishnamurthy Subramanian, Eugene Bosmans, Vincent C. Lombardi, Kenny L. De Meirleir. Biomedical approach in autism spectrum disorders—the importance of assessing inflammation[J]. AIMS Molecular Science, 2018, 5(3): 173-182. doi: 10.3934/molsci.2018.3.173

    Related Papers:

    [1] Morteza Fotouhi, Andreas Minne, Henrik Shahgholian, Georg S. Weiss . Remarks on the decay/growth rate of solutions to elliptic free boundary problems of obstacle type. Mathematics in Engineering, 2020, 2(4): 698-708. doi: 10.3934/mine.2020032
    [2] Daniela De Silva, Ovidiu Savin . Uniform density estimates and $ \Gamma $-convergence for the Alt-Phillips functional of negative powers. Mathematics in Engineering, 2023, 5(5): 1-27. doi: 10.3934/mine.2023086
    [3] Catharine W. K. Lo, José Francisco Rodrigues . On an anisotropic fractional Stefan-type problem with Dirichlet boundary conditions. Mathematics in Engineering, 2023, 5(3): 1-38. doi: 10.3934/mine.2023047
    [4] Filippo Gazzola, Gianmarco Sperone . Remarks on radial symmetry and monotonicity for solutions of semilinear higher order elliptic equations. Mathematics in Engineering, 2022, 4(5): 1-24. doi: 10.3934/mine.2022040
    [5] Donatella Danielli, Rohit Jain . Regularity results for a penalized boundary obstacle problem. Mathematics in Engineering, 2021, 3(1): 1-23. doi: 10.3934/mine.2021007
    [6] Aleksandr Dzhugan, Fausto Ferrari . Domain variation solutions for degenerate two phase free boundary problems. Mathematics in Engineering, 2021, 3(6): 1-29. doi: 10.3934/mine.2021043
    [7] Hugo Tavares, Alessandro Zilio . Regularity of all minimizers of a class of spectral partition problems. Mathematics in Engineering, 2021, 3(1): 1-31. doi: 10.3934/mine.2021002
    [8] Tatsuya Miura . Polar tangential angles and free elasticae. Mathematics in Engineering, 2021, 3(4): 1-12. doi: 10.3934/mine.2021034
    [9] Matteo Novaga, Marco Pozzetta . Connected surfaces with boundary minimizing the Willmore energy. Mathematics in Engineering, 2020, 2(3): 527-556. doi: 10.3934/mine.2020024
    [10] Daniela De Silva, Giorgio Tortone . Improvement of flatness for vector valued free boundary problems. Mathematics in Engineering, 2020, 2(4): 598-613. doi: 10.3934/mine.2020027
  • Autism spectrum disorders (ASD) are severe heterogeneous neurodevelopmental disorders characterized by dysfunctions in social interaction and communication skills, repetitive and stereotypic verbal and nonverbal behaviors. Published findings have identified widespread changes in the immune systems of children with autism, at the systemic and cellular levels, suggesting that autism may, in fact, be a systemic disorder with connections to abnormal immune responses. Evaluating autism is hindered by a lack of specific biomarkers, making these pathologies difficult to diagnose. A critical priority for the future of ASD management is the identification of potential targets for the development of diagnostic and therapeutic strategies. The purpose of this brief report is to raise awareness regarding the involvement of different inflammatory processes in ASD and the need to assess them as a part of a biomedical evaluation. An extensive analysis of biomarkers relating to inflammation, immune dysfunctions, intestinal dysfunctions and infections will assist in the management of the autistic patient through a more personalized therapy.


    Volatility forecasting is essential in asset pricing, portfolio allocation and risk management research. Early volatility forecasting was based on economic models. The most famous economic models are the ARCH model [1] and the GARCH model [2], which can capture volatility clustering and heavy-tail features. However, they fail to capture asymmetry, such as leverage effects. The leverage effect is due to the fact that negative returns have a more significant impact on future volatility than positive returns. To overcome this drawback, the exponential GARCH (EGARCH) model [3] and GJR model [4] were proposed. In the following years, new volatility models based on the GARCH model emerged, such as the stochastic volatility model [5] proposed by Hull and White and the realized volatility model [6] offered by Blair et al. They formed a class of GARCH-type volatility models for financial markets.

    The traditional GARCH model has strict constraints and requires the financial time series to satisfy the stationarity condition. It usually assumes conditional variances have a linear relationship with previous errors and previous variances. However, many financial time series show certain nonstationary and nonlinear characteristics in practice. Consequently, some extended model from GARCH is necessary to study the volatility of these time series.

    With the development of computer and big data technologies, machine learning brings new ideas to volatility forecasting [7,8,9,10,11,12,13,14,15,16,17,18,19,20,21]. Especially artificial neural networks(ANN) have shown an outstanding performance. It derives its computational ideas from biological neurons and is now widely used in various fields.

    In financial risk analysis, researchers have utilized neural networks to study the volatility of financial markets. Hamid and Iqbal [22] apply ANN to predict the S&P500 index implied volatility, finding that ANN's forecasting performance surpasses that of the American option pricing model. Livieris proposes an artificial neural network prediction model for forecasting gold prices and trends [23]. Additionally, Dunis and Huang [24] explore neural network regression (NNR), recurrent neural networks, and their collaborative NNR-RNN models for predicting and trading the volatility of daily exchange rates of GBP/USD and USD/JPY, with results indicating that RNNs have the best volatility forecasting performance. Beyond the direct application of neural networks, researchers have investigated a series of mixture models [25,26,27,28,29,30,31] that combine ANNs and GARCH models. Liu et al. [32] introduce a volatility forecasting model based on the recurrent neural network (RNN) and the GARCH model. Experiments reveal that such mixture models enhance the predictive capabilities of traditional GARCH models, capturing normality, skewness and kurtosis of financial volatility more accurately.

    This study employs a mixture model (DeepAR-GMM-GARCH) that combines the deep autoregressive network, the Gaussian mixture model and the GARCH model for probabilistic volatility forecasting. First, we discuss the design of the mixture model; Second, the article presents the model's inference and the training algorithm of the model; Third, we conduct a simulation experiment using artificial data and compare the outcomes with traditional GARCH models, finding that our model yields smaller RMSE and MAE; Last, we investigate the correlation between the square of extreme values and the square of returns for the CSI300 index. The empirical data is partitioned into training and test sets. After training and testing, we analyze the prediction results and observe that our proposed model outperforms other models in both in-sample and out-of-sample analyses.

    The key offerings presented in this article can be summarized as follows: Initially, this article introduces a novel conditional volatility probability prediction model, which addresses the leptokurtic and heavy-tail traits of conventional financial volatility. This model is built upon a deep autoregressive network combined with a Gaussian mixture distribution. Subsequently, we incorporate extreme values into the mixture model via the neural network. It is discovered that the inclusion of extreme values enhances the accuracy of volatility predictions.

    The structure of this paper is as follows: Section 2 outlines the GARCH model and the deep autoregressive network. Section 3 delves into the mixture model, elaborating on inference, prediction and the relevant algorithm. Section 4 encompasses the simulation studies that we propose. Lastly, Section 5 focuses on the empirical analysis of our proposed model.

    Scholars usually believe that stock price or stock index returns are nonlinear, asymmetric, heavy-tailed, returns are generally uncorrelated. Aggregation characterises volatility, which was first found by Engle (1982) and Bollerslev (1986) in ARCH and GARCH models. The GARCH model is defined as follows:

    $ rt=εtht,ht=α0+qi=1αir2ti+pj=1βjhtj,
    $
    (2.1)

    where $ h_{t} $ is the conditional heteroscedastic variance of return series $ r_{t} $.

    Although there are many criteria of GARCH($ p $, $ q $) models to find $ p $ and $ q $, it is sufficient to apply the GARCH($ 1 $, $ 1 $) model to characterize the conditional volatilities.

    The DeepAR model [33], illustrated in Figure 1, is a time series forecasting model that employs a deep autoregressive recurrent network architecture. Distinct from other time series forecasting models, DeepAR generates probabilistic predictions.

    Figure 1.  The network structure of the DeepAR model with $ p $ inputs and $ q $ outputs.

    Consider a time series $ [x_{1}, \dots, x_{t_{0}}, x_{t_{0}+1}, \dots, x_{T}] : = x_{1:T} $. Given its past time series $ [x_{1}, \dots, x_{t_{0}-2}, x_{t_{0}-1}] : = x_{1:t_{0}-1} $, our objective is to predict the future time series $ [x_{t_0}, \dots, x_{t_{0}+(T-1)}, x_{t_{0}+T}] : = x_{t_{0}:t_ {0}+T} $. The DeepAR model constructs the conditional distribution $ P_{\Theta}(x_{t_{0}: T} |x_{1:t_{0}-1}) $ using a latent factor $ z $, which is implemented by a deep recurrent network architecture. This conditional distribution, $ P_{\Theta}(x_{t_{0}: T} |x_{1:t_{0}-1}) $, comprises a product of likelihood factors ($ z $)

    $ PΘ(xt0:Tx1:t01)=Tt=t0PΘ(xtx1:t1)=Tt=t0p(xtθ(zt,Θ)).
    $
    (2.2)

    The likelihood $ p(x_{t}|\theta(z_{t})) $ is a fixed distribution with parameters determined by a function $ \theta(z_{t}, \Theta) $ of the network output $ z_{t} $. As suggested by the model's authors, Gaussian likelihood is appropriate for real-valued data.

    Forecasting the probability of volatility in finance and economics is an important problem. There are mainly two methods to tackle this. First, statistical models such as ARCH and the GARCH models are usually adopted. These models are specifically designed to capture the dynamic nature of volatility over time and help to predict future levels of volatility based on past patterns.

    Another strategy involves using machine learning models, such as neural networks, which can analyze vast amounts of data and uncover patterns that may not be readily apparent to human analysts. A case in point is the DeepAR model is a series-to-series probabilistic forecasting model. The advantages of the DeepAR model are: it makes probabilistic forecasting and allows to introduce additional covariates. Due to these advantages, it can be used to predict financial volatility($ h_{t} $) based on the series $ r_{t}^{2} $. However, the DeepAR model usually assumes that $ p(x_{t}|\theta(z_{t})) $ (given in (2.2)) follows a Gaussian distribution, which may be unreasonable due to the non-negative, leptokurtic and heavy-tail characteristics of traditional financial volatility. To avoid this problem, people use the gaussian mixture distribution to describe the density of $ p(ln(x_{t})|\theta(z_{t})) $, see references [34]. Motivated by the above results, this paper propose an improved mixture model: DeepAR-GMM-GARCH.

    The conditional distribution of $ ln(h_{t}) $ can be expressed as:

    $ P(ln(ht)|r21:t1,x1:t1),
    $
    (3.1)

    where $ h_{t} $ represents the future volatility at time $ t $, $ [r_{1}, ..., r_{t-2}, r_{t-1}]: = r_{1:t-1} $ denotes the past return series during the $ [1:t-1] $ period, and $ {\bf x}_{1:t-1} $ refers to the covariate, which is observable at all times. The past time horizon is represented by $ [1:t-1] $.

    The proposed hybrid model assumes that the conditional density for logarithm of the volatility is given by $ p(ln(h_{t})|r_{1:t-1}^{2}, {\bf x}_{1:t-1}) $, which includes a set of latent factors, denoted as $ z_{t} $. A recurrent neural network with hyperparameters($ \Theta_{1} $), specifically an LSTM, encodes the squared returns $ r_{t}^{2} $, the input features $ {\bf x}_{t} $ and the previous latent factors $ z_{t-1} $, generating the updated latent factors $ z_{t} $. The likelihood $ p(ln(h_{t})|\theta (z_{t})) $ follows a Gaussian mixture distribution with parameters determined by a function $ \theta(z_{t}, \Theta_{2}) $ of the network output $ z_{t} $. The network architecture of the DeepAR-GMM-GARCH model is depicted in Figure 2.

    Figure 2.  The network structure of the DeepAR-GMM-GARCH model with $ m $ inputs and one output.

    Due to the complex interplay between volatility and the factors that influence it, this paper's central model component declares that the volatility $ h_{t} $ of a time series at time $ t $ is derived from the latent variable $ z_{t-1} $ at time $ t-1 $, the square of return $ r_{t-1}^{2} $ and the covariates $ {\bf x}_{t-1} $. $ p(\ln(h_{t})|\theta(z_{t})) $ follows a Gaussian mixture distribution composed of K components. In the empirical analysis, $ {\bf x}_{t-1} $ will be substituted with a vector of extreme values. A nonlinear mapping function $ g $ is used to establish this relationship. The DeepAR-GMM-GARCH model proposed in this paper is as follows.

    $ zt=g(zt1,r2t1,xt1,Θ1),μk,t=log(1+exp(wTk,μzt+bk,μ)),σk,t=log(1+exp(wTk,σzt+bk,σ)),πk,t=log(1+exp(wTk,πzt+bk,π)),P(ln(ht)|zt,Θ2)Ki=1πkN(μk,t,σk,t),rt=εtht,Ki=1πk=1.
    $
    (3.2)

    The model can be viewed as a structure for nonlinear volatility prediction models since the conditional distribution of the perturbation $ \varepsilon_{t} $ in the model can be selected as $ N(0, 1) $ and $ T(0, 1, v) $. Consequently, this gives rise to two distinct models, referred to as DeepAR-GMM-GARCH and DeepAR-GMM-GARCH-t.

    Assuming that the distribution of $ p(ln(h_{t})|\theta(z_{t})) $ follows a Gaussian distribution, model (3.2) will be reduced to a more simple version:

    $ zt=g(zt1,r2t1,xt1,Θ1),μt=log(1+exp(wTμzt+bμ)),σt=log(1+exp(wTσzt+bσ)),P(ln(ht)|zt,Θ2)N(μt,σ2t),rt=εtht.
    $
    (3.3)

    For similarity, we call the above as DeepAR-GARCH model.

    For a given time series, our goal is to estimate the parameters $ \Theta_{1} $ of the LSTM cells and the parameters $ \Theta_{2} $ in function $ \theta $ which applies an affine transformation followed by a softplus activation. We employ a quasi-maximum likelihood estimation method with the likelihood function: $ \Theta = argmax\sum_{i}logp(\widetilde{h_{i}}|\Theta_{1}, \Theta_{2}) $. Inferring from this likelihood function necessitates taking into account the latent variable $ z_{t} $.

    The flowchart for the training algorithm of the model is shown below. First, we utilize the BIC criterion to identify the number of classifications, $ K $, for all samples. Each data point is assigned a label from 1 to $ K $, and each cluster $ k $ has its mean vector and covariance matrix. Based on these findings, we establish the initial $ \pi_{k} $ as the proportion of data points labelled as $ k $ and set the initial mean vector $ \mu_{k} $ and covariance matrix $ \sum_{k} $ to the mean vector and covariance matrix within cluster $ k $. As a result, we obtain the parameter values $ (\theta = {\widetilde \pi_{k, 0}, \widetilde \mu_{k, 0}, \widetilde \sigma_{k, 0}^{2}}) $ from the initial cluster and use them to pre-train the DeepAR-GMM-GARCH model. This approach allows our model to converge quickly. Next, we partition the training sample data into multiple batches, select one sample from a batch and use the sample $ (r_{t0-m}^{2}, \dots, r_{t0-1}^{2}) $ as the input for the DeepAR-GMM-GARCH model. The model calculates a set of $ \widetilde \pi_{k, t}, \widetilde \mu_{k, t}, \widetilde \sigma_{k, t}^{2} $, after which we sample from this Gaussian mixture model, compute the loss, and update the parameters through gradient descent. Since direct differentiation of the sampling is infeasible, we apply the reparameterization trick to adjust the model's parameters. We continue this training process until the end of the training cycle. Last, we input the training set sample into our trained model for prediction evaluation. The model sequentially calculates the parameters for both the latent variable and the mixed Gaussian model and then proceeds with sampling. Ultimately, we provide the prediction results derived from the sampling outcomes.

    The training algorithm is shown in the Algorithm $ 1 $.

    Algorithm 1 Training Procedure for DeepAR-GMM-GARCH Mixture Model
    1: for each batch do
    2:  for each $ t \in [t0-m, t0-1] $ do
    3:    if $ t $ is $ t0-m $ then
    4:      $ z_{t-1} = 0 $
    5:    else {$ t $ is not $ t0-m $}
    6:      $ z_{t} = g(z_{i-1}, r_{i-1}^{2}, x_{t-1}, \Theta_{1}) $
    7:    end if
    8:    for each $ k \in [1, K] $ do
    9:      $ \widetilde \mu_{k, t} = log(1+exp(w_{k, \mu}^{T}z_{t}+b_{k, \mu})) $
    10:      $ \widetilde \sigma_{k, t}^{2} = log(1+exp(w_{k, \sigma}^{T}z_{t}+b_{k, \sigma})) $
    11:      $ \widetilde \pi_{k, t} = log(1+exp(w_{k, \pi}^{T}z_{t}+b_{k, \pi})) $
    12:    end for
    13:    sample $ ln(\widetilde{h_{t}}) \sim GMM(\widetilde \pi_{k, t}, \widetilde \mu_{k, t}, \widetilde \sigma_{k, t}^{2}) $
    14:  end for
    15:  compute Loss, model parameters $ \Theta_{1} $, $ \Theta_{2} $ adjust using gradient descent method.
    16: end for

    During the training process, the definition of the loss function determines the prediction quality of the model. We use the the average negative loglikelihood function as the loss function $ Loss = L_{h} $. In the GARCH model, we usually assume that $ \varepsilon_{t} $ obeys a Gaussian distribution or a Student distribution, two loss functions are as follows:

    (1) When $ \varepsilon_{t} \sim N(0, 1) $, the loss function is:

    $ Lh=1NNt=1[log(˜htr2t2˜ht)].
    $
    (3.4)

    (2) When $ \varepsilon_{t} \sim t(0, 1, v) $, the loss function is:

    $ Lh=1NNt=1[log(˜ht)+12(v+1)log(1+r2t˜ht(v2))].
    $
    (3.5)

    To calculate the above loss functions, we need to get samples for $ \widetilde{h_{t}} $ based on algorithm1 given in Section 3.2.1. In practice, if $ \varepsilon_{t} $ follows other distributions, based on the idea of QMLE, we still can use the loss function given in (3.4) see Liu and So, 2020.

    Experiments are carried out on volatility inference using simulated time series. These series exhibit flexibility, with both volatility and mixing coefficients changing over time, as detailed below:

    $ rt=εtht,εtN(0,1),p(ht|Ft1)=η1,tϕ(μt,σ21,t)+η2,tϕ(μt,σ22,t),μt=a0+a1ht1,σ21,t=α01+α11r2t1+β1σ21,t1,σ22,t=α02+α12r2t1+β2σ22,t1,π1,t=c0+c1ht1,η1,t=exp(π1,t)/(1+exp(π1,t)),
    $
    (4.1)

    where $ F_{t-1} $ denotes the information set through time $ t-1 $ and $ \phi $ is the Gaussian density function. $ \eta_{1, t} $ and $ \eta_{2, t} $ are mixing coefficients of two gaussian distribution and satisfy: $ \eta_{2, t} = (1-\eta_{1, t}) $. When generating the simulation data, we set: $ \alpha_{01} = 0. 01 $, $ \alpha_{11} = 0. 1 $, $ \beta_{1} = 0. 15 $, $ \alpha_{02} = 0. 04 $, $ \alpha_{12} = 0. 15 $, $ \beta_{2} = 0. 82 $, $ c_{0} = 0. 02, c_{1} = 0. 90, a_{0} = 0. 02, a_{1} = 0. 6 $. The time series has initial values:$ r_{0} = 0. 1, \sigma_{0}^{2} = 0 $, $ h_{0} = 0 $. The sample sizes of T = 500, 1000 and 1500 are considered, and the replication time is 1000.

    For the series simulated from (4.1), we apply three models to forecast their volatility, namely, the GARCH model with $ \epsilon_{t} \sim N(0, 1) $(GARCH-n), the GARCH model with$ \epsilon_{t} \sim T(0, 1, v) $(GARCH-t) and the DeepAR-GMM-GARCH model. Using MCMC sampling method, the degree of freedom for the Student-t distribution is determined to be 6. For the DeepAR-GMM-GARCH model, we set a recurrent neural network with three LSTM layers and 24 hidden nodes. For the input nodes m in Figure 2, we use the Grid Search Algorithm to find their optimal values. We use BIC rule to choose K based on the Mclust package [35]. Our model's hyperparameters are trained by the software Optuna, a commonly applied automatic hyperparameters optimization software.

    Table 1 informs the three volatility forecasting models' in-sample errors (RMSE and MAE). The GARCH-n and GARCH-t show similar performance, and our DeepAR-GMM-GARCH model is outstanding; All the models get decreased RMSE as sample size increase. These results imply the proposed estimation can be asymptotically convergent. Table 2 reports the average out-of-sample errors of the three volatility forecasting models. Similar to the in-sample results, our DeepAR-GMM-GARCH model is superior to the GARCH-n and GARCH-t models.

    Table 1.  The average in-sample errors of the GARCH-n, the GARCH-t and the DeepAR-GMM-GARCH models.
    sample size Modle RMSE MAE
    $ T=500 $ $ \text{GARCH-n} $ $ 0. 1364 $ $ 0. 0628 $
    $ \text{GARCH-t} $ $ 0. 1211 $ $ 0. 0591 $
    $ \text{DeepAR-GMM-GARCH} $ $ \underline{0. 1068} $ $ \underline{0. 0548} $
    $ T=1000 $ $ \text{GARCH-n} $ $ 0. 0621 $ $ 0. 0437 $
    $ \text{GARCH-t} $ $ 0. 0578 $ $ 0. 0419 $
    $ \text{DeepAR-GMM-GARCH} $ $ \underline{0. 0398} $ $ \underline{0. 0331} $
    $ T=1500 $ $ \text{GARCH-n} $ $ 0. 0604 $ $ 0. 0428 $
    $ \text{GARCH-t} $ $ 0. 0652 $ $ 0. 0401 $
    $ \text{DeepAR-GMM-GARCH} $ $ \underline{0. 0300} $ $ \underline{0. 0325} $
    Note: Number of replications = 1000.

     | Show Table
    DownLoad: CSV
    Table 2.  The average out-of-sample errors of the GARCH-n, the GARCH-t and the DeepAR-GMM-GARCH models.
    sample size Modle RMSE MAE
    $ T=500 $ $ \text{GARCH-n} $ $ 0. 3564 $ $ 0. 3028 $
    $ \text{GARCH-t} $ $ 0. 3271 $ $ 0. 3091 $
    $ \text{DeepAR-GMM-GARCH} $ $ \underline{0. 2761} $ $ \underline{0. 2311} $
    $ T=1000 $ $ \text{GARCH-n} $ $ 0. 2619 $ $ 0. 2117 $
    $ \text{GARCH-t} $ $ 0. 2318 $ $ 0. 2033 $
    $ \text{DeepAR-GMM-GARCH} $ $ \underline{0. 2091} $ $ \underline{0. 1834} $
    $ T=1500 $ $ \text{GARCH-n} $ $ 0. 2241 $ $ 0. 2179 $
    $ \text{GARCH-t} $ $ 0. 2213 $ $ 0. 1971 $
    $ \text{DeepAR-GMM-GARCH} $ $ \underline{0. 1911} $ $ \underline{0. 1722} $
    Note: Number of replications = 1000.

     | Show Table
    DownLoad: CSV

    Comprehensive stock index represents the average of the economic performance of the whole financial market. In this section, we study the China Shanghai Shenzhen index(CSI 300 index) daily OHLC data. The OHLC data contains daily high, low, open and close prices. Scholars have pointed out that combining the open, high, low and close prices can obtain more effective volatility estimates. Hence, We also intruduce OHLC data to our mixture model.

    The data of the CSI 300 index studied in this paper are from January 4, 2010, to December 30, 2021, with a total of 2916 trading days. Let $ r_{t} $ be the returns of the corresponding series, which are calculated using the closing price $ C_{t} $ series of the CSI300 index.

    $ rt=100logCt+1Ct.
    $
    (5.1)

    The time series of $ r_{t} $ and $ r_{t}^{2} $ are ploted in Figure 3. We can find that $ r_{t}^{2} $ displays a significant volatility clustering characteristic, and the amplitude of volatility is gradually decreasing.

    Figure 3.  The time series of $ r_{t} $ and $ r_{t}^{2} $ for the CSI 300 index, as shown in (5.1), spans from January 4, 2010, to December 30, 2021.

    The collected data is divided into training data and test data. Table 3 describes the statistical characteristics of training and test data, respectively. The mean of $ r_{t} $ is relatively small, only 0.007465. The standard deviation is 2.251006, which displays that the degree of variation is large. The skewness is less than 0, and the kurtosis is greater than 3, indicating that the sequence is of left deviation and heavy tail. The test data has the same characteristics as the training data, such as data dispersion, large volatility, left deviation and higher protrusion than the normal distribution. This may imply that the normal distribution may not be suitable for our data, and other heavy-tailed distributions, such as t distribution or mixed normal distribution, could be more suitable.

    Table 3.  Descriptive statistics for the training and test sets of CSI 300 returns.
    Data Set Period Mean Std. Skew. Kurt.
    $ \text{training data} $ $ 04 / 01 / 2010 $ to $ 29 / 12 / 2017 $ $ 0.007465 $ $ 2.251006 $ $ -0.755580 $ $ 5.136707 $
    $ \text{test data} $ $ 02 / 01 / 2018 $ to $ 30 / 12 / 2021 $ $ 0.019125 $ $ 1.717302 $ $ -0.427819 $ $ 3.248347 $

     | Show Table
    DownLoad: CSV

    Besides the close price($ C_{t} $), we also introduce the high price($ H_{t} $), the open price($ O_{t} $) and the low price($ L_{t} $). Define $ u_{t} = (H_{t}-O_{t})^{2} $, $ d_{t} = (L_{t}-O_{t})^{2} $, $ c_{t} = (C_{t}-O_{t})^{2} $.

    The correlation matrix below (5.2) shows the correlation coefficients between $ u_{t} $, $ d_{t} $, $ c_{t} $ and $ r_{t}^{2} $. It can be found that there is large correlation coefficient between the pair ($ u_{t} $, $ r_{t}^{2} $), ($ d_{t} $, $ r_{t}^{2} $) and ($ c_{t} $, $ r_{t}^{2} $). From a common sense, large values for $ u_{t} $, $ d_{t} $ and $ c_{t} $, usually means large volatility($ r_{t}^{2} $). However, classical volatility models do not take such extreme values into account. Consequently, it is reasonable to use neural network together with $ u_{t} $, $ d_{t} $ and $ c_{t} $ to forecast volatility, because neural network can introduce additional covariates and capture the complex relation between different covariates.

    $ [r2tutdtctr2t1.0000.3940.4750.716ut0.3941.0000.0120.556dt0.4750.0121.0000.690ct0.7160.5560.6901.000]
    $
    (5.2)

    This paper uses four evaluation indicators to measure the predictive performance of the model, they are: NMAE, HR, linear correlation coefficient and rank correlation coefficient, which is defined as follows:

    $ NMAE=Nt=1|r2t+1˜ht+1|Nt=1|r2t+1r2t|,
    $
    (5.3)
    $ HR=1NNt=1θt,θt={1:(˜ht+1r2t)(r2t+1r2t)00: else ,
    $
    (5.4)

    where N represents the number of predicted samples. Both NMAE and HR values range between 0 and 1. The smaller the values of these two indicators, the better the model's performance.

    Scholars usually use high-frequency data volatility estimates as a proxy for actual volatility to evaluate forecasting models. We also use realized volatility($ \sigma_{RV, t}^{2} $) as a proxy for actual volatility, calculated by summing up the squares of intra-day returns every 5 minutes.

    $ σ2RV,t=48i=1[logrt,ilogrt,i1]2.
    $
    (5.5)

    We focus on the out-of-sample predictive performance of the models, the correlation between realized volatilities $ \sigma_{RV, t+1}^{2} $ and predicted volatilities $ \widetilde{h}_{t+1} $ is measured only on the test set. We calculated Pearson's coefficient

    $ r=Ni=1(σ2RV,t+1σ2RV)(˜ht+1˜h)Ni=1(σ2RV,t+1σ2RV)2Ni=1(˜ht+1˜h)2,
    $
    (5.6)

    where $ \sigma_{RV}^{2} $ and $ \widetilde{h} $ denote the respective mean values, and Spearman's rank order correlation coefficient $ r_{s} $. $ r_{s} $ is also calculated using Eq (5.6). However, the actual volatilities are replaced by their ranks. Spearman's rank order correlation coefficient is considered more robust than Pearson's coefficient. $ r $ and $ r_{s} $ are both between $ -1 $ and $ 1 $. A value of $ r $($ r_{s} $) around $ 0 $ means that the realized volatilities and predicted volatilities are uncorrelated.

    Simulation experiments demonstrate that our proposed model exhibits greater prediction accuracy than the GARCH model. In this section, to highlight the advantages of our model, we compare it to the classic GARCH model, ANN-GARCH (an existing neural network GARCH model) and the DeepAR-GARCH model using empirical data.

    Among the four models, the GARCH and ANN-GARCH models predict conditional volatility, whereas the DeepAR-GARCH and DeepAR-GMM-GARCH models provide probabilistic forecasts for conditional volatility. To facilitate comparison, we will calculate the mean and quantiles of the probability density function for conditional volatility derived from the DeepAR-GARCH and DeepAR-GMM-GARCH models.

    The estimated parameters of the GARCH models(GARCH-n and GARCH-t) are summarized in Tables 4 and 5. For the GARCH-t model, the degrees of freedom parameter $ v $ is estimated at around $ 6 $. The GARCH-n and GARCH-t models are all nearly estimated with higher values of $ \beta_{1} $ and lower values of $ \alpha_{1} $. The sum of $ \alpha_{1} $ and $ \beta_{1} $ is almost 1, Which implies the sequence may be non-stationary. Therefore, our model without the stationary constraint is more suitable. The ANN-GARCH model employs a three-layer ANN structure, featuring two input nodes, 24 nodes for the hidden layer and a single output node. Likewise, the DeepAR-GARCH and DeepAR-GMM-GARCH models also have a three-layer design, consisting of 14 input nodes, 24 nodes for the hidden layer and an output layer with two output nodes and five output nodes.

    Table 4.  Parameter estimation of GARCH-n model.
    Data Set $a_{0}$ $ \alpha_{1} $ $ \beta_{1} $
    $ CSI300 $ $ 1.8365e-04 $ $ 0.1000 $ $ 0.8800 $

     | Show Table
    DownLoad: CSV
    Table 5.  Parameter estimation of GARCH-t model.
    Data Set $a_{0}$ $ \alpha_{1} $ $ \beta_{1} $ $ \nu $
    $ CSI300 $ $ 1.8103e-04 $ $ 0.1000 $ $ 0.8400 $ $ 6.4625 $

     | Show Table
    DownLoad: CSV

    Table 6 lists the performance of five volatility prediction models. In the in-sample study, the DeepAR-GMM-GARCH model has the smallest HR and loss, and the DeepAR-GARCH model has the smallest NMAE. The volatility prediction performance of the DeepAR-GMM-GARCH model is better than the traditional GARCH and DeepAR models.

    Table 6.  In-sample forecasting results for the GARCH-n, GARCH-t, DeepAR-GARCH and DeepAR-GMM-GARCH models.
    Data Set Model Loss NMAE HR
    In-sample $ \text{GARCH-n} $ $ 1. 401 $ $ 0. 763 $ $ 0. 704 $
    $ \text{GARCH-t} $ $ 1. 331 $ $ 0. 761 $ $ 0. 637 $
    $ \text{ANN-GARCH} $ $ 1. 603 $ $ 0. 827 $ $ 0. 690 $
    $ \text{DeepAR-GARCH} $ $ 1. 541 $ $ {\bf 0. 748} $ $ 0. 717 $
    $ \text{DeepAR-GMM-GARCH} $ $ {\bf 1. 311} $ $ 0. 751 $ $ {\bf 0.630} $

     | Show Table
    DownLoad: CSV

    In Figure 4, we display a portion of the forecasting results from various models and compare them with $ r_{t}^{2} $. As shown in (a), the forecasting results from the GARCH models differ from $ r_{t}^{2} $. The GARCH models fail to capture significant changes in $ r_{t}^{2} $. From (b), (c), it is evident that the neural network models capture the trend of $ r_{t}^{2} $ and more accurately predict large fluctuations, with the DeepAR-GMM-GARCH model demonstrating the best performance. In (d), we observe that the estimated 90% quantiles from the DeepAR-GMM-GARCH model appear to be closer to the observations ($ r_{t}^{2} $).

    Figure 4.  Subplots $ (a), (b) $, $ (c) $ and $ (d) $ are the plots of the in-sample volatility forecasting results of the classic GARCH, ANN-GARCH, DeepAR-GARCH and DeepAR-GMM-GARCH models. (a) is the comparison of volatility forecasting results from the GARCH models. (b) is the comparison of volatility forecasting results between the ANN-GARCH model and $ r_{t}^{2} $. (c) is the comparison of probabilistic forecasting models between the DeepAR-GARCH model and DeepAR-GMM-GARCH model. (d) is the comparison of 90% quantile forecasting results between the DeepAR-GARCH model and DeepAR-GMM-GARCH model.

    To sum up, from the estimation results of Table 6 and the plots in Figure 4, it is shown that introducing the extreme values($ u_{t} $, $ d_{t} $, $ c_{t} $) can help to improve the forecasting accuracy of the mixture volatility models. Hence the proposed approach is of particular practical value.

    In the out-of-sample analysis, as discussed in Section 5.1, the test set comprises the time series of 972 trading days subsequent to the respective training set.

    Table 7 presents the performance of the models using common error measures (loss function, NMAE and HR). The DeepAR-GARCH model attains a lower loss function value compared to the neural network models on the test set. The neural network models display lower NMAE and HR values than the GARCH model on the training set. The DeepAR-GMM-GARCH models exhibit the lowest NMAE and HR values on the test data set.

    Table 7.  Out-of-sample forecasting results for the GARCH-n, GARCH-t, ANN-GARCH, DeepAR-GARCH and DeepAR-GMM-GARCH models.
    Data Set Model Loss NMAE HR
    Out-sample $ \text{GARCH-n} $ $ 2. 320 $ $ 0. 917 $ $ 0. 868 $
    $ \text{GARCH-t} $ $ 2. 008 $ $ 0. 915 $ $ 0. 859 $
    $ \text{ANN-GARCH} $ $ 2. 517 $ $ 0. 903 $ $ 0. 783 $
    $ \text{DeepAR-GARCH} $ $ {\bf 1. 916} $ $ 0. 929 $ $ 0. 801 $
    $ \text{DeepAR-GMM-GARCH} $ $ 2. 100 $ $ {\bf 0.790} $ $ {\bf 0.722} $

     | Show Table
    DownLoad: CSV

    In Figure 5, we plot part of the forecasting results from the five models mentioned with out-sample data set and compare it with $ r_{t}^{2} $. It can be seen from (a) that the GARCH models do not capture significant changes of $ r_{t}^{2} $, the same as the in-sample results. For (b), (c), The neural network models capture most of the fluctuations of $ r_{t}^{2} $ well, the DeepAR-GMM-GARCH model performing the best.

    Figure 5.  Subplots $ (a), (b) $, $ (c) $ and $ (d) $ are the plots of the out-sample volatility forecasting results of the classic GARCH, ANN-GARCH, DeepAR-GARCH and DeepAR-GMM-GARCH models. (a) is the comparison of volatility forecasting results from the GARCH models. (b) is the comparison of volatility forecasting results between the ANN-GARCH model and realized volatility($ r_{t}^{2} $). (c) is the comparison of probabilistic forecasting models between the DeepAR-GARCH model and DeepAR-GMM-GARCH model. (d) is the comparison of 90% quantile forecasting results between the DeepAR-GARCH model and DeepAR-GMM-GARCH model.

    From (d), we could find that The estimated 90% quantiles from the DeepAR-GMM-GARCH model seem to be more closely aligned with the observations ($ r_{t}^{2} $).

    Section 5.2 mentions that the linear correlation $ r $ and the rank correlation $ r_{s} $ are two measures for comparing realized volatilities. The linear correlation r and the rank correlation $ r_{s} $ between predicted and realized volatilities of the test set are reported in Table 8. On average, the DeepAR-GMM-GARCH model shows the best performance of all models. It obtains the highest rank correlation on the test set. Rank correlation is more robust than linear correlation since it detects correlations nonparametrically.

    Table 8.  A comparison of linear correlation (r) and rank correlation ($ r_s $) between the realized volatility and volatility predicted by GARCH-n, GARCH-t, DeepAR-GARCH, ANN-GARCH and DeepAR-GMM-GARCH models for test data set on the CSI300 index. The best model (the highest correlation) is underlined.
    Out-sample GARCH-n GARCH-t ANN-GARCH DeepAR-GARCH DeepAR-GMM-GARCH
    $ r $ $ 0. 381 $ $ 0.420 $ $ 0.490 $ $ 0. 473 $ $ \underline{0. 504} $
    $ r_s $ $ 0. 477 $ $ 0.502 $ $ 0.500 $ $ 0. 516 $ $ \underline{0. 527} $

     | Show Table
    DownLoad: CSV

    This paper studies a mixture volatility forecasting model based on the autoregressive neural work and the GARCH model to obtain more precise forecasting for the conditional volatility model. The inference, loss functions and training algorithm of the mixture model are given. The simulation results show that our model performs better with less error than the classic GARCH models. The empirical study based on the CSI300 index shows that our model can significantly improve the forecasting accuracy with extreme values compared to the usual models.

    Our research findings can offer valuable insights into the prediction of volatility uncertainty. In future studies, our model can be employed for various high-frequency volatility analysis, where it is anticipated to exhibit enhanced performance.

    This work is partially supported by Guangdong Basic and Applied Basic Research Foundation (2022A1515010046) and Funding by Science and Technology Projects in Guangzhou (SL2022A03J00654).

    The authors declare no conflict of interest.

    [1] Christensen DL, Bilder DA, Zahorodny W, et al. (2016) Prevalence and Characteristics of Autism Spectrum Disorder Among Children Aged 8 Years-Autism and Developmental Disabilities Monitoring Network, 11 Sites, United States, 2012. MMWR Surveill Summ 65: 1–23.
    [2] Zablotsky B, Bramlett M, Blumberg SJ (2015) Factors associated with parental ratings of condition severity for children with autism spectrum disorder. Dis Health J 8: 626–634. doi: 10.1016/j.dhjo.2015.03.006
    [3] Hertz-Picciotto I, Delwiche L (2009) The rise in autism and the role of age at diagnosis. Epidemiology 20: 84–90. doi: 10.1097/EDE.0b013e3181902d15
    [4] First MB (2013) Diagnostic and statistical manual of mental disorders, 5th edition, and clinical utility. J Nerv Ment Dis 201: 727–729. doi: 10.1097/NMD.0b013e3182a2168a
    [5] Siniscalco D, Cirillo A, Bradstreet JJ, et al. (2013) Epigenetic findings in autism: New perspectives for therapy. Int J Environ Res Public Health 10: 4261–4273. doi: 10.3390/ijerph10094261
    [6] Siniscalco D (2013) Current findings and research prospective in autism spectrum disorders. Autism-Open Access S2: e001.
    [7] Siniscalco D (2014) The searching for autism biomarkers: A commentary on: A new methodology of viewing extra-axial fluid and cortical abnormalities in children with autism via transcranial ultrasonography. Front Hum Neurosci 8: 240.
    [8] Ashwood P, Krakowiak P, Hertz-Picciotto I, et al. (2011) Altered T cell responses in children with autism. Brain Behav Immun 25: 840–849. doi: 10.1016/j.bbi.2010.09.002
    [9] Suzuki K, Matsuzaki H, Iwata K, et al. (2011) Plasma cytokine profiles in subjects with high-functioning autism spectrum disorders. PLoS One 6: e20470. doi: 10.1371/journal.pone.0020470
    [10] Okada K, Hashimoto K, Iwata Y, et al. (2007) Decreased serum levels of transforming growth factor-beta1 in patients with autism. Prog Neuro-Psychopharmacol Biol Psychiatry 31: 187–190. doi: 10.1016/j.pnpbp.2006.08.020
    [11] Ashwood P, Enstrom A, Krakowiak P, et al. (2008) Decreased transforming growth factor beta1 in autism: A potential link between immune dysregulation and impairment in clinical behavioral outcomes. J Neuroimmunol 204: 149–153. doi: 10.1016/j.jneuroim.2008.07.006
    [12] Al-Ayadhi LY, Mostafa GA (2012) Elevated serum levels of interleukin-17A in children with autism. J Neuroinflammation 9: 158.
    [13] Inga Jacome MC, Morales Chacon LM, Vera CH, et al. (2016) Peripheral inflammatory markers contributing to comorbidities in autism. Behav Sci 6: 29.
    [14] Siniscalco D, Schultz S, Brigida AL, et al. (2018) Inflammation and neuro-immune dysregulations in autism spectrum disorders. Pharmaceuticals 11: E56. doi: 10.3390/ph11020056
    [15] El-Ansary A, Al-Ayadhi L (2012) Lipid mediators in plasma of autism spectrum disorders. Lipids Health Dis 11: 160. doi: 10.1186/1476-511X-11-160
    [16] Skorupka C, Amet L (2017) Autisme, un nouveau regard: Causes et solutions (French), French: Editions Mosaïque-Santé.
    [17] Rossignol DA, Frye RE (2012) Mitochondrial dysfunction in autism spectrum disorders: A systematic review and meta-analysis. Mol Psychiatry 17: 290–314. doi: 10.1038/mp.2010.136
    [18] Horvath K, Perman JA (2002) Autism and gastrointestinal symptoms. Curr Gastroenterol Rep 4: 251–258. doi: 10.1007/s11894-002-0071-6
    [19] Adams JB, Johansen LJ, Powell LD, et al. (2011) Gastrointestinal flora and gastrointestinal status in children with autism-comparisons to typical children and correlation with autism severity. BMC Gastroenterol 11: 22. doi: 10.1186/1471-230X-11-22
    [20] Finegold SM, Dowd SE, Gontcharova V, et al. (2010) Pyrosequencing study of fecal microflora of autistic and control children. Anaerobe 16: 444–453. doi: 10.1016/j.anaerobe.2010.06.008
    [21] Iovene MR, Bombace F, Maresca R, et al. (2017) Intestinal dysbiosis and yeast isolation in stool of subjects with autism spectrum disorders. Mycopathologia 182: 349–363. doi: 10.1007/s11046-016-0068-6
    [22] Critchfield JW, van Hemert S, Ash M, et al. (2011) The potential role of probiotics in the management of childhood autism spectrum disorders. Gastroenterol Res Pract 2011: 161358.
    [23] Siniscalco D, Antonucci N (2013) Involvement of dietary bioactive proteins and peptides in autism spectrum disorders. Curr Protein Pept Sci 14: 674–679.
    [24] Bransfield RC, Wulfman JS, Harvey WT, et al. (2008) The association between tick-borne infections, Lyme borreliosis and autism spectrum disorders. Med Hypotheses 70: 967–974. doi: 10.1016/j.mehy.2007.09.006
    [25] Kuhn M, Grave S, Bransfield R, et al. (2012) Long term antibiotic therapy may be an effective treatment for children co-morbid with Lyme disease and autism spectrum disorder. Med Hypotheses 78: 606–615. doi: 10.1016/j.mehy.2012.01.037
    [26] Vanuytsel T, Vermeire S, Cleynen I (2013) The role of Haptoglobin and its related protein, Zonulin, in inflammatory bowel disease. Tissue Barriers 1: e27321. doi: 10.4161/tisb.27321
    [27] Uribarri J, Oh MS, Carroll HJ (1998) D-lactic acidosis. A review of clinical presentation, biochemical features, and pathophysiologic mechanisms. Medicine 77: 73–82.
    [28] Nicolson GL, Gan R, Nicolson NL, et al. (2007) Evidence for Mycoplasma ssp., Chlamydia pneunomiae, and human herpes virus-6 coinfections in the blood of patients with autistic spectrum disorders. J Neurosci Res 85: 1143–1148.
    [29] Careaga M, Van de Water J, Ashwood P (2010) Immune dysfunction in autism: A pathway to treatment. Neurotherapeutics 7: 283–292. doi: 10.1016/j.nurt.2010.05.003
  • Reader Comments
  • © 2018 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(8213) PDF downloads(2194) Cited by(2)

Figures and Tables

Figures(2)  /  Tables(1)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog