Loading [MathJax]/jax/output/SVG/jax.js
Theory article

VaR calculation by binary response models

  • Received: 04 January 2024 Revised: 31 May 2024 Accepted: 26 June 2024 Published: 15 July 2024
  • JEL Codes: G12, G17

  • The original Risk-Metrics method is underpinned by the assumption that daily asset returns are conditional Gaussian independently identically distributed (iid) random variables with a mean of zero. In this paper, a new method to calculate Value at Risk (VaR) was suggested to overcome the shortcoming of Risk-Metrics by employing binary response models to compute probability forecasts of the portfolio return by exceeding a grid of candidate quantile values. From those values, the VaR quantile value was selected. The proposed model was called BRV (Binary Response VaR method). Consistent application of BRV to the Dow Jones Industrial Average (INDEXDJX: DJI) and Dow Jones U.S. Marine Transportation Index (DJUSMT) time series proved that it was more accurate than the Risk-Metric system. This method not only worked similar to quantile regression but had the advantage that conventional maximum likelihood methods could be used for parameter estimation and inference. The BRV method was the best performing method for computing the daily VaR at both the 95% and 99% confidence levels over the period 02/01/06–31/12/08. The BRV and the QR (quantile regression) methods performed similarly, but the BRV method had the practical advantage that conventional maximum likelihood (ML) technique could be used for parameter estimation and robust inference.

    Citation: Kasra Pourkermani. VaR calculation by binary response models[J]. Data Science in Finance and Economics, 2024, 4(3): 350-361. doi: 10.3934/DSFE.2024015

    Related Papers:

    [1] Markus Haas . The Cowles–Jones test with unspecified upward market probability. Data Science in Finance and Economics, 2023, 3(4): 324-336. doi: 10.3934/DSFE.2023019
    [2] Moses Khumalo, Hopolang Mashele, Modisane Seitshiro . Quantification of the stock market value at risk by using FIAPARCH, HYGARCH and FIGARCH models. Data Science in Finance and Economics, 2023, 3(4): 380-400. doi: 10.3934/DSFE.2023022
    [3] Yi Chen, Zhehao Huang . Measuring the effects of investor attention on China's stock returns. Data Science in Finance and Economics, 2021, 1(4): 327-344. doi: 10.3934/DSFE.2021018
    [4] Nitesha Dwarika . Asset pricing models in South Africa: A comparative of regression analysis and the Bayesian approach. Data Science in Finance and Economics, 2023, 3(1): 55-75. doi: 10.3934/DSFE.2023004
    [5] Michael Jacobs, Jr . Benchmarking alternative interpretable machine learning models for corporate probability of default. Data Science in Finance and Economics, 2024, 4(1): 1-52. doi: 10.3934/DSFE.2024001
    [6] Yue Yuin Lim, Sie Long Kek, Kok Lay Teo . Efficient state estimation strategies for stochastic optimal control of financial risk problems. Data Science in Finance and Economics, 2022, 2(4): 356-370. doi: 10.3934/DSFE.2022018
    [7] Nitesha Dwarika . The risk-return relationship in South Africa: tail optimization of the GARCH-M approach. Data Science in Finance and Economics, 2022, 2(4): 391-415. doi: 10.3934/DSFE.2022020
    [8] Xinying Zhang, Chuanjun Zhao, Xianwei Zhou, Xiaojun Wu, Ying Li, Meiling Wu . Capital market and public health emergencies in Chinese sports industry based on a market model. Data Science in Finance and Economics, 2023, 3(2): 112-132. doi: 10.3934/DSFE.2023007
    [9] Dominic Joseph . Estimating credit default probabilities using stochastic optimisation. Data Science in Finance and Economics, 2021, 1(3): 253-271. doi: 10.3934/DSFE.2021014
    [10] Shuanglian Chen, Cunyi Yang, Khaldoon Albitar . Is there any heterogeneous impact of mandatory disclosure on corporate social responsibility dimensions? Evidence from a quasi-natural experiment in China. Data Science in Finance and Economics, 2021, 1(3): 272-297. doi: 10.3934/DSFE.2021015
  • The original Risk-Metrics method is underpinned by the assumption that daily asset returns are conditional Gaussian independently identically distributed (iid) random variables with a mean of zero. In this paper, a new method to calculate Value at Risk (VaR) was suggested to overcome the shortcoming of Risk-Metrics by employing binary response models to compute probability forecasts of the portfolio return by exceeding a grid of candidate quantile values. From those values, the VaR quantile value was selected. The proposed model was called BRV (Binary Response VaR method). Consistent application of BRV to the Dow Jones Industrial Average (INDEXDJX: DJI) and Dow Jones U.S. Marine Transportation Index (DJUSMT) time series proved that it was more accurate than the Risk-Metric system. This method not only worked similar to quantile regression but had the advantage that conventional maximum likelihood methods could be used for parameter estimation and inference. The BRV method was the best performing method for computing the daily VaR at both the 95% and 99% confidence levels over the period 02/01/06–31/12/08. The BRV and the QR (quantile regression) methods performed similarly, but the BRV method had the practical advantage that conventional maximum likelihood (ML) technique could be used for parameter estimation and robust inference.



    The Value at Risk (VaR) system falls into three major classes: The parametric method introduced by the Morgan Risk-Metric system, the non-parametric method based on historical simulation, and the semi-parametric method based on Extreme Value Theory (EVT) tail distribution. These methods are based on the good approximation of probability distributions extracted from an asset market price. The regulatory capital requirement for market risk is typically determined using the VaR for aggregate trading portfolio over a ten-day horizon and with a 99% confidence interval, and by the performance of the banks VaR models in backtesting exercises (Zumbach, 2007). Backtest is the best way to check the risk model performance. In backtesting, the estimated VaR is compared with the actual return over the same period. The VaR exceedance occurs when the return is more negative than the VaR. In order to backtest the accuracy for the estimated VaRs, we compute the empirical failure rates. By definition, the failure rate is the number of times returns exceed the forecasted VaR. If the model is correctly specified, the failure rate should be equal to the specified VaR level. In this paper, the backtesting VaR relies on the Christoffersen (2006) and Kupiec (1995) proposed system. Dumitrescu et al. (2012) have also proposed a backtest based on the non-linear regression model. Backtesting is a formal statistical framework that consists of verifying if actual trading losses are in line with model generated VaR forecasts and relies on testing over VaR violations. A violation is said to occur when the realized trading loss exceeds the VaR forecast.

    The Risk Metrics method to compute VaR is set out in Zumbach (2006) and Jorge and Jerry (2001). The original Risk Metrics method is underpinned by the assumption that daily asset returns are conditional, Gaussian independently and identically distributed (iid) random variables with a mean of zero. Under this assumption, VaR at the one-day horizon can be computed by multiplying the relevant quantile of the standard normal distribution by the one day ahead forecast of the conditional standard deviation of the portfolio return and multiplying the result by the market-to-market value of the portfolio. A convenient consequence of the iid assumption is that VaR for longer horizons can be computed by multiplying the daily VaR by the square root of the time horizon on days. Other methods to compute VaR are the historical simulation method (HS) where relevant quantile from the empirical distribution of simulated returns is applied (see Jorion, 2007 for details on both methods). Emenogu et al. (2020) discovered that the persistence of the GARCH models is robust, with the exception of a few cases where IGARCH and EGARCH were unstable. The SGARCH and GJRGARCH models also failed to converge for t-student innovation, and the mean reverting number of days for returns varied between models. Altun (2020) also found that GARCH models listed under the TSLx innovation distribution produce more accurate VaR forecasts than other competing models. Slim et al. (2017) claimed that in developed markets, the related models show signs of long memories, suggesting that the FIGARCH model is preferable to the GARCH and GJR models. In frontier and emerging markets, the GJR and GARCH are the most important specifications for capturing risk. This means that when analyzing frontier markets, risk managers should favor models that account for asymmetry.

    In practice, the Risk Metrics and HS methods to computing VaR have been the most popular methods within the industry. Note, however, that the HS method is highly restrictive since it ignores any additional explanatory information that might be thought to have an impact on the relevant return quantile (such as financial or macroeconomics information). This information can be incorporated into the VaR computed using the Risk Metrics method but only through the conditional mean or the conditional variance of return. The other weakness of the original Risk Metrics method is that it assumes normality of the returns. In practice, empirical evidence suggests that for many financial prices, the conditional distributions is fat-tailed.

    VaR can also be computed using quantile regression (QR) (Jorge and Jerry, 2001) (see for example (Taylor 1999, 2018)). When QR is used, the relevant return quantile can be modelled as a function of contemporaneous or lagged explanatory variables. In addition to this, QR based methods to VaR are more flexible than the Risk Metrics method since they do not require returns to be conditionally Gaussian. However, QR has several drawbacks. For example, the linearity assumption is highly questionable when data is heteroscedastic, (see Kupiec 1995, Peracchi 2002) and while nonlinear QR estimation techniques have been proposed, the asymptotic theory is not well developed. Furthermore, with both linear and nonlinear QR, the presence of heteroscedasticity can lead to estimated quantile that cross each other, and in both cases, robust inference typically requires bootstrapping, which increases the computational cost of the method. Tabasi et al. (2019) used GARCH models to model the volatility-clustering feature and found that using the t-student distribution function instead of the Normal distribution function improved model parameter estimation. Nieto and Ruiz (2016) compared the forecasting potential of various GARCH-based VaR models to their alternatives in an updated report. Surprisingly, they found that forecasting outcomes are affected by the number of out-of-sample observations as well as the time span being studied. They concluded that no single model outperforms another. Furthermore, only the asymmetric EGARCH-based model with skewed Student's-t distribution can be approved under the various model tests. Thavaneswaran et al. (2020) have introduced a volatility estimator applying an estimating function approach.

    I propose an alternative parametric method to computing VaR. This alternative method allows the practitioner to utilize additional explanatory information to forecast the relevant quantile. However, relative to QR-based methods, the method proposed here has the significant practical advantage that the conditional maximum likelihood (ML) technique can be used for parameter estimation and robust inference.

    The proposed method exploits the inverse relationship between the conditional quantile function (QF) and the conditional cumulative distribution (CDF), utilizing a technique for estimating the conditional CDF developed by Foresi and Peracchi (1995). I assume the market-to-market value of the relevant portfolio is one; hence, the daily VaR depends only on a one day ahead forecast of the relevant quantile for portfolio returns. Rather than directly forecast the VaR quantile, here it is proposed that the forecast is obtained indirectly using binary response models to compute probability forecasts over a grid of candidate quantile values. The candidate quantile value with an associated probability forecast closest to the desired probability (e.g. p = 0.01 for VaR at the 99% confidence level) is used as the VaR. This method is equivalent to forecasting points on the left tail of the conditional CDF and then inverting at the required VaR probability (Jelito and Pitera 2021).

    Binary response models have previously been shown to be useful for forecasting the direction of asset returns. For example, Christoffersen and Diebold (2006) use a Logit model that conditions volatility as a predictor to forecast the direction of a time series index. Using binary response models to estimate points on the conditional CDF for stock returns has not previously been used for computing VaR. For brevity, we will refer to the method proposed here as the BRV (Binary Response VaR method). Ugurlu (2023) has also proposed a coherent multivariate average Var to quantify the total risk.

    I compare the empirical performance of the BRV method with the orthodox Risk Metrics and HS method and a QR method in Monte Carlo simulations and an empirical application. The empirical application involves recursively computing daily VaR for two stock market indices, Dow Jones Industrial Average (INDEXDJX: DJI), and Dow Jones U.S. Marine Transportation Index (DJUSMT) over a three-year period 02/01/06–31/12/18 using the previous five years of daily data for parameter estimation at each day. The results are analyzed using tests for correct unconditional coverage and independence of the VaR exceedances. Computing VaR over this period is challenging as the period begins with benign market conditions and ends with extremely volatile conditions associated with the global financial crisis that began in 2007. I found that the BRV method and QR method clearly dominate the Risk Metrics and HS method over this period. In particular, it appears that the Risk Metrics and HS method consistently underestimate the population VaR over this period since the proportion of VaR exceedances is too large for the given confidence levels. In contrast, exceedances when the BRV method is used are much closer to the expected number. Underestimating the population VaR can lead to serious penalties for banks operating in countries where the Basel ll Capital Accord has been implemented. Hence, this result is of practical importance.

    Define the probability of the log portfolio returns Rt exceeding a threshold ri conditional on a known k×1 vector of predictors Xt1 as,

    Pi,t=Pt(Rtri|Xt1)=E(Yi,t|Xt1) (1)

    where <ri<(i=1,2.,N)and Yi,t is a binary indicator,

    Yi,t=1,ifRtri0,ifRt>ri (2)

    Note that: Pi,t can be interpreted as the value of the conditional CDF for Rtevaluated at ri.

    The VaR return quantile at the confidence level (1 – p) ×100%, 0 < p < 1, is the value ri.

    Such that Pt(Riri|Xt1)=p. Let this be denoted by Qi(p).In practice the aim when computing VaR is to forecast the future value QT+h(p). Throughout this paper we focus on computing daily VaR, so T denotes the current day and h = 1.

    The BRV method to forecasting QT+1(p). proposed here has three clear steps:

    (ⅰ) Estimate multiple binary response models with the binary indicator (2) as the dependent variable over a range of candidate VaR return quantile, ri, using conventional ML (the form of the link function in the binary response model and the location and number of values for ri will be discussed below);

    (ⅱ) project the estimated binary response models forward to compute a one step ahead forecast of the probability of exceeding

    ri,ˆpi,T+1=PT+1(RT+1ri|XT); (3)

    (ⅲ) as a forecast of the VaR quantile QT+1(p) use the threshold ri from (ⅰ) that minimizes the distance between the probability forecast ˆpi,T+1 and the desired VaR probability p,

    ˆri=argmin{ˆpi,T+1p} (4)

    The use of binary response models to estimate the conditional CDF for stock returns was proposed by Foresi and Peracchi (1995). Let <r1<r2<<rN< be N feasible values of the return over the conditional CDF for Rt. Foresi and Peracchi (1995) show that points on the conditional CDF correspondence to ri(i = 1, 2, ..., N) can be estimated using a functional form that best approximates the population conditional CDF, Ft,as the link function in separate binary response models with the binary indicator (2) as the dependent variable. The "best approximation" is formalized as the approximation that minimizes the Kullback–Leibler divergence. Under weak regularity conditions, the parameters of the binary respons model can be consistently estimated by ML giving the estimated point ˆFi,t.

    Clearly, the conditional CDF should satisfy the standard condition,

    0<Fi,t<1 (5)

    An attractive approach of the Foresi and Peracchi (1995) technique is that since it involves modelling the log-odds in [Fi,t/(1Fi,t)]rather than Fi,t directly, (5) is automatically satisfied. Foresi and Peracchi (1995) use a semi-parametric Logit model in their empirical work. In principle, any twice-differentiable CDF can be used. Note that monotonicity of the estimated CDF,

    0<ˆFi,t<<ˆFN,t<1 (6)

    will not necessarily be satisfied if the ML estimation is unrestricted. Whether monotonicity is satisfied or not depends on numerous factors, including the sample size and the spread of the ri values. In practice, even if monotonicity is violated, this might not have a serious detrimental impact on the practical performance of the technique. Monotonicity can be incorporated into the estimation algorithm if in practice it is a serious problem.

    Foresi and Peracchi (1995) focus on estimating points on the conditional CDF using binary response models. Here, estimated binary response models are used to produce forecasts of the probability of exceeding candidate quantiles in the left tail of the conditional CDF (i.e., forecasts of points on the left tail of the conditional CDF). In the simulation and empirical work here, for simplicity, the cumulative normal and logistic CDFs are used. For example, when the logistic is used,

    pi,t=11+exp(ˊXt1βi) (7)

    Where βi is a k×1 vector of parameters and Xt1is the vector of predictors in (1). The conditional CDF Fi,t can then be estimated by replacing B in (7) with the ML estimator ˆβi.1

    1 The asymptotic properties of the ML estimator for a Logit model are well-known and they are omitted.

    Step (ⅱ) of the BRV method is to use the estimated parameters from (ⅰ) to compute a one-step ahead probability forecast for each candidate threshold ri. Therefore, when the logistic CDF is used,

    ˆpi,T+1=11+exp(ˊXTˆβi) (8)

    where XT is the vector of predictors at time T. Step (ⅲ) of the BRV method involves finding the relevant VaR ˆri=argmin{ˆpi,T+1p}. This can be done using a linear computer grid search. To implement this method, the practitioner needs to decide on a functional form for the link function and on the total number of thresholds ri to use the size of N) and on their location and spacing. In the empirical application below, for simplicity, the logistic CDF is used as a link function. Similar link functions can also be used (e.g., normal CDF, Student t CDF, Generalized Extreme Value CDF, semi-parametric link functions, etc.) and, in practice, backtesting over a historical sample period could be employed to select the best performing link function from a set of candidate functions. In simulations and empirical applications discussed below, I found a grid of ri values in backtesting, starting with the third value of the order statistics for the historical returns followed by the 1st, 3rd, 5th, 10th and 15th percentiles, which produce good results (thus N = 6). Cubic spline interpolation is then used to increase the number of forecasts and thresholds.

    Clearly, if the link function has exactly the same form as the population conditional CDF for returns (e.g., normal-normal or logistic-logistic) and appropriate regressors are employed, then as T, the Foresi and Peracchi (1995) method provides a consistent estimator of points on the left tail of the conditional CDF, providing that the regularity conditions required for ML to be a consistent estimator in this instance, are satisfied. To illustrate this in action, assume the following Data Generating Process (DGP) for log returns,

    Rt=γ1X1,t+γ2X2,t+ϵt, (9)
    ϵtL(0,1) (10)
    X1,t=θ1X1,t1+V1,t,V1,tN(0,1) (11)
    X2,t=θ2X2,t1+V2,t,V2,tN(0,1) (12)

    where L(0, 1) and N(0, 1) denote a logistic and normal distribution with mean of zero and a variance of one. Therefore, conditional on the X variables generated by AR(1) models, returns have a logistic distribution. We simulate representative series of returns from this general DGP. For one set the following parameter values are used; γ1=0.50,γ2=0,θ1=0.30For the other set; γ1=0.50,γ2=0.50,θ1=0.30,θ2=0.30.Therefore in the first set of simulations the model contains a single stationary regressor, while in the second set the model contains two stationary regressors. Observations from (9)–(12) are simulated for the following sample sizes, T = 100,200,500, 1000, 10000. For each series, we then estimate the left tail of the conditional CDF using the method of Foresi and Peracchi (1995) employing the logistic CDF as the link function over a grid of ri values starting with the third value of the order statistics for the historical returns followed by the 1st, 3rd, 5th, 10th and 15th percentile values.2 In both cases, a constant and the correct explanatory variables are used in the link function.

    2Note that here the model is not predictive since the explanatory variables are current values which we make no attempt to forecast. In the Monte Carlo simulations and empirical application discussed below, which involves forecasting, lagged values of the explanatory variables are used.

    For each replication VaR is computed using QR, along with the original Risk Metrics and HS methods, the true volatility is used when computing the Risk Metrics VaR.

    To assess the finite-sample performance of each method, for each replication the estimated unconditional coverage is computed ˆp. The 5th, 25th, 50th, 75th and 95th percentiles of the empirical distribution of ˆp are reported in Table 1 for the 95% confidence level and Table 2 for the 99% confidence levels. In Table 1 and 2, these are reported in four rows: The first row is Gaussian GARCH data generating process (DGP), the second row is threshold GARCH data generating process, the third row contains autoregressive Gaussian GARCH data generating process, and the fourth-row reports autoregressive threshold GARCH date generation process. The selection of a confidence level for an interval determines the probability that the confidence interval produced will contain the true parameter value. Common choices for the confidence level are 0.95, and 0.99. These levels correspond to percentages of the area of the normal density curve. For example, a 95% confidence interval covers 95% of the normal curve-the probability of observing a value outside of this area is less than 0.05. Because the normal curve is symmetric, half of the area is in the left tail of the curve, and the other half of the area is in the right tail of the curve.

    Table 1.  Simulated ˆp distribution 95% confidence.
    Percentile
    Gaussian-GARCH DGP
    5th 25th 50th 75th 95th
    1-BRV 0.021 0.030 0.042 0.063 0.074
    t-GARCH DGP
    2-BRV 0.014 0.036 0.049 0.079 0.128
    AR-Gaussian- GARCH DGP
    3-BRV 0.019 0.028 0.062 0.064 0.090
    AR-t GARCH DGP
    4-BRV 0.022 0.023 0.042 0.082 0.122

     | Show Table
    DownLoad: CSV
    Table 2.  Simulated ˆp distribution 99% confidence.
    Percentile
    Gaussian-GARCH DGP
    5th 25th 50th 75th 95th
    BRV 0.000 0.007 0.011 0.014 0.026
    t-GARCH DGP
    BRV 0.000 0.003 0.013 0.016 0.030
    AR-Gaussian-GARCH DGP
    BRV 0.000 0.005 0.013 0.016 0.024
    AR-t GARCH DGP
    BRV 0.000 0.004 0.013 0.021 0.040

     | Show Table
    DownLoad: CSV

    The second-fourth simulation experiments are the same as the first but with different DGPs for the returns, allowing for serial correlation and conditionally non-Gaussian returns. The various DGPs for all the experiments are given below:

    DGP 1. Gaussian-GARCH

    Rt=ϵth1/2t,ht=0.1+0.1R2t1+0.8ht1,ϵtN(0,1) (13)

    DGP 2. t-GARCH.

    Rt=vth1/2t,ht=0.1+0.1R2t1+0.8ht1,vtt(5) (14)

    DGP 3. AR-Gaussian-GARCH

    Rt=0.3Rt1+ut,ut=ϵtR1/2t,ϵtN(0,1) (15)
    ht=0.1+0.1u2t1+0.8ht1 (16)

    DGP 4. AR – t – GARCH

    Rt=0.3Rt1+ut,ut=vtR1/2t,vtt(5) (17)
    ht=0.1+0.1u2t1+0.8ht1 (18)

    For DGPs 1 and 3, a normal CDF is used as the link function when using the BRV method (i.e. Probit models are estimated). For DGPs 2 and 4, a logistic CDF is used as the link function (Logit models are estimated). For DGPs 1 and 2, just a constant is included as a predictor in the relevant link function, and for DGPs 3 and 4, the link function also contains a lag of returns (hence, the estimated models are correctly specified).

    For the 95% confidence level, it can be seen in Table 1 that the empirical distribution of ˆp for both the BRV and QR methodes are virtually identical. They both have good levels of unconditional coverage given the small size of the backtesting period (250 observations). The distribution of ˆp is centered close to the population value of p = 0.05, irrespective of the DGP. In both cases, the performance of these methods is similar, irrespective of whether returns are conditionally Gaussian or non-Gaussian. It can be shown that the logistic CDF closely approximates a student t CDF with 9 degrees of freedom (see Mudholkar and George (1978). Hence, the BVR method using Logit models is clearly well-suited to computing VaR if the conditional distribution for returns is thought to be fat tailed. The results in Table 2 show that at the 99% confidence level, the BRV and QR methods also produce very similar results. Again, the distributions of ˆp are centered close to the population value of p = 0.01 and both have a similar variance.

    In contrast, however, it can be seen that at both confidence levels, the Risk Metrics method significantly underestimates the population VaR for DGPs 2 and 4 (non-Gaussian returns). For DGP 3, (Gaussian returns but with serial correlation), the Risk Metrics method, which ignores serial correlation, slightly underestimates the population VaR. The HS method gives similar results to the BRV and QR methods for all the DGPs at both confidence levels, and there is no distinguishable difference in the performance of the HS method for the DGPs with and without serial correlation.

    In this section, I discuss an empirical application involving DJI and DJUSMT series. The application involves recursively computing the daily VaR at the 95% and 99% confidence levels using the BRV, QR, Risk Metrics, and HS methods for every trading day over the three-year period 02/01/06–31/12/08 (755 days), using a five-year window of historical data for parameter estimation (approximately 1250 observations). For example, VaR on 02/01/06 is computed using data from 31/12/00–31/12/05. Note that the parameters of the BRV and QR models are re-estimated each day. The conditional standard deviation h1/2t1 is chosen as an important predictor following the evidence in Christoffersen and Diebold (2006) on its ability to forecast the direction of stock returns, and a positive relationship with pi,t is expected. TBt1 is included to allow for present value effects, and Vt1 is included to capture market sentiment, and again for both, a positive relationship with pi,t is expected. Conventional ML is used for parameter estimation in the BRV method and the interior point algorithm of Koenker and Park (1978) is used for parameter estimation in the QR method. In the Risk Metrics method, an EWMA volatility forecast with a weight of 0.94 is used, which is the default choice for the Risk Metrics method applied to daily data.

    Prior to discussing the backtesting results, as an example, the estimated Logit model parameters and robust t-statistics computed using Huber-White robust standard errors are given in Table 3 for each index and stock at three points over the backtesting period (the points are 29/12/06, 31/12/07, and 30/12/08). In each case, the 1st percentile of the order statistics is used to define the threshold. On the basis of the robust t – statistics, I found clear evidence that all three explanatory variables are statistically significant at one or more of these points, and that when statistically significant, the signs of the estimated parameters are consistent with our expectations. Note that in Table 3, ˆβ0 is the estimated constant, which are −13.162 for DJI and −10.708 for DJUSMT, and ˆβ1 is the estimated parameter on TBt1, which are 0.518 for DJI and −0.018 for DJUSMT. ˆβ2 is the estimated parameter on Vt1, which are 0.119 for DJI and 0.107 for DJUSMT. ˆβ3 is the estimated parameter on h1/2t1, which are 2.617 for DJI and 2.551 for DJUSMT. Robust t-statistics computed using Huber-White standard errors are in parentheses.

    Table 3.  Logit model estimation results.
    Sample end ˆβ0 ˆβ1 ˆβ2 ˆβ3
    DJI
    1-31/12/08 −13.162 0.518 0.119 2.617
    DJUSMT
    2-31/12/08 −10.708 −0.018 0.107 2.551

     | Show Table
    DownLoad: CSV

    As one might expect, the exact statistical significance of the estimated parameters varies depending on the index or stock and sample period, but there are some patterns. For example, for each series, I found that h1/2t1 is strongly statistically significant, but that the statistical significance of TBt1 and Vt1 varies over the sample. Note that I do not eliminate statistically insignificant predictors prior to computing VaR using the BRV method; however, this method could be taken in future research to allow for structural change.

    To summarize the backtesting results, for each method and at each confidence level, the estimated unconditional coverage ˆp is reported for comparison with the population unconditional coverage p. Christoffersen (1998) proposes a complete methodology for evaluating the number of exceedances and their independence. The independence test rationale dictates that, if the violations are dependent, then the transition probabilities would not be equal. Finally, Christoffersen (1998) proposes a joint test that combines both hypotheses (Conditional Coverage CC hypothesis). Correct unconditional coverage LRuc and independence of the VaR exceedances LRind are proposed by Christoffersen (1998). The LRuc and LRindtests utilize the fact that if the VaR method is perfect, then VaR exceedances should be unpredictable and so a binary indicator of exceedances (the hit indicator),

    Ht+1=1,ifRt+1<ˆQ(p)t+10,ifRt+1ˆQ(p)t+1 (19)

    where ˆQ(p)t+1 is the forecast of the relevant return quantile, should be an independent Bernoulli random variable. The LRuc and LRindtests are straightforward to compute and have a x2 (1) asymptotic distribution (see Christoffersen, 1998, for further details).

    The backtesting results for the DJI at the 95% and 99% confidence levels, respectively, are given in Table 4. The results for the method that is optimal on the basis of the estimated unconditional coverage ˆp relative to the population value p are bolded. For DJI, at the 95% confidence level, the BRV result is ˆp = 0.05, suggesting that the population VaR over this period is estimated extremely well (p = 0.05). In contrast, the Risk Metrics, HS, and QR results are ˆp = 0.061, ˆp = 0.110, and ˆp = 0.052, respectively, suggesting that the population VaR is underestimated by each of these methods. The LRac test rejects the null hypothesis of correct unconditional coverage for the Risk Metrics and HS methodes at either the 5% or 1% significance levels. For DJI at the 99% confidence level, the Risk Metrics result is ˆp = 0.029 while the HS result is ˆp = 0.045, and the QR result is ˆp = 0.034.

    Table 4.  Backtesting results: DJI.
    ˆp LRuc LRind
    95% confidence
    Risk Metrics 0.061 5.092 ** 1.086
    BRV 0.050 0.002 0.681
    QR 0.052 1.037** 0.272
    HS 0.110 30.091 ** 0.469
    99% confidence
    Risk Metrics Metrics 0.029 18.439 ***** 1.382
    BRV 0.010 0.024 0.634
    QR 0.034 5.095 ** 0.561
    HS 0.045 28.333 ** 1.104
    Note that: *, **, and *** indicate a rejection of the null hypothesis at the 10%, 5%, and 1% levels, respectively.

     | Show Table
    DownLoad: CSV

    On the basis of the estimated unconditional coverage ˆp, in both cases, the BRV method is superior to any of the other methodes considered. For DJUSMT, at the 95% confidence level, the BRV result is ˆp = 0.05, suggesting that the population VaR over this period is estimated extremely well (p = 0.05). In contrast, the Risk Metrics, HS, and QR results are ˆp = 0.065, ˆp = 0.088, and ˆp = 0.032, respectively, suggesting that the population VaR is underestimated by each of these methods. The LRac test rejects the null hypothesis of correct unconditional coverage for the Risk Metrics and HS methodes at either the 5% or 1% significance levels. For DJUSMT at the 99% confidence level, the Risk Metrics result is ˆp = 0.024 while the HS result is ˆp = 0.053, and the QR result is ˆp = 0.19. Therefore, again, in all three cases, these results suggest that the population VaR is underestimated. Furthermore, in all three cases, LRuc rejects the null hypothesis of correct unconditional coverage at conventional significance levels. The BRV result is ˆp = 0.015, which is much closer to the desired level of coverage.

    The results for the DJI and DJUSMT indices in Table 5 show that at the 95% confidence level, the QR method is preferred on the basis of the estimated unconditional coverage with the BRV method being the next most accurate. At the 99% confidence level, the BRV method is preferred. Again, for the other method considered ˆp>p and the rejections obtained from LRuc, it is suggested that the population VaR is underestimated. In this case, rejections are also obtained from LRind for the BRV, QR, and HS methods at the 95% confidence level, suggesting mis-specified models.

    Table 5.  Backtesting results: DJUSMT.
    ˆp LRuc LRind
    95% confidence
    Risk Metrics Metrics 0.065 3.239 * 0.705
    BRV 0.021 0.543 5.111 **
    QR 0.032 0.067 0.265
    HS 0.088 21.81 *** 7.002 **
    99% confidence
    Risk Metrics Metrics 0.023 8.817 *** 0.830
    BRV 0.015 0.894 0.365
    QR 0.19 6.655*** 0.574
    HS 0.053 36.339*** 9.55**

     | Show Table
    DownLoad: CSV

    The optimal method on the basis of the estimated unconditional coverage ˆp are either the BRV method (optimal in four out of six cases) or the QR method (optimal in the remaining two cases).

    Underestimating the population VaR can lead to serious penalties for banks in countries where the Basel Ⅱ Capital Accord has been implemented and is a well-known weakness of the Risk Metrics method if the population conditional distribution is fat-tailed.

    I propose an alternative parametric method to computing the widely used financial risk measure VaR. The BRV method involves using binary response models to compute probability forecasts of the portfolio return exceeding a grid of candidate quantile values. The candidate quantile value associated with a probability forecast closest to the desired VaR probability is chosen as the VaR. The performance of the BRV method is impressive relative to the orthodox Risk Metrics and HS methods and a QR-based method, both in Monte Carlo simulation experiments and an empirical application involving DJI and DJUSMT indexes. In the empirical application, the BRV method is the most accurate method in most cases on the basis of the estimated unconditional coverage. Note in particular that the BRV method is the best performing method for computing the daily VaR at both the 95% and 99% confidence levels over the turbulent period 02/01/06–31/12/08. The BRV and QR methods perform similarly, but relative to QR, the BRV method has the practical advantage that conventional ML methods can be used for parameter estimation and robust inference.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The authors declare no conflict of interest.



    [1] Altun M (2020) A new approach to Value-at-Risk: GARCH-TSLx model with inference. Comm Stat Simu Comput 49: 3134–3151. https://doi.org/10.1080/03610918.2018.1535069 doi: 10.1080/03610918.2018.1535069
    [2] Christoffersen PF (1998) Evaluating Interval Forecasts. Int Econ Review 39: 841–862. https://doi.org/10.2307/2527341 doi: 10.2307/2527341
    [3] Christoffersen PF, Diebold FX (2006) Financial Asset Returns, Direction-of-Change Forecasting and Volatility Dynamics. Manag Sci 52: 1273–1287. https://doi.org/10.1287/mnsc.1060.0520 doi: 10.1287/mnsc.1060.0520
    [4] Dumitrescu EI, Hurlin C, Pham V (2012) Backtesting Value-at-Risk: From Dynamic Quantile to Dynamic Binary Tests. Halshs. https://shs.hal.science/halshs-00671658
    [5] Emenogu NG, Adenomon MO, Nweze NO (2020) On the volatility of daily stock returns of Total Nigeria Plc: evidence from GARCH models, value-at-risk and backtesting. Financ Innov 6: 18. https://doi.org/10.1186/s40854-020-00178-1 doi: 10.1186/s40854-020-00178-1
    [6] Engle RF, Manganelli S (2004) CAViaR: Conditional Autoregressive Value at Risk by Regression Quantiles. J Bus Econ Stat 22: 367–381. http://www.jstor.org/stable/1392044
    [7] Foresi S, Peracchi F (1995) The Conditional Distribution of Excess Returns: An Empirical Analysis. J Amer Statist Assoc 90: 451–466. https://doi.org/10.1080/01621459.1995.10476537 doi: 10.1080/01621459.1995.10476537
    [8] Huber PJ (1967) The Behavior of Maximum Likelihood Estimates Under Nonstandard Conditions. Proceedings of the fifth Berkeley symposium on mathematical statistics and probability 1: 221–233.
    [9] Jelito D, Pitera M (2021) New fat-tail normality test based on conditional second moments with applications to finance. Stat Pap 62: 2083–2108. https://doi.org/10.1007/s00362-020-01176-2 doi: 10.1007/s00362-020-01176-2
    [10] Jorge M, Jerry X (2001) Return to Risk Metrics: The Evolution of a Standard. Risk Metrics Group Inc.
    [11] Jorion P (2007) Value at Risk: The New Benchmark for Managing Financial Risk. 3. McGraw-Hill, New York.
    [12] Koenker R, Bassett G (1978) Regression Quantiles. Econometrica 46: 33–50. http://dx.doi.org/10.2307/1913643 doi: 10.2307/1913643
    [13] Koenker R, Park BJ (1966) An interior point algorithm for nonlinear quantile regression. J Econometrics 71: 265–283. https://doi.org/10.1016/0304-4076(96)84507-6 doi: 10.1016/0304-4076(96)84507-6
    [14] Kupiec P (1995) Techniques for Verifying the Accuracy of Risk Measurement Models. J Deriv 3: 73–84. http://dx.doi.org/10.3905/jod.1995.407942 doi: 10.3905/jod.1995.407942
    [15] Mudhokar GS, George O (1978) A remark on the shape of the logistic distribution. Biometrika 65: 667–668. https://doi.org/10.1093/biomet/65.3.667 doi: 10.1093/biomet/65.3.667
    [16] Nieto MR, Ruiz E (2016) Frontiers in VaR forecasting and backtesting. Intl J Forecasting 32: 475–501. https://doi.org/10.1016/j.ijforecast.2015.08.003 doi: 10.1016/j.ijforecast.2015.08.003
    [17] Peracchi F (2002) On estimating conditional quantiles and distribution functions. Comput Statis Data Anal 38: 433–447. https://doi.org/10.1016/S0167-9473(01)00070-6 doi: 10.1016/S0167-9473(01)00070-6
    [18] Slim S, Koubaa Y, BenSaida A (2017) Value-at-Risk under Lévy GARCH models: Evidence from global stock markets. J Int Financ Mark I 46: 30–53 https://doi.org/10.1016/j.intfin.2016.08.008 doi: 10.1016/j.intfin.2016.08.008
    [19] Tabasi H, Yousefi V, Tamošaitienė J, et al. (2019) Estimating Conditional Value at Risk in the Tehran Stock Exchange Based on the Extreme Value Theory Using GARCH Models. Adm Sci 9: 40. https://doi.org/10.3390/admsci9020040 doi: 10.3390/admsci9020040
    [20] Taylor JW (1999) A quantile regression method to estimating the distribution of multi period returns. J Deriv 7: 64–78. https://doi.org/10.3905/jod.1999.319106 doi: 10.3905/jod.1999.319106
    [21] Taylor JW (2008) Using Exponentially Weighted Quantile Regression to Estimate Value at Risk and Expected Shortfall. J Financ Economet 6: 382–406. https://doi.org/10.1093/jjfinec/nbn007 doi: 10.1093/jjfinec/nbn007
    [22] Thavaneswaran A, Paseka A, Frank J (2020) Generalized value at risk forecasting. Commun Stat-Theor M 49: 4988–4995. https://doi.org/10.1080/03610926.2019.1610443 doi: 10.1080/03610926.2019.1610443
    [23] Ugurlu K (2023) A new coherent multivariate average-value-at-risk. Optimization 72: 493–519. https://doi.org/10.1080/02331934.2021.1970755 doi: 10.1080/02331934.2021.1970755
    [24] White H (1980) A Heteroskedasticity-Consistent Covariance Matrix Estimator and a Direct Test for Heteroskedasticity. Econometrica 48: 817–838. https://doi.org/10.2307/1912934 doi: 10.2307/1912934
    [25] Zumbach G (2006) The Risk Metrics 2006 Methodology. Available at SSRN. http://dx.doi.org/10.2139/ssrn.1420185
  • This article has been cited by:

    1. Badreddine Slime, Jaspreet Singh Sahni, Modeling default risk charge (DRC) with intensity probability theory, 2025, 10, 2473-6988, 2958, 10.3934/math.2025137
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1171) PDF downloads(57) Cited by(1)

Figures and Tables

Tables(5)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog