The critical discovery of the presenilins and their association with familial Alzheimer’s disease (AD) prompted an intensive research effort to understand the molecular mechanisms of that disease. The presenilins were subsequently found to be the catalytic component of the multi-protein enzyme complex, γ-secretase, the enzyme that is known to act on the amyloid precursor protein (APP) to generate amyloid beta (Aβ) peptides that comprise the neuritic plaques implicated in AD pathology. Here, we discuss the background of γ-secretase- mediated proteolysis of APP and its association with familial AD. We discuss the association of neuroinflammation with AD, focusing on the link between the innate immune response, the clearance of the Aβ peptides and disease progression. Currently, there are limited treatments for AD that strive to ameliorate the symptoms of the disease but do not address the molecular basis of the disease. The greater understanding of γ- secretase functions has provided new insights into potential therapeutics for AD, a number of which are in clinical trials.
1.
Introduction
Estimating foreign exchange rate (FX) volatility is a core risk management activity for financial institutions, corporates and regulators. The subject has been extensively investigated among both practitioners and scientific researchers, and several alternative models exist. Among the most prominent are the models belonging to GARCH and stochastic volatility classes. However, the true value of volatility cannot be directly observed. Hence, volatility must be estimated, inevitably with error. This constitutes a fundamental problem in implementing parametric models, especially in the context of high-frequency data. Andersen and Bollerslev (1998) proposed using realized volatility, as derived from high-frequency data, to accurately measure the true latent integrated volatility. This approach has gained attention for volatility modeling in markets where tick-level data is available (Andersen et al., 2013). Andersen et al. (2003) suggest fractionally integrated ARFIMA models in this context. Still, the long-memory HAR (heterogeneous autoregressive) model of Corsi (2009) is arguably the most widely used to capture the high persistence typically observed in realized volatility of financial prices. The HAR model is relatively simple and easy to estimate. In empirical applications, the model tends to perform better than GARCH and stochastic volatility models possibly due to the sensitivity of tightly parameterized volatility models to minor model misspecifications (Sizova., 2011). Although realized volatility $ (RV) $ is a consistent estimator of the true latent volatility, it is subject to measurement error in empirical finite samples. Hence, $ RV $ will not only reflect the true latent integrated volatility $ (IV) $, but also additional measurement errors. Bollerslev et al. (2016) propose utilizing higher-order realized moments of the realized distribution to approximate these measurement errors. More specifically, Bollerslev et al. (2016) propose the HARQ-model, which augments the HAR model with realized quarticity as an additional covariate.
The empirical performance of the HARQ and related extensions has been extensively studied. The focus has predominantly been on equity markets. A majority of the studies analyze U.S. data; see Bollerslev et al. (2016); Clements, A. and Preve, D. (2021); Pascalau and Poirier (2023); Andersen et al. (2023) and others. Liu et al. (2018) and Wang et al. (2020) investigate Chinese equity markets, whereas Liang et al. (2022); Ma et al. (2019) analyse international data. Bitcoin and electricity markets have attracted some attention; see, for instance, Shen et al. (2020); Qieu et al. (2021), and Qu et al. (2018).
Empirical applications of the HARQ model in the context of foreign exchange rate risk are sparse. Lyócsa et al. (2016) find that the standard HAR model rarely is outperformed by less parsimonious specifications on CZKEUR, PLZEUR, and HUFEUR data. Plíhal et al. (2021) and Rokicka and Kudła (2020) estimate the HARQ model on EURUSD and EURGBP data, respectively. Their focus is different from ours, as they investigate the incremental predictive power of implied volatility for a broad class of HAR models. In a similar vein, Götz (2023) and Lyócsa et al. (2024) utilize the HARQ model for the purpose of estimating foreign exchange rate tail risk.
Using updated tick-level data from two major currency pairs, EURUSD and USDJPY, this paper documents the relevance of realized quarticity for improving volatility estimates across varying forecasting horizons. These results are robust across estimation windows, evaluation metrics, and model specifications.
2.
Materials and method
2.1. Data
We use high-frequency intraday ticklevel spot data, publicly available at DukasCopy* The sample period is 1. January 2010 to 31. December 2022. Liu et al. (2015) investigate the optimal intraday sampling frequency across a significant number of asset classes and find that 5-min intervals usually outperform others. Hence, as common in the literature, we estimate the realized volatility from 5-minute returns.
*This data source is also used by Plíhal et al. (2021), Risstad et al. (2023) and Lyócsa et al. (2024), among others.
To filter tick-level data, we follow a two-step cleaning procedure based on the recommendations by Barndorff-Nielsen et al. (2009). Initially, we eliminate data entries that exhibit any of the following issues: (i) absence of quotes, (ii) a negative bid-ask spread, (iii) a bid-ask spread exceeding 50 times the median spread of the day, or (iv) a mid-quote deviation beyond ten mean absolute deviations from a centered mean (computed excluding the current observation from a window of 25 observations before and after). Following this, we calculate the mid-quotes as the average of the bid and ask quotes and then resample the data at 5-minute intervals.
We compute the consistent estimator of the true latent time-$ t $ variance from
where $ M = 1/\Delta $, and the $ \Delta $-period intraday return is $ r_{t, i} \equiv \log(S_{t-1+i\times\Delta}) - \log(S_{t-1+(i-1)\times\Delta}) $, where $ S $ is the spot exchange rate. Analogously, the multi$ (h) $-period realized variance estimator is
Setting $ h = 5 $ and $ h = 22 $ yields weekly and monthly estimates, respectively.
Table 1 displays descriptive statistics for daily realized variances, as computed from (1).
2.2. The HARQ models
To represent the long-memory dynamic dependencies in volatility, Corsi (2009) proposed using daily, weekly, and monthly lags of realized volatility as covariates. The original HAR model is defined as
where $ RV $ is computed from (1) and (2). If the variables in (2.2) contain measurement errors, the beta coefficients will be affected. Bollerslev et al. (2016) suggests two measures to alleviate this. First, they include a proxy for measurement error as an additional explanatory variable. Furthermore, they directly adjust the coefficients in proportion to the magnitude of the measurement errors:
where realized quarticity $ RQ $ is defined as
The full HARQ model in (2.2) adjusts the coefficients on all lags of $ RV $. A reasonable conjecture is that measurement errors in realized volatilities tend to diminish at longer forecast horizons, as these errors are diversified over time. This suggests that measurement errors in daily lagged realized volatilities are likely to be relatively more important. Motivated by this Bollerslev et al. (2016) specify the HARQ model as
Although there is no reason to expect that autoregressive models of order one will be able to accurately capture long memory in realized volatility, we estimate AR(1) models as a point of reference. The AR and ARQ models are defined as
and
in equations (6) and (7), respectively.
3.
Results and discussion
Due to noisy data and related estimation errors, forecasts from realized volatility models might occasionally appear as unreasonably high or low. Thus, in line with Swanson et al. (1997) and Bollerslev et al. (2016), we filter forecasts from all models so that any forecast outside the empirical distribution of the estimation sample is replaced by the sample mean.
3.1. In-sample estimation results
Table 2 reports in-sample parameter estimates for the ARQ, HARQ, and HARQ-F models, along with the benchmark AR and ARQ models, for one-day ahead EURUSD (upper panel) and USDJPY (lower panel) volatility forecasts. Robust standard errors (s.e.) are computed as proposed by White (1980). $ R^{2} $, $ MSE $, and $ QLIKE $ are displayed at the bottom of each panel.
The coefficients $ \beta_{1 Q} $ are negative and exhibit strong statistical significance, aligning with the hypothesis that $ RQ $ represents time-varying measurement error. When comparing the autoregressive (AR) coefficient of the AR model to the autoregressive parameters in the ARQ model, the AR coefficient is markedly lower, reflecting the difference in in persistence between the models.
In the comparative analysis of the HAR and HARQ models applied to both currency pairs, the HAR model assigns more emphasis to the weekly and monthly lags, which are generally less sensitive to measurement errors. In contrast, the HARQ model typically assigns a higher weight to the daily lag. However, when measurement errors are substantial, the HARQ model reduces the weight on the daily lag to accommodate the time-varying nature of the measurement errors in the daily realized volatility ($ RV $). The flexible version of this model, the HARQ-F, allows for variability in the weekly and monthly lags, resulting in slightly altered parameters compared to the standard HARQ model. Notably, the coefficients $ \beta_{2 Q} $ and $ \beta_{3 Q} $ in the HARQ-F model are statistically significant, and this model demonstrates a modest enhancement in in-sample fit relative to the HARQ model.
3.2. Out-of-sample forecasting results
To further assess the out-of-sample performance of the HARQ model, we consider three alternative HAR type specifications. More specifically, we include both the HAR-with-Jumps (HAR-J) and the Continuous-HAR (CHAR) proposed by Andersen et al. (2007), as well as the SHAR model proposed by Patton and Sheppard (2015), in the forecasting comparisons. Based on the Bi-Power Variation (BPV) measure of Barndorff-Nielsen and Shephard (2004), HAR-J and CHAR decompose the total variation into a continuous and a discontinuous (jump) part.
The HAR-J model augments the standard HAR model with a measure of the jump variation;
where $ J_{t} \equiv \max \left[R V_{t}-B P V_{t}, 0\right] $, and the $ B P V $ measure is defined as,
with $ \mu_{1} = \sqrt{2 / \pi} = \mathbb{E}(|Z|) $, and $ Z $ is a standard normal random variable.
The CHAR model includes measures of the continuous component of the total variation as covariates;
Inspired by the semivariation measures of Barndorff-Nielsen et al. (2008), Patton and Sheppard (2015) propose the SHAR model, which, in contrast to the HAR model, effectively allows for asymmetric responses in volatility forecasts from negative and positive intraday returns. More specifically, when $ R V_{t}^{-} \equiv \sum_{i = 1}^{M} r_{t, i}^{2} \mathbb{I}_{\left\{r_{t, i} < 0\right\}} $ and $ R V_{t}^{+} \equiv \; \sum_{i = 1}^{M} r_{t, i}^{2} \mathbb{I}_{\left\{r_{t, i} > 0\right\}} $, the SHAR model is defined as:
To evaluate model performance, we consider the mean squared error (MSE) and the QLIKE loss, which, according to Patton (2011), both are robust to noise. MSE is defined as
where $ F_{t} $ refers to the one-period direct forecast. QLIKE is defined as
3.2.1. Daily forecasting horizon
Table 3 contains one-day-ahead forecasts for EURUSD and USDJPY. The table reports model performance, expressed as model loss normalized by the loss of the HAR model. Each row reflects a combination of estimation window and loss function. The lowest ratio on each row, highlighting the best, performing model, is in bold. We evaluate the models using both a rolling window $ (RW) $ and an expanding window $ (EW) $. In both cases, forecasts are derived from model parameters re-estimated each day with a fixed length $ RW $ comprised of the previous 1000 days, as well as an $ EW $ using all of the available observations. The sample sizes for $ EW $ thus range from 1000 to 3201 days. The results are consistent in that the HARQ-F model is the best performer for both currency pairs and across loss functions and estimation windows. The HARQ model is closest to HARQ-F. Neither HAR-J, CHAR, nor SHAR appear to consistently improve upon the standard HAR model.
Judging from Table 3, it is beneficial to include RQ as an explanatory variable when RV is measured inaccurately. However, precise measurement of RV becomes more difficult when RV is high, inducing a positive correlation between RV and RQ. At the same time, high $ RV $ often coincides with jumps. To clarify whether the performance of RQ-based models is due to jump dynamics, Table 4 further segments the results in Table 3 into forecasts for days when the previous day's $ RQ $ was very high (Top 5% RQ, Table 4b) and the remaining sample (Bottom 95% RQ, Table 4a). As this breakdown shows, the RQ-based models perform relatively well also during periods of non-extreme heteroscedasticity of RQ.
3.2.2. Longer forecast horizons
In practitioner applications, longer forecasts than one day are often of interest. We now extend our analysis to weekly and monthly horizons, using direct forecasts. The daily forecast analysis in subsubsection 3.2.1 indicates the lag order of RQ plays an important role in forecast accuracy. Hence, following Bollerslev et al. (2016), we consider the HARQ-h model, and adjust the lag corresponding to the specific forecast horizon only. Specifically, for the weekly and monthly forecasts analysed here, the relevant HARQ-h specifications become
and
respectively.
Table 5 presents in-sample parameter estimates across model specifications. The patterns observed here closely resemble those of the daily estimates detailed in Table 2. All coefficients on RQ ($ \beta_{1Q}, \beta_{2Q}, \beta_{3Q} $) are negative, except for the $ (h = 22) $ lag statistically significant. This indicates that capturing measurement errors is relevant also for forecast horizons beyond one day. The HARQ model consistently allocates greater weight to the daily lag compared to the standard HAR model. Similarly, the HARQ-h model predominantly allocates its weight towards the time-varying lag. The weights of the HARQ-F model on the different lags are relatively more stable when compared to the HARQ-h model.
Table 6 and Table 7 detail the out-of-sample performance for weekly and monthly forecasts, respectively. Notably, the HAR-J, CHAR, and SHAR models generally fail to demonstrate consistent improvements over the basic HAR model. This is a sharp contrast to the RQ-augmented models. The HARQ-F model outperforms the HAR model both for EURUSD and USDJPY for nearly all instances. Also, HARQ-h delivers forecasts that are relatively consistent with the HAR model. Judging from both weekly and monthly results, the inherent flexibility of the HARQ-F is beneficial also for longer-term forecasts. We note that, at the monthly forecasting horizon for USDJPY, there is some variability as to preferred Q-specifications. Also, in some monthly instances, the Diebold-Mariano null hypothesis of equal predictability cannot be rejected. This is not unreasonable, since the number of independent monthly observations naturally becomes lower than for corresponding shorter forecasting horizons, leading to higher parameter uncertainty and related noise in volatility estimates.
3.3. Robustness
3.3.1. Alternative HARQ Specifications
The intention of the HARQ model is to capture the heteroskedastic measurement error of realized variance. The HARQ model in (5) approximates this through the square root of $ RQ $. Bollerslev et al. (2016) argues that this encounters possible issues with numerical stability. Still, this specification is somewhat ad-hoc and a number of reasonable alternatives exist. To clarify whether the performance of the HARQ model is sensitive to the definition of $ RQ $, we follow Bollerslev et al. (2016) and substitute $ R Q, R Q^{-1 / 2}, R Q^{-1} $, and $ \log (R Q) $ in place of $ R Q^{1/2} $. Furthermore, we augment the standard HAR and HARQ models with $ R Q^{1 / 2} $ as an additional explanatory variable, which allows the $ \operatorname{HAR}(\mathrm{Q}) $ model intercept to be time-varying.
Table 8 reports the out-of-sample forecast results from the alternative HARQ specifications. We normalize all losses by those of the HARQ model based on $ R Q^{1/2} $.
The two rightmost columns of Table 8 reveal that including $ R Q^{1/2} $ as an explanatory variable in the HAR and HARQ models does not lead to improved forecasts. Similarly, applying alternative RQ transformations does not appear to be helpful. Overall, we conclude that the HARQ model demonstrates greater stability and is generally favored over the alternative specifications.
3.3.2. Alternative Q-Models
HARQ is essentially an expansion of the HAR model. In a similar vein, the other benchmark volatility models can be extended accordingly. Following Bollerslev et al. (2016), from the HAR-J model defined in (3.2), we construct the HARQ-J model;
Furthermore, from the CHAR model defined in (3.2), we construct the CHARQ model;
Lastly, from the SHAR model defined in (3.2), we construct the SHARQ model;
Table 9 compares out-of-sample forecast results from each of the alternative Q-models (HARQ-J, CHARQ, and SHARQ), to their non-Q adjusted baseline specification. We also include the HARQ model. For both currencies, the enhancements seen in the HARQ-J and CHARQ models align with those observed in the basic HARQ model. This is in contrast to the SHARQ model, which is outperformed by SHAR. Bollerslev et al. (2016) report similar results.
3.3.3. Subsample analysis
Recent history contains two independent events that separately have induced turbulence in the global macroeconomy and financial markets. One is the outbreak of COVID-19 in March 2020; another is the Russian invasion of Ukraine in the second half of 2022, as illustrated in Figure 1.
To analyze this period of extreme market conditions in isolation, we perform a sub-sample analysis covering 2020–2022. Table 10 contains out-of-sample results for day-ahead volatility forecasts. Reassuringly, the overall results remain intact, in that the HARQ-F model is the best performing model also when this extreme period is considered in isolation.
4.
Conclusions
This study uses updated tick-level data from two major currency pairs, EURUSD and USDJPY, covering January 2010 to December 2022, to investigate the relevance of realized quarticity for out-of-sample volatility forecasts. We find that realized quarticity effectively captures noise caused by measurement errors, as evidenced by increased precision in daily, weekly, and monthly volatility estimates from models augmented with realized quarticity as an additionally explanatory variable. These results are robust across estimation windows, evaluation metrics, and model specifications. As such, the results conform to comparable studies from other markets, predominantly on equity indices and single stocks. This paper also complements the relatively scarce body of literature on foreign exchange markets in this context.
A myriad of volatility models based on the HAR framework have been proposed. Still, simple linear HAR specifications have proven remarkably difficult to beat, as shown by Audrino et al. (2024) and Branco et al. (2024). In a recent survey, Gunnarsson et al. (2024) report promising results for machine learning models and volatility forecasting across asset classes. The FX implied volatility surface contains a rich set of relevant predictive information across forecasting horizons and quantiles (de Lange et al., 2022). Thus, combining implied volatilities and high-frequency data using machine learning models, along the lines of Blom et al. (2023), appears as an interesting avenue for future research.
Rarely, one single model dominates others in terms of statistical and economic criteria. To this end, investigating ensemble models where high-frequency models are combined with other volatility model classes, such as time series models and stochastic volatility models-possibly including jump-processes, should be of interest. The recently developed rough-path volatility models based on fractional Brownian motion (Salmon and SenGuptz, 2021; Bayer et al., 2023) appear particularly relevant in this context.
Use of AI tools declaration
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
Author contributions
M.R.: Conceptualization, Methodology, Software, Formal analysis, Writing - Original Draft, Writing - Review & Editing.
M.H.: Data Curation, Writing - Original Draft, Writing - Review & Editing.
Acknowledgments
We would like to thank Andrew Patton for making the Matlab code from Bollerslev et al. (2016) available at https://public.econ.duke.edu/ap172/. Furthermore, we are grateful for insightful comments from the Editor and two anonymous reviewers, which helped us improve the paper.
Conflict of interest
The authors declare no conflicts of interest.