[go: up one dir, main page]

Academia.eduAcademia.edu
DOI: xxxx Risk Analysis, Vol. xx, No. x, 2012 Quantile Uncertainty and Value-at-Risk Model Risk ⋆ Carol Alexander,1 José Marı́a Sarabia2 This paper develops a methodology for quantifying model risk in quantile risk estimates. The application of quantile estimates to risk assessment has become common practice in many disciplines, including hydrology, climate change, statistical process control, insurance and actuarial science and the uncertainty surrounding these estimates has long been recognized. Our work is particularly important in finance, where quantile estimates (called Value-at-Risk) have been the cornerstone of banking risk management since the mid 1980’s. A recent amendment to the Basel II Accord recommends additional market risk capital to cover all sources of ‘model risk’ in the estimation of these quantiles. We provide a novel and elegant framework whereby quantile estimates are adjusted for model risk, relative to a benchmark which represents the state of knowledge of the authority that is responsible for model risk. A simulation experiment in which the degree of model risk is controlled illustrates how to quantify model risk and compute the required regulatory capital add-on for banks. An empirical example based on real data shows how the methodology can be put into practice, using only two time series (daily Value-at-Risk and daily profit and loss) from a large bank. We conclude with a discussion of potential applications to non-financial risks. KEY WORDS: Basel II, maximum entropy, model risk, quantile, risk capital, value-at-risk 1. INTRODUCTION This paper focuses on the model risk of quantile risk assessments with particular reference to ‘Value-at-Risk’ (VaR) estimates, which are derived from quantiles of portfolio profit and loss (P&L) distributions. VaR corresponds to an amount that could be lost, with a specified probability, if the portfolio remains unmanaged over a specified time horizon. It has become the global standard for assessing risk in all types of financial firms: in fund management, where portfolios with long-term VaR 1 ICMA Centre, Henley Business School at the University of Reading, Reading, RG6 6BA, UK; c.alexander@icmacentre.rdg.ac.uk 2 Department of Economics, University of Cantabria, Avda. de los ⋆ Castros s/n, 39005-Santander, Spain; sarabiaj@unican.es Address correspondence to Carol Alexander, Chair of Risk Management, ICMA Centre, Henley Business School at the University of Reading, Reading, RG6 6BA, UK; Phone: +44 118 3786431 objectives are actively marketed; in the treasury divisions of large corporations, where VaR is used to assess position risk; and in insurance companies, who measure underwriting and asset management risks in a VaR framework. But most of all, banking regulators remain so confident in VaR that its application to computing market risk capital for banks, used since the 1996 amendment to the Basel I Accord,3 will soon be extended to include stressed VaR under an amended Basel II and the new Basel III Accords.4 The finance industry’s reliance on VaR has been supported by decades of academic research. Especially during the last ten years there has been an explosion of articles published on this subject. Popular topics include the introduction of new VaR 3 See 4 See Basel Committee on Banking Supervision. (1) Basel Committee on Banking Supervision. (2,3) 1 0272-4332/xx/0100-0001$22.00/1 ✐ C 2012 Society for Risk Analysis 2 models,5 and methods for testing their accuracy.6 However, the stark failure of many banks to set aside sufficient capital reserves during the banking crisis of 2008 sparked an intense debate on using VaR models for the purpose of computing the market risk capital requirements of banks. Turner (18) is critical of the manner in which VaR models have been applied and Taleb (19) even questions the very idea of using statistical models for risk assessment. Despite the warnings of Turner, Taleb and early critics of VaR models such as Beder, (20) most financial institutions continue to employ them as their primary tool for market risk assessment and economic capital allocation. For internal, economic capital allocation purposes VaR models are commonly built using a ‘bottom-up’ approach. That is, VaR is first assessed at an elemental level, e.g. for each individual trader’s positions, then is it progressively aggregated into desk-level VaR, and VaR for larger and larger portfolios, until a final VaR figure for a portfolio that encompasses all the positions in the firm is derived. This way the traders’ limits and risk budgets for desks and broader classes of activities can be allocated within a unified framework. However, this bottom-up approach introduces considerable complexity to the VaR model for a large bank. Indeed, it could take more than a day to compute the full (often numerical) valuation models for each product over all the simulations in a VaR model. Yet, for regulatory purposes VaR must be computed at least daily, and for internal management intra-day VaR computations are frequently required. To reduce complexity in the internal VaR system simplifying assumptions are commonly used, in the data generation processes assumed for financial asset returns and interest rates and in the valuation models used to mark complex products to market every day. For instance, it is very common to apply normality assumptions in VaR models, along 5 Historical simulation (4) is the most popular approach amongst banks (5) but data-intensive and prone to pitfalls. (6) Other popular VaR models assume normal risk factor returns with the RiskMetrics covariance matrix estimates. (7) More complex VaR models are proposed by Hull and White, (8) Mittnik and Paolella, (9) , Ventner and de Jongh (10) , Angelidis et al. (11) , Hartz et al. (12) , Kuan et al. (13) and many others. 6 The coverage tests introduced by Kupiec (14) are favoured by banking regulators, and these are refined by Christoffersen. (15) However Berkowitz et al. (16) demonstrate that more sophisticated tests such as the conditional autoregressive test of Engle and Manganelli (17) may perform better. Alexander & Sarabia with lognormal, constant volatility approximations for exotic options prices and sensitivities.7 Of course, there is conclusive evidence that financial asset returns are not well represented by normal distributions. However, the risk analyst in a large bank may be forced to employ this assumption for pragmatic reasons. Another common choice is to base VaR calculations on simple historical simulation. Many large commercial banks have legacy systems that are only able to compute VaR using this approach, commonly basing calculations on at least 3 years of daily data for all traders’ positions. Thus, some years after the credit and banking crisis, vastly over-inflated VaR estimates were produced by these models long after the markets returned to normal. The implicit and simplistic assumption that history will repeat itself with certainty – that the banking crisis will recur within the risk horizon of the VaR model – may well seem absurd to the analyst, yet he is constrained by the legacy system to compute VaR using simple historical simulation. Thus, financial risk analysts are often required to employ a model that does not comply with their views on the data generation processes for financial returns, and data that they believe are inappropriate.8 Given some sources of uncertainty a Bayesian methodology (21,22) provides an alternative framework to make probabilistic inferences about VaR, assuming that VaR is described in terms of a set of unknown parameters. Bayesian estimates may be derived from posterior parameter densities and posterior model probabilities which are obtained from the prior densities via Bayes theorem, assuming that both the model and its parameters are uncertain. Our method shares ideas with the Bayesian approach, in the sense that we use a ‘prior’ distribution for α̂, in order to obtain a posterior distribution for the quantile. The problem of quantile estimation under model and parameter uncertainty has also been studied from a classical (i.e. non-Bayesian) point of view. Modarres, Nayak and Gastwirth (23) considered the accuracy of upper and extreme tail estimates of 7 Indeed, model risk frequently spills over from one business line to another, e.g. normal VaR models are often employed in large banks simply because they are consistent with the geometric Brownian motion assumption that is commonly applied for option pricing and hedging. 8 Banking regulators recommend 3-5 years of data for historical simulation and require at least 1 year of data for constructing the covariance matrices used in other VaR models. Quantile Uncertainty and Value-at-Risk Model Risk three right skewed distributions (log-normal, loglogistic and log-double exponential) under model and parameter uncertainty. These authors examined and compared performances of the maximum likelihood and non-parametric estimators based on the empirical or a quasi-empirical quantile function, assuming four different scenarios: the model is correctly specified, the model is mis-specified, the best model is selected using the data and no form is assumed for the model. Giorgi and Narduzzi (24) have studied quantile estimation for a self-similar time series and uncertainty that affects their estimates. Figlewski (25) deals with estimation error in the assessment of financial risk exposure. This author finds that, under stochastic volatility, estimation error can increase the probabilities of multi-day events such as three 1% tail events in a row, by several orders of magnitude. Empirical findings are also reported using 40 years of daily S&P 500 returns. The term ‘model risk’ is commonly applied to encompass various sources of uncertainty in statistical models, including model choice and parameter uncertainty. In July 2009, revisions to the Basel II market risk framework added the requirement that banks set aside additional reserves to cover all sources of model risk in the internal models used to compute the market risk capital charge.9 Thus, the issue of model risk in internal risk models has recently become very important to banks. Financial risk research has long recognized the importance of model risk. However, following some early work (26,27,28,29,30,31) surprizingly few papers deal explicitly with VaR model risk. Early work (32,33) investigated sampling error and Kerkhof et al. (34) quantify the adjustment to VaR that is necessary for some econometric models to pass regulatory backtests. Quantile-based risk assessment has also been applied to numerous problems in insurance and actuarial science: see Reiss and Thomas, (35) Cairns, (36) Matthys et al., (37) Dowd and Blake (38) and many others. However, a general methodology for assessing quantile model risk in finance has yet to emerge. This paper introduces a new framework for measuring quantile model risk and derives an elegant, intuitive and practical method for computing the risk capital add-on to cover VaR model risk. In addition to the computation of a model risk ‘addon’ for a given VaR model and given portfolio, our 9 See Basel Committee on Banking Supervision, Section IV. (3) 3 approach can be used to assess which, of the available VaR models, has the least model risk relative to a given portfolio. Similarly, given a specific VaR model, our approach can assess which portfolio has the least model risk. However, outside of a simulation environment, the concept of a ‘true’ model against which one might assess model risk is meaningless. All we have is some observable data and our beliefs about the conditional and/or unconditional distribution of the random variable in question. As a result, model risk can only be assessed relative to some benchmark model, which itself is a matter for subjective choice. In the following: the definition of model risk and a benchmark for assessing model risk is discussed in Section 2; Section 3 gives a formal definition of quantile model risk and outlines a framework for its quantification. We present a statistical model for the probability α̂ that is assigned, under the benchmark distribution, to the α quantile of the model distribution. Our idea is to endogenize model risk by using a distribution for α̂ to generate a distribution for the quantile. The mean of this model-risk-adjusted quantile distribution detects any systematic bias in the model’s α quantile, relative to the α quantile of the benchmark distribution. A suitable quantile of the model-risk-adjusted distribution determines an uncertainty buffer which, when added to the bias-adjusted quantile gives a model-riskadjusted quantile that is no less than the α quantile of the benchmark distribution at a pre-determined confidence level, this confidence level corresponding to a penalty imposed for model risk; Section 4 presents a numerical example on the application of our framework to VaR model risk, in which the degree of model risk is controlled by simulation; Section 5 illustrates how the methodology could be implemented by a manager or regulator having access to only two time series from the bank: its aggregate daily trading P&L and its corresponding 1% VaR estimates, derived from the usual ‘bottom up’ VaR aggregation framework; Section 6 discusses the relevance of the methodology to non-financial problems; and Section 7 summarizes and concludes. 2. MODEL RISK AND THE BENCHMARK We distinguish two sources of model risk: model choice, i.e. inappropriate assumptions about the form of the statistical model for the random variable; and parameter uncertainty, i.e. estimation error in the parameters of the chosen model. One never knows the 4 ‘true’ model except in a simulation environment, so assumptions about the form of statistical model must be made. Parameter uncertainty includes sampling error (parameter values can never be estimated exactly because only a finite set of observations on a random variable are available) and optimization error (e.g. different numerical algorithms typically produce slightly different estimates based on the same model and the same data). We remark that there is no consensus on the sources of model risk. For instance, Cont (39) points out that both these sources could be encompassed within a universal model, and Kerkhof et al. (34) distinguish ‘identification risk’ as an additional source. Model risk in finance has been approached in two different ways: examining all feasible models and evaluating the discrepancy in their results, or specifying a benchmark model against which model risk is assessed. Papers on the quantification of valuation model risk in the risk-neutral measure exemplify each approach: Cont (39) quantifies the model risk of a complex product by the range of prices obtained under all possible valuation models that are calibrated to market prices of liquid (e.g. vanilla) options; Hull and Suo (40) define model risk relative to the implied price distribution, i.e. a benchmark distribution implied by market prices of vanilla options. In the context of VaR model risk the benchmark approach, which we choose to follow, is more practical than the former. Some authors identify model risk with the departure of a model from a ‘true’ dynamic process: see Branger and Schlag (41) for instance. Yet, outside of an experimental or simulation environment, we never know the ‘true’ model for sure. In practice, all we can observe are realizations of the data generation processes for the random variables in our model. It is futile to propose the existence of a unique and measurable ‘true’ process because such an exercise is beyond our realm of knowledge. However, we can observe a maximum entropy distribution (MED). This is based on a ‘state of knowledge’, i.e. no more and no less than the information available regarding the random variable’s behaviour. This information includes the observable data that are thought to be relevant plus any subjective beliefs. Since neither the choice of sample nor the beliefs of the modeller can be regarded as objective, the MED is subjective. For our application to VaR we consider two perspectives on the MED, the internal perspective where the MED would be set by the risk analyst himself, or else by Alexander & Sarabia the Chief Risk Officer of the bank, and the external perspective where the MED would be set by the regulator. Shannon (43) defined the entropy of a probability density function g(x), x ∈ R as Z g(x) log g(x)dx. H(g) = −Eg [log g(x)] = − R This is a measure of the uncertainty in a probability distribution and its negative is a measure of information.10 The maximum entropy density is the function f (x) that maximizes H(g), subject to a set of conditions on g(x) which capture the testable information.11 The criterion here is to be as vague as possible (i.e. to maximize uncertainty) given the constraints imposed by the state of knowledge. This way, the MED represents no more (and no less) than the information available. If this information consists only of a historical sample on X of size n then, in addition to the normalization condition, there are n conditions on g(x), one for each data point. In this case, the MED is just the empirical distribution based on that sample. Otherwise, the testable information consists of fewer conditions, which capture only that sample information which is thought to be relevant, and any other conditions imposed by subjective beliefs. Our recommendation is that banks assess their VaR model risk by comparing their aggregate VaR figure, which is typically computed using the bottomup approach, with the VaR obtained using the MED in a ‘top-down’ approach, i.e. calibrated directly to the bank’s aggregate daily trading P&L. Typically this P&L contains marked-to-model prices for illiquid products, in which case their valuation model risk is not quantified in our framework. From the banking regulator’s perspective what matters is not the ability to aggregate and disaggregate VaR in a bottom-up framework, but the adequacy of a bank’s total market risk capital 10 For instance, if g is normal with variance σ 2 , H(g) = so the entropy increases as σ increases and there is more uncertainty and less information in the distribution. As σ → 0 and the density collapses the Dirac function at 0, there is no uncertainty but −H(g) → ∞ and there is maximum information. However, there is no universal relationship between variance and entropy and where their orderings differ entropy is the superior measure of information. See Ebrahimi, Maasoumi and Soofi (42) for further insight. 11 A piece of information is testable if it can be determined whether F is consistent with it. One of piece of information is always a normalization condition. 1 (1 + log(2π) + log(σ)), 2 Quantile Uncertainty and Value-at-Risk Model Risk reserves, which are derived from the aggregate market VaR. Therefore, regulators only need to define a benchmark VaR model to apply to the bank’s aggregate daily P&L. This model will be the MED of the regulator, i.e. the model that best represents the regulator’s state of knowledge regarding the accuracy of VaR models. Following the theoretical work of Shannon, (43) Zellner, (44) Jaynes (45) and many others it is common to assume the testable information is given by a set of moment functions derived from a sample, in addition to the normalization condition. When only the first two standard moments (mean and variance) are deemed relevant, the MED is a normal distribution. (43) More generally, when the testable information contains the first N sample moments, f (x) takes an exponential form. This is found by maximizing entropy subject to the conditions Z xn g(x)dx, n = 0, . . . , N, µn = R where µ0 = 1 and µn , n = 1, ..., N are the moments of the distribution. The solution is à n=N ! X n f (x) = exp − λn x , n=0 where the parameters λ0 , . . . λn are obtained by solving the system of non-linear equations à n=N ! Z X µn = xn exp − λn xn dx, n = 0, . . . , N. n=0 Rockinger and Jondeau, (46) Wu, (47) Chan (48,49) and others have applied a simple four-moment MED to various econometric and risk management problems. Perhaps surprizingly, since tail weight is an important aspect of financial asset returns distributions, none consider the tail weight that is implicit in the use of an MED based on standard sample moments. But simple moment-based MEDs are only well-defined when N is even. For any odd value of N there will be an increasing probability weight in one of the tails. Also, the four-moment MED has lighter tails than a normal distribution, due to the presence of the term exp[−λ4 x4 ] with nonzero λ4 in f (x). Indeed, the more moments included in the conditions, the thinner the tail of the MED. Because financial asset returns are typically heavytailed it is likely that this property will carry over to a bank’s aggregate daily P&L, in which case we would not advocate the use of simple moment-based MEDs. 5 Park and Bera (50) address the issue of heavy tails in financial data by introducing additional parameters into the moment functions, thus extending the family of moment-based MEDs. Even with just two (generalized) moment conditions based on one additional parameter they show that many heavy-tailed distributions are MEDs, including the Student t and generalized error distributions that are commonly applied to VaR analysis – see Jorion (32) and Lee et al. (51) for example. Since our paper concerns the estimation of low-probability quantiles we shall utilize these distributions as MEDs in our empirical study of Section 5. There are advantages in choosing a parametric MED for the benchmark. VaR is a quantile of a forward-looking P&L distribution, but to base parameter estimates entirely on historical data limits beliefs about the future to experiences from the past. Parametric distributions are frequently advocated for VaR estimation, and stress testing in particular, because the parameters estimated from historical data may be changed subjectively to accommodate beliefs about the future P&L distribution. We distinguish two types of parametric MEDs. Unconditional MEDs are based on the independent and identically distributed (i.i.d.) assumption. However, since Mandelbrot (52) it has been observed that financial asset returns typically exhibit a ‘volatility clustering’ effect, thus violating the i.i.d. assumption. Therefore it may be preferable to assume the stochastic process for returns has time-varying conditional distributions that are MEDs. Volatility clustering is effectively captured by the flexible and popular class of generalized conditional heteroskedastic models (GARCH) models introduced by Bollerslev (53) and since extended in numerous ways by many authors. Berkowitz and O’Brien (54) found that most bottom-up internal VaR models produced VaR estimates that were too large, and insufficiently risk-sensitive, compared with topdown GARCH VaR estimates derived directly from aggregate daily P&L. Thus, from the regulator’s perspective, a benchmark for VaR model risk based on a GARCH process for aggregate daily P&L with conditional MEDs would seem appropriate. Filtered historical simulation of aggregate daily P&L would be another popular alternative, especially when applied with a volatility filtering that increases its risk sensitivity: see Barone-Adesi et al. (55) and Hull and White. (56) Alexander and Sheedy (57) demonstrated empirically that GARCH volatility filtering 6 Alexander & Sarabia combined with historical simulation can produce very accurate VaR estimates, even at extreme quantiles. By contrast, the standard historical simulation approach, which is based on the i.i.d. assumption, failed many of their backtests. The α quantile of a continuous distribution F of a real-valued random variable X with range R is denoted (1) In financial applications the probability α is often predetermined. Frequently it will be set by senior managers or regulators and small or large values corresponding to extreme quantiles are very commonly used. For instance, regulatory market risk capital is based on VaR models with α = 1% and a risk horizon of 10 trading days. In our statistical framework F is identified with the unique MED based on a state of knowledge K which contains all testable information on F . We characterise a statistical model as a pair {F̂ , K̂} where F̂ is a distribution and K̂ is a filtration which encompasses both the model choice and its parameter values. The model provides an estimate F̂ of F , and uses this to compute the α quantile. That is, instead of (1) we use qαF̂ = F̂ −1 (α). (2) Quantile model risk arises because {F̂ , K̂} 6= {F, K}. Firstly, K̂ 6= K, e.g. K may include the belief that only the last six months of data are relevant to the quantile today; yet K̂ may be derived from an industry standard that must use at least one year of observed data in K̂;12 and secondly, F̂ is not, typically, the MED even based on K̂, e.g. the execution of firm-wide VaR models for a large commercial bank may present such a formidable time challenge that F̂ is based on simplified data generation processes, as discussed in the introduction. In the presence of model risk the α quantile of the model is not the α quantile of the MED, i.e. qαF̂ 6= qαF . The model’s α quantile qαF̂ is at a different quantile of F and we use the notation α̂ for this quantile, i.e. qαF̂ = qα̂F , or equivalently, 12 As is the case under current banking regulations for the use of VaR to estimate risk capital reserves - see Basel Committee on Banking Supervision.(1) (3) In the absence of model risk α̂ = α for every α. Otherwise, we can quantify the extent of model risk by the deviation of α̂ from α, i.e. the distribution of the quantile probability errors e(α|F, F̂ ) = α̂ − α. 3. MODELLING QUANTILE MODEL RISK qαF = F −1 (α). α̂ = F (F̂ −1 (α)). (4) If the model suffers from a systematic, measurable bias at the α quantile then the mean error ē(α|F, F̂ ) should be significantly different from zero. A significant and positive (negative) mean indicates a systematic over (under) estimation of the α quantile of the MED. Even if the model is unbiased it may still lack efficiency, i.e. the dispersion of e(α|F, F̂ ) may be high. Several measures of dispersion may be used to quantify the efficiency of the model, including the root mean squared error (RMSE), the mean absolute error (MAE) and the range. We now regard α̂ = F (F̂ −1 (α)) as a random variable with a distribution that is generated by our two sources of model risk, i.e. model choice and parameter uncertainty. Because α̂ is a probability it has range [0, 1], so the α quantile of our model, adjusted for model risk, falls into the category of generated random variables. For instance, if α̂ is parameterized by a beta distribution B(a, b) with density (0 < u < 1) fB (u; a, b) = B(a, b)−1 [ua−1 (1 − u)b−1 ], (5) a, b ≥ 0, where B(a, b) is the beta function, then the α quantile of our model, adjusted for model risk, is a beta-generated random variable: Q(α|F, F̂ ) = F −1 (α̂), α̂ ∼ B(a, b). Beta generated distributions were introduction by Eugene et al. (58) and Jones. (59) They may be characterized by their density function (−∞ < x < ∞) gF (x) = B(a, b)−1 f (x)[F (x)]a−1 [1 − F (x)]b−1 , where f (x) = F ′ (x). Several other distributions D[0, 1] with range the unit interval are available for generating the model-risk-adjusted quantile distribution; see Zografos and Balakrishnan(59) for example. Hence, in the most general terms the model-riskadjusted VaR is a random variable with distribution: Q(α|F, F̂ ) = F −1 (α̂), α̂ ∼ D[0, 1]. (6) The mean E[Q(α|F, F̂ )] of Q(α|F, F̂ ) quantifies any systematic bias in the quantile estimates: e.g. if the MED has heavier tails than the model then Quantile Uncertainty and Value-at-Risk Model Risk extreme quantiles qαF̂ will be biased: if α is close to zero then E[Q(α|F, F̂ )] > qαF and if α is close to one then E[Q(α|F, F̂ )] < qαF . This bias can be removed by adding the difference qαF − E[Q(α|F, F̂ )] to the model’s α quantile qαF̂ so that the bias-adjusted quantile has expectation qαF . The bias-adjusted α quantile estimate could still be far away from the maximum entropy α quantile: the more dispersed the distribution of Q(α|F, F̂ ), the greater the potential for qαF̂ to deviate from qαF . Because financial regulators require VaR estimates to be conservative, we adjust for the inefficiency of the VaR model by introducing an uncertainty buffer to the bias-adjusted α quantile by adding a quantity equal to the difference between the mean of Q(α|F, F̂ ) and G−1 F (y), the y quantile of Q(α|F, F̂ ), to the bias-adjusted α quantile estimate. This way, we become (1 − y)% confident that the model-riskadjusted α quantile is no less than qαF . Finally, our point estimate for the model-riskadjusted α quantile becomes: F̂ qα + F − E[Q(α|F, F̂ )]} + {E[Q(α|F, F̂ )] − G−1 {qα F (y)} = F F̂ − G−1 + qα qα F (y), (7) where {qαF − E[Q(α|F, F̂ )]} is the ‘bias adjustment’ and {E[Q(α|F, F̂ )] − G−1 F (y)} is the ‘uncertainty buffer’. The total model-risk adjustment to the quantile estimate is thus qαF − G−1 F (y), and the computation of E[Q(α|F, F̂ )] could be circumvented if the decomposition into bias and uncertainty components is not required. The confidence level 1 − y reflects a penalty for model risk which could be set by the regulator. When X denotes daily P&L and α is small (e.g. 1%), typically all three terms on the right hand side of (7) will be negative. But the α% daily VaR is −qαF̂ , so the model-risk-adjusted VaR estimate becomes −qαF̂ − qαF + G−1 F (y). The add-on to the F daily VaR estimate, G−1 F (y) − qα , will be positive unless VaR estimates are typically much greater than the benchmark VaR. In that case there should be a negative bias adjustment, and this could be large enough to outweigh the uncertainty buffer, especially when y is large, i.e. when we require only a low degree of confidence for the model-risk-adjusted VaR to exceed the benchmark VaR. 4. NUMERICAL EXAMPLE We now describe an experiment in which a portfolio’s returns are simulated based on a known 7 data generation process. This allows us to control the degree of VaR model risk and to demonstrate that our framework yields intuitive and sensible results for the bias and inefficiency adjustments described above. Recalling that the popular and flexible class of GARCH models was advocated by Berkowitz and O’Brien (54) for top-down VaR estimation we assume that our conditional MED for the returns Xt at time t is N (0, σt2 ), where σt2 follows an asymmetric GARCH process. The model falls into the category of maximum entropy ARCH models introduced by Park and Bera,(49) where the conditional distribution is normal. Thus it has only two constraints, on the conditional mean and variance. First the return xt from time t to t + 1 and its variance σt2 are simulated using: 2 σt2 = ω+α(xt−1 −λ)2 +βσt−1 , xt |It ∼ N (0, σt2 ),(8) where ω > 0, α, β ≥ 0, α + β ≤ 1 and It = (xt−1 , xt−2 , . . .).13 For the simulated returns the parameters of (8) are assumed to be: ω = 1.5 × 10−6 , α = 0.04, λ = 0.005, β = 0.95, (9) and so the steady-state annualized volatility of the portfolio return is 25%.14 Then the MED at time t is Ft = F (Xt |Kt ), i.e. the conditional distribution of the return Xt given the state of knowledge Kt , which comprises the observed returns It and the knowledge that Xt |It ∼ N (0, σt2 ). At time t, a VaR model provides a forecast F̂t = F̂ (Xt |K̂t ) where K̂ comprises It plus the model Xt |It ∼ N (0, σ̂t2 ). We now consider three different models for σ̂t2 . The first model has the correct choice of model but uses incorrect parameter values: instead of (9) the fitted model is: 2 , σ̂t2 = ω̂ + α̂(xt−1 − λ̂)2 + β̂ σ̂t−1 (10) with ω̂ = 2 × 10−6 , α̂ = 0.0515, λ̂ = 0.01, β̂ = 0.92. (11) The steady-state volatility estimate is therefore correct, but since α̂ > α and β̂ < β the fitted volatility process is more ‘jumpy’ than the simulated variance 13 We employ the standard notation α for the GARCH return parameter here; this should not be confused with the notation α for the quantile of the returns distribution, which is also standard notation in the VaR model literature. 14 The steady-state variance is σ̄ 2 = (ω + αλ2 )/(1 − α − β) and for the annualization we have assumed returns are daily, and that there are 250 business days per year. 8 Alexander & Sarabia generation process. In other words, compared with σt , σ̂t has a greater reaction but less persistence to innovations in the returns, and especially to negative returns since λ̂ > λ. The other two models are chosen because they are commonly adopted by financial institutions, having been popularized by the ‘RiskMetrics’ methodology introduced by JP Morgan in the mid1990’s – see RiskMetrics. (7) The second model uses a simplified version of (8) with: ω̂ = λ̂ = 0, α̂ = 0.06, β̂ = 0.94. (12) This is the RiskMetrics exponentially weighted moving average (EWMA) estimator in which a steady-state volatility is not defined. The third model is the RiskMetrics ‘Regulatory’ estimator in which: 250 1 X 2 α̂ = λ̂ = β̂ = 0, ω̂ = x . 250 i=1 t−i (13) 10,000 is A time series of 10,000 returns {xt }t=1 simulated from the ‘true’ model (8) with parameters (9). Then, for each of the three models defined above we use this time series to (a) estimate the daily VaR, which when expressed as a percentage of the portfolio value is given by Φ−1 (α)σ̂t , and (b) compute the probability α̂t associated with this quantile under the simulated returns distribution Ft = F (Xt |Kt ). Because Φ−1 (α̂t )σt = Φ−1 (α)σ̂t , this is given by · ¸ σ̂t −1 α̂t = Φ Φ (α) . (14) σt Now for each VaR model we use the simulated distribution to estimate α̂ at every time point, using (14). For α = 0.1%, 1% and 5%, Table I reports the mean of α̂ and the RMSE between α̂ and α. The closer α̂ is to α, the smaller the RMSE and the less model risk there is in the VaR model. The Regulatory model yields an α̂ with the highest RMSE, for every α, so this has the greatest degree of model risk. The AGARCH model, which we already know has the least model risk of the three, produces a distribution for α̂ that has mean closest to the true α and the smallest RMSE. These observations are supported by Figure 1, which depicts the empirical distribution of α̂ and Figure 2, which shows the empirical densities of the model-risk-adjusted VaR estimates F −1 (α̂), taking α = 1% for illustration. Here and henceforth VaR is stated as a percentage of the portfolio value, multiplied by 100. A point estimate for model-risk-adjusted VaR (RaVaR, for short) is computed using (7). Because we have a conditional MED the benchmark VaR (BVaR, for short) depends on the time it is measured, and so does the RaVaR. For illustration, we select a point when the simulated volatility is at its steady-state value of 25% – so the BVaR is 4.886, 3.678 and 2.601 at the 0.1%, 1% and 5% levels, respectively. Drawing at random from the points when the simulated volatility was 25%, we obtain AGARCH, EWMA and Regulatory volatility forecasts of 27.02%, 23.94% and 28.19% respectively. These volatilities determine the VaR estimates that we shall now adjust for model risk. Table II summarizes the bias and the uncertainty buffer, for different levels of α, based on the empirical distribution of Q(α|F, F̂ ).15 It reveals a general tendency for the EWMA model to slightly underestimate VaR and the other models to slightly overestimate VaR. Yet the bias is relatively small, since all models assume the same normal form as the MED and the only difference between them is their volatility forecast. Although the bias tends to increase as α decreases it is not significant for any model.16 Beneath the bias we report the 5% quantile of the model-risk-adjusted VaR distributions, since we shall first compute the RaVaR so that it is no less than the BVaR with 95% confidence. Following the framework introduced in the previous section we now define: RaVaR(y) = VaR + (BVaR − E[Q(α|F, F̂ )]) + + (E[Q(α|F, F̂ )] − G−1 F (y)), where {BVaR − E[Q(α|F, F̂ )]} is the ‘bias adjustment’ and {E[Q(α|F, F̂ )] − G−1 F (y)} is the ‘uncertainty buffer’. Table III sets out the RaVaR computation for y = 5%. The model’s volatility forecasts are in the first row and the corresponding VaR estimates are in the first row of each cell, for α = 0.1%, 1% and 5% respectively. The (small) bias is corrected by adding the bias from Table II to each VaR estimate. The main source of model risk here concerns the potential for a large (positive or negative) errors in the quantile probabilities, i.e. the dispersion of the densities in Figure 2. To adjust for this we add to the bias15 Similar results based on the fitted distributions are not reported for brevity. 16 Standard errors of Q(α|F, F̂ ) are not reported, for brevity. They range between 0.157 for the AGARCH at 5% to 0.891 for the Regulatory model at 0.1%, and are directly proportional to the degree of model risk just like the standard errors on the quantile probabilities given in Table I . Quantile Uncertainty and Value-at-Risk Model Risk adjusted VaR the uncertainty buffer given in Table II . This gives the RaVaR estimates shown in the third row of each cell. Since risk capital is a multiple of VaR, the percentage increase resulting from replacing VaR by RaVaR(y) is: 9 Table I . Sample statistics for quantile probabilities. The mean of α̂ and the RMSE between α̂ and α. The closer α̂ is to α, the smaller the RMSE and the less model risk there is in the VaR model. α AGARCH EWMA Regulatory 0.10% Mean RMSE 0.11% 0.07% 0.16% 0.14% 0.23% 0.37% (15) 1% Mean RMSE 1.03% 0.42% 1.25% 0.64% 1.34% 1.22% The penalty (15) for model risk depends on α, except in the case that both the MED and VaR model are normal, and on the confidence level (1 − y)%. Table IV reports the percentage increase in risk capital due to model risk when RaVaR is no less than the BVaR with (1 − y)% confidence. We consider y = 5%, 15% and 25%, with smaller values of y corresponding to a stronger condition on the model-risk adjustment. We also set α = 1% because risk capital is based on the VaR at this level of significance under the Basel Accords. We also take the opportunity here to consider two further scenarios, in order to verify the robustness of our qualitative conclusions. The first row of each section of Table IV reports the volatilities estimated by each VaR model at a point in time when the benchmark model has volatility 25%. Thus for scenario 1, upon which the results have been based up to now, we have the volatilities 27.02%, 23.94% and 28.19% respectively. For scenario 2 the three volatilities are 28.32%, 27.34% and 22.40%, i.e. the AGARCH and EWMA models over-estimate and the Regulatory model under-estimates the benchmark model’s volatility. For scenario 3 the AGARCH model slightly underestimates the benchmark’s volatility and the other two models over-estimate it. The three rows in each section of the Table IV give the percentage increase in risk capital that would be required were the regulator to choose 95%, 85% or 75% confidence levels for the RaVaR. Clearly, for each model and each scenario, the add-on for VaR model risk increases with the degree of confidence that the regulator requires for the RaVaR to be at least as great as the BVaR. At the 95% confidence level, a comparison of the first row of each section of the table shows that risk capital would be increased by roughly 8–10% when based on the AGARCH model, whereas it would be increased by about 13– 14.5% under the EWMA model and by roughly 21– 27% under the Regulatory model. The same ordering of the RaVaRs applies to each scenario, and at every confidence level. That is, the model-risk adjustment 5% Mean RMSE 4.97% 1.03% 5.44% 1.31% 5.27% 2.66% % risk capital increase = G−1 F (y) BVaR − VaR . results in an increase in risk capital that is positively related to the degree of model risk, as it should be. Finally, comparison between the three scenarios shows that the add-on will be greater on days when the model under-estimates the VaR than it is on days when it over-estimates VaR, relative to the benchmark. Yet even when the model VaR is greater than the benchmark VaR the add-on is still positive. This is because the uncertainty buffer remains large relative to the bias adjustment, even at the 75% level of confidence. However, if regulators were to require a lower confidence for the uncertainty buffer, such as only 50% in this example, then it could happen that the model-risk add-on becomes negative. 5. EMPIRICAL ILLUSTRATION How could the authority responsible for model risk, such as a bank’s local regulator or its Chief Risk Officer, implement the proposed adjustment for model risk in practice? The required inputs to a model-risk-adjusted VaR calculation are two daily time series that the bank will have already been recording to comply with Basel regulations: one series is the aggregate daily trading P&L and the other is the aggregated 1% daily VaR estimates corresponding to this trading activity. From the regional office of a large international banking corporation we have obtained data on aggregate daily P&L and the corresponding aggregate VaR for each day, the VaR being computed in a bottom-up framework based on standard (un-filtered) historical simulation. The data span the period 3 Sept 2003 to 18 March 2009, thus including the banking crisis during the last quarter of 2008. In this section the bank’s daily VaR will be compared with a top-down VaR estimate based on a benchmark VaR model tuned to the aggregate daily trading P&L, and a 10 Alexander & Sarabia Table II . Components of the model-risk adjustment. The bias and the 95% uncertainty buffer, for different levels of α, derived from the mean and 5% quantile of the empirical distribution of Q(α|F, F̂ ). The bias is the difference between the benchmark VaR (which is 4.886, 3.678 and 2.601 at the 0.1%, 1% and 5% levels) and the mean. The uncertainty buffer is the difference between the mean and the 5% quantile. UB means ‘Uncertainty Buffer’ and Q is the quantile. AGARCH EWMA Regulatory Mean 5% Q Bias UB 4.919 4.447 -0.033 0.472 4.793 4.177 0.093 0.615 4.961 3.912 -0.075 1.049 Mean 5% Q Bias UB 3.703 3.348 -0.025 0.355 3.608 3.145 0.070 0.463 3.735 2.945 -0.056 0.789 Mean 5% Q Bias UB 2.618 2.366 -0.017 0.252 2.551 2.224 0.050 0.327 2.641 2.082 -0.040 0.559 α 0.1% 1% 5% Table III . Computation of 95% RaVaR. The volatility σ̂t determines the VaR estimates for α = 0.1%, 1% and 5% respectively, as Φ−1 (α) σ̂t . Adding the bias shown in Table II gives the bias-adjusted VaR. Adding the uncertainty buffer given in Table II to the bias-adjusted VaR yields the model-risk-adjusted VaR estimates (RaVAR) shown in the third row of each cell. BA means ‘Bias-Adjusted’. AGARCH EWMA Regulatory α Volatility 27.02% 23.94% 28.19% 0.10% VaR BA VaR RaVaR 5.277 5.244 5.716 4.678 4.772 5.387 5.509 5.434 6.483 1% VaR BA VaR RaVaR 3.972 3.948 4.303 3.522 3.592 4.055 4.147 4.091 4.880 5% VaR BA VaR RaVaR 2.809 2.791 3.043 2.490 2.540 2.867 2.932 2.892 3.451 model-risk-adjusted VaR will be derived for each day between 3 September 2007 and 18 March 2009. When the data are not i.i.d. the benchmark should be a conditional MED rather than an unconditional MED. To illustrate this we compute the time series of 1% quantile estimates based on alternative benchmarks. First we employ the Student Table IV . Percentage increase in risk capital from model-risk adjustment of VaR. The percentage increase from VaR to RaVaR based on three scenarios for each model’s volatility estimate at time t. In each case the benchmark model’s conditional volatility was 25%. AGARCH EWMA Regulatory Scenario 1 27.02% 23.94% 28.19% 95% 85% 75% 8.40% 5.23% 3.08% 14.46% 9.45% 6.53% 21.29% 13.93% 9.14% Scenario 2 28.32% 27.34% 22.40% 95% 85% 75% 8.02% 4.98% 2.94% 12.66% 8.27% 5.71% 26.79% 17.53% 11.50% Scenario 3 23.18% 26.34% 28.66% 95% 85% 75% 9.80% 6.09% 3.59% 13.14% 8.59% 5.93% 20.94% 13.70% 8.99% t distribution, which maximizes the Shannon entropy subject to the moment constraint17 µ ¶ µ 2¶ 1 + ν2 ν E[log(ν 2 +(X/λ)2 )] = log(ν 2 )+ψ −ψ . 2 2 Secondly we consider the AGARCH process (8) which has a normal conditional distribution for the errors. We also considered taking the generalized error distribution (GED), introduced by Nelson, (61) as an unconditional MED benchmark. The GED has the probability density (−∞ < x < ∞) ¶ µ 1 ν −1/ν f (x; λ, ν) = exp − |x/λ|ν , 2Γ(1 + 1/ν)λ ν and maximizes the Shannon entropy subject to ν −1/ν . the moment constraint E[ν −1 |X/λ|ν ] = 2Γ(1+1/ν) This is more flexible than the (always heavy-tailed) Student t because when ν < 2 (ν > 2) the GED has heavier (lighter) tails than the normal distribution. We also considered using a Student t conditional MED with the AGARCH process, and a symmetric GARCH process, where λ = 0 in (8), with both Student t and normal conditional distributions. However, the unconditional GED produced results similar to (and just as bad as) the unconditional 17 Here ψ(z) denotes the digamma function, λ is a scale parameter and ν the corresponding shape parameter. See Park and Bera,(49) Table 1 for the moment condition. Quantile Uncertainty and Value-at-Risk Model Risk Student t. Also all four conditional MEDs produced quite similar results. Our choice of the AGARCH process with normal errors was based on the successful results of the conditional and unconditional coverage tests that are commonly applied to test for VaR model specification – see Christoffersen. (15) For reasons of space, none of these results are reported but they are available from the authors on request. We estimate the two selected benchmark model parameters using a ‘rolling window’ framework that is standard practice for VaR estimation. Each sample contains n consecutive observations on the bank’s aggregate daily P&L, and a sample is rolled forward one day at a time, each time re-estimating the model parameters. Figure 3 compares the 1% quantile of the Student t distribution with the 1% quantile of the normal AGARCH process on the last day of each rolling window. Also shown is the bank’s aggregate P&L for the day corresponding to the quantile estimate, between 3 September 2007 and 18 March 2008. The effect of the banking crisis is evidenced by the increase in volatility of daily P&L which began with the shock collapse of Lehmann Brothers in mid September 2008. Before this time the 1% quantiles of the unconditional Student t distribution were very conservative predictors of daily losses, because the rolling windows included the commodity crisis of 2006. Yet at the time of the banking crisis the Student t model clearly underestimated the losses that were being experienced. Even worse, from the bank’s point of view, the Student t model vastly over-estimated the losses made during the aftermath of the crisis in early 2009 and would have led to crippling levels of risk capital reserves. Even though we used n = 200 for fitting the unconditional Student t distribution and a much larger sample, with n = 800, for fitting the normal AGARCH process it is clear that the GARCH process captures the strong volatility clustering of daily P&L far better than the unconditional MED. True, the AGARCH process often just misses a large unexpected loss, but because it has the flexibility to respond the very next day, the AGARCH process rapidly adapts to changing market conditions just as a VaR model should. In an extensive study of the aggregate P&L of several large commercial banks, Berkowitz and O’Brien (54) found that GARCH models estimated on aggregate P&L are far more accurate predictors of aggregate losses than the bottom-up VaR figures that most banks use for regulatory capital calculations. Figure 3 verifies this finding by also depicting the 1% daily VaR reported by the bank, multiplied by 11 −1 since it is losses rather than profits that VaR is supposed to cover. This time series has many features in common with the 1% quantiles derived from the Student t distribution. The substantial losses of up to $80m per day during the last quarter of 2008 were not predicted by the bank’s VaR estimates, yet following the banking crisis the bank’s VaR was far too conservative. We conclude, like Berkowitz and O’Brien, (54) that unconditional approaches are much less risk sensitive than GARCH models and for this reason we choose the normal AGARCH rather than the Student t as our benchmark for model risk assessment below. Figure 4 again depicts −1× the bank’s aggregate daily VaR, denoted −VaRt in the formula below. Now our bias adjustment is computed daily using an empirical model-risk-adjusted VaR distribution based on the observations in each rolling window. Under the normal AGARCH benchmark, for a sample starting at T and ending at T + n, the daily P&L distribution at time t, with T ≤ t ≤ T + n is N(0, σ̂t2 ) where σ̂t is the time-varying standard deviation of the¡ AGARCH ¢model. Following (3) we set α̂t = Φ − σ̂t−1 VaRt for each day in the window and then we use the empirical distribution of α̂t for T ≤ t ≤ T + n to generate the model-risk-adjusted VaR distribution (6). Then, following (7), the bias adjustment at T + n is the difference between the mean of the model-riskadjusted quantile distribution and the benchmark VaR at T + n. Since the bank’s aggregate VaR is very conservative at the beginning of the period but not large enough during the crisis, in Figure 4 a positive bias reduces the VaR prior to the crisis but during the crisis a negative bias increases the VaR. Having applied the bias adjustment we then set y = 25% in (7) to derive the uncertainty buffer corresponding to a 75% confidence that the RaVaR will be no less than the BVaR. This is the difference between the mean and the 25%-quantile of the model-risk-adjusted VaR distribution. Adding this uncertainty buffer to the bias-adjusted VaR we obtain the 75% RaVaR depicted in Figure 4 which is given by (7). This is more variable than the bank’s original aggregate VaR, but risk capital is based on an average VaR figure over the last 60 days (or the previous VaR, if this is greater) so the adjustment need not induce excessive variation in risk capital, which would be difficult to budget. 12 Alexander & Sarabia Fig. 1. Density of quantile probabilities. Empirical distribution of α̂t derived from (14) with α = 1%, based on 10,000 daily returns simulated using (8) with parameters (9). 0.25 AGARCH 0.2 EWMA Regulatory 0.15 0.1 0.05 0 0% 1% 2% 3% 4% 5% 6% Fig. 2. Density of model-risk-adjusted daily VaR. Empirical densities of the model-risk-adjusted VaR estimates F −1 (α̂) with α = 1%, based on the 10,000 observations on α̂ whose density is shown in Figure 1. 0.35 0.3 AGARCH EWMA 0.25 Regulatory 0.2 0.15 0.1 0.05 0 2 2.5 3 3.5 4 4.5 5 5.5 6 6. APPLICATION TO NON-FINANCIAL PROBLEMS Quantile-based risk assessment has become standard practice in a wide variety of non-financial disciplines, especially in environmental risk assessment and in statistical process control. For instance, applications to hydrology are studied by Arsenault and Ashkar (62) and Chebana and Ouarda, (63) and other environmental applications of quantile risk assessments include climate change (Katz et al., (64) and Diodato and Bellocchi, (65) ) wind power (66) and nuclear power (67) . In statistical process control, quantiles are used for computing capability indices, (68) for measuring efficiency (69) and for reliability analysis. (70) In these contexts the uncertainty surrounding quantile-based risk assessments has been the subject of many papers (36,71,72,73,74,75) . Both model choice and parameter uncertainty has been considered. For instance, Vermaat and Steerneman (76) discuss modified quantiles based on extreme value distributions in reliability analysis, and Sveinsson et al. (77) examine the errors induced by using a sample limited to a single site in a regional frequency analysis. As in banking, regulations can be a key driver for the accurate assessment of environmental risks such as radiation from nuclear power plants. Nevertheless, health or safety regulations are unlikely to extend as far as requirements for regular monitoring and reporting of quantile risks in the foreseeable future. The main concern about the uncertainty surrounding quantile risk assessment is more likely to come from senior management, in recognition that inaccurate risk assessment could jeopardize the reputation of the firm, profits to shareholders and/or the safety of the public. The question then arises: if it is a senior manager’s knowledge that specifies the benchmark distribution for model risk assessment, why should this benchmark distribution not be utilized in practice? As exemplified by the work of Sveinsson et al., (77) the time and expense of utilizing a complete sample of data may not be feasible except on a few occasions where a more detailed risk analysis is performed, possibly by an external consultant. In this case the most significant source of model risk in regular risk assessments would stem from parameter uncertainty. Model choice might also be a source of risk when realistic model assumptions would lead to systems that are too costly and time-consuming to employ on a daily basis. For instance, Merrik et al. (2005) point out that the use of Bayesian simulation for modelling large and complex maritime risk systems should be considered state-of-the-art, rather than standard practice. Also in reliability modelling, individual risk assessments for various components are typically aggregated to derive the total risk for the system. A full account of component default codependencies may require a lengthy scenario analyses based on a complex model (e.g. multivariate copulas with non-standard marginal distributions). This type of risk assessment might not be feasible every day, but if it could be performed on an occasional basis then it could be used as a benchmark for adjusting everyday quantile estimates for model risk. Generalizations and extensions to higher dimensions of the benchmark model could be implemented. A multivariate parametric MED for the benchmark model can be obtained using similar arguments to Quantile Uncertainty and Value-at-Risk Model Risk Fig. 3. Daily P&L, daily VaR and two potential benchmark VaRs. The bank’s daily P&L is depicted by the grey line and it’s ‘bottom-up’ daily VaR estimates by the black line. The dotted and dashed lines are the Student t (unconditional MED) benchmark VaR and the normal AGARCH (conditional MED) benchmark VaR. 80 13 Fig. 4. Aggregate VaR, bias adjustment and 75% RaVaR. The bank’s daily VaR estimates are repeated (black line) and compared with the bias-adjustment (grey line) and the final model-risk-adjusted VaR at the 75% confidence level (dotted line) based on the normal AGARCH benchmark model. 80 60 60 40 40 20 20 0 0 -20 -20 -40 -40 -60 -60 -80 -80 -100 -100 Daily VaR Daily P&L Student t Normal AGARCH Bias 75% RaVaR Daily VaR those used in the univariate case. In an engineering context, Kapur (78) have considered several classes of multivariate MED. Zografos (79) characterized Pearson’s type II and VII multivariate distributions, Aulogiaris and Zografos (80) the symmetric Kotz and Burr multivariate families and Bhattacharya (81) the class of multivariate Liouville distributions. Closed expressions for entropies in several multivariate distributions have been provided by Ahmed and Gokhale (82) and Zografos and Nadarajah. (83) A major difference between financial and nonfinancial risk assessment is the availability of data. For instance, in the example described in the previous section the empirical distributions for model-risk-adjusted quantiles were derived from several years of regular output from the benchmark model. Clearly, the ability to generate the adjusted quantile distribution from a parametric distribution for α̂, such as the beta distribution (5), opens the methodology to applications where relatively few observations on the benchmark quantile are available, but there are enough to estimate the parameters of a distribution for α̂. 7. SUMMARY This paper concerns the model risk of quantilebased risk assessments, with a focus on the risk of producing inaccurate VaR estimates because of an inappropriate choice of VaR model and/or inaccuracy in the VaR model parameter estimates. We develop a statistical methodology that provides a practical solution to the problem of quantifying the regulatory capital that should be set aside to cover this type of model risk, under the July 2009 Basel II proposals. We argue that there is no better choice of model risk benchmark than a maximum entropy distribution since, by definition, this embodies the entirety of information and beliefs, no more and no less. In the context of the model risk capital charge under the Basel II Accord the benchmark could be specified by the local regulator; more generally it should be specified by any authority that is concerned with model risk, such as the Chief Risk Officer. Then VaR model risk is assessed using a topdown approach to compute the benchmark VaR from the bank’s total daily P&L, and comparing this with the bank’s aggregate daily VaR, which is typically derived using a computationally intensive bottomup approach that necessitates many approximations and simplifications. The main ideas are as follows: in the presence of model risk an α quantile is at a different quantile of the benchmark model, and has an associated tail probability under the benchmark that is stochastic. Thus, the model-risk-adjusted quantile becomes a generated random variable and its distribution quantifies the bias and uncertainty due to model risk. A significant bias arises if the aggregate VaR estimates tend to be consistently above or below the benchmark VaR, and this is reflected in a significant difference between the mean of the modelrisk-adjusted VaR distribution and the benchmark VaR. Even when the bank’s VaR estimates have an insignificant bias, an adjustment for uncertainty is required because the difference between the bank’s VaR and the benchmark VaR could vary 14 considerably over time. The bias and uncertainty in the VaR model, relative to the benchmark, determine a risk capital adjustment for model risk whose size will also depend on the confidence level regulators require for the adjusted risk capital to be no less than the risk capital based on the benchmark model. Our framework was validated and illustrated by a numerical example which considers three common VaR models in a simulation experiment where the degree of model risk has been controlled. A further empirical example describes how the model-risk adjustment could be implemented in practice given only two time series, on the bank’s aggregate VaR and its aggregate daily P&L, which are in any case reported daily under banking regulations. Further research would be useful on backtesting the model-risk-adjusted estimates relative to commonly-used VaR models, such as the RiskMetrics models considered in this paper. Where VaR models are failing regulatory backtests and thus being heavily penalized or even disallowed, the top-down modelrisk-adjustment advocated in this paper would be very much more cost effective than implementing a new or substantially modified bottom-up VaR system. There is potential for extending the methodology to the quantile-based metrics that are commonly used to assess non-financial risks in hydrology, climate change, statistical process control and reliability analysis. In the case that relatively few observations on the model and benchmark quantiles are available, the approach should include a parameterization the model-risk-adjusted quantile distribution, for instance as a beta-generated distribution. Alexander & Sarabia 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. ACKNOWLEDGMENTS 19. The authors would like to thank to the associate editor and two anonymous reviewers for many constructive comments that improved the original version. The second author thanks Ministerio de Economı́a y Competitividad, Spain (ECO201015455) for partial support. 20. 21. 22. 23. 24. REFERENCES 1. Basel Committee on Banking Supervision. Amendment to the capital accord to incorporate market risks. Bank for International Settlements, Basel, 1996. 2. Basel Committee on Banking Supervision. International convergence of capital measurement and capital stan- 25. 26. 27. dards: A revised framework. Bank for International Settlements, Basel, 2006. Basel Committee on Banking Supervision. Revisions to the Basel II market risk framework. Bank for International Settlements, Basel, 2009. Hendricks D. Evaluation of Value-at-Risk models using historical data. FRBNY Economic Policy Review, 1996; April:39-69. Perignon C, Smith D. The level and quality of Value-atRisk disclosure by commercial banks. Journal of Banking and Finance, 2010; 34(2):362-377. Pritsker M. The hidden dangers of historical simulation. Journal of Banking and Finance, 2006; 30:561-582. RiskMetrics. Technical Document. Available from the RiskMetrics website, 1997. Hull J, White A. Value at risk when daily changes in market variables are not normally distributed. Journal of Derivatives, 1998; 5(3):9-19. Mittnik S, Paolella M. Conditional density and Value-atRisk: prediction of Asian currency exchange rates. Journal of Forecasting, 2000; 19:313-333. Venter J, de Jongh P. Risk estimation using the normal inverse Gaussian distribution. Journal of Risk, 2002; 4:123. Angelidis T, Benos A, Degiannakis S. The use of GARCH models in VaR estimation. Statistical Methodology, 2004; 1(2):105-128. Hartz C, Mittnik S, Paolella M. Accurate Value-at-Risk forecasting based on the normal-GARCH model. Computational Statistics and Data Analysis, 2006; 51(4):22952312. Kuan C-M, Yeh J-H, Hsu Y-C. Assessing Value-atRisk with CARE: the conditional autoregressive expectile models. Journal of Econometrics, 2009; 150(2):261-270. Kupiec P. Techniques for verifying the accuracy of risk measurement models. Journal of Derivatives, 1995; 3(2):73-84. Christoffersen P. Evaluating interval forecasts. International Economic Review, 1998; 39:841-862. Berkowitz J, Christoffersen P, Pelletier D. Evaluating VaR models with desk-level data. Management Science, Management Science, 2011; 57:2213-2227. Engle R, Manganelli S. CAViaR: Conditional autoregressive Value at Risk by regression quantile. Journal of Business and Economic Statistics, 2004; 22:367-381. Turner L. The Turner review: A regulatory response to the global banking crisis. Financial Services Authority, London, 2009. Taleb N. The black swan: The impact of the highly improbable. Penguin, 2007. Beder T. VAR: Seductive but dangerous. Financial Analysts Journal, 1995; 51(5):12-24. Bernardo JM, Smith AFM. Bayesian Theory. Wiley, Chichester, UK, 1994. O’Hagan A. Kendall’s Advanced Theory of Statistics. Vol 2B: Bayesian Inference. Edward Arnold, London, 1994. Modarres R, Nayak TK, Gastwirth JL. Estimation of upper quantiles model and parameter uncertainty. Computational Statistics and Data Analysis, 2002; 39:529-554. Giorgi G, Narduzzi C. Uncertainty of quantile estimates in the measurement of self-similar processes. In: International Workshop on Advanced Methods for Uncertainty Estimation Measurements, AMUEM 2008, Sardagna, Trento, Italy, 2008. Figlewski S. Estimation error in the assessment of financial risk exposure. SSRN-id424363, version of June 29, 2003. Derman E. Model risk. Risk, 1996; 9(5):139-145. Simons K. Model error - evaluation of various finance Quantile Uncertainty and Value-at-Risk Model Risk 28. 29. 30. 31. 32. 33. 34. 35. 36. 37. 38. 39. 40. 41. 42. 43. 44. 45. 46. 47. 48. 49. 50. models. New England Economic Review, 1997; NovDec:17-28. Crouhy M, Galai D, Mark R. Model risk. Journal of Financial Engineering, 1998; 7:267-288. Green T, Figlewski S. Market risk and model risk for a financial institution writing options. Journal of Finance, 1999; 54:1465-1499. Kato T, Yoshiba T. Model risk and its control. Monetary and Economic Studies, 2000; December:129-157. Rebonato R. Managing Model Risk. In Volume 2 of Mastering Risk, C. Alexander (ed.), Pearson UK, 2001. Jorion P. Measuring the risk in value at risk. Financial Analysts Journal, 1996; 52:47-56. Talay D, Zheng Z. Worst case model risk management. Finance and Stochastics, 2002; 6:517-537. Kerkhof J, Melenberg B, Schumacher H. Model risk and capital reserves. Journal of Banking and Finance, 2010; 34:267-279. Reiss R, Thomas M. Statistical Analysis of Extreme Values with Applications to Insurance, Finance, Hydrology and Other Fields. Birkhauser Verlag, Basel, 1997. Cairns A. A discussion of parameter and model uncertainty in insurance. Insurance: Mathematics and Economics, 2000; 27:313-330. Matthys G, Delafosse E, Guillou A, Beirlant J. Estimating catastrophic quantile levels for heavy-tailed distributions. Insurance: Mathematics and Economics, 2004; 34:517-537. Dowd K, Blake D. After VaR: The theory, estimation and insurance applications of quantile-based risk measures. Journal of Risk and Insurance, 2009; 73(2):193-229. Cont R. Model uncertainty and its impact on the pricing of derivative instruments. Mathematical Finance, 2006; 16(3):519-547. Hull J, Suo W. A methodology for assessing model risk and its application to the implied volatility function model. Journal of Financial and Quantitative Analysis, 2002; 37(2):297-318. Branger N, Schlag C. Model risk: A conceptual framework for risk measurement and hedging. Working Paper, EFMA Basel Meetings. Available from SSRN, 2004. Ebrahimi N, Maasoumi E, Soofi E. Ordering univariate distributions by entropy and variance. Journal of Econometrics, 1999; 90:317-336. Shannon C. The mathematical theory of communication, Bell System Technical Journal July-Oct. Reprinted in: C.E. Shannon and W. Weaver, The Mathematical Theory of Communication University of Illinois Press, 1948; Urbana, IL 3-91. Zellner A. Maximal data information prior distributions. In: A. Aykac and C. Brumat, eds., New methods in the applications of Bayesian methods. North-Holland, 1977. Jaynes E. Papers on Probability, Statistics and Statistical Physics. Edited by R.D. Rosenkrantz, Reidel Publishing Company, Dordrecht, 1983. Rockinger M, Jondeau E. Entropy densities with an application to autoregressive conditional skewness and kurtosis. Journal of Econometrics, 2002; 106:119-142. Wu X. Calculation of maximum entropy densities with application to income distribution. Journal of Econometrics, 2003; 115(2):347-354 Chan F. Modelling time-varying higher moments with maximum entropy density. Mathematics and Computers in Simulation, 2009; 79(9):2767-2778. Chan F. Forecasting value-at-risk using maximum entropy. 18th World IMACS / MODSIM Congress, Cairns, Australia, 2009. Park S, Bera A. Maximum entropy autoregressive conditional heteroskedasticity model. Journal of Econometrics, 2009; 150(2):219-230. 15 51. Lee M-C, Su J-B, Liu H-C. Value-at-risk in US stock indices with skewed generalized error distribution. Applied Financial Economics Letters, 2008; 4(6):425-431. 52. Mandelbrot B. The variation of certain speculative prices. Journal of Business, 1963; 36:394-419. 53. Bollerslev T. A conditionally heteroskedastic time series model for speculative prices and rates of return. Review of Economics and Statistics, 1987; 69(3):542-547. 54. Berkowitz J, O’Brien J. How accurate are Value-at-Risk models at commercial banks? Journal of Finance, 2002; 55:1093-1111. 55. Barone-Adesi G, Bourgoin F, Giannopoulos K. Don’t look back. Risk, 1998; 11(8):100-103. 56. Hull J, White A. Incorporating volatility updating into the historical simulation method for Value-at-Risk. Journal of Risk, 1998; 1:5-19. 57. Alexander C, Sheedy E. Developing a stress testing framework based on market risk models. Journal of Banking and Finance, 2008; 32(10):2220-2236. 58. Eugene N, Lee C, Famoye F. The beta normal distribution and its applications. Communications in Statistics, Theory and Methods, 2002; 31:497-512. 59. Jones M. Families of distributions arising from distributions of order statistics. Test, 2004; 13(1):1-43. 60. Zografos K, Balakrishnan, N. On families of beta- and generalized gamma-generated distributions and associated inference. Statistical Methodology, 2009; 6(4):344-362. 61. Nelson DB. Conditional heteroskedasticity in asset returns: A new approach. Econometrica, 1991; 59:347-370. 62. Arsenault M, Ashkar F. Regional flood quantile estimation under linear transformation of the data. Water Resources Research, 2000; 36(6):1605-1610. 63. Chebana F, Ouarda T. Multivariate quantiles in hydrological frequency analysis. Environmetrics, 2011; 22(1):63-78. 64. Katz R, Parlange M, Naveau P. Statistics of extremes in hydrology. Advances in Water Resources, 2002; 25:12871304. 65. Diodato N, Bellocchi G. Assessing and modelling changes in rainfall erosivity at different climate scales. Earth Surface Processes and Landforms, 2009; 34(7):969-980. 66. Bremnes J. Probabilistic wind power forecasts using local quantile regression. Wind Energy, 2004; 7(1):47-54. 67. Rumyantsev A. Safety prediction in nuclear power. Atomic Energy, 2007; 102(2):94-99. 68. Anghel C. Statistical process control methods from the viewpoint of industrial application. Economic Quality Control, 2001; 16(1):49-63. 69. Wheelock D, Wilson P. Non-parametric, unconditional quantile estimation for efficiency analysis with an application to Federal Reserve check processing operations. Journal of Econometrics, 200; 145:209-225. 70. Unnikrishnan Nair N, Sankaran P. Quantile-based reliability analysis. Communications in Statistics - Theory and Methods, 2009; 38(2):222-232. 71. Mkhandi S, Kachroo R, Guo S. Uncertainty analysis of flood quantile estimates with reference to Tanzania. Journal of Hydrology, 1996; 185, 317-333. 72. Pinson P, Kariniotakis G. On-line assessment of prediction risk for wind power production forecasts. Wind Energy, 2004; 7(2):119-132. 73. Noon B, Cade B, Flather C. Quantile regression reveals hidden bias and uncertainty in habitat models. Ecology, 2009; 86(3):786-800. 74. Merrick J, Van Dorp R, Dinesh V. Assessing uncertainty in simulation-based maritime risk assessment. Risk Analysis, 2005; 25(3):731-743. 75. Peng L, Qi Y. Confidence regions for high quantiles of a heavy tailed distribution. Annals of Statistics, 2006; 34(4):1964-1986. 16 76. Vermaat M, Does R, Steerneman A. A modified quantile estimator using extreme-value theory with applications. Economic Quality Control, 2005; 20(1):31-39. 77. Sveinsson O, Salas J, Boes D. Uncertainty of quantile estimators using the population index flood method. Water Resources Research, 2003; 39(8), 1206 doi:10.1029/2002WR001594 78. Kapur JN. Maximum Entropy Models in Engineering. John Wiley, New York, 1989. 79. Zografos K. On maximum entropy characterization of Pearson’s type II and VII multivariate distributions. Journal of Multivariate Analysis, 1999; 71:67-75. 80. Aulogiaris G, Zografos K. A maximum entropy characterization of symmetric Kotz type and multivariate Burr distributions. Test, 2004; 13:65-83. 81. Bhattacharya B. Maximum entropy characterizations of the multivariate Liouville distributions. Journal of Multivariate Analysis, 2006; 97:1272-1283. 82. Ahmed NA, Gokhale DV. Entropy expressions and their estimators for multivariate distributions. IEEE Transactions on Information Theory, 1989; 35:688-692. 83. Zografos K, Nadarajah S. Expressions for Renyi and Shannon entropies for multivariate distributions. Statistics and Probability Letters, 2005; 71:71-84. Alexander & Sarabia