Conference PaperPDF Available

Value- at-Risk vs Conditional Value-at-Risk in Risk Management and Optimization

Authors:

Abstract and Figures

From the mathematical perspective considered in this tutorial, risk management is a procedure for shaping a risk distribution. Popular functions managing risk are value at-risk (VaR) and conditional value-at-risk (CVaR). The problem of choice between VaR and CVaR, especially in financial risk management, has been quite popular in academic literature. Reasons affecting the choice between VaR and CVaR are based on the differences in mathematical properties, stability of statistical estimation, simplicity of optimization procedures, acceptance by regulators, etc. This tutorial presents our personal experience working with these key percentile risk measures. We try to explain strong and weak features of these risk measures and illustrate them with several examples. We demonstrate risk management/optimization case studies conducted with the Portfolio Safeguard package.
Content may be subject to copyright.
INFORMS 2008 c
2008 INFORMS |isbn 978-1-877640-23-0
doi 10.1287/educ.1080.0052
Value-at-Risk vs. Conditional Value-at-Risk in
Risk Management and Optimization
Sergey Sarykalin
Gobal Fraud Risk Management, American Express, Phoenix, Arizona 85021,
sergey.sarykalin@aexp.com
Gaia Serraino and Stan Uryasev
Department of Industrial and Systems Engineering, Risk Management and Financial
Engineering Lab, University of Florida, Gainesville, Florida 32611
{serraino@ufl.edu, uryasev@ufl.edu}
Abstract From the mathematical perspective considered in this tutorial, risk management is a
procedure for shaping a risk distribution. Popular functions managing risk are value-
at-risk (VaR) and conditional value-at-risk (CVaR). The problem of choice between
VaR and CVaR, especially in financial risk management, has been quite popular in
academic literature. Reasons affecting the choice between VaR and CVaR are based on
the differences in mathematical properties, stability of statistical estimation, simplic-
ity of optimization procedures, acceptance by regulators, etc. This tutorial presents
our personal experience working with these key percentile risk measures. We try to
explain strong and weak features of these risk measures and illustrate them with sev-
eral examples. We demonstrate risk management/optimization case studies conducted
with the Portfolio Safeguard package.
Keywords VaR; CVaR; risk measures; deviation measures; risk management; optimization; Port-
folio Safeguard package
1. Introduction
Risk management is a broad concept involving various perspectives. From the mathematical
perspective considered in this tutorial, risk management is a procedure for shaping a loss
distribution (for instance, an investor’s risk profile). Among the vast majority of recent inno-
vations, only a few have been widely accepted by practitioners, despite their active interest
in this area. Conditional value-at-risk (CVaR), introduced by Rockafellar and Uryasev [19],
is a popular tool for managing risk. CVaR approximately (or exactly, under certain con-
ditions) equals the average of some percentage of the worst-case loss scenarios. CVaR risk
measure is similar to the value-at-risk (VaR) risk measure, which is a percentile of a loss dis-
tribution. VaR is heavily used in various engineering applications, including financial ones.
VaR risk constraints are equivalent to the so-called chance constraints on probabilities of
losses. Some risk communities prefer VaR others prefer chance (or probabilistic) functions.
There is a close correspondence between CVaR and VaR: with the same confidence level,
VaR is a lower bound for CVaR. Rockafellar and Uryasev [19, 20] showed that CVaR is
superior to VaR in optimization applications. The problem of the choice between VaR and
CVaR, especially in financial risk management, has been quite popular in academic litera-
ture. Reasons affecting the choice between VaR and CVaR are based on the differences in
mathematical properties, stability of statistical estimation, simplicity of optimization pro-
cedures, acceptance by regulators, etc. Conclusions made from this properties may often be
quite contradictive. Plenty of relevant material on this subject can be found at the website
http://www.gloriamundi.org. This tutorial should not be considered as a review on VaR and
270
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 271
CVaR: many important findings are beyond the scope of this tutorial. Here we present only
our personal experience with these key percentile risk measures and try to explain strong
and weak features of these two risk measures and illustrate them with several examples. We
demonstrate risk management/optimization case studies with the Portfolio Safeguard (PSG)
package by American Optimal Decisions (an evaluation copy of PSG can be requested at
http://www.AOrDa.com).
Key observations presented in this tutorial are as follows:
CVaR has superior mathematical properties versus VaR. CVaR is a so-called “coherent
risk measure”; for instance, the CVaR of a portfolio is a continuous and convex function
with respect to positions in instruments, whereas the VaR may be even a discontinuous
function.
CVaR deviation (and mixed CVaR deviation) is a strong “competitor” to the standard
deviation. Virtually everywhere the standard deviation can be replaced by a CVaR deviation.
For instance, in finance, a CVaR deviation can be used in the following concepts: the Sharpe
ratio, portfolio beta, one-fund theorem (i.e., optimal portfolio is a mixture of a risk-free
asset and a master fund), market equilibrium with one or multiple deviation measures, and
so on.
Risk management with CVaR functions can be done quite efficiently. CVaR can be
optimized and constrained with convex and linear programming methods, whereas VaR is
relatively difficult to optimize (although significant progress was made in this direction; for
instance, PSG can optimize VaR for quite large problems involving thousands of variables
and hundreds of thousands of scenarios).
VaR risk measure does not control scenarios exceeding VaR (for instance you can signif-
icantly increase the largest loss exceeding VaR, but the VaR risk measure will not change).
This property can be both good and bad, depending upon your objectives:
The indifference of VaR risk measure to extreme tails may be a good property if poor
models are used for building distributions. VaR disregards some part of the distribution for
which only poor estimates are available. VaR estimates are statistically more stable than
CVaR estimates. This actually may lead to a superior out-of-sample performance of VaR
versus CVaR for some applications. For instance, for a portfolio involving instruments with
strong mean reverting properties, VaR will not penalize instruments with extremely heavy
losses. These instruments may perform very well at the next iteration. In statistics, it is well
understood that estimators based on VaR are “robust” and may automatically disregard
outliers and large losses, which may “confuse” the statistical estimation procedure.
The indifference of VaR to extreme tails may be quite an undesirable property, allow-
ing to take high uncontrollable risks. For instance, so-called “naked” option positions involve
a very small chance of extremely high losses; these rare losses may not be picked by VaR.
CVaR accounts for losses exceeding VaR. This property may be good or bad, depending
upon your objectives:
CVaR provides an adequate picture of risks reflected in extreme tails. This is a very
important property if the extreme tail losses are correctly estimated.
CVaR may have a relatively poor out-of-sample performance compared with VaR if
tails are not modelled correctly. In this case, mixed CVaR can be a good alternative that
gives different weights for different parts of the distribution (rather than penalizing only
extreme tail losses).
Deviation and risk are quite different risk management concepts. A risk measure evalu-
ates outcomes versus zero, whereas a deviation measure estimates wideness of a distribution.
For instance, CVaR risk may be positive or negative, whereas CVaR deviation is always
positive. Therefore, the Sharpe-like ratio (expected reward divided by risk measure) should
involve CVaR deviation in the denominator rather than CVaR risk.
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
272 Tutorials in Operations Research, c
2008 INFORMS
Figure 1. Risk functions: graphical representation of VaR, VaR Deviation, CVaR, CVaR Devia-
tion, Max Loss, and Max Loss Deviation.
VaR, CVaR, deviations
Frequency
VaR
Probability
1–α
Max
Loss
CVaR
CVaR Deviation
Max Loss Deviation
Loss
Mean
VaR Deviation
2. General Picture of VaR and CVaR
2.1. Definitions of VaR and CVaR
This section gives definitions of VaR and CVaR and discusses their use and basic properties.
We refer to Figure 1 for their graphical representation.
Let Xbe a random variable with the cumulative distribution function FX(z)=P{Xz}.
Xmay have meaning of loss or gain. In this tutorial, Xhas meaning of loss and this impacts
the sign of functions in definition of VaR and CVaR.
Definition 1 (Value-at-Risk). The VaR of Xwith confidence level α]0,1[ is
VaRα(X) = min{z|FX(z)α}.(1)
By definition, VaRα(X)isalowerα-percentile of the random variable X. Value-at-risk is
commonly used in many engineering areas involving uncertainties, such as military, nuclear,
material, airspace, finance, etc. For instance, finance regulations, like Basel I and Basel II,
use VaR deviation measuring the width of daily loss distribution of a portfolio.
For normally distributed random variables, VaR is proportional to the standard deviation.
If XN(µ, σ2) and FX(z) is the cumulative distribution function of X, then (see Rockafellar
and Uryasev [19]),
VaRα(X)=F1
X(α)=µ+k(α)σ, (2)
where k(α)=2 erf 1(2α1) and erf(z)=(2/π)z
0et2dt.
The ease and intuitiveness of VaR are counterbalanced by its mathematical properties.
As a function of the confidence level, for discrete distributions, VaRα(X) is a nonconvex,
discontinuous function. For a discussion of numerical difficulties of VaR optimization, see,
for example, Rockafellar [17] and Rockafellar and Uryasev [19].
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 273
Definition 2 (Conditional Value-at-Risk). An alternative percentile measure of risk
is conditional value-at-risk (CVaR). For random variables with continuous distribution func-
tions, CVaRα(X) equals the conditional expectation of Xsubject to XVa Rα(X). This
definition is the basis for the name of conditional value-at-risk. The term conditional value-
at-risk was introduced by Rockafellar and Uryasev [19]. The general definition of conditional
value-at-risk (CVaR) for random variables with a possibly discontinuous distribution func-
tion is as follows (see Rockafellar and Uryasev [20]).
The CVaR of Xwith confidence level α]0,1[ is the mean of the generalized α-tail
distribution:
CVaRα(X)=
−∞
zdFα
X(z),(3)
where
Fα
X(z)=
0,when z<VaRα(X),
FX(z)α
1α,when zVaRα(X).
Contrary to popular belief, in the general case, CVaRα(X) is not equal to an average of
outcomes greater than VaRα(X). For general distributions, one may need to split a prob-
ability atom. For example, when the distribution is modelled by scenarios, CVaR may be
obtained by averaging a fractional number of scenarios. To explain this idea in more detail,
we further introduce alternative definitions of CVaR. Let CVaR+
α(X), called “upper CVaR,”
be the conditional expectation of Xsubject to X>VaRα(X):
CVaR+
α(X)=E[X|X>Va R α(X)].
CVaRα(X) can be defined alternatively as the weighted average of VaRα(X) and
CVaR+
α(X), as follows. If FX(VaRα(X)) <1, so there is a chance of a loss greater than
VaRα(X), then
CVaRα(X)=λα(X)VaRα(X)+(1λα(X))CVaR+
α(X),(4)
where
λα(X)=FX(VaRα(X)) α
1α,(5)
whereas if FX(VaRα(X)) = 1, so that VaRα(X) is the highest loss that can occur, then
CVaRα(x)=VaR
α(x).(6)
The definition of CVaR as in Equation (4) demonstrates that CVaR is not defined as a
conditional expectation. The function CVaR
α(X)=E[X|XVaRα(X)], called “lower
CVaR,” coincides with CVaRα(X) for continuous distributions; however, for general distri-
butions it is discontinuous with respect to αand is not convex. The construction of CVaRα
as a weighted average of VaRαand CVaR+
α(X) is a major innovation. Neither VaR nor
CVaR+
α(X) behaves well as a measure of risk for general loss distributions (both are discon-
tinuous functions), but CVaR is a very attractive function. It is continuous with respect to
αand jointly convex in (X, α). The unusual feature in the definition of CVaR is that VaR
atom can be split. If FX(x) has a vertical discontinuity gap, then there is an interval of
confidence level αhaving the same VaR. The lower and upper endpoints of that interval are
α=FX(VaR
α(X)) and α+=FX(VaRα(X)), where FX(VaR
α(X)) = P{X<Va R α(X)}.
When FX(VaR
α(X)) <α<F
X(VaRα(X)) <1, the atom VaRα(X) having total probability
α+αis split by the confidence level αin two pieces with probabilities α+αand αα.
Equation (4) highlights this splitting.
The definition of CVaR is illustrated further with the following examples. Suppose we have
six equally likely scenarios with losses f1···f6. Let α=2
3(see Figure 2). In this case, αdoes
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
274 Tutorials in Operations Research, c
2008 INFORMS
Figure 2. CVaR Example 1: computation of CVaR when αdoes not spilt the atom.
CVaR
6
1
6
1
VaRLoss CVaR+
6
1
6
1
6
1
6
1
f1f2f3
f4f5f6
CVaR
not split any probability atom. Then VaRα(X)<CVaR
α(X)<CVaRα(X)=CVaR
+
α(X),
λα(X)=(FX(VaRα(X)) α)/(1 α) = 0 and CVaRα(X)=CVaR
+
α(X)=1
2f5+1
2f6, where
f5and f6are losses number five and six, respectively. Now, let α=7
12 (see Figure 3).
In this case, αdoes split the VaRα(X) atom, λα(X)=(FX(VaRα(X)) α)/(1 α)>0, and
CVaRα(X) is given by
CVaRα(X)= 1
5VaRα(X)+ 4
5CVaR+
α(X)=1
5f4+2
5f5+2
5f6.
In the last case, we consider four equally likely scenarios and α=7
8splits the last atom
(see Figure 4). Now VaRα(X)=CVaR
α(X)=CVaR
α(X), upper CVaR, CVaR+
α(X)isnot
defined, λα(X)=(FX(VaRα(X)) α)/(1 α)>0, and CVaRα(X)=VaR(X)=f4. The
PSG package defines the CVaR function for discrete distributions equivalently to (4) through
the lower CVaR and upper CVaR. Suppose that VaRα(X) atom having total probability
α+αis split by the confidence level αin two pieces with probabilities α+αand αα.
Then,
CVaRα(X)= α+α
α+α
1α
1αCVaR
α(X)+ αα
α+α
1α+
1αCVaR+
α(X),(7)
where
CVaR
α(X)=E[X|XVaRα(X)],CVaR+
α(X)=E[X|X>Va R α(X)].(8)
Pflug [15] followed a different approach and suggested to define CVaR via an optimization
problem, which he borrowed from Rockafellar and Uryasev [19]:
CVaRα(X) = min
CC+1
1αE[XC]+,where [t]+= max{0,t}.(9)
One more equivalent representation of CVaR was given by Acerbi [1], who showed that
CVaR is equal to “expected shortfall” defined by
CVaRα(X)= 1
αα
0
VaRβ(X).
Figure 3. CVaR Example 2: computation of CVaR when αsplits the atom.
CVaR
CVaR
VaRLoss ½f5f6 = CVaR+
6
1
6
1
6
1
6
1
6
1
6
112
1
f1f2f3
f4f5f6
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 275
Figure 4. CVaR Example 3: computation of CVaR when αsplits the last atom.
CVaR
VaRLoss
4
1
4
1
8
1
4
1
4
1
f1f2f3f4
For normally distributed random variables, a CVaR deviation is proportional to the standard
deviation. If XN(µ, σ2), then (see Rockafellar and Uryasev [19])
CVaRα(X)=E[X|XVa R α(X)] = µ+k1(α)σ, (10)
where
k1(α)=2πexp(erf1(2α1))2(1 α)1
and erf(z)=(2/π)z
0et2dt.
2.2. Risk Measures
Axiomatic investigation of risk measures was suggested by Artzner et al. [3]. Rockafellar [17]
defined a functional R:L2]−∞,]asacoherent risk measure in the extended sense if
R1: R(C)=Cfor all constant C;
R2: R((1 λ)X+λX)(1 λ)R(X)+λR(X) for λ]0,1[ (convexity);
R3: R(X)≤R(X) when XX(monotonicity);
R4: R(X)0 when XkX20 with R(Xk)0 (closedness).
A functional R:L2]−∞,] is called a coherent risk measure in the basic sense if it
satisfies axioms R1, R2, R3, R4, and additionally the axiom
R5: R(λX)=λR(X) for λ>0 (positive homogeneity).
A functional R:L2]−∞,] is called an averse risk measure in the extended sense if it
satisfies axioms R1, R2, R4, and
R6: R(X)>EX for all nonconstant X(aversity).
Aversity has the interpretation that the risk of loss in a nonconstant random variable X
cannot be acceptable; i.e., R(X)<0, unless EX < 0.
A functional R:L2]−∞,] is called an averse risk measure in the basic sense if it
satisfies R1, R2, R4, R6, and also R5.
Examples of coherent measures of risk are R(X)=µX =E[X]orR(X) = sup X. However,
R(X)=µ(X)+λσ(X) for some λ>0 is not a coherent measure of risk because it does not
satisfy the monotonicity axiom R3.
R(X)=VaR
α(X) is not a coherent nor an averse risk measure. The problem lies in the
convexity axiom R2, which is equivalent to the combination of positive homogeneity and
subadditivity, this last defined as R(X+X)≤R(X)+R(X). Although positive homo-
geneity is obeyed, the subadditivity is violated. It has been proved, for example, in Acerbi
and Tasche [2], Pflug [15], and Rockafellar and Uryasev [20], that for any probability level
α]0,1[, R(X)=CVaR
α(X) is a coherent measure of risk in the basic sense. CVaRα(X)is
also an averse measure of risk for α]0,1]. An averse measure of risk might not be coherent;
a coherent measure might not be averse.
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
276 Tutorials in Operations Research, c
2008 INFORMS
2.3. Deviation Measures
In this section, we refer to Rockafellar [17] and Rockafellar et al. [24]. A functional D:L2
[0,] is called a a deviation measure in the extended sense if it satisfies
D1: D(C) = 0 for constant C, but D(X)>0 for nonconstant X;
D2: D((1 λ)X+λX)(1 λ)D(X)+λD(X) for λ]0,1[ (convexity);
D3: D(X)dwhen XkX20 with D(Xk)d(closedness).
A functional is called a deviation measure in the basic sense when it satisfies axioms
D1, D2, D3, and, furthermore,
D4: D(λX)=λD(X) for λ>0 (positive homogeneity).
A deviation measure in the extended or basic sense is called a coherent measure in the
extended or basic sense if it additionally satisfies
D5: D(X)sup XE[X] for all X(upper range boundedness).
An immediate example of a deviation measure in the basic sense is the standard deviation
σ(X)=(E[XEX]2)1/2,
which satisfies axioms D1, D2, D3, D4, but not D5. In other words, standard deviation is not
a coherent deviation measure. Here are more examples of deviation measures in the basic
sense:
Standard semideviations:
σ+(X)=(E[max{XEX,0}]2)1/2,
σ(X)=(E[max{EX X, 0}]2)1/2;
Mean Absolute Deviation:
MAD(X)=E[|XEX|].
Moreover, we define the α-value-at-risk deviation measure and the α-conditional value-at-
risk deviation measure as
αVaR
α(X)=VaR
α(XEX) (11)
and
αCVaR
α(X)=CVaR
α(XEX),(12)
respectively. The VaR deviation measure VaR
α(X) is not a deviation measure in the general
or basic sense because the convexity axiom D2 is not satisfied. The CVaR deviation measure
CVaR
α(X) is a coherent deviation measure in the basic sense.
2.4. Risk Measures vs. Deviation Measures
Rockafellar et al. (originally in [24], then in [17]), obtained the following result:
Theorem 1. A one-to-one correspondence between deviation measures Din the extended
sense and averse risk measures Rin the extended sense is expressed by the relations
R(X)=D(X)+EX,
D(X)=R(XEX);
additionally,
Ris coherent ↔D is coherent.
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 277
Moreover, the positive homogeneity is preserved:
Ris positively homogeneous ↔Dis positively homogeneous.
In other words, for an averse risk measures Rin the basic sense and a deviation measures
Din the basic sense the one-to-one correspondence is valid, and, additionally, coherent R
coherent D.
With this theorem we obtain that for the standard deviation, σ(X), which is a deviation
measure in the basic sense, the counterpart is the standard risk EX +σ(X), which is a risk
averse measure in the basic sense. For CVaR deviation, CVaR
α(X), which is a coherent
deviation measure in the basic sense, the counterpart is CVaR risk, CVaRα(X), which is a
risk-averse coherent measure in the basic sense.
Another coherent deviation measure in the basic sense is the so-called Mixed Deviation
CVaR, which we think is the most promising for risk management purposes. Mixed Deviation
CVaR is defined as
Mixed-CVaR
α(X)=
K
k=1
λkCVaR
αk(X)
for λk0, K
k=1 λk= 1, and αkin ]0,1[. The counterpart to the Mixed Deviation CVaR is
the Mixed CVaR, which is the coherent averse risk measure in the basic sense, defined by
Mixed-CVaRα(X)=
K
k=1
λkCVaRαk(X).
3. VaR and CVaR in Optimization and Statistics
3.1. Equivalence of Chance and VaR Constraints
For this section, we refer to Rockafellar [18]. Several engineering applications deal with
probabilistic constraints such as the reliability of a system or the system’s ability to meet
demand; in portfolio management, often it is required that portfolio loss at a certain future
time is, with high reliability, at most equal to a certain value. In these cases an optimization
model can be set up so that constraints are required to be satisfied with some probability
level rather than almost surely. Let x∈nand let ωΩ be a random event ranging over the
set Ω of all random events. For a given x, we may require that most of the time some random
functions fi(x, ω), i=1,...,m, satisfy the inequalities fi(x, ω)0, i=1,...,m; that is, we
may want that
Prob{fi(x, ω)0}≥pifor i=1,...,m, (13)
where 0 pi1. Requiring this probability to be equal to 1 is the same as requiring that
fi(x, ω)0 almost surely. In most applications, this approach can lead to modelling and
technical problems. In modelling, there is little guidance on what level of pito set; moreover,
one has to deal with the issue of constraint interactions and decide whether, for instance,
it is better to require Prob{f1(x, ω)0}≥p1=0.99 and Prob{f2(x, ω)0}≥p2=0.95 or
to work with a joint condition like Prob{fi(x, ω )0}≥p. Dealing numerically with the
functions Fi(x) = Prob{fi(x, ω)0}leads to the task of finding the relevant properties
of Fi; a common difficulty is that the convexity of fi(x, ω ) with respect to xmay not carry
over to the convexity of Fi(x) with respect to x.
Chance constraints and percentiles of a distribution are closely related. Let VaRα(x)be
the VaRαof a loss function f(x, ω ); that is,
VaRα(x) = min{: Prob{f(x, ω )}≥α}.(14)
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
278 Tutorials in Operations Research, c
2008 INFORMS
Then, the following constraints are equivalent:
Prob{f(x, ω)}≥αProb{f(x, ω)>}≤1αVaRα(x). (15)
Generally, VaRα(x) is nonconvex with respect to x; therefore, VaRα(x)and
Prob{f(x, ω)}≥αmay be nonconvex constraints.
3.2. CVaR Optimization
For this section, we refer to Rockafellar and Uryasev [19] and Uryasev [29]. Nowadays, VaR
has achieved the high status of being written into industry regulations (for instance, in
regulations for finance companies). It is difficult to optimize VaR numerically when losses
are not normally distributed. Only recently VaR optimization was included in commercial
packages such as PSG. As a tool in optimization modeling, CVaR has superior properties
in many respects. CVaR optimization is consistent with VaR optimization and yields the
same results for normal or elliptical distributions (see definition of elliptical distribution
in Embrechts et al. [6]); for models with such distributions, working with VaR, CVaR, or
minimum variance (Markowitz [11]) is equivalent (see Rockafellar and Uryasev [19]). Most
importantly, CVaR can be expressed by a minimization formula suggested by Rockafellar
and Uryasev [19]. This formula can be incorporated into the optimization problem with
respect to decision variables xX∈
n, which are designed to minimize risk or shape
it within bounds. Significant shortcuts are thereby achieved while preserving the crucial
problem features like convexity. Let us consider that a random loss function f(x, y) depends
upon the decision vector xand a random vector yof risk factors. For instance, f(x, y)=
(y1x1+y2x2) is the negative return of a portfolio involving two instruments. Here, x1,x
2
are positions and y1,y
2are rates of returns of two instruments in the portfolio. The main
idea in Rockafellar and Uryasev [19] is to define a function that can be used instead of
CVaR:
Fα(x, ζ)=ζ+1
1αE{[f(x, y)ζ]+}.(16)
The authors proved that
1. Fα(x, ζ) is convex with respect to (w.r.t.) α;
2. VaRα(x) is a minimum point of function Fα(x, ζ ) w.r.t. ζ;
3. minimizing Fα(x, ζ ) w.r.t. ζgives CVaRα(x):
CVaRα(x) = min
αFα(x, ζ).(17)
In optimization problems, CVaR can enter into the objective or constraints or both. A big
advantage of CVaR over VaR in that context is the preservation of convexity; i.e., if f(x, y)
is convex in x, then CVaRα(x)isconvexinx. Moreover, if f(x, y)isconvexinx, then
the function Fα(x, ζ)isconvexinbothxand ζ. This convexity is very valuable because
minimizing Fα(x, ζ) over (x, ζ )X×results in minimizing CVaRα(x):
min
xXCVaRα(x) = min
(x, ζ)X× Fα(x, ζ ).(18)
In addition, if (x) minimizes Fαover X×, then not only does xminimize CVaRα(x)
over Xbut also
CVaRα(x)=Fα(x).
In risk management, CVaR can be utilized to “shape” the risk in an optimization model.
For that purpose, several confidence levels can be specified. Rockafellar and Uryasev [19]
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 279
showed that for any selection of confidence levels αiand loss tolerances ωi,i=1,...,l, the
problem
min
xXg(x)
s.t. CVaRαi(x)ωi,i=1,...,l
(19)
is equivalent to the problem
min
x, ζ1,...,ζl,X××··· g(x)
s.t. Fαi(x, ζi)ωi,i=1,...,l.
(20)
When Xand gare convex and f(x, y)isconvexinx, the optimization problems (18)
and (19) are ones of convex programming and, thus, especially favorable for computation.
When Yis a discrete probability space with elements yk,k=1,...,N having probabilities
pk,k=1,...,N,wehave
Fαi(x, ζi)=ζi+1
1αi
N
k=1
pk[f(x, yk)ζi]+.(21)
The constraint Fα(x, ζ)ωcan be replaced by a system of inequalities by introducing
additional variables ηk:
ηk0,f(x, yk)ζηk0,k=1,...,N,
ζ+1
1α
N
k=1
pkηkω. (22)
The minimization problem in (19) can be converted into the minimization of g(x) with
the constraints Fαi(x, ζi)ωibeing replaced as presented in (22). When fis linear in x,
constraints (22) are linear.
3.3. Generalized Regression Problem
In linear regression, a random variable Yis approximated in terms of random variables
X1,X
2,...,X
nby an expression c0+c1X1+···+cnXn. The coefficients are chosen by
minimizing the mean square error:
min
c0,c1,...,cn
E(Y[c0+c1X1+···+cnXn])2.(23)
Mean square error minimization is equivalent to minimizing the standard deviation with the
unbiasedness constraint (see, Rockafellar et al. [21, 26]):
min σ(Y[c0+c1X1+···+cnXn])
s.t. E[c0+c1X1+···+cnXn]=EY. (24)
Rockafellar et al. [21, 26] considered a general axiomatic setting for error measures and
corresponding deviation measures. They defined an error measure as a functional E:L2(Ω)
[0,] satisfying the following axioms:
E1: E(0) = 0, E(X)>0 for X=0, E(C)<for constant C;
E2: E(λX)=λE(X) for λ>0 (positive homogeneity);
E3: E(X+X)≤E(X)+E(X) for all Xand X(subadditivity);
E4: {X∈L2(Ω) |E(X)c}is closed for all c<(lower semicontinuity).
For an error measure E, the pro jected deviation measure Dis defined by the equation
D(X) = minCE(XC), and the statistic S(X) is defined by S(X) = argminCE(XC).
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
280 Tutorials in Operations Research, c
2008 INFORMS
Their main finding is that the general regression problem
min
c0,c1,...,cnE(Y[c0+c1X1+···+cnXn]) (25)
is equivalent to
min
c1,...,cnD(Y[c1X1+···+cnXn])
s.t.c
0∈S(Y[c1X1+···+cnXn]).
The equivalence of optimization problems (23) and (24) is a special case of this theorem.
This leads to the identification of a link between statistical work on percentile regression
(see Koenker and Basset [7]) and CVaR deviation measure: minimization of the Koenker
and Bassett [7] error measure is equivalent to minimization of CVaR deviation. Rockafellar
et al. [26] show that when the error measure is the Koenker and Basset [7] function,
Eα
KB(X)=E[max{0,X}+(α11) max{0,X}], the projected measure of deviation is
D(X)=CVaR
α(X) = CVaRα(XEX), with the corresponding averse measure of risk and
associated statistic given by
R(X) = CVaRα(X),
S(X)=VaR
α(X).
Then,
min
C∈(E[XC]++(α11)E[XC])=CVaR
α(X),
arg min
C∈
(E[XC]++(α11)E[XC])=VaR
α(X).
A similar result is available for the “mixed Koenker and Bassett error measure” and the
corresponding mixed deviation CVaR; see Rockafellar et al. [26].
3.4. Stability of Estimation
The need to estimate VaR and CVaR arises typically when we are interested in estimating
tails of distributions. It is of interest, in this respect, to compare the stability of estimates of
VaR and CVaR based on a finite number of observations. The common flaw in such compar-
isons is that some confidence level is assumed and estimations of VaRα(X) and CVaRα(X)
are compared with the common value of confidence level α, usually, 90%, 95%, and 99%. The
problem with such comparisons is that VaR and CVaR with the same confidence level mea-
sure “different parts” of the distribution. In reality, for a specific distribution, the confidence
levels α1and α2for comparison of VaR and CVaR should be found from the equation
VaRα1(X) = CVaRα2(X).(26)
For instance, in the credit risk example in Serraino et al. [27], we find that CVaR with
confidence level α=0.95 is equal to VaR with confidence level α=0.99. The paper by
Yamai and Yoshiba [30] can be considered a “good” example of a “flawed” comparison of
VaR and CVaR estimates. Yamai and Yoshiba [30] examine VaR and CVaR estimations for
the parametrical family of stable distributions. The authors ran 10,000 simulations of size
1,000 and compared standard deviations of VaR and CVaR estimates normalized by their
mean values. Their main findings are as follows. VaR estimators are generally more stable
than CVaR estimators with the same confidence level. The difference is most prominent for
fat-tailed distributions and is negligible when the distributions are close to normal. A larger
sample size increases the accuracy of CVaR estimation. We provide here two illustrations of
Yamai and Yoshiba’s [30] results of these estimators’ performances.
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 281
In the first case, the distribution of an equity option portfolio is modelled. The portfolio
consists of call options based on three stocks with joint log-normal distribution. VaR and
CVaR are estimated at the 95% confidence level on 10,000 sets of Monte Carlo simulations
with a sample size of 1,000. The resulting loss distribution for the portfolio of at-the-money
options is quite close to normal; estimation errors of VaR and CVaR are similar. The result-
ing loss distribution for the portfolio of deep out-of-the-money options is fat tailed; in this
case, the CVaR estimator performed significantly worse than the VaR estimator.
In the second case, estimators are compared on the distribution of a loan portfolio, con-
sisting of 1,000 loans with homogeneous default rates of 1% through 0.1%. Individual loan
amounts obey the exponential distribution with an average of $100 million. Correlation
coefficients between default events are homogeneous at levels 0.00, 0.03, and 0.05. Results
show that estimation errors of CVaR and VaR estimators are similar when the default rate
is higher; for lower default rates, the CVaR estimator gives higher errors. Also, the higher
the correlation between default events, the more the loan portfolio distribution becomes fat
tailed, and the higher is the error of CVaR estimator relative to VaR estimator.
As we pointed out, these numerical experiments compare VaR and CVaR with the same
confidence level, and some other research needs to be done to compare stability of estimators
for the same part of the distribution.
3.5. Decomposition According to Contributions of Risk Factors
This subsection discusses decomposition of VaR and CVaR risk according to risk factor
contributions; see, for instance, Tasche [28] and Yamai and Yoshiba [30]. Consider a portfolio
loss X, which can be decomposed as
X=
n
i=1
Xizi,
where Xiare losses of individual risk factors and ziare sensitivities to the risk factors, i=
1,...,n. The following decompositions of VaR and CVaR hold for continuous ditributions:
VaRα(X)=
n
i=1
VaRα(X)
∂zi
zi=E[Xi|X=VaR
α(X)]zi,(27)
CVaRα(X)=
n
i=1
CVaRα(X)
∂zi
zi=E[Xi|XVaRα(X)]zi.(28)
When a distribution is modelled by scenarios, it is much easier to estimate quantities
E[Xi|XVaRα(X)] in the CVaR decomposition than quantities E[Xi|X=VaR
α(X)]
in the VaR decomposition. Estimators of VaRα(X)/∂ziare less stable than estimators of
CVaRα(X)/∂ zi.
3.6. Generalized Master-Fund Theorem and CAPM
The one-fund theorem is a fundamental result of the classical portfolio theory. It establishes
that any mean-variance-efficient portfolio can be constructed as a combination of a single
master-fund portfolio and a risk-free instrument. Rockafellar et al. [22] investigated in detail
the consequences of substituting the standard deviation in the setting of classical theory with
general deviation measures, in cases where rates of return may have discrete distributions,
mixed discrete-continuous distributions (which can arise from derivatives, such as options)
or continuous distributions. Their main result is that the one-fund theorem holds regardless
of the particular choice of the deviation measure. The optimal risky portfolio needs not
always be unique and it might not always be expressible by a master fund as traditionally
conceived, even when only the standard deviation is involved. The authors introduce the
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
282 Tutorials in Operations Research, c
2008 INFORMS
concepts of a basic fund as the portfolio providing the minimum portfolio deviation δfor
a gain of exactly 1 over the risk-free rate, with the corresponding basic deviation, defined
as the minimum deviation amount. Then they establish the generalized one-fund theorem,
according to which for any level of risk premium ∆ over the risk-free rate, the solution of
the portfolio optimization problem is given by investing an amount ∆ in the basic fund
and an amount 1 ∆ in the risk-free rate. In particular, the price of the basic fund can be
positive, negative, or equal to zero, leading respectively to a long position, short position, or
no investment in the basic fund. Rockafellar et al., first in Rockafellar et al. [23] and further
in Rockafellar et al. [25], developed these concepts to obtain the generalized capital asset
pricing (CAPM) relations. They find that the generalized CAPM equilibrium holds under
the following assumptions: there are several groups of investors, each with utility function
Uj(ERj,Dj(Rj)) based on deviation measure Dj; the utility functions depend on mean
and deviation and they are concave w.r.t. mean and deviation, increasing w.r.t. mean, and
decreasing w.r.t. deviation; investors maximize their utility functions sub ject to the budget
constraint. The main finding is that an equilibrium exists w.r.t. Diin which each group of
investors has its own master fund and investors invest in the risk-free asset and their own
master funds. A generalized CAPM holds and takes the form
¯rij r0=βij rjM r0),
βij =cov(Gj,r
ij )
D(rjM),
where
¯rij is expected return of asset iin group j,
r0is risk-free rate,
¯rjM is expected return of market fund for investor group j,
Gjis the risk identifier for the market fund j.
From this general statement we can obtain that in classical framework when all investors
are interested in standard deviation, βiis defined as
βi=cov(ri,r
M)
σ2(rM).
Similarly, when all investors are interested in semideviation D(X)=σ(X), then
βi=cov(ri,r
M)
σ2
(rM),
and when D(X) = CVaR
α(X) with continuously distributed random values, then
βi=E[ri¯ri|−rMVa Rα(rM)]
CVaR
α(rM).
It is interesting to observe in the last case that “beta” picks up only events when market is
in α·100% of its highest losses; i.e., rMVaRα(rM).
4. Comparative Analysis of VaR and CVaR
4.1. VaR Pros and Cons
4.1.1. Pros. VaR is a relatively simple risk management notion. Intuition behind
α-percentile of a distributions is easily understood and VaR has a clear interpretation: how
much you may lose with certain confidence level. VaR is a single number measuring risk,
defined by some specified confidence level, e.g., α=0.95. Two distributions can be ranked
by comparing their VaRs for the same confidence level. Specifying VaR for all confidence
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 283
levels completely defines the distribution. In this sense, VaR is superior to the standard
deviation. Unlike the standard deviation, VaR focuses on a specific part of the distribution
specified by the confidence level. This is what is often needed, which made VaR popular in
risk management, including finance, nuclear, airspace, material science, and various military
applications.
One of important properties of VaR is stability of estimation procedures. Because VaR
disregards the tail, it is not affected by very high tail losses, which are usually difficult to
measure. VaR is estimated with parametric models; for instance, covariance VaR based on
the normal distribution assumption is very well known in finance, with simulation models
such as historical or Monte Carlo or by using approximations based on second-order Taylor
expansion.
4.1.2. Cons. VaR does not account for properties of the distribution beyond the confi-
dence level. This implies that VaRα(X) may increase dramatically with a small increase
in α. To adequately estimate risk in the tail, one may need to calculate several VaRs with
different confidence levels. The fact that VaR disregards the tail of the distribution may
lead to unintentional bearing of high risks. In a financial setting, for instance, let us consider
the strategy of “naked” shorting deep out-of-the-money options. Most of the time, this will
result in receiving an option premium without any loss at expiration. However, there is a
chance of a big adverse market movement leading to an extremely high loss. VaR cannot
capture this risk.
Risk control using VaR may lead to undesirable results for skewed distributions. Later on
we will demonstrate this phenomenon with the case study comparing risk profiles of VaR
and CVaR optimization. In this case, the VaR optimal portfolio has about 20% longer tail
than the CVaR optimal portfolio, as measured by the max loss of those portfolios.
VaR is a nonconvex and discontinuous function for discrete distributions. For instance,
in financial setting, VaR is a nonconvex and discontinuous function w.r.t. portfolio posi-
tions when returns have discrete distributions. This makes VaR optimization a challenging
computational problem. Nowadays there are codes, such as PSG, that can work with VaR
functions very efficiently. PSG can optimize portfolios with a VaR performance function and
also shape distributions of the portfolio with multiple VaR constraints. For instance, in port-
folio optimization it is possible to maximize expected return with several VaR constraints
at different confidence levels.
4.2. CVaR Pros and Cons
4.2.1. Pros. CVaR has a clear engineering interpretation. It measures outcomes that hurt
the most. For example, if Lis a loss then the constraint CVaRα(L)¯
Lensures that the
average of (1α)% highest losses does not exceed ¯
L.
Defining CVaRα(X) for all confidence levels αin (0,1) completely specifies the distribution
of X. In this sense, it is superior to standard deviation.
Conditional value at risk has several attractive mathematical properties. CVaR is a coher-
ent risk measure. CVaRα(X) is continuous with respect to α. CVaR of a convex combina-
tion of random variables CVaRα(w1X1+···+wnXn) is a convex function with respect to
(w1,...,w
n). In financial setting, CVaR of a portfolio is a convex function of portfolio posi-
tions. CVaR optimization can be reduced to convex programming, in some cases to linear
programming (i.e., for discrete distributions).
4.2.2. Cons. CVaR is more sensitive than VaR to estimation errors. If there is no good
model for the tail of the distribution, CVaR value may be quite misleading; accuracy of
CVaR estimation is heavily affected by accuracy of tail modelling. For instance, historical
scenarios often do not provide enough information about tails; hence, we should assume
a certain model for the tail to be calibrated on historical data. In the absence of a good
tail model, one should not count on CVaR. In financial setting, equally weighted portfolios
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
284 Tutorials in Operations Research, c
2008 INFORMS
may outperform CVaR-optimal portfolios out of sample when historical data have mean
reverting characteristics.
4.3. What Should You Use, VaR or CVaR?
VaR and CVaR measure different parts of the distribution. Depending on what is needed,
one may be preferred over the other.
Let us illustrate this topic with financial applications of VaR and CVaR and examine
the question of which measure is better for portfolio optimization. A trader may prefer
VaR to CVaR, because he may like high uncontrolled risks; VaR is not as restrictive as
CVaR with the same confidence level. Nothing dramatic happens to a trader in case of
high losses. He will not pay losses from his pocket; if fired, he may move to some other
company. A company owner will probably prefer CVaR; he has to cover large losses if they
occur; hence, he “really” needs to control tail events. A board of directors of a company
may prefer to provide VaR-based reports to shareholders and regulators because it is less
than CVaR with the same confidence level. However, CVaR may be used internally, thus
creating asymmetry of information between different parties.
VaR may be better for optimizing portfolios when good models for tails are not available.
VaR disregards the hardest to measure events. CVaR may not perform well out of sample
when portfolio optimization is run with poorly constructed set of scenarios. Historical data
may not give right predictions of future tail events because of mean-reverting characteristics
of assets. High returns typically are followed by low returns; hence, CVaR based on history
may be quite misleading in risk estimation.
If a good model of tail is available, then CVaR can be accurately estimated and CVaR
should be used. CVaR has superior mathematical properties and can be easily handled in
optimization and statistics.
When comparing stability of estimation of VaR and CVaR, appropriate confidence levels
for VaR and CVaR must be chosen, avoiding comparison of VaR and CVaR for the same
level of αbecause they refer to different parts of the distribution.
5. Numerical Results
This section reports results of several case studies exemplifying how to use VaR and CVaR
in optimization settings.
5.1. Portfolio Safeguard
We used PSG to do the case studies. We posted MATLAB files to run these case studies
in a MATLAB environment on the MathWorks website (http://www.mathworks.com), in
the file exchange-optimization area. PSG is designed to solve a wide range of risk manage-
ment and stochastic optimization problems. PSG has built-in algorithms for evaluating and
optimizing various statistical and deterministic functions. For instance, PSG includes the
statistical functions Mean, Variance, Standard Deviation, VaR, CVaR, CDaR, MAD, Max-
imum Loss, Partial Moment, Probability of Exceeding a Threshold, and the deterministic
functions Cardinality, Fixed Charge, and Buyin. For a complete list of functions, see Table
A.1 in Appendix I. Required data inputs are matrices of Scenarios or Covariance Matrices
on which statistical functions are evaluated. PSG uses a new design for defining optimiza-
tion problems. A function is defined by just providing an underlying Matrix to a Function.
Figure 5 illustrates PSG procedure for function definition (Mean and Max Loss Functions are
defined: Matrix of Scenarios Loss Vector Functions). With PSG design all needed data,
including names of variables, are taken from a Matrix of Scenarios or Covariance Matrix.
PSG can calculate values and sensitivities of defined functions on decision vectors. Also,
you can place functions to objective and constraint and define an optimization problem.
Linear combinations of functions can be placed in the constraints. PSG includes many case
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 285
Figure 5. Example of evaluation of two PSG functions: Mean and Max Loss.
INPUT DATA: Matrix of scenarios
xy
Scenario benchmark Scenario probability
Loss vector
Loss function Scenario probability
Function evaluation
Function value
Mean
Max Loss
θ11 θ01
θ21
θ12
θ22 θ02
θ01 –(θ11 *X+θ21 *Y)
p1*[θ01 –(θ11 *X+θ21 *Y)] + p2*[θ02 –(θ12 *X+θ22 *Y)]
max[θ01 –(θ11 *X+θ21*Y)], [θ02 –(θ12 *X+θ22 *Y)]
θ02 –(θ12 *X+θ22 *Y)
p1
p2
p1
p2
Notes. These functions are completely defined by the matrix of scenarios. Both names of variables and
function values are calculated using matrix of scenarios.
studies, most of them motivated by financial applications (such as collateralized debt obli-
gation structuring, portfolio management, portfolio replication, and selection of insurance
contracts) that are helpful in learning the software. To build a new optimization problems,
you can simply make a copy of the case study and change input matrices and parameters.
PSG can solve problems with simultaneous constraints on various risks at different time
intervals (e.g., multiple constraints on standard deviation obtained by resampling, combined
with VaR and drawdown constraints), thus allowing robust decision making. It has built-in
efficient algorithms for solving large-scale optimization problems (up to 1,000,000 scenarios
and up to 10,000 decision variables in callable MATLAB and C++ modules). PSG offers
several tools for analyzing solutions, generated by optimization or through other procedures.
Among these tools are sensitivities of risk measures to changes in decision variables, and
the incremental impact of decision variables on risk measures and various functions of risk
measures. Analysis can reveal decision variables having the biggest impact on risk and other
functions, such as expected portfolio return. For visualization of these and other characteris-
tics, PSG provides tools for building and plotting various characteristics combining different
functions, points, and variables. It is user friendly and callable from MATLAB and C/C++
environment.
5.2. Risk Control Using VaR (PSG MATLAB Environment)
Risk control using VaR may lead to paradoxical results for skewed distributions. In this case,
minimization of VaR may lead to a stretch of the tail of the distribution exceeding VaR. The
purpose of VaR minimization is to reduce extreme losses. However, VaR minimization may
lead to an increase in the extreme losses that we try to control. This is an undesirable feature
of VaR optimization. Larsen et al. [9] showed this phenomenon on a credit portfolio with the
loss distribution created with a Monte Carlo simulation of 20,000 scenarios of joint credit
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
286 Tutorials in Operations Research, c
2008 INFORMS
Figure 6. VaR minimization.
VaR
CVaR
320
290
260
230
200
170
0 5 10 15 20 25 30 35 40 45 50 55
MUSD
Iteration
Notes. 99%-VaR and 99%-CVaR for algorithm A2 in Larsen et al. [9]. The algorithm lowered VaR at the
expense of an increase in CVaR.
states. The distribution is skewed with a long fat right tail. A more detailed description of
this portfolio can be found in Bucay and Rosen [4] and Mausser and Rosen [13, 14]. Larsen
et al. [9] suggested two heuristic algorithms for optimization of VaR. These algorithms are
based on the minimization of CVaR. The minimization of VaR leads to about 16% increase of
the average loss for the worst 1% scenarios, compared with the worst 1% scenarios in CVaR
minimum solution. These numerical results are consistent with the theoretical results: CVaR
is a coherent, whereas VaR is not a coherent, measure of risk. Figure 6 reproduces results
from Larsen et al. [9] showing how iteratively VaR is decreasing and CVaR is increasing.
We observed a similar performance of VaR on a small portfolio consisting of 10 bonds
and modelled with 1,000 scenarios. We solved two optimization problems with PSG. In the
first one, we minimized 99%-CVaR deviation of losses subject to constraints on budget and
required return. In the second one, we minimized 99%-VaR deviation of losses subject to
the same constraints.
Problem 1: min CVaR
α(x)
s.t.
n
i=1
rixi¯r,
n
i=1
xi=1.
(29)
Problem 2: min VaR
α(x)
s.t.
n
i=1
rixi¯r,
n
i=1
xi=1.
(30)
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 287
Table 1. Value of risk functions.
min CVaR
0.99 min VaR
0.99 Ratio
CVaR0.99 0.0073 0.0083 1.130
CVaR
0.99 0.0363 0.0373 1.026
VaR0.99 0.0023 0.0005 0.231
VaR
0.99 0.0313 0.0295 0.944
Max loss = CVaR10.0133 0.0148 1.116
Max loss deviation = CVaR
10.0423 0.0438 1.036
Notes. Column “min CVaR0.99
α” reports the value of risk functions for the portfolio obtained by
minimizing 99%-CVaR deviation, column “minVaR
α” reports the value of risk functions for the
portfolio obtained by minimizing 99%-VaR deviation, and column “Ratio” contains the ratio of
every cell of column “minVaR
α” to every cell of column “minCVaR
α.”
where xis the vector of portfolio weights, riis the rate of return of asset iris the lower
bound on estimated portfolio return. We then evaluated different risk functions at the opti-
mal points for Problems 1 and 2. Results are shown in Table 1.
Suppose that we start with the portfolio having minimal 99%-CVaR deviation. Minimiza-
tion of 99%-VaR deviation leads to 13% increase in 99%-CVaR, compared with 99%-CVaR
in the optimal 99%-CVaR deviation portfolio. We found that even in a problem with a rela-
tively small number of scenarios, if the distribution is skewed, minimization of VaR deviation
may lead to a stretch of the tail compared with the CVaR optimal portfolio. This result is
quite important when we look at financial risk management regulations like Basel II that
are based on minimization of VaR deviation.
5.3. Linear Regression-Hedging: VaR, CVaR, Mean Absolute, and
Standard Deviations (PSG MATLAB Environment)
This case study investigates performance of optimal hedging strategies based on different
deviation measures measuring quality of hedging. The objective is to build a portfolio of
financial instruments that mimics the benchmark portfolio. Weights in the replicating port-
folio are chosen such that the deviation between the value of this replicating portfolio and the
value of the benchmark is minimized. Benchmark value and replicating financial instruments
values are random valves. Determining the optimal hedging strategy is a linear regression
problem where the response is the benchmark portfolio valve, the predictors are the repli-
cating financial instrument valves, and the coefficients of the predictors to be determined
are the portfolio weights. Let ˆ
θbe the replicating portfolio value, θ0be the benchmark
portfolio value, θ1,...,θ
Ibe replicating instrument values, and x1,...,x
Ibe their weights.
The replicating portfolio value can be expressed as follows:
ˆ
θ=x1θ1+···+xIθI.(31)
The coefficients x1,...,x
Ishould be chosen to minimize a replication error function depend-
ing upon the residual θ0ˆ
θ. The intercept in this case is absent. According to the equivalence
of (25) and (26), an error minimization problem is equivalent to the minimization of the
appropriate deviation; see Rockafellar et al. [21] and [26]. This case study considers hedging
pipeline risk in the mortgage underwriting process. Hedging instruments are 5% Mortgage-
Backed Securities (MBS) forward, 5.5% MBS, and call options on 10-year treasury note
futures. Changes of values of the benchmark and the hedging instruments are driven by
changes in the mortgage rate. We minimize five different deviation measures: Standard Devi-
ation, Mean Absolute Deviation, CVaR Deviation, Two-Tailed 75%-VaR, and Two-Tailed
90%-VaR. We tested in-sample and out-of-sample performance of the hedging strategies.
On our set of scenarios, we found that Two-Tailed 90%-VaR has the best out-of-sample
performance, whereas the standard deviation has the worst out-of-sample performance.
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
288 Tutorials in Operations Research, c
2008 INFORMS
We think that the out-of-sample performance of hedging strategies based on different
deviation measures significantly depends on the skewness of the distribution. In this case,
the distribution of residuals is quite skewed.
We use here PSG definitions of Loss and Deviation functions. The Loss Function is defined
as follows:
Loss Function = L(x, θ)=L(x1,...,x
I
0,...,θ
I)=θ0
I
i=1
θixi.(32)
The loss function has jscenarios, L(x, θ1),...,L(x, θJ), each with probability pj,j=1,...,J.
Here are deviation measures considered in this case study:
Mean Absolute Deviation = MAD(L(x, θ)),(33)
Standard Deviation = σ(L(x, θ )),(34)
α%-VaR Deviation = VaR
α(L(x, θ)),(35)
α%-CVaR Deviation = CVaR
α(L(x, θ)),(36)
Two-Tail α%-VaR Deviation = TwoTailVaR
α(L(x, θ))
=VaR
α(L(x, θ))+VaR
α(L(x, θ)).(37)
We solved the following minimization problems:
Minimize 90%-CVaR Deviation
min
xCVaR
0.9(L(x, θ)) (38)
Minimize Mean Absolute Deviation
min
xMAD(L(X,θ)) (39)
Minimize Standard Deviation
min
xσ(L(x, θ)) (40)
Minimize Two-Tail 75%-VaR Deviation
min
xTwoTailVaR
0.75(L(X, θ)) (41)
Minimize Two-Tail 90%-VaR Deviation
min
xTwoTailVaR
0.9(L(X,θ)).(42)
The data set for the case study includes 1,000 scenarios of value changes for each hedging
instrument and for the benchmark. For the out-of-sample testing, we partitioned the 1,000
scenarios into 10 groups with 100 scenarios in each group. Each time, we selected one group
for the out-of-sample test and we calculated optimal hedging positions based on the remain-
ing nine groups containing 900 scenarios. For each group of 100 scenarios, we calculated the
ex ante losses (i.e., underperformances of hedging portfolio versus target) with the optimal
hedging positions obtained from 900 scenarios. We repeated the procedure 10 times, once
for every out-of-sample group with 100 scenarios. To estimate the out-of-sample perfor-
mance, we aggregated the out-of-sample losses from the 10 runs and obtained a combined
set including 1,000 out-of-sample losses. Then, we calculated five deviation measures on the
out-of-sample 1,000 losses: Standard Deviation, Mean Absolute Deviation, CVaR Deviation,
Two-Tail 75%-VaR Deviation, and Two-Tail 90%-VaR Deviation. In addition, we calculated
three downside risk measures: 90%-CVaR, 90%-VaR, and Max Loss = 100%-CVaR on the
out-of-sample losses. Tables 2 and 3 show the results.
By minimizing the Two-Tail 90%-VaR Deviation, we obtained the best values for all three
considered downside risk measures (negative loss indicates gain). Minimization of CVaR
deviation lead to good results, whereas minimization of standard deviation gave the worst
level for three downside risk measures.
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 289
Table 2. Out-of-sample performance of various deviations on optimal hedging portfolios.
Optimal points CVaR
0.9MAD σTwoTailVaR
0.75 TwoTailVaR
0.9
CVaR
0.90.690 0.815 1.961 0.275 1.122
MAD 1.137 0.714 1.641 0.379 1.880
σ1.405 0.644 1.110 0.979 1.829
TwoTailVaR
0.75 1.316 0.956 1.955 0.999 1.557
TwoTailVaR
0.90.922 0.743 1.821 0.643 1.256
Notes. Each row reports values of five different deviation measures evaluated at optimal hedging points
obtained with five hedging strategies.
5.4. Example of Equivalence of Chance and VaR Constraints
(PSG MATLAB Environment)
This case study illustrates the equivalence between chance constraints and VaR constraints,
as explained in §3.1. We will illustrate numerically the equivalence
Prob{L(x, θ)>}≤1αVa R α(L(x, θ)) , (43)
where
Loss Function L(x, θ )=L(x1,...,x
I
1,...,θ
I)=
I
i=1
θixi.(44)
The case study is based on a data set including 1,000 return scenarios for 10 clusters of loans.
Here, I= 10 is the number of instruments, θ1,...,θ
Iare rates of returns of instruments,
x1,...,x
Iare instrument weights, and L(x, θ ) is a portfolio loss. We solved two portfolio
optimization problems. In both cases we maximized the estimated return of the portfolio. In
the first problem, we imposed a constraint on probability; in the second problem, we imposed
an equivalent constraint on VaR. In particular, in the first problem we require the 95%-VaR
of the optimal portfolio to be at most equal to the constant , whereas in the second problem
we require the probability of losses greater than to be lower than 1 α=10.95=0.05.
We expected to obtain at optimality the same objective function value and similar optimal
portfolios for the two problems. Problem formulations are as follows:
Problem 1: max E[L(x, θ)]
s.t. Prob{L(x, θ)>}≤1α=0.05,
vixiui,i=1,...,I,
I
i=1
xi=1.
(45)
Table 3. Out-of-sample performance of various downside risks on optimal hedging portfolios.
Optimal points Max Loss CVaR0.9VaR0.9
CVaR
0.918.01 18.05 18.08
MAD 16.49 17.44 17.88
σ13.31 15.29 15.60
TwoTailVaR
0.75 15.31 16.19 16.71
TwoTailVaR
0.918.02 18.51 18.66
Notes. Each row reports values of three different risk measures evaluated at optimal hedging points obtained
with five hedging strategies (negative loss indicates gain).
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
290 Tutorials in Operations Research, c
2008 INFORMS
Table 4. Chance vs. VaR constraints.
Optimal weights Prob 0.05 VaR
x10.051 0.051
x20.055 0.055
x30.071 0.071
x40.053 0.053
x50.079 0.079
x60.289 0.289
x70.020 0.020
x80.300 0.300
x90.063 0.063
x10 0.020 0.020
Notes. Optimal portfolios obtained on the same data set when we
maximized return with the chance constraint (the first column)
and with the VaR constraint (the second column).
Problem 2: max E[L(x, θ)]
s.t. VaRα(L(x, θ )) ,
vixiui,i=1,...,I,
I
i=1
xi=1.
(46)
In the problems, viis the lower bound on position for asset i, and uiis the upper bound
on position for asset i. We also have the budget constraint: sum of weights is equal to 1.
The two problems at optimality selected the same portfolios and have the same objective
function value of 120.19. Table 4 shows the optimal points.
5.5. Portfolio Rebalancing Strategies: Risk vs. Deviation
(PSG MATLAB Environment)
In this case study we consider a portfolio rebalancing problem. A portfolio manager allocates
his wealth to different funds periodically solving an optimization problem and in each time
period building the portfolio that minimizes a certain risk function, given budget constraints
and bounds on each exposure. We solved the following problem:
min R(x, θ)kE[L(x, θ)] (47)
s.t.
I
i=1
xi= 1 (48)
vi(xi)uii=1...,I, (49)
where we denote by R(x, θ) a risk function, E[L(x, θ)] is the expected portfolio return,
θ1,...,θ
Iare rates of returns of instruments, and x1,...,x
Iare instrument weights. The
scenario data set is composed of 46 monthly return observations for seven managed funds;
we solved the first optimization problem using the first 10 scenarios, then we rebalanced the
portfolio monthly. We used as risk functions VaR, CVaR, VaR Deviation, CVaR Deviation,
and Standard Deviation.
We then evaluated Sharpe ratio and mean value of each sequence of portfolios obtained
by succesively solving the optimization problem with a given ob jective function. Results are
reported in Tables 5 and 6 for different values of the parameter k. Both in terms of Sharpe
ratio and mean portfolio value we found a good performance of VaR and VaR deviation
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 291
Table 5. Out-of-sample Sharpe ratio.
kVaR CVaR VaR Deviation CVaR Deviation Standard Deviation
1 1.2710 1.2609 1.2588 1.2693 1.2380
3 1.2711 1.2667 1.2762 1.2652 1.2672
5 1.2712 1.2666 1.2721 1.2743 1.2628
Notes. Sharpe ratio for the rebalancing strategy when different risk functions are used in the
objective function with several values of parameter k.
Table 6. Out-of-sample portfolio mean return.
kVaR CVaR VaR Deviation CVaR Deviation Standard Deviation
1 0.2508 0.2556 0.2445 0.2567 0.2425
3 0.2645 0.2575 0.2631 0.2598 0.2576
5 0.2663 0.2612 0.2662 0.2617 0.2532
Notes. Portfolio mean return for the rebalancing strategy when different risk functions are used
in the objective function with several values of parameter k.
minimization, whereas standard deviation minimization gives inferior results. Overall, we
rebalanced the portfolios 37 times for each objective function. Results depend on the scenario
data set and on the parameter k; thus, we cannot conclude that minimization of a certain
risk function is always the best choice. However, we observe that in the presence of mean
reversion, the tails of historical distribution are not good predictors of the tails in the
future. In this case, VaR disregarding the tails may lead to a good out-of-sample portfolio
performance. In fact, VaR disregards the unstable part of the distribution.
Appendix I
Table A.1. PSG functions.
Function group Full name PSG version
Deterministic Functions
Linear group Variable Function 1.1
Linear Function 1.1
Linear Multiple 1.1
Nonlinear group Polynomial Absolute 1.1
Relative Entropy 1.1
Maximum Positive 1.2 Beta
Maximum Negative 1.2 Beta
CVaR Positive 1.2 Beta
CVaR Negative 1.2 Beta
VaR Positive 1.2 Beta
VaR Negative 1.2 Beta
Cardinality group Cardinality Positive 1.2 Beta
Cardinality Negative 1.2 Beta
Cardinality 1.2 Beta
Buyin Positive 1.2 Beta
Buyin Negative 1.2 Beta
Buyin 1.2 Beta
Fixed Charge Positive 1.2 Beta
Fixed Charge Negative 1.2 Beta
Fixed Charge 1.2 Beta
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
292 Tutorials in Operations Research, c
2008 INFORMS
Table A.1. Continued.
Function group Full name PSG version
Risk functions
Average group Average Gain 1.1
Average Loss 1.1
VaR group VaR Deviation for Gain 1.1
VaR Deviation for Loss 1.1
VaR Risk for Gain 1.1
VaR Risk for Loss 1.1
CVaR group CVaR Ceviation for Gain 1.1
CVaR Deviation for Loss 1.1
CVaR Risk for Gain 1.1
CVaR Risk for Loss 1.1
CDaR group CDaR Deviation for Gain 1.1
CDaR Deviation for Gain Multiple 1.1
CDaR Deviation 1.1
CDaR Deviation Multiple 1.1
Drawdown Deviation Average for Gain 1.1
Drawdown Deviation Average for Gain Multiple 1.1
Drawdown Deviation Average 1.1
Drawdown Deviation Average Multiple 1.1
Drawdown Deviation Maximum for Gain 1.1
Drawdown Deviation Maximum for Gain Multiple 1.1
Drawdown Deviation Maximum 1.1
Drawdown Deviation maximum Multiple 1.1
Maximum group Maximum Deviation for Gain 1.1
Maximum Deviation for Loss 1.1
Maximum Risk for Gain 1.1
Maximum Risk for Loss 1.1
Mean Abs group Mean Absolute Deviation 1.1
Mean Absolute Penalty 1.1
Mean Absolute Risk for Gain 1.1
Mean Absolute Risk for Loss 1.1
Partial moment group Partial Moment Gain Deviation 1.1
Partial Moment Loss Deviation 1.1
Partial Moment Penalty for Gain 1.1
Partial Moment Penalty for Loss 1.1
Probability group Probability Exceeding Deviation for Gain 1.1
Probability Exceeding Deviation for Gain Multiple 1.1
Probability Exceeding Deviation for Loss 1.1
Probability Exceeding Deviation for Loss Multiple 1.1
Probability Exceeding Penalty for Gain 1.1
Probability Exceeding Penalty for Gain Multiple 1.1
Probability Exceeding Penalty for Loss 1.1
Probability Exceeding Penalty for Loss Multiple 1.1
Standard group Standard Deviation 1.1
Standard Gain 1.1
Standard Penalty 1.1
Standard Risk 1.1
Mean Square Penalty 1.1
Variance 1.1
Utility group Exponential Utility 1.2 Beta
Logarithmic Utility 1.2 Beta
Power Utility 1.2 Beta
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
Tutorials in Operations Research, c
2008 INFORMS 293
References
[1] C. Acerbi. Spectral measures of risk: A coherent representation of subjective risk aversion.
Journal of Banking and Finance 26:1505–1518, 2002.
[2] C. Acerbi and D. Tasche. On the coherence of expected shortfall. Journal of Banking and
Finance 26:1487–1503, 2002.
[3] P. Artzner, F. Delbaen, J.-M. Eber, and D. Heath. Coherent measures of risk. Mathematical
Finance 9:203–227, 1999.
[4] N. Bucay and D. Rosen. Credit risk of an international bond portfolio: A case study. ALGO
Research Quarterly 2(1):9–29.
[5] A. Chekhlov, S. Uryasev, and M. Zabarankin. Portfolio optimization with drawdown con-
straints. B. Scherer, ed. Asset and Liability Management Tools. Risk Books, London, 263–278,
2003.
[6] P. Embrechts, A. J. McNeil, and D. Straumann. Correlation and dependence in risk manage-
ment: Properties and pitfalls. M. Dempster, ed. Risk Management: Value at Risk and Beyond.
Cambridge University Press, Cambridge, UK, 2002.
[7] R. Koenker and G. W. Bassett. Regression quantiles. Econometrica 46:33–50, 1978.
[8] H. Konno and H. Yamazaki. Mean absolute deviation portfolio optimization model and its
application to Tokyo stock market. Management Science 37:519–531, 1991.
[9] N. Larsen, H. Mausser, and S. Uryasev. Algorithms for optimization of value-at-risk.
P. Pardalos and V. K. Tsitsiringos, eds. Financial Engineering, e-Commerce and Supply Chain.
Kluwer Academic Publishers, Dordrecht, The Netherlands, 129–157, 2000.
[10] A. Lo. The three P’s of total risk management. Financial Analysts Journal 55:13–26, 1999.
[11] H. M. Markowitz. Portfolio selection. Journal of Finance 7(1):77–91, 1952.
[12] H. Mausser and D. Rosen. Beyond VaR: From measuring risk to managing risk. ALGO Research
Quarterly 1(2):5–20, 1999.
[13] H. Mausser and D. Rosen. Applying scenario optimization to portfolio credit risk. ALGO
Research Quarterly 2(2):19–33, 1999.
[14] H. Mausser and D. Rosen. Efficient risk/return frontiers for credit risk. ALGO Research Quar-
terly 2(4):35–48, 1999.
[15] G. C. Pflug. Some remarks on the value-at-risk and the conditional value-at-risk. S. P. Uryasev,
ed. Probabilistic Constrainted Optimization: Methodology and Applications. Kluwer, Norwell,
MA, 278–287, 2000.
[16] R. T. Rockafellar. Convex Analysis. Princeton University Press, Princeton, NJ, 1970.
[17] R. T. Rockafellar. Coherent approaches to risk in optimization under uncertainty. INFORMS
Tutorials in Operations Research. Institute for Operations Research and the Management Sci-
ences, Hanover, MD, 38–61, 2007.
[18] R. T. Rockafellar. Optimization Under Uncertainty, Lectures Notes. http://www.math.
washington.edu/rtr/uncertainty.pdf, 2001.
[19] R. T. Rockafellar and S. P. Uryasev. Optimization of conditional value-at-risk. Journal of Risk
2:21–42, 2000.
[20] R. T. Rockafellar and S. P. Uryasev. Conditional value-at-risk for general loss distributions.
Journal of Banking and Finance 26:1443–1471, 2002.
[21] R. T. Rockafellar, S. Uryasev, and M. Zabarankin. Deviation measures in generalized linear
regression. Research Report 2002-9, Department of Industrial and Systems Engineering, Uni-
versity of Florida, Gainesville, 2002.
[22] R. T. Rockafellar, S. Uryasev, and M. Zabarankin. Master funds in portfolio analysis with
general deviation measures. Journal of Banking and Finance 30:743–778, 2005.
[23] R. T. Rockafellar, S. Uryasev, and M. Zabarankin. Optimality conditions in portfolio analysis
with generalized deviation measures. Mathematical Programming 108(2–3):515–540, 2006.
[24] R. T. Rockafellar, S. Uryasev, and M. Zabarankin. Generalized deviations in risk analysis.
Finance and Stochastics 10:51–74, 2006.
[25] R. T. Rockafellar, S. Uryasev, and M. Zabarankin. Equilibrium with investors using a diversity
of deviation measures. Journal of Banking and Finance 31:1443–1471, 2007.
[26] R. T. Rockafellar, S. Uryasev, and M. Zabarankin. Risk tuning with generalized linear regres-
sion. Mathematics of Operations Research, forthcoming.
Sarykalin et al.: VaR vs.CVaR in Risk Management and Optimization
294 Tutorials in Operations Research, c
2008 INFORMS
[27] G. Serraino, U. Theiler, and S. Uryasev. Risk return optimization with different risk aggregation
strategies. G. N. Gregoriov, ed. Groundbreaking resarch and developments in VaR, forthcoming.
Bloomberg Press, New York, 2009.
[28] D. Tasche. Risk contribution and performance measurement. Working paper, Technical Uni-
versity of Munich, Munich, Germany.
[29] S. Uryasev. Conditional value-at-risk: Optimization algorithms and applications. Financial
Engineering News (14 February):1–5, 2000.
[30] Y. Yamai and T. Yoshiba. Comparative analysis of expected shortfall and value-at-risk:
Their estimation error, decomposition and optimization. Monetary and Economic Studies
20(1):57–86, 2002.
... To evaluate the numerical performance of our algorithms we consider Conditional Value at Risk (CVaR) as the risk-measure, see, for example, [26], [22], [29]. CVaR is a popular mathematical tool for managing risk, among others, in finance for portfolio optimization and asset liability management [25]. ...
... CVaR is a popular mathematical tool for managing risk, among others, in finance for portfolio optimization and asset liability management [25]. Unlike more classical riskmeasures such as the Value at Risk (VaR) [20], CVaR is a coherent risk-measure, which makes the optimization more stable and efficient [29] and allows us to deploy our framework. Our goal however is not that of comparing different risk measures, but rather that of empirically studying the performance of the discussed algorithms in terms of numerical stability and convergence. ...
Preprint
Full-text available
Inspired by semismooth Newton methods, we propose a general framework for designing solution methods with convergence guarantees for risk-averse Markov decision processes. Our approach accommodates a wide variety of risk measures by leveraging the assumption of Markovian coherent risk measures. To demonstrate the versatility and effectiveness of this framework, we design three distinct solution methods, each with proven convergence guarantees and competitive empirical performance. Validation results on benchmark problems demonstrate the competitive performance of our methods. Furthermore, we establish that risk-averse policy iteration can be interpreted as an instance of semismooth Newton's method. This insight explains its superior convergence properties compared to risk-averse value iteration. The core contribution of our work, however, lies in developing an algorithmic framework inspired by semismooth Newton methods, rather than evaluating specific risk measures or advocating for risk-averse approaches over risk-neutral ones in particular applications.
... The CVaR formulation used here provides a reasonable approximation for a discrete distribution. Thus, uncertainty is modeled using limited scenarios, each reflecting the density function (Sarykalin et al. 2008). The first part of the expression represents the goal function's expected value across all scenarios., i.e., expected total cost, system reliability, and carbon emission. ...
Article
Full-text available
Energy plays a pivotal role in addressing the climate crisis, with fossil fuel combustion being a significant source of greenhouse gas emissions. To combat this, industrial nations are urged to transition to renewable energy sources like solar photovoltaic systems. However, optimizing photovoltaic systems requires consideration of technical, economic, and environmental factors. In this study, we devised a two-phase stochastic programming method for optimizing a photovoltaic system connected to a power grid. Our focus is on incorporating economic and environmental uncertainties into risk scenarios. The aim is to ascertain the optimal number of photovoltaic array installations, accounting for importing energy from the grid and exporting surplus energy to external utilities. Environmental factors and market change parameters are dynamically determined during decision-making, while model parameters are deterministic. The initial optimization stage determines the necessary number of photovoltaic modules. Given the scenario-driven nature of photovoltaic module selection, energy allocation to clients, external utilities, and grid interactions vary accordingly. Our proposed approach employs risk metrics to address environmental, economic, and reliability considerations in decision-making. A case study for sizing a photovoltaic system to power a housing area at King Fahd University of Petroleum and Minerals demonstrates the effectiveness of our approach. Sensitivity analysis highlights key trade-offs, showing the proposed system can break even at $1.79 million per year, generate 388.53 megawatt-hours of electricity annually, and reduce carbon dioxide emissions by 50.52 kg per year. Our study underscores the importance of integrating renewable energy solutions into the global energy mix to mitigate climate change impacts.
... This process is challenging when n δ is large, when parameter dependencies are strong, when the dataset is incomplete (the dependency between disjoint subsets of data cannot be measured) or when the data are scarce. Poorly chosen uncertainty models might lead to designs that grossly underperform in practice [48,49]. Nonprobabilistic models and mixtures of non-probabilistic and probabilistic models have been used when the data are scarce [50,51,52,53]. ...
Preprint
Full-text available
This paper proposes risk-averse and risk-agnostic formulations to robust design in which solutions that satisfy the system requirements for a set of scenarios are pursued. These scenarios, which correspond to realizations of uncertain parameters or varying operating conditions, can be obtained either experimentally or synthetically. The proposed designs are made robust to variations in the training data by considering perturbed scenarios. This practice allows accounting for error and uncertainty in the measurements, thereby preventing data overfitting. Furthermore, we use relaxation to trade-off a lower optimal objective value against lesser robustness to uncertainty. This is attained by eliminating a given number of optimally chosen outliers from the dataset, and by allowing for the perturbed scenarios to violate the requirements with an acceptably small probability. For instance, we can pursue a riskier design that attains a lower objective value in exchange for a few scenarios violating the requirements, or we might seek a more conservative design that satisfies the requirements for as many perturbed scenarios as possible. The design of a flexible wing subject to aeroelastic constraints is used for illustration.
Article
To meet decarbonization targets, nations around the globe have made ambitious commitments to expand forested land. Operationalizing these commitments requires choosing a planting strategy: How many trees should be planted, of which species, and where? Given those choices must be made now but have long-term consequences, such decisions are plagued by uncertainty. For example, species that are well suited to present conditions may perform poorly under future climates, yet those future climates are themselves highly uncertain. Using the exemplar of the United Kingdom, a nation committed to achieving net zero emissions by midcentury, we quantify key uncertainties pertaining to coevolving climate and economic conditions and examine how modern methods of decision-making under uncertainty can advise on planting choices. Our analysis reveals that the best planting strategy assuming a “high-emissions” future is radically different to that for a future that remains on a “near-historic” path. Planting for the former while experiencing the latter results in substantial net costs to UK society. Assimilating uncertainty into decision-making identifies planting strategies that diversify risk and significantly reduce the probability of high-cost outcomes. Importantly, our research reveals that the scope for mitigating risk through choice of planting strategy is relatively limited. Despite this persistent risk, we find that tree planting remains a highly cost-effective carbon removal solution when compared to alternative technologies, even when those alternatives are assumed to be riskless.
Article
Full-text available
Rail transportation of hazardous material (Hazmat) involves low-probability, and high-consequence risks, requiring strategies that mitigate extreme accident impacts while ensuring fair risk distribution. To address this, we introduce conditional value-at-risk with equity (CVaRE) into railway Hazmats risk assessment, enabling flexible decision-making that balances risk minimization and equity considerations. Unlike conventional models that focus solely on risk reduction, CVaRE incorporates a risk equity goal, ensuring a more balanced distribution of risk across transportation routes. This study develops a novel CVaRE model that replaces fixed threshold constraints with a dynamic risk equity goal, providing greater flexibility in risk distribution adjustments. A k-shortest path-based algorithm was designed to balance extreme risk minimization with equitable risk allocation in route selection. A case study on the Yangtze River Delta railway network validates the model, demonstrating that moderate cost increases can significantly reduce extreme accident risks while achieving fairer risk distribution. Findings also show that direct transportation improves risk equity over transfer-based routes, highlighting the importance of strategic route planning. This research offers practical decision support for transport companies and regulators, helping optimize routes while ensuring cost efficiency and regulatory compliance. It also provides a scientific foundation for risk-equity-based policies, promoting safer, more sustainable Hazmat railway transportation.
Book
O objetivo deste estudo é avaliar os custos e os benefícios de diferentes regras de compartilhamento do risco de demanda para concessões de infraestrutura rodoviária, sob o ponto de vista da atratividade econômica dos projetos para os concessionários e da redução do preço para os usuários. Simula-se a trajetória de tráfego de uma rodovia, por meio de modelo estocástico embasado no movimento geométrico browniano (MGB), utilizando-se dados e parâmetros do edital de concessão da BR-040. A partir dessas projeções, são avaliados os impactos de diferentes modelos de regras de compartilhamento do risco de demanda sobre a estrutura financeira dos projetos. Demonstra-se a vantagem dessas regras sobre o modelo sem compartilhamento de risco, com o objetivo de ampliar a viabilidade financeira dos projetos e o potencial de redução tarifária aos usuários.
Article
Full-text available
Current risk-management practices are based on probabilities of extreme dollar losses (e.g., measures like Value at Risk), but these measures capture only part of the story. Any complete risk-management system must address two other important factors—prices and preferences. Together with probabilities, these compose the three P's of “Total Risk Management.” This article describes how the three P's interact to determine sensible risk profiles for corporations and for individuals—guidelines for how much risk to bear and how much to hedge. By synthesizing existing research in economics, psychology, and decision sciences and through an ambitious research agenda to extend this synthesis into other disciplines, a complete and systematic approach to rational decision making in an uncertain world is within reach.
Article
Full-text available
In this paper we study both market risks and nonmarket risks, without complete markets assumption, and discuss methods of measurement of these risks. We present and justify a set of four desirable properties for measures of risk, and call the measures satisfying these properties “coherent.” We examine the measures of risk provided and the related actions required by SPAN, by the SEC/NASD rules, and by quantile-based methods. We demonstrate the universality of scenario-based methods for providing coherent measures. We offer suggestions concerning the SEC method. We also suggest a method to repair the failure of subadditivity of quantile-based methods.
Article
Decisions often need to be made before all the facts are in. A facility must be built to withstand storms, floods, or earthquakes of magnitudes that can only be guessed from historical records. A portfolio must be purchased in the face of only statistical knowledge, at best, about how markets will perform. In optimization, this implies that constraints may need to be envisioned in terms of safety margins instead of exact requirements. But what does that really mean in model formulation? What guidelines make sense, and what are the consequences for optimization structure and computation? The idea of a coherent measure of risk in terms of surrogates for potential loss, which has been developed in recent years for applications in financial engineering, holds promise for a far wider range of applications in which the traditional approaches to uncertainty have been subject to criticism. The general ideas and main facts are presented here with the goal of facilitating their transfer to practical work in those areas.
Article
Linear regression is traditionally based on the minimization of variance, or equivalently, standard deviation, but other approaches are possible in which standard deviation is replaced by something more general. A one-to-one correspondence is now known between risk measures, such as have been introduced for various applications in finance, and a large class of deviation measures characterized by simple axioms. Included in that class are asymmetric measures coming from conditional value- at-risk and other currently attractive notions. This paper looks at deviation in that wide sense, formulating the associated problem of regression and investigating the existence and uniqueness of the coecients that constitute a solution. Such coecients are characterized in ways that provide a key to their computation.
Article
A simple minimization problem yielding the ordinary sample quantiles in the location model is shown to generalize naturally to the linear model generating a new class of statistics we term "regression quantiles." The estimator which minimizes the sum of absolute residuals is an important special case. Some equivariance properties and the joint aymptotic distribution of regression quantiles are established. These results permit a natural generalization to the linear model of certain well-known robust estimators of location. Estimators are suggested, which have comparable efficiency to least squares for Gaussian linear models while substantially out-performing the least-squares estimator over a wide class of non-Gaussian error distributions.
Article
A new approach to optimizing or hedging a portfolio of financial instruments to reduce risk is presented and tested on applications. It focuses on minimizing Conditional Value-at-Risk (CVaR) rather than minimizing Value-at-Risk (VaR), but portfolios with low CVaR necessarily have low VaR as well. CVaR, also called Mean Excess Loss, Mean Shortfall, or Tail VaR, is anyway considered to be a more consistent measure of risk than VaR. Central to the new approach is a technique for portfolio optimization which calculates VaR and optimizes CVaR simultaneously. This technique is suitable for use by investment companies, brokerage firms, mutual funds, and any business that evaluates risks. It can be combined with analytical or scenario-based methods to optimize portfolios with large numbers of instruments, in which case the calculations often come down to linear programming or nonsmooth programming. The methodology can be applied also to the optimization of percentiles in contexts outside of finance.
Article
The purpose of this paper is to demonstrate that a portfolio optimization model using the L1 risk (mean absolute deviation risk) function can remove most of the difficulties associated with the classical Markowitz's model while maintaining its advantages over equilibrium models. In particular, the L1 risk model leads to a linear program instead of a quadratic program, so that a large-scale optimization problem consisting of more than 1,000 stocks may be solved on a real time basis. Numerical experiments using the historical data of NIKKEI 225 stocks show that the L1 risk model generates a portfolio quite similar to that of the Markowitz's model within a fraction of time required to solve the latter.
Article
Generalized measures of deviation are considered as substitutes for standard deviation in a framework like that of classical portfolio theory for coping with the uncertainty inherent in achieving rates of return beyond the risk-free rate. Such measures, derived for example from conditional value-at-risk and its variants, can reflect the different attitudes of different classes of investors. They lead nonetheless to generalized one-fund theorems in which a more customized version of portfolio optimization is the aim, rather than the idea that a single “master fund” might arise from market equilibrium and serve the interests of all investors.The results that are obtained cover discrete distributions along with continuous distributions. They are applicable therefore to portfolios involving derivatives, which create jumps in distribution functions at specific gain or loss values, well as to financial models involving finitely many scenarios. Furthermore, they deal rigorously with issues that come up at that level of generality, but have not received adequate attention, including possible lack of differentiability of the deviation expression with respect to the portfolio weights, and the potential nonuniqueness of optimal weights.The results also address in detail the phenomenon that if the risk-free rate lies above a certain threshold, the usually envisioned master fund must be replaced by one of alternative type, representing a “net short position” instead of a “net long position” in the risky instruments. For nonsymmetric deviation measures, the second type need not just be the reverse of the first type, and there can sometimes even be an interval for the risk-free rate in which no master fund of either type exists. A notion of basic fund, in place of master fund, is brought in to get around this difficulty and serve as a single guide to optimality regardless of such circumstances.
Article
We study a space of coherent risk measures Mφ obtained as certain expansions of coherent elementary basis measures. In this space, the concept of “risk aversion function” φ naturally arises as the spectral representation of each risk measure in a space of functions of confidence level probabilities. We give necessary and sufficient conditions on φ for Mφ to be a coherent measure. We find in this way a simple interpretation of the concept of coherence and a way to map any rational investor's subjective risk aversion onto a coherent measure and vice-versa. We also provide for these measures their discrete versions M(N)φ acting on finite sets of N independent realizations of a r.v. which are not only shown to be coherent measures for any fixed N, but also consistent estimators of Mφ for large N.