# 8 Other risk measures¶

In the definitions (2.1)-(2.4) of the classical MVO problem the variance (or standard deviation) of portfolio return is used as the measure of risk, making computation easy and convenient. If we assume that the portfolio return is normally distributed, then the variance is in fact the optimal risk measure, because then MVO can take into account all information about return. Moreover, if \(T\) is large compared to \(N\), the sample mean and covariance matrix are maximum likelihood estimates (MLE), implying that the optimal portfolio resulting from estimated inputs will also be a MLE of the true optimal portfolio [Lau01].

Empirical observations suggest however, that the distribution of linear return is often skewed, and even if it is elliptically symmetric, it can have fat tails and can exhibit tail dependence [1]. As the distribution moves away from normality, the performance of a variance based portfolio estimator can quickly degrade.

No perfect risk measure exists, though. Depending on the distribution of return, different measures can be more appropriate in different situations, capturing different characteristics of risk. Return of diversified equity portfolios are often approximately symmetric over periods of institutional interest. Options, swaps, hedge funds, and private equity have return distributions that are unlikely to be symmetric. Also less symmetric are distributions of fixed-income and real estate index returns, and diversified equity portfolios over long time horizons [MM08].

The measures presented here are expectations, meaning that optimizing them would lead to stochastic optimization problems in general. As a simplification, we consider only their sample average approximation by assuming that a set of linear return scenarios are given.

## 8.1 Deviation measures¶

This class of measures quantify the variability around the mean of the return distribution, similar to variance. If we associate investment risk with the uncertainty of return, then such measures could be ideal.

Let \(X, Y\) be random variables of investment returns, and let \(D\) be a mapping. The general properties that deviation measures should satisfy are [RUZ06]:

Positivity: \(D(X) \geq 0\) with \(D(X) = 0\) only if \(X = c \in \mathbb{R}\),

Translation invariance: \(D(X + c) = D(X),\ c \in \mathbb{R}\),

Subadditivity: \(D(X + Y) \leq D(X) + D(Y)\),

Positive homogeneity: \(D(cX) = cD(X),\ c \geq 0\)

Note that positive homegeneity and subadditivity imply the *convexity* of \(D\).

### 8.1.1 Robust statistics¶

A group of deviation measures are robust statistics. Maximum likelihood estimators are highly sensitive to deviations from the assumed distribution. Thus if the return is not normally distributed, robust portfolio estimators can be an alternative. These are less efficient than MLE for normal distribution, but their performance degrades less quickly under departures from normality. One suitable class of estimators to express portfolio risk are M-estimators [VD09]:

where \(\delta\) is a convex loss function with unique minimum at zero. It has the sample average approximation \(\min_q \frac{1}{T}\sum_{k=1}^T\delta(\mathbf{x}^\mathsf{T}\mathbf{r}_k - q)\).

#### 8.1.1.1 Variance¶

Choosing \(\delta_\mathrm{var}(y) = \frac{1}{2}y^2\) we get \(D_{\delta_\mathrm{var}}\), a different expression for the portfolio variance. The optimal \(q\) in this case is the portfolio mean return \(\mu_\mathbf{x}\). This gives us a way to perform standard mean–variance optimization directly using scenario data, without the need for a sample covariance matrix:

By defining new variables \(t_k^+ = \mathrm{max}\left(\mathbf{x}^\mathsf{T}\mathbf{r}_k - q, 0\right)\) and \(t_k^- = \mathrm{max}\left(-\mathbf{x}^\mathsf{T}\mathbf{r}_k + q, 0\right)\), then using Sec. 10.1.1.1 (Maximum function) we arrive at a QO model:

where \(\mathbf{R} = [\mathbf{r}_1, \dots, \mathbf{r}_T]\) is the scenario matrix, where each column is a scenario. The norms can be further modeled by the rotated quadratic cone; see in Sec. 10.1.1.8 (Squared Euclidean norm). While this formulation does not need covariance estimation, it can still suffer from computational inefficiency because the number of variables and the number of constraints depend on the sample size \(T\).

#### 8.1.1.2 \(\alpha\)-shortfall¶

Choosing \(\delta_{\mathrm{sf},\alpha}(y) = \alpha y - \mathrm{min}(y, 0)\) or equivalently \(\delta_{\mathrm{sf},\alpha}(y) = \alpha \mathrm{max}(y, 0) + (1-\alpha) \mathrm{max}(-y, 0)\) for \(\alpha\in (0, 1)\) will give the \(\alpha\)*-shortfall* \(D_{\delta_{\mathrm{sf},\alpha}}\) studied in [Lau01]. The optimal \(q\) will be the \(\alpha\)-quantile of the portfolio return. \(\alpha\)-shortfall is a deviation measure with favorable robustness properties when the portfolio return has an asymmetric distribution, fat tailed symmetric distribution, or exhibits multivariate tail-dependence. The \(\alpha\) parameter adjusts the level of asymmetry in the \(\delta\) function, allowing us to penalize upside and downside returns differently.

Portfolio optimization using sample \(\alpha\)-shortfall can be formulated as

By defining new variables \(t_k^- = \mathrm{max}(-\mathbf{x}^\mathsf{T}\mathbf{r}_k + q, 0)\), and model accoring to Sec. 10.1.1.1 (Maximum function) we arrive at an LO model:

A drawback of this model is that the number of constraints depends on the sample size \(T\), resulting in computational inefficiency for large number of samples.

For elliptically symmetric return distributions the \(\alpha\)-shortfall is equivalent to the variance in the sense that the corresponding sample portfolio estimators will be estimating the same portfolio. In fact for normal distribution, the \(\alpha\)*-shortfall* is proportional to the portfolio variance: \(D_{\delta_{\mathrm{sf},\alpha}} = \frac{\phi(q_\alpha)}{\alpha}\sqrt{D_{\delta_{\mathrm{var}}}}\), where \(\phi(q_\alpha)\) is the value of the standard normal density function at its \(\alpha\)-quantile.

However when return is symmetric but not normally distributed, the sample portfolio estimators for \(\alpha\)-shortfall can have less estimation error than sample portfolio estimators for variance.

#### 8.1.1.3 MAD¶

A special case of \(\alpha\)-shortfall is for \(\alpha=0.5\). The function \(\delta_{\mathrm{MAD}}(y)=\frac{1}{2}|y|\) gives us \(D_{\delta_{\mathrm{MAD}}}\), which is called the *mean absolute deviation* (MAD) measure or the \(L_1\)*-risk*. For this case the optimal \(q\) will be the median of the portfolio return, and the sample MAD portfolio optimization problem can be formulated as

After modeling the absolute value based on Sec. 10.1.1.3 (Absolute value) we arrive at the following LO:

where \(\mathbf{R}\) is the return data matrix with one observation \(\mathbf{r}_k,\ k = 1,\dots,T\) in each column. Note that the number of constraints in this LO problem again depends on the sample size \(T\).

For normally distributed returns, the MAD is proportional to the variance: \(D_{\delta_{MAD}} = \sqrt{\frac{2}{\pi}}\sqrt{D_{\delta_{\mathrm{var}}}}\).

The \(L_1\)-risk can also be applied without minimizing over \(q\). We can just let \(q\) to be the sample portfolio mean instead, i. e., \(q = \mathbf{x}^\mathsf{T}\EMean\) [KY91].

#### 8.1.1.4 Risk measure from the Huber function¶

Another robust portfolio estimator can be obtained for the case of symmetric return distributions using the Huber function

yielding the risk measure \(D_{\delta_{\mathrm{H}}}\). A different form of the Huber function is \(\delta_\mathrm{H}(y) = \mathrm{min}_u u^2 + 2c|y-u|\), which leads to a QO formulation:

Modeling the absolute value based on Sec. 10.1.1.3 (Absolute value) we get

Note that the size of this problem depends on the number of samples \(T\).

### 8.1.2 Downside deviation measures¶

In financial context many distributions are skewed. Investors might only be concerned with negative deviations (losses) relative to the expected portfolio return \(\mu_\mathbf{x}\), or in general with falling short of some benchmark return \(r_\mathrm{bm}\). In these cases downside deviation measures are more appropriate.

A class of downside deviation measures are *lower partial moments* of order \(n\):

where \(r_\mathrm{bm}\) is some given target return. The discrete version of this measure is \(\sum_{k=1}^T p_k \mathrm{max}(r_\mathrm{bm} - \mathbf{x}^\mathsf{T}\mathbf{r}_k, 0)^n,\) where \(\mathbf{r}_k\) is a scenario of the portfolio return \(R_\mathbf{x}\) occurring with probability \(p_k\).

We have the following special cases:

\(\mathrm{LPM}_0\) is the probability of loss relative to the target return \(r_\mathrm{bm}\). \(\mathrm{LPM}_0\) is an incomplete measure of risk, because it does not provide any indication of how severe the shortfall can be, should it occur. Therefore it is best used as a constraint while optimizing for a different risk measure. This way \(\mathrm{LPM}_0\) can still provide information about the risk tolerance of the investor.

\(\mathrm{LPM}_1\) is the

*expected loss*, also called*target shortfall*.\(\mathrm{LPM}_2\) is called

*target semi-variance*.

While lower partial moments only consider outcomes below the target \(r_\mathrm{bm}\), the optimization still uses the entire distribution of \(R_\mathbf{x}\). The right tail of the distribution (representing the outcomes above the target) is captured in the mean \(\mu_\mathbf{x}\) of the distribution.

The \(\mathrm{LPM}_n\) optimization problem can be formulated as [Lau01]

If we define the new variables \(t_k^- = \mathrm{max}(r_\mathrm{bm} - \mathbf{x}^\mathsf{T}\mathbf{r}_k, 0)\), then for \(n=1\) we arrive at an LO and for \(n=2\) we arrive at a QO problem:

where \(\mathbf{t}^-\) is \(T\) dimensional vector. Thus there will be \(N+T\) variables, and the number of constraints will also depend on the sample size \(T\).

Contrary to the variance, \(\mathrm{LPM}_n\) measures are consistent with more general classes of investor utility functions, and assume less about the return distribution. Thus they better reflect investor preferences and are valid under a broader range of conditions [Har91].

However, LPMs are only useful for skewed distributions. If the return distribution is symmetric, \(\mathrm{LPM}_1\) and \(\mathrm{LPM}_2\) based portfolio estimates will be equivalent to MAD and variance based ones. In particular the MAD for a random variable with density \(f(x)\) will be the same as \(\mathrm{LPM}_1\) of a random variable with density \(f(x)+f(-x)\). Also, the mean-LPM optimization model is very sensitive to the changes in the target value.

## 8.2 Tail risk measures¶

Tail risk measures try to capture the risk of extreme events. The measures described below are commonly defined for random variables treating loss as positive number, so to apply them on security returns \(R\) we have to flip the sign and define loss as \(L=-R\).

Let \(X, Y\) be random variables of investment losses, and let \(\tau\) be a mapping. The formal properties that a reasonable tail risk measure should satisfy are the following [FS02]:

Monotonicity: If \(X \leq Y\), then \(\tau(X) \leq \tau(Y)\),

Translation invariance: \(\tau(X + m) = \tau(X) + m,\ m \in \mathbb{R}\),

Subadditivity: \(\tau(X + Y) \leq \tau(X) + \tau(Y)\),

Positive homogeneity: \(\tau(cX) = c\tau(X),\ c \geq 0\)

If \(\tau\) satisfies these properties then it is called *coherent risk measure*. If we relax condition 3 and 4 to the property of *convexity*, then we get the more general class of *convex risk measures*. Not all risk measures used in practice satisfy these properties, violation of condition 1 and/or 2 is typical. However, in the context of portfolio selection these are less critical, the property of convexity is most important to achieve risk diversification [Lau01].

### 8.2.1 Value-at-Risk¶

Denote the \(\alpha\)-quantile of a random variable \(L\) by \(q_\alpha(L) = \mathrm{inf}\{x|\mathbb{P}(L \leq x) \geq \alpha\}\). If \(L\) is the loss over a given time horizon, then the *value-at-risk* (VaR) of \(L\) with confidence level \(\alpha\) (or risk level \(1-\alpha\)) is defined to be

This is the amount of loss (a positive number) over the given time horizon that will not be exceeded with probability \(\alpha\). However, VaR does not give information about the magnitude of loss in case of the \(1-\alpha\) probability event when losses are greater than \(q_{\alpha}(L)\). Also it is not a coherent risk measure (it does not respect convexity property), meaning it can discourage diversification. [2] In other words, portfolio VaR may not be lower than the sum of the VaRs of the individual securities.

### 8.2.2 Conditional Value-at-Risk¶

A modification of VaR that is a coherent risk measure is *conditional value-at-risk* (CVaR). CVaR is the average of the \(1-\alpha\) fraction of worst case losses, i. e., the losses equal to or exceeding \(\mathrm{VaR}_\alpha(L)\). Its most general definition (applicable also for loss distributions with possible discontinuities) can be found in [RU02]. Let \(\lambda_\alpha = \frac{1}{1-\alpha}(\mathbb{P}(L \leq q_{\alpha}(L)) - \alpha)\). Then the CVaR of \(L\) with confidence level \(\alpha\) will be

which is a linear combination of VaR and the quantity called *mean shortfall*. The latter is also not coherent on its own.

If the distribution function \(\mathbb{P}(L \leq q)\) is continuous at \(q_{\alpha}(L)\) then \(\lambda_\alpha = 0\). If there is discontinuity and we have to account for a probability mass concentrated at \(q_{\alpha}(L)\), then \(\lambda_\alpha\) is nonzero in general. This is often the case in practice, when the loss distribution is discrete, for example for scenario based approximations.

#### 8.2.2.1 CVaR for discrete distribution¶

Suppose that the loss distribution is described by points \(q_1 < \dots < q_T\) with nonzero probabilities \(p_1, \dots, p_T\), and \(i_\alpha \in \{1,\dots,T\}\) is the index such that \(\sum_{i=1}^{i_\alpha-1}p_i < \alpha \leq \sum_{i=1}^{i_\alpha}p_i\). Then \(\mathrm{VaR}_\alpha(L) = q_{i_\alpha}\) and

where \(\lambda_\alpha = \frac{1}{1-\alpha}(\sum_{i=1}^{i_\alpha} p_i - \alpha)\). As a special case, if we assume that \(p_i = \frac{1}{T}\), i. e., we have a sample average approximation, then the above formula simplifies with \(i_\alpha = \lceil\alpha T\rceil\).

It can be seen that \(\mathrm{VaR}_\alpha(L) \leq \mathrm{CVaR}_\alpha(L)\) always holds. CVaR is also consistent with second-order stochastic dominance (SSD), i. e., the CVaR efficient portfolios are the ones actually preferred by some wealth-seeking risk-averse investors.

#### 8.2.2.2 Portfolio optimization with CVaR¶

If we substitute portfolio loss scenarios into formula (8.11), we can see that the quantile \((-\mathbf{R}^\mathsf{T}\mathbf{x})_{i_\alpha}\) will depend on \(\mathbf{x}\). It follows that the ordering of the scenarios and the index \(i_\alpha\) will also depend on \(\mathbf{x}\), making it difficult to optimize. However, note that formula (8.11) is actually the linear combination of largest elements of the vector \(\mathbf{q}\). We can thus apply Sec. 10.1.1.5 (Linear combination of largest elements) to get the dual form of \(\mathrm{CVaR}_\alpha(L)\), which is an LO problem:

Note that problem (8.12) is equivalent to

This convex function (8.13) is exactly the one found in [RU00], where it is also proven to be valid for continuous probability distributions as well.

Now we can substitute the portfolio return into \(\mathbf{q}\), and optimize over \(\mathbf{x}\) to find the portfolio that minimizes CVaR:

Because CVaR is represented as a convex function in formula (8.13), we can also formulate an LO to maximize expected return, while limiting risk in terms of CVaR:

The drawback of optimizing CVaR using problems (8.14) or (8.15) is that both the number of variables and the number of constraints depend on the number of scenarios \(T\). This can make the LO model computationally expensive for very large number of samples. For example if the distribution of return is not known, we might need to obtain or simulate a substantial amount of samples, depending on the confidence level \(\alpha\).

### 8.2.3 Entropic Value-at-Risk¶

A more general risk measure in this class is the *entropic value-at-risk* (EVaR). EVaR is also a coherent risk measure, with additional favorable monotonicity properties; see in [AJ12]. It is defined as the tightest upper bound on VaR obtained from the Chernoff inequality:

where \(M_L(s) = \mathbb{E}(\mathrm{e}^{sL})\) is the moment generating function of \(L\). The EVaR incorporates losses both less and greater than the VaR simultaneously, thus it always holds that \(\mathrm{CVaR}_\alpha(L) \leq \mathrm{EVaR}_\alpha(L)\).

#### 8.2.3.1 EVaR for discrete distribution¶

Based on the definition (8.16), the discrete version of the EVaR will be

We can make formula (8.17) convex by substituting a new variable \(t = \frac{1}{s}\):

We can transform formula (8.18) into a conic optimization problem by substituting the first term of the objective with a new variable \(z\) and adding the new constraint \(z \geq \mathrm{EVaR}_\alpha(L)\). Then we apply the rule Sec. 10.1.1.14 (Perspective of log-sum-exp):

#### 8.2.3.2 Portfolio optimization with EVaR¶

Now we can substitute the portfolio return into \(\mathbf{q}\), and optimize over \(\mathbf{x}\) to find the portfolio that minimizes EVaR:

Because EVaR is represented as a convex function (8.18), we can also formulate a conic problem to maximize expected return, while limiting risk in terms of EVaR:

A disadvantage of the EVaR conic model is that it still depends on \(T\) in the number of exponential cone contraints.

Note that if we assume the return distribution to be a Gaussian mixture, we can find a different approach to computing EVaR. See in Sec. 8.3.2 (EVaR using Gaussian mixture return).

### 8.2.4 Relationship between risk measures¶

Suppose that \(\tau(X)\) is a coherent tail risk measure that additionally satisfies \(\tau(X) > -\mathbb{E}(X)\). Suppose also that \(D(X)\) is a deviation measure that additionally satisfies \(D(X) \leq \mathbb{E}(X)\) for all \(X \geq 0\). Then these subclasses of risk measures can be related through the following identities:

For example, \(\alpha\)-shortfall and CVaR is related this way. Details can be found in [RUZ06].

### 8.2.5 Practical considerations¶

Many risk measures in this chapter can be computed in practice through scenario based approximations. The relevant optimization problems can then be conveniently formulated as LO or QO problems. The advantage of these models is that they do not need a covariance matrix estimate. The simplification of the problem structure, however, comes at the cost of increasing the problem dimension by introducing a number of new variables and constraints proportional to the number of scenarios \(T\).

If \(T\) is large, this can lead to computational burden not only because of the increased number of variables and constraints, but also because the scenario matrix \(\mathbf{R}\) also depending on \(T\) becomes very non-sparse, making the solution process less efficient.

If \(T\) is too small, specifically \(T < N\), then there will be fewer scenarios then parameters to estimate. This makes the portfolio optimization problem very imprecise, and prone to overfit the data. Also depending on other constraints, solution of LO problems can have computational difficulties. To mitigate this, we can use regularization methods, such as \(L_1\) (lasso) or \(L_2\) (ridge) penalty terms in the objective:

where \(\mathcal{R}_\mathbf{x}\) is the portfolio risk, \(\mathbf{x}_\mathrm{pri}\) is a prior portfolio, and \(\lambda\) is the regularization strength parameter. The penalty term will ensure that the extreme deviations from the prior are unlikely. Thus \(\lambda\) basically sets the investor confidence in the portfolio \(\mathbf{x}_\mathrm{pri}\). Using \(p = 1, 2\) we arrive at LO and QO problems respectively. See details in [Lau01].

## 8.3 Expected utility maximization¶

Apart from selecting different risk measures, we can also approach portfolio optimization through the use of utility functions. We specify a concave and increasing utility function that measures the investors preference for each specific outcome. Then the objective is to maximize the expected utility under the return distribution.

Expected utility maximization can take into account any type of return distribution, while MVO works only with the first two moments, therefore it is accurate only for normally distributed return. MVO is also equivalent with expected utility maximization if the utility function is quadratic, because it models the investors indifference about higher moments of return. The only advantage of MVO in this comparison is that it works without having to discretize the return distribution and work with scenarios.

### 8.3.1 Optimal portfolio using gaussian mixture return¶

In [LB22] an expected utility maximization approach is taken, assuming that the return distribution is a Gaussian mixture (GM). The benefits of a GM distribution are that it can approximate any continuous distribution, including skewed and fat-tailed ones. Also its components can be interpreted as return distributions given a specific market regime. Moreover, the expected utility maximization using this return model can be formulated as a convex optimization problem without needing return scenarios, making this approach as efficient and scalable as MVO.

We denote security returns having Gaussian mixture (GM) distribution with \(K\) components as

where \(\mu_i\) is the mean, \(\Sigma_i\) is the covariance matrix, and \(\pi_i\) is the probability of component \(i\). As special cases, for \(K=1\) we get the normal distribution \(\mathcal{N}(\mu_1, \Sigma_1)\), and for \(\Sigma_i = 0,\ i = 1,\dots,K\) we get a scenario distribution with \(K\) return scenarios \(\{\mu_1, \dots, \mu_K\}\).

Using definition (8.24), the distribution of portfolio return \(R_\mathbf{x}\) will also be GM:

where \(\mu_{\mathbf{x},i} = \mu_i^\mathsf{T}\mathbf{x}\), and \(\sigma_{\mathbf{x},i}^2 = \mathbf{x}^\mathsf{T}\Sigma_i\mathbf{x}\).

To select the optimal portfolio we use the exponential utility function \(U_\delta(x) = 1-e^{-\delta x}\), where \(\delta > 0\) is the risk aversion parameter. This choice allows us to write the expected utility of the portfolio return \(R_\mathbf{x}\) using the moment generating function:

Thus maximizing the function (8.26) is the same as minimizing the moment generating function of the portfolio return, or equivalently, we can minimize its logarithm, the cumulant generating function:

If \(R_\mathbf{x}\) is assumed to be a GM random variable, then its cumulant generating function will be the following:

The function (8.28) is convex in \(\mathbf{x}\) because it is a composition of the convex and increasing log-sum-exp function and a convex quadratic function. Note that for \(K=1\), we get back the same quadratic utility objective as in version (2.3) of the MVO problem.

Assuming we have the GM distribution parameter estimates \(\EMean_i,\ECov_i,\ i = 1, \dots, K\), we can apply Sec. 10.1.1.9 (Quadratic form) and Sec. 10.1.1.12 (Log-sum-exp) to arrive at the conic model of the utility maximization problem (8.27):

where \(q_i\) is an upper bound for the quadratic term \(\frac{1}{2}\mathbf{x}^\mathsf{T}\ECov_i\mathbf{x}\), and \(\mathbf{G}_i\) is such that \(\ECov_i = \mathbf{G}_i\mathbf{G}_i^\mathsf{T}\).

### 8.3.2 EVaR using Gaussian mixture return¶

We introduced Entropic Value-at-Risk (EVaR) in Sec. 8.2.3 (Entropic Value-at-Risk). EVaR can also be expressed using the cumulant generating function \(K_L(s) = \log(M_L(s))\) of the loss \(L\):

After substituting the return \(R = -L\) instead of the loss, we see that \(K_{-R}(s) = K_R(-s)\). Assuming now that \(R = R_\mathbf{x}\) is the portfolio return, we get the following optimization problem:

We can find a connection between EVaR computation (8.31) and maximization of expected exponential utility (8.27). Suppose that the pair \((\mathbf{x}^*, s^*)\) is optimal for problem (8.31). Then \(\mathbf{x}^*\) also optimal for problem (8.27), with risk aversion parameter \(\delta = s^*\).

By assuming a GM distribution for security return, we can optimize problem (8.31) without needing a scenario distribution. First, to formulate it as a convex optimization problem, define the new variable \(t = \frac{1}{s}\):

We can observe that the first term in formula (8.32) is the perspective of \(K_{R_\mathbf{x}}(-1)\). Therefore by substituting the cumulant generating function (8.28) of the GM portfolio return, we get a convex function in the portfolio \(\mathbf{x}\):

Assuming we have the GM distribution parameter estimates \(\EMean_i,\ECov_i,\ i = 1, \dots, K\), we can apply Sec. 10.1.1.9 (Quadratic form) and Sec. 10.1.1.14 (Perspective of log-sum-exp) to arrive at the conic model of the EVaR minimization problem (8.31):

where \(q_i\) is an upper bound for the quadratic term \(\frac{1}{2}\mathbf{x}^\mathsf{T}\ECov_i\mathbf{x}\), and \(\mathbf{G}_i\) is such that \(\ECov_i = \mathbf{G}_i\mathbf{G}_i^\mathsf{T}\).

The huge benefit of this EVaR formulation is that its size does not depend on the number of scenarios, because it is derived without using a scenario distribution. It depends only on the number of GM components \(K\).

## 8.4 Example¶

This example shows how can we compute the CVaR efficient frontier using the dual form of CVaR in **MOSEK** Fusion.

### 8.4.1 Scenario generation¶

The input data is again obtained the same way as detailed in Sec. 3.4.2 (Data collection), but we do only the steps until Sec. 3.4.3.3 (Projection of invariants). This way we get the expected return estimate \(\EMean^\mathrm{log}\) and the covariance matrix estimate \(\ECov^\mathrm{log}\) of yearly logarithmic returns. These returns have approximately normal distribution, so we can easily generate \(T\) number of return scenarios from them, using the numpy default random number generator. Here we choose \(T=99999\). This number ensures to have a nonzero \(\lambda_\alpha\) in formula (8.11), which is the most general case.

```
# Number of scenarios
T = 99999
# Generate logarithmic return scenarios assuming normal distribution
R_log = np.random.default_rng().multivariate_normal(m_log, S_log, T)
```

Next, we convert the received logarithmic return scenarios to linear return scenarios using the inverse of formula (3.2).

```
# Convert logarithmic return scenarios to linear return scenarios
R = np.exp(scenarios_log) - 1
R = R.T
```

We transpose the resulting matrix just to remain consistent with the notation in this chapter, namely that each column of \(\mathbf{R}\) is a scenario. We also set the scenario probabilities to be \(\frac{1}{T}\):

```
# Scenario probabilities
p = np.ones(T) / T
```

### 8.4.2 The optimization problem¶

The optimization problem we solve here resembles problem (2.12), but we will change the risk measure from portfolio standard deviation to portfolio CVaR:

Applying the dual CVaR formula (8.13), we get:

We know that formula (8.13) equals CVaR only if it is minimal in \(t\). This will be satisfied in the optimal solution of problem (8.36), because the CVaR term is implicitly forced to become smaller, and \(t\) is independent from \(\mathbf{x}\).

Now we model the maximum function, and arrive at the following LO model of the mean-CVaR efficient frontier:

### 8.4.3 The Fusion model¶

Here we show the Fusion model of problem (8.37). We give the constraints modeling the maximum function in a separate function called `maximum`

:

```
# max(x, y) <= t
def maximum(M, x, y, t):
M.constraint(Expr.sub(t, x), Domain.greaterThan(0.0))
M.constraint(Expr.sub(t, y), Domain.greaterThan(0.0))
def EfficientFrontier(N, T, m, R, p, alpha, deltas):
with Model("Case study") as M:
# Settings
#M.setLogHandler(sys.stdout)
# Variables
# The variable x is the fraction of holdings relative to the initial capital.
# It is constrained to take only positive values.
x = M.variable("x", N, Domain.greaterThan(0.0))
# Budget constraint
M.constraint('budget', Expr.sum(x), Domain.equalsTo(1.0))
# Auxiliary variables.
t = M.variable("t", 1, Domain.unbounded())
u = M.variable("u", T, Domain.unbounded())
# Constraint modeling maximum
maximum(M, Expr.sub(Expr.mul(-R.T, x), Expr.repeat(t, T, 0)), Expr.constTerm(T, 0.0), u)
# Objective
delta = M.parameter()
cvar_term = Expr.add(t, Expr.mul(1/(1-alpha), Expr.dot(p, u)))
M.objective('obj', ObjectiveSense.Maximize, Expr.sub(Expr.dot(m, x), Expr.mul(delta, cvar_term)))
# Create DataFrame to store the results.
columns = ["delta", "obj", "return", "risk"] + df_prices.columns.tolist()
df_result = pd.DataFrame(columns=columns)
for d in deltas:
# Update parameter
delta.setValue(d)
# Solve optimization
M.solve()
# Save results
portfolio_return = m @ x.level()
portfolio_risk = t.level()[0] + 1/(1-alpha) * np.dot(p, u.level())
row = pd.Series([d, M.primalObjValue(), portfolio_return, portfolio_risk] + list(x.level()), index=columns)
df_result = df_result.append(row, ignore_index=True)
return df_result
```

Next, we compute the efficient frontier. We select the confidence level \(\alpha = 0.95\). The following code produces the optimization results:

```
alpha = 0.95
# Compute efficient frontier with and without shrinkage
deltas = np.logspace(start=-1, stop=2, num=20)[::-1]
df_result = EfficientFrontier(N, T, m, R, p, alpha, deltas)
```

On Fig. 8.1 we can see the risk-return plane, and on Fig. 8.2 the portfolio composition for different levels of risk.

Footnotes