Entropic value at risk

In financial mathematics and stochastic optimization, the concept of risk measure is used to quantify the risk involved in a random outcome or risk position. Many risk measures have hitherto been proposed, each having certain characteristics. The entropic value at risk (EVaR) is a coherent risk measure introduced by Ahmadi-Javid,[1][2] which is an upper bound for the value at risk (VaR) and the conditional value at risk (CVaR), obtained from the Chernoff inequality. The EVaR can also be represented by using the concept of relative entropy. Because of its connection with the VaR and the relative entropy, this risk measure is called "entropic value at risk". The EVaR was developed to tackle some computational inefficiencies[clarification needed] of the CVaR. Getting inspiration from the dual representation of the EVaR, Ahmadi-Javid[1][2] developed a wide class of coherent risk measures, called g-entropic risk measures. Both the CVaR and the EVaR are members of this class.

Definition

Let ( Ω , F , P ) {\displaystyle (\Omega ,{\mathcal {F}},P)} be a probability space with Ω {\displaystyle \Omega } a set of all simple events, F {\displaystyle {\mathcal {F}}} a σ {\displaystyle \sigma } -algebra of subsets of Ω {\displaystyle \Omega } and P {\displaystyle P} a probability measure on F {\displaystyle {\mathcal {F}}} . Let X {\displaystyle X} be a random variable and L M + {\displaystyle \mathbf {L} _{M^{+}}} be the set of all Borel measurable functions X : Ω R {\displaystyle X:\Omega \to \mathbb {R} } whose moment-generating function M X ( z ) {\displaystyle M_{X}(z)} exists for all z 0 {\displaystyle z\geq 0} . The entropic value at risk (EVaR) of X L M + {\displaystyle X\in \mathbf {L} _{M^{+}}} with confidence level 1 α {\displaystyle 1-\alpha } is defined as follows:

EVaR 1 α ( X ) := inf z > 0 { z 1 ln ( M X ( z ) α ) } . {\displaystyle {\text{EVaR}}_{1-\alpha }(X):=\inf _{z>0}\left\{z^{-1}\ln \left({\frac {M_{X}(z)}{\alpha }}\right)\right\}.} (1)

In finance, the random variable X L M + , {\displaystyle X\in \mathbf {L} _{M^{+}},} in the above equation, is used to model the losses of a portfolio.

Consider the Chernoff inequality

Pr ( X a ) e z a M X ( z ) , z > 0. {\displaystyle \Pr(X\geq a)\leq e^{-za}M_{X}(z),\quad \forall z>0.} (2)

Solving the equation e z a M X ( z ) = α {\displaystyle e^{-za}M_{X}(z)=\alpha } for a , {\displaystyle a,} results in

a X ( α , z ) := z 1 ln ( M X ( z ) α ) . {\displaystyle a_{X}(\alpha ,z):=z^{-1}\ln \left({\frac {M_{X}(z)}{\alpha }}\right).}

By considering the equation (1), we see that

EVaR 1 α ( X ) := inf z > 0 { a X ( α , z ) } , {\displaystyle {\text{EVaR}}_{1-\alpha }(X):=\inf _{z>0}\{a_{X}(\alpha ,z)\},}

which shows the relationship between the EVaR and the Chernoff inequality. It is worth noting that a X ( 1 , z ) {\displaystyle a_{X}(1,z)} is the entropic risk measure or exponential premium, which is a concept used in finance and insurance, respectively.

Let L M {\displaystyle \mathbf {L} _{M}} be the set of all Borel measurable functions X : Ω R {\displaystyle X:\Omega \to \mathbb {R} } whose moment-generating function M X ( z ) {\displaystyle M_{X}(z)} exists for all z {\displaystyle z} . The dual representation (or robust representation) of the EVaR is as follows:

EVaR 1 α ( X ) = sup Q ( E Q ( X ) ) , {\displaystyle {\text{EVaR}}_{1-\alpha }(X)=\sup _{Q\in \Im }(E_{Q}(X)),} (3)

where X L M , {\displaystyle X\in \mathbf {L} _{M},} and {\displaystyle \Im } is a set of probability measures on ( Ω , F ) {\displaystyle (\Omega ,{\mathcal {F}})} with = { Q P : D K L ( Q | | P ) ln α } {\displaystyle \Im =\{Q\ll P:D_{KL}(Q||P)\leq -\ln \alpha \}} . Note that

D K L ( Q | | P ) := d Q d P ( ln d Q d P ) d P {\displaystyle D_{KL}(Q||P):=\int {\frac {dQ}{dP}}\left(\ln {\frac {dQ}{dP}}\right)dP}

is the relative entropy of Q {\displaystyle Q} with respect to P , {\displaystyle P,} also called the Kullback–Leibler divergence. The dual representation of the EVaR discloses the reason behind its naming.

Properties

  • The EVaR is a coherent risk measure.
  • The moment-generating function M X ( z ) {\displaystyle M_{X}(z)} can be represented by the EVaR: for all X L M + {\displaystyle X\in \mathbf {L} _{M^{+}}} and z > 0 {\displaystyle z>0}
M X ( z ) = sup 0 < α 1 { α exp ( z EVaR 1 α ( X ) ) } . {\displaystyle M_{X}(z)=\sup _{0<\alpha \leq 1}\{\alpha \exp(z{\text{EVaR}}_{1-\alpha }(X))\}.} (4)
  • For X , Y L M {\displaystyle X,Y\in \mathbf {L} _{M}} , EVaR 1 α ( X ) = EVaR 1 α ( Y ) {\displaystyle {\text{EVaR}}_{1-\alpha }(X)={\text{EVaR}}_{1-\alpha }(Y)} for all α ] 0 , 1 ] {\displaystyle \alpha \in ]0,1]} if and only if F X ( b ) = F Y ( b ) {\displaystyle F_{X}(b)=F_{Y}(b)} for all b R {\displaystyle b\in \mathbb {R} } .
  • The entropic risk measure with parameter θ , {\displaystyle \theta ,} can be represented by means of the EVaR: for all X L M + {\displaystyle X\in \mathbf {L} _{M^{+}}} and θ > 0 {\displaystyle \theta >0}
θ 1 ln M X ( θ ) = a X ( 1 , θ ) = sup 0 < α 1 { EVaR 1 α ( X ) + θ 1 ln α } . {\displaystyle \theta ^{-1}\ln M_{X}(\theta )=a_{X}(1,\theta )=\sup _{0<\alpha \leq 1}\{{\text{EVaR}}_{1-\alpha }(X)+\theta ^{-1}\ln \alpha \}.} (5)
  • The EVaR with confidence level 1 α {\displaystyle 1-\alpha } is the tightest possible upper bound that can be obtained from the Chernoff inequality for the VaR and the CVaR with confidence level 1 α {\displaystyle 1-\alpha } ;
VaR ( X ) CVaR ( X ) EVaR ( X ) . {\displaystyle {\text{VaR}}(X)\leq {\text{CVaR}}(X)\leq {\text{EVaR}}(X).} (6)
  • The following inequality holds for the EVaR:
E ( X ) EVaR 1 α ( X ) esssup ( X ) {\displaystyle {\text{E}}(X)\leq {\text{EVaR}}_{1-\alpha }(X)\leq {\text{esssup}}(X)} (7)
where E ( X ) {\displaystyle {\text{E}}(X)} is the expected value of X {\displaystyle X} and esssup ( X ) {\displaystyle {\text{esssup}}(X)} is the essential supremum of X {\displaystyle X} , i.e., inf t R { t : Pr ( X t ) = 1 } {\displaystyle \inf _{t\in \mathbb {R} }\{t:\Pr(X\leq t)=1\}} . So do hold EVaR 0 ( X ) = E ( X ) {\displaystyle {\text{EVaR}}_{0}(X)={\text{E}}(X)} and lim α 0 EVaR 1 α ( X ) = esssup ( X ) {\displaystyle \lim _{\alpha \to 0}{\text{EVaR}}_{1-\alpha }(X)={\text{esssup}}(X)} .

Examples

Comparing the VaR, CVaR and EVaR for the standard normal distribution
Comparing the VaR, CVaR and EVaR for the uniform distribution over the interval (0,1)

For X N ( μ , σ 2 ) , {\displaystyle X\sim N(\mu ,\sigma ^{2}),}

EVaR 1 α ( X ) = μ + σ 2 ln α . {\displaystyle {\text{EVaR}}_{1-\alpha }(X)=\mu +\sigma {\sqrt {-2\ln \alpha }}.} (8)

For X U ( a , b ) , {\displaystyle X\sim U(a,b),}

EVaR 1 α ( X ) = inf t > 0 { t ln ( t e t 1 b e t 1 a b a ) t ln α } . {\displaystyle {\text{EVaR}}_{1-\alpha }(X)=\inf _{t>0}\left\lbrace t\ln \left(t{\frac {e^{t^{-1}b}-e^{t^{-1}a}}{b-a}}\right)-t\ln \alpha \right\rbrace .} (9)

Figures 1 and 2 show the comparing of the VaR, CVaR and EVaR for N ( 0 , 1 ) {\displaystyle N(0,1)} and U ( 0 , 1 ) {\displaystyle U(0,1)} .

Optimization

Let ρ {\displaystyle \rho } be a risk measure. Consider the optimization problem

min w W ρ ( G ( w , ψ ) ) , {\displaystyle \min _{{\boldsymbol {w}}\in {\boldsymbol {W}}}\rho (G({\boldsymbol {w}},{\boldsymbol {\psi }})),} (10)

where w W R n {\displaystyle {\boldsymbol {w}}\in {\boldsymbol {W}}\subseteq \mathbb {R} ^{n}} is an n {\displaystyle n} -dimensional real decision vector, ψ {\displaystyle {\boldsymbol {\psi }}} is an m {\displaystyle m} -dimensional real random vector with a known probability distribution and the function G ( w , . ) : R m R {\displaystyle G({\boldsymbol {w}},.):\mathbb {R} ^{m}\to \mathbb {R} } is a Borel measurable function for all values w W . {\displaystyle {\boldsymbol {w}}\in {\boldsymbol {W}}.} If ρ = EVaR , {\displaystyle \rho ={\text{EVaR}},} then the optimization problem (10) turns into:

min w W , t > 0 { t ln M G ( w , ψ ) ( t 1 ) t ln α } . {\displaystyle \min _{{\boldsymbol {w}}\in {\boldsymbol {W}},t>0}\left\{t\ln M_{G({\boldsymbol {w}},{\boldsymbol {\psi }})}(t^{-1})-t\ln \alpha \right\}.} (11)

Let S ψ {\displaystyle {\boldsymbol {S}}_{\boldsymbol {\psi }}} be the support of the random vector ψ . {\displaystyle {\boldsymbol {\psi }}.} If G ( . , s ) {\displaystyle G(.,{\boldsymbol {s}})} is convex for all s S ψ {\displaystyle {\boldsymbol {s}}\in {\boldsymbol {S}}_{\boldsymbol {\psi }}} , then the objective function of the problem (11) is also convex. If G ( w , ψ ) {\displaystyle G({\boldsymbol {w}},{\boldsymbol {\psi }})} has the form

G ( w , ψ ) = g 0 ( w ) + i = 1 m g i ( w ) ψ i , g i : R n R , i = 0 , 1 , , m , {\displaystyle G({\boldsymbol {w}},{\boldsymbol {\psi }})=g_{0}({\boldsymbol {w}})+\sum _{i=1}^{m}g_{i}({\boldsymbol {w}})\psi _{i},\qquad g_{i}:\mathbb {R} ^{n}\to \mathbb {R} ,i=0,1,\ldots ,m,} (12)

and ψ 1 , , ψ m {\displaystyle \psi _{1},\ldots ,\psi _{m}} are independent random variables in L M {\displaystyle \mathbf {L} _{M}} , then (11) becomes

min w W , t > 0 { g 0 ( w ) + t i = 1 m ln M g i ( w ) ψ i ( t 1 ) t ln α } . {\displaystyle \min _{{\boldsymbol {w}}\in {\boldsymbol {W}},t>0}\left\lbrace g_{0}({\boldsymbol {w}})+t\sum _{i=1}^{m}\ln M_{g_{i}({\boldsymbol {w}})\psi _{i}}(t^{-1})-t\ln \alpha \right\rbrace .} (13)

which is computationally tractable. But for this case, if one uses the CVaR in problem (10), then the resulting problem becomes as follows:

min w W , t R { t + 1 α E [ g 0 ( w ) + i = 1 m g i ( w ) ψ i t ] + } . {\displaystyle \min _{{\boldsymbol {w}}\in {\boldsymbol {W}},t\in \mathbb {R} }\left\lbrace t+{\frac {1}{\alpha }}{\text{E}}\left[g_{0}({\boldsymbol {w}})+\sum _{i=1}^{m}g_{i}({\boldsymbol {w}})\psi _{i}-t\right]_{+}\right\rbrace .} (14)

It can be shown that by increasing the dimension of ψ {\displaystyle \psi } , problem (14) is computationally intractable even for simple cases. For example, assume that ψ 1 , , ψ m {\displaystyle \psi _{1},\ldots ,\psi _{m}} are independent discrete random variables that take k {\displaystyle k} distinct values. For fixed values of w {\displaystyle {\boldsymbol {w}}} and t , {\displaystyle t,} the complexity of computing the objective function given in problem (13) is of order m k {\displaystyle mk} while the computing time for the objective function of problem (14) is of order k m {\displaystyle k^{m}} . For illustration, assume that k = 2 , m = 100 {\displaystyle k=2,m=100} and the summation of two numbers takes 10 12 {\displaystyle 10^{-12}} seconds. For computing the objective function of problem (14) one needs about 4 × 10 10 {\displaystyle 4\times 10^{10}} years, whereas the evaluation of objective function of problem (13) takes about 10 10 {\displaystyle 10^{-10}} seconds. This shows that formulation with the EVaR outperforms the formulation with the CVaR (see [2] for more details).

Generalization (g-entropic risk measures)

Drawing inspiration from the dual representation of the EVaR given in (3), one can define a wide class of information-theoretic coherent risk measures, which are introduced in.[1][2] Let g {\displaystyle g} be a convex proper function with g ( 1 ) = 0 {\displaystyle g(1)=0} and β {\displaystyle \beta } be a non-negative number. The g {\displaystyle g} -entropic risk measure with divergence level β {\displaystyle \beta } is defined as

ER g , β ( X ) := sup Q E Q ( X ) {\displaystyle {\text{ER}}_{g,\beta }(X):=\sup _{Q\in \Im }{\text{E}}_{Q}(X)} (15)

where = { Q P : H g ( P , Q ) β } {\displaystyle \Im =\{Q\ll P:H_{g}(P,Q)\leq \beta \}} in which H g ( P , Q ) {\displaystyle H_{g}(P,Q)} is the generalized relative entropy of Q {\displaystyle Q} with respect to P {\displaystyle P} . A primal representation of the class of g {\displaystyle g} -entropic risk measures can be obtained as follows:

ER g , β ( X ) = inf t > 0 , μ R { t [ μ + E P ( g ( X t μ + β ) ) ] } {\displaystyle {\text{ER}}_{g,\beta }(X)=\inf _{t>0,\mu \in \mathbb {R} }\left\lbrace t\left[\mu +{\text{E}}_{P}\left(g^{*}\left({\frac {X}{t}}-\mu +\beta \right)\right)\right]\right\rbrace } (16)

where g {\displaystyle g^{*}} is the conjugate of g {\displaystyle g} . By considering

g ( x ) = { x ln x x > 0 0 x = 0 + x < 0 {\displaystyle g(x)={\begin{cases}x\ln x&x>0\\0&x=0\\+\infty &x<0\end{cases}}} (17)

with g ( x ) = e x 1 {\displaystyle g^{*}(x)=e^{x-1}} and β = ln α {\displaystyle \beta =-\ln \alpha } , the EVaR formula can be deduced. The CVaR is also a g {\displaystyle g} -entropic risk measure, which can be obtained from (16) by setting

g ( x ) = { 0 0 x 1 α + otherwise {\displaystyle g(x)={\begin{cases}0&0\leq x\leq {\frac {1}{\alpha }}\\+\infty &{\text{otherwise}}\end{cases}}} (18)

with g ( x ) = 1 α max { 0 , x } {\displaystyle g^{*}(x)={\tfrac {1}{\alpha }}\max\{0,x\}} and β = 0 {\displaystyle \beta =0} (see [1][3] for more details).

For more results on g {\displaystyle g} -entropic risk measures see.[4]

Disciplined Convex Programming Framework

The disciplined convex programming framework of sample EVaR was proposed by Cajas[5] and has the following form:

EVaR α ( X ) = { min z , t , u t + z ln ( 1 α T ) s.t. z j = 1 T u j ( X j t , z , u j ) K exp j = 1 , , T {\displaystyle {\begin{aligned}{\text{EVaR}}_{\alpha }(X)&=\left\{{\begin{array}{ll}{\underset {z,\,t,\,u}{\text{min}}}&t+z\ln \left({\frac {1}{\alpha T}}\right)\\{\text{s.t.}}&z\geq \sum _{j=1}^{T}u_{j}\\&(X_{j}-t,z,u_{j})\in K_{\text{exp}}\;\forall \;j=1,\ldots ,T\\\end{array}}\right.\end{aligned}}} (19)

where z {\displaystyle z} , t {\displaystyle t} and u {\displaystyle u} are variables; K exp {\displaystyle K_{\text{exp}}} is an exponential cone;[6] and T {\displaystyle T} is the number of observations. If we define w {\displaystyle w} as the vector of weights for N {\displaystyle N} assets, r {\displaystyle r} the matrix of returns and μ {\displaystyle \mu } the mean vector of assets, we can posed the minimization of the expected EVaR given a level of expected portfolio return μ ¯ {\displaystyle {\bar {\mu }}} as follows.

min w , z , t , u t + z ln ( 1 α T ) s.t. μ w τ μ ¯ i = 1 N w i = 1 z j = 1 T u j ( r j w τ t , z , u j ) K exp j = 1 , , T w i 0 ; i = 1 , , N {\displaystyle {\begin{aligned}&{\underset {w,\,z,\,t,\,u}{\text{min}}}&&t+z\ln \left({\frac {1}{\alpha T}}\right)\\&{\text{s.t.}}&&\mu w^{\tau }\geq {\bar {\mu }}\\&&&\sum _{i=1}^{N}w_{i}=1\\&&&z\geq \sum _{j=1}^{T}u_{j}\\&&&(-r_{j}w^{\tau }-t,z,u_{j})\in K_{\text{exp}}\;\forall \;j=1,\ldots ,T\\&&&w_{i}\geq 0\;;\;\forall \;i=1,\ldots ,N\\\end{aligned}}} (20)

Applying the disciplined convex programming framework of EVaR to uncompounded cumulative returns distribution, Cajas[5] proposed the entropic drawdown at risk(EDaR) optimization problem. We can posed the minimization of the expected EDaR given a level of expected return μ ¯ {\displaystyle {\bar {\mu }}} as follows:

min w , z , t , u , d t + z ln ( 1 α T ) s.t. μ w τ μ ¯ i = 1 N w i = 1 z j = 1 T u j ( d j R j w τ t , z , u j ) K exp j = 1 , , T d j R j w τ j = 1 , , T d j d j 1 j = 1 , , T d j 0 j = 1 , , T d 0 = 0 w i 0 ; i = 1 , , N {\displaystyle {\begin{aligned}&{\underset {w,\,z,\,t,\,u,\,d}{\text{min}}}&&t+z\ln \left({\frac {1}{\alpha T}}\right)\\&{\text{s.t.}}&&\mu w^{\tau }\geq {\bar {\mu }}\\&&&\sum _{i=1}^{N}w_{i}=1\\&&&z\geq \sum _{j=1}^{T}u_{j}\\&&&(d_{j}-R_{j}w^{\tau }-t,z,u_{j})\in K_{\text{exp}}\;\forall \;j=1,\ldots ,T\\&&&d_{j}\geq R_{j}w^{\tau }\;\forall \;j=1,\ldots ,T\\&&&d_{j}\geq d_{j-1}\;\forall \;j=1,\ldots ,T\\&&&d_{j}\geq 0\;\forall \;j=1,\ldots ,T\\&&&d_{0}=0\\&&&w_{i}\geq 0\;;\;\forall \;i=1,\ldots ,N\\\end{aligned}}} (21)

where d {\displaystyle d} is a variable that represent the uncompounded cumulative returns of portfolio and R {\displaystyle R} is the matrix of uncompounded cumulative returns of assets.

For other problems like risk parity, maximization of return/risk ratio or constraints on maximum risk levels for EVaR and EDaR, you can see [5] for more details.

The advantage of model EVaR and EDaR using a disciplined convex programming framework, is that we can use softwares like CVXPY [7] or MOSEK[8] to model this portfolio optimization problems. EVaR and EDaR are implemented in the python package Riskfolio-Lib.[9]

See also

References

  1. ^ a b c d Ahmadi-Javid, Amir (2011). "An information-theoretic approach to constructing coherent risk measures". 2011 IEEE International Symposium on Information Theory Proceedings. St. Petersburg, Russia: Proceedings of IEEE International Symposium on Information Theory. pp. 2125–2127. doi:10.1109/ISIT.2011.6033932. ISBN 978-1-4577-0596-0. S2CID 8720196.
  2. ^ a b c d Ahmadi-Javid, Amir (2012). "Entropic value-at-risk: A new coherent risk measure". Journal of Optimization Theory and Applications. 155 (3): 1105–1123. doi:10.1007/s10957-011-9968-2. S2CID 46150553.
  3. ^ Ahmadi-Javid, Amir (2012). "Addendum to: Entropic Value-at-Risk: A New Coherent Risk Measure". Journal of Optimization Theory and Applications. 155 (3): 1124–1128. doi:10.1007/s10957-012-0014-9. S2CID 39386464.
  4. ^ Breuer, Thomas; Csiszar, Imre (2013). "Measuring Distribution Model Risk". arXiv:1301.4832v1 [q-fin.RM].
  5. ^ a b c Cajas, Dany (24 February 2021). "Entropic Portfolio Optimization: a Disciplined Convex Programming Framework". doi:10.2139/ssrn.3792520. S2CID 235319743. {{cite journal}}: Cite journal requires |journal= (help)
  6. ^ Chares, Robert (2009). "Cones and interior-point algorithms for structured convex optimization involving powers andexponentials". S2CID 118322815. {{cite web}}: Missing or empty |url= (help)
  7. ^ "CVXPY".
  8. ^ "MOSEK".
  9. ^ "Riskfolio-Lib". GitHub. 18 July 2022.