Skip to main content
Log in

Portfolio selection in quantile decision models

  • Research Article
  • Published:
Annals of Finance Aims and scope Submit manuscript

Abstract

This paper develops a model for optimal portfolio allocation for an investor with quantile preferences, i.e., who maximizes the \(\tau \)-quantile of the portfolio return, for \(\tau \in (0,1)\). Quantile preferences allow to study heterogeneity in individuals’ portfolio choice by varying the quantiles, and have a solid axiomatic foundation. Their associated risk attitude is captured entirely by a single dimensional parameter (the quantile \(\tau \)), instead of the utility function. We formally establish the properties of the quantile model. The presence of a risk-free asset in the portfolio produces an all-or-nothing optimal response to the risk-free asset that depends on investors’ quantile preference. In addition, when both assets are risky, we derive conditions under which the optimal portfolio decision has an interior solution that guarantees diversification vis-à-vis fully investing in a single risky asset. We also derive conditions under which the optimal portfolio decision is characterized by two regions: full diversification for quantiles below the median and no diversification for upper quantiles. These results are illustrated in an exhaustive simulation study and an empirical application using a tactical portfolio of stocks, bonds and a risk-free asset. The results show heterogeneity in portfolio diversification across risk attitudes.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Notes

  1. Some studies suggesting that individuals did not always employ objective probabilities resulted in, among others, Prospect Theory (Kahneman and Tversky 1979), Rank-Dependent Expected Utility Theory (Quiggin 1982), Cumulative Prospect Theory (Tversky and Kahneman 1992), Regret (Bell 1982), and Ambiguity Aversion (Gilboa and Schmeidler 1989). Rabin (2000) criticized EU theory arguing that EU would require unreasonably large levels of risk aversion to explain the data from some small-stakes laboratory experiments. See also Simon (1979), Tversky and Kahneman (1981), Payne et al. (1992) and Baltussen and Post (2011) as examples providing experimental evidence on the failure of the EU paradigm.

  2. See, e.g., Duffie and Pan (1997) and Jorion (2007). The VaR measure is one of the main practical tools for reporting the exposure to risk by financial institutions.

  3. See, e.g., Apiwatcharoenkul et al. (2016) and Fanchi and Christiansen (2017).

  4. Rostek (2010) discusses other advantages of the QP, such as robustness, ability to deal with categorical (instead of continuous) variables, and the flexibility of offering a family of preferences indexed by quantiles.

  5. Intuitively, the monotonicity of quantiles allows one to avoid modeling individuals’ utility function. This is because the maximization problem is invariant to monotonic transformations of the distribution of portfolio returns.

  6. Our model can deal with short sale, but we leave this to future work.

  7. For instance, if the preference is given by EU, the belief is captured by the probability while the tastes by the utility function over outcomes or consequences (such as monetary payoffs). Beliefs and tastes are not completely separated, however, because if we take a monotonic transformation of the utility function, which maintains the same tastes over consequences, we may end up with a different preference. That is, the pair beliefs and tastes come together and are stable only under affine transformations of the utility function. In other words, the EU preferences, as many other preferences, do not allow a separation of tastes and beliefs.

  8. See, e.g., Damodaran (2010) for an argument that every asset carries some risk.

  9. In the context of risk management, Theorem 2 provides theoretical support to the lack of subadditivity of VaR measures in general settings, see Artzner et al. (1999).

  10. The inferior limit m may be taken over a limited region, not over the whole support. That is, we can accommodate cases in which the support is infinite so that \(f(x,y) \rightarrow 0\) when \(y \rightarrow \infty \).

  11. Of course, \(w^{*}=0\), when \({\underline{y}}>{\underline{x}}>-\infty \) and \({\underline{y}}- {\underline{x}} \geqslant \frac{M}{2m}\).

  12. The inferior limit m may be taken over a limited region, not over the whole support. That is, we can accommodate cases in which the support is infinite so that \(f(x,y) \rightarrow 0\) when \(y \rightarrow \infty \).

  13. The two illustrations (a) and (b) correspond, respectively, to \(\tau \leqslant \Pr \left( \left[ \frac{X+Y}{2} \leqslant q \right] \right) \) and \(\tau > \Pr \left( \left[ \frac{X+Y}{2} \leqslant q \right] \right) \). In the first case (a), the line \(wX+(1-w)Y=q\) is below the dashed blue line joining (1, 0) to (0, 1). In case (b) \(\tau >\Pr \left( \left[ \frac{X+Y}{2} \leqslant q \right] \right) \), the line \(wX+(1-w)Y=q\) is above the dashed blue line.

  14. The simulation algorithm reflects this issue by selecting either \(w=0\) or \(w=1\) for \(\tau _0=0.5\).

  15. When \({\mathscr {I}}_{X}\) or \({\mathscr {I}}_{Y}\) are not \({\mathbb {R}}\), then we have to consider limits that make the derivatives more complex.

  16. The case of two risky assets with different means is available from the authors upon request.

References

  • Apiwatcharoenkul, W., Lake, L., Jensen, J.: Uncertainty in proved reserves estimates by decline curve analysis. In: SPE/IAEE Hydrocarbon Economics and Evaluation Symposium, Society of Petroleum Engineers (2016)

  • Artzner, P., Delbaen, F., Eber, J., Heath, D.: Coherent measures of risk. Math Finance 9, 203–228 (1999)

    Article  Google Scholar 

  • Baltussen, G., Post, G.T.: Irrational diversification: an examination of individual portfolio choice. J Financ Quant Anal 46, 1463–1491 (2011)

    Article  Google Scholar 

  • Basak, S., Shapiro, A.: Value-at-risk based risk management: optimal policies and asset prices. Rev Financ Stud 14, 371–405 (2001)

    Article  Google Scholar 

  • Bassett, G.W., Koenker, R., Kordas, G.: Pessimistic portfolio allocation and choquet expected utility. J Financ Econom 2, 477–492 (2004)

    Google Scholar 

  • Bell, D.E.: Regret in decision making under uncertainty. Oper Res 30, 961–981 (1982)

    Article  Google Scholar 

  • Bhattacharya, D.: Inferring optimal peer assignment from experimental data. J Am Stat Assoc 104, 486–500 (2009)

    Article  Google Scholar 

  • Brown, D.B., Sim, M.: Satisficing measures for analysis of risky positions. Manag Sci 55, 71–84 (2009)

    Article  Google Scholar 

  • Campbell, J.: Household finance. J Financ Econ 61, 1553–1604 (2006)

    Google Scholar 

  • Campbell, J.Y.: Financial Decisions and Markets: A Course in Asset Pricing. Princeton: Princeton University Press (2018)

  • Campbell, R., Huisman, R., Koedijk, K.: Optimal portfolio selection in a value-at-risk framework. J Bank Finance 25, 1789–1804 (2001)

    Article  Google Scholar 

  • Chambers, C.P.: Ordinal aggregation and quantiles. J Econ Theory 137, 416–431 (2007)

    Article  Google Scholar 

  • Chambers, C.P.: An axiomatization of quantiles on the domain of distribution functions. Math Finance 19, 335–342 (2009)

    Article  Google Scholar 

  • Cochrane, J.H.: Asset Pricing, Revised. Princeton: Princeton University Press (2005)

  • Damodaran, A.: Into the abyss: what if nothing is risk free? Available at SSRN: https://ssrn.com/abstract=1648164. https://doi.org/10.2139/ssrn.1648164 (2010)

  • de Castro, L., Galvao, A.F.: Dynamic quantile models of rational behavior. Econometrica 87, 1893–1939 (2019)

    Article  Google Scholar 

  • de Castro, L., Galvao, A.F.: Static and dynamic quantile preferences. Econ Theory (2021) (Forthcoming)

  • de Castro, L., Galvao, A.F., Noussair, C.N., Qiao, L.: Do people maximize quantiles? Games Econ Behav 132, 22–40 (2022)

  • Duffie, D., Pan, J.: An overview of value at risk. J Deriv 4, 7–49 (1997)

    Article  Google Scholar 

  • Engle, R.F., Manganelli, S.: CAViaR: conditional autoregressive value at risk by regression quantiles. J Bus Econ Stat 22, 367–381 (2004)

    Article  Google Scholar 

  • Fanchi, J.R., Christiansen, R.L.: Introduction to Petroleum Engineering. New York: Wiley (2017)

  • Fishburn, P.C.: Mean-risk analysis with risk associated with below-target returns. Am Econ Rev 67, 116–126 (1977)

    Google Scholar 

  • Föllmer, H., Leukert, P.: Quantile hedging. Finance Stoch 3, 251–273 (1999)

    Article  Google Scholar 

  • Gaivoronski, A., Pflug, G.: Value-at-risk in portfolio optimization: properties and computational approach. J Risk 7(2), 1–31 (2005)

    Article  Google Scholar 

  • Garlappi, L., Uppal, R., Wang, T.: Portfolio selection with parameter and model uncertainty: a multi-prior approach. Rev Financ Stud 20, 41–81 (2007)

    Article  Google Scholar 

  • Ghirardato, P., Marinacci, M.: Ambiguity made precise: a comparative foundation. J Econ Theory 102, 251–289 (2002)

    Article  Google Scholar 

  • Ghirardato, P., Maccheroni, F., Marinacci, M.: Certainty independence and the separation of utility and beliefs. J Econ Theory 120, 129–136 (2005)

    Article  Google Scholar 

  • Gilboa, I., Schmeidler, D.: Maxmin expected utility with a non-unique prior. J Math Econ 18, 141–153 (1989)

    Article  Google Scholar 

  • Giovannetti, B.C.: Asset pricing under quantile utility maximization. Rev Financ Econ 22, 169–179 (2013)

    Article  Google Scholar 

  • He, X.D., Zhou, X.Y.: Portfolio choice via quantiles. Math Finance Int J Math Stat Financ Econ 21, 203–231 (2011)

    Google Scholar 

  • Ibragimov, R., Walden, D.: The limits of diversification when losses may be large. J Bank Finance 31, 2551–2569 (2007)

    Article  Google Scholar 

  • Jorion, P.: Value at Risk. The New Benchmark for Managing Financial Risk, vol. 81, 3rd edn. New York: McGraw-Hill (2007)

  • Kahneman, D., Tversky, A.: Prospect theory: an analysis of decision under risk. Econometrica 47, 263–292 (1979)

    Article  Google Scholar 

  • Kempf, A., Ruenzi, S.: Tournaments in mutual-fund families. Rev Financ Stud 21, 1013–1036 (2007)

    Article  Google Scholar 

  • Krokhmal, P., Uryasev, S., Palmquist, J.: Portfolio optimization with conditional value-at-risk objective and constraints. J Risk 4(2), 43–68 (2001)

    Article  Google Scholar 

  • Kulldorff, M.: Optimal control of favorable games with a time limit. SIAM J Control Optim 31, 52–69 (1993)

    Article  Google Scholar 

  • Lintner, J.: The valuation of risky assets and the selection of risky investments in the portfolios and capital budgets. Rev Econ Stat 47, 13–37 (1965)

    Article  Google Scholar 

  • Manski, C.: Ordinal utility models of decision making under uncertainty. Theor Decis 25, 79–104 (1988)

    Article  Google Scholar 

  • Markowitz, H.: Portfolio selection. J Finance 7, 77–91 (1952)

    Google Scholar 

  • Mendelson, H.: Quantile-preserving spread. J Econ Theory 42, 334–351 (1987)

    Article  Google Scholar 

  • Mitton, T., Vorkink, K.: Equilibrium underdiversification and the preference for skewness. Rev Financ Stud 20, 1255–1288 (2007)

    Article  Google Scholar 

  • Payne, J.W., Bettman, J.R., Johnson, E.J.: Behavioral decision research: a constructive processing perspective. Annu Rev Psychol 43, 87–131 (1992)

    Article  Google Scholar 

  • Porter, R.B.: Semivariance and stochastic dominance: a comparison. Am Econ Rev 64, 200–204 (1974)

    Google Scholar 

  • Quiggin, J.: A theory of anticipated utility. J Econ Behav Organ 3, 323–343 (1982)

    Article  Google Scholar 

  • Rabin, M.: Risk aversion and expected-utility theory: a calibration theorem. Econometrica 68, 1281–1292 (2000)

    Article  Google Scholar 

  • Rachev, S.T., Stoyanov, S., Fabozzi, F.J.: Advanced Stochastic Models, Risk Assessment, and Portfolio Optimization: The Ideal Risk, Uncertainty, and Performance Measures. New York: Wiley (2007)

  • Rostek, M.: Quantile maximization in decision theory. Rev Econ Stud 77, 339–371 (2010)

    Article  Google Scholar 

  • Rothschild, M., Stiglitz, J.E.: Increasing risk: I. A definition. J Econ Theory 2, 225–243 (1970)

    Article  Google Scholar 

  • Savage, L.: The Foundations of Statistics. New York: Wiley (1954)

  • Sharpe, W.F.: Capital asset prices: a theory of market equilibrium under conditions of risk. J Finance 19, 425–462 (1964)

    Google Scholar 

  • Simon, H.A.: Rational decision making in business organizations. Am Econ Rev 69, 493–513 (1979)

    Google Scholar 

  • Tobin, J.: Liquidity preference as behavior towards risk. Rev Econ Stud 67, 65–86 (1958)

    Article  Google Scholar 

  • Tversky, A., Kahneman, D.: The framing of decisions and the psychology of choice. Science 211, 453–458 (1981)

    Article  Google Scholar 

  • Tversky, A., Kahneman, D.: Advances in prospect theory: cumulative representation of uncertainty. J Risk Uncertain 5, 297–323 (1992)

    Article  Google Scholar 

  • Wu, G., Xiao, Z.: An analysis of risk measures. J Risk 4(4), 53–75 (2002)

    Article  Google Scholar 

  • Yaari, M.E.: The dual theory of choice under risk. Econometrica 55, 95–115 (1987)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

The authors would like to express their appreciation to the Editor-in-Chief Anne Villamil, an anonymous reviewer, Nabil Al-Najjar, Hide Ichimura, Derek Lemoine, Richard Peter, Tiemen Woutersen, María Florencia Gabrielli and seminar participants at the University of Iowa, INSPER, JOLATE Bahía Blanca, RedNIE, Southampton Econometrics Conference, and Universidad de Buenos Aires for helpful comments and discussions. Luciano de Castro acknowledges the support of the National Council for Scientific and Technological Development – CNPq. Computer programs to replicate the numerical analyses are available from the authors.

Appendices

Appendix

A Proofs

This Appendix collects the proofs for the results in the main text.

1.1 Proof of Lemma 3.1

Proof

The objective function of the primal problem in (4) is

$$\begin{aligned} \max _{w\in [0,1]^{n}} \text {Q}_{\tau }\left[ u( S_{w} ) \right] . \end{aligned}$$

Noticing that \(u(\cdot )\) is continuous and increasing, and that the quantile is invariant with respect to monotone transformations, then the above maximization argument is given by

$$\begin{aligned} \arg \max _{w} \text {Q}_{\tau }\left[ u( S_{w} ) \right]&= \arg \max _{w} u \left( \text {Q}_{\tau }\left[ S_{w} \right] \right) \\&= \arg \max _{w} \text {Q}_{\tau }\left[ S_{w} \right] . \end{aligned}$$

\(\square \)

1.2 Proof of Lemma 3.2

Proof

It is sufficient to show that \( w \mapsto \text {Q}_{\tau }\left[ w X + (1-w) Y \right] =\text {Q}_{\tau }[S_{w}] \) is continuous. But this follows from Assumption 1, which implies that the CDF of \(S_{w}=w X + (1-w) Y \) is strictly increasing, thus making its quantile continuous. \(\square \)

1.3 Proof of Theorem 1

Proof

(1) By appealing to translations if necessary, we may assume without loss of generality that \({\underline{x}}={\underline{y}}=0\), that is, if this is not the case, we may consider the random variables \(X' = X - {\underline{x}}\) and \(Y'= Y - {\underline{y}}\). We may also assume without loss of generality that \({\overline{x}}, {\overline{y}}< \infty \), for if \({\overline{x}}=\infty \) or \({\overline{y}}=\infty \), we may truncate the distribution at large \({\hat{x}}, {\hat{y}}\) so that \(\Pr ( \{ (x,y): x> {\hat{x}} \text { or } y> {\hat{y}} \} ) < \epsilon \) for some small \(\epsilon >0\). Thus, without loss of generality, we may assume that there exists \(m,M \in {\mathbb {R}}_{+}\) such that \(0< m< f(x,y)< M < \infty \) for all \((x,y) \in [0,{\overline{x}} ] \times [0,{\overline{y}} ]\).

We will prove that we can choose \(\tau \) small enough so that \(w=1\) is dominated by \(w = \frac{1}{2}\). The argument that \(w=0\) is also dominated by \(w= \frac{1}{2}\) for sufficiently small \(\tau \) is analogous. For this, we will assume that \(w=1\) is better than \(w= \frac{1}{2}\) and find a contradiction for \(\tau \) small enough. Let \(x_{\tau }\) denote the \(\tau \)-quantile of X and assume that \(q=q\left( \frac{1}{2}, \tau \right) \leqslant x_{\tau }\). See Fig. 11 where \(q < x_{\tau }\). From the definition, the probability under the red line \(\frac{X+Y}{2} \leqslant 2\) is \(\tau \), which is the same of the probability of the vertical band from 0 to \(x_{\tau }\). Therefore, the blue area in the graph (both light and dark blue together) have to be the same as the dark red that corresponds to the small triangle from \(x_{\tau }\) to 2q in the horizontal axis and from 0 to \(2q-x_{\tau }\). To better follow the argument, let \(A_{1}\) correspond to the probability in the dark blue area and \(A_{2}\), the probability of the light blue area. Let \(A_{3}\) correspond to the probability of the dark red area and \(A_{4}\), the probability of the light red area. Therefore, as observed above, by definition,

$$\begin{aligned} A_{1}+A_{2}= A_{3}. \end{aligned}$$
(14)
Fig. 11
figure 11

Illustration of the proof

We will show, however, that we can choose \(\tau \) sufficiently small so that

$$\begin{aligned} A_{1}>A_{3}+A_{4}. \end{aligned}$$
(15)

Observe that (15) implies

$$\begin{aligned} A_{1}+A_{2}>A_{1}> A_{3}+A_{4}>A_{3}, \end{aligned}$$

which contradicts (14). In order to obtain (15), observe that

$$\begin{aligned} A_{1}&= \int _{0}^{x_{\tau }} \left[ \int _{2x_{\tau }}^{{\overline{y}}} f(x,y) dy \right] dx \\ A_{3} + A_{4}&= \int _ {x_{\tau }}^{2x_{\tau }} \left[ \int _{0}^{x_{\tau }} f(x,y) dy \right] dx . \end{aligned}$$

Since \(m< f(x,y) < M\) for \((x,y) \in [0,{\overline{x}}]^{2}\), we have

$$\begin{aligned} A_{1}&> \int _{0}^{x_{\tau }} \left[ \int _{2x_{\tau }}^{{\overline{y}}} m dy \right] dx = m x_{\tau } \left( {\overline{y}} - 2x_{\tau } \right) \\ A_{3} + A_{4}&< \int _ {x_{\tau }}^{2x_{\tau }} \left[ \int _{0}^{x_{\tau }} M dy \right] dx = M x_{\tau }^{2}. \end{aligned}$$

Therefore, we obtain (15) if

$$\begin{aligned}&m x_{\tau } \left( {\overline{y}} - 2x_{\tau } \right)> M x_{\tau }^{2} \\&\quad \iff x_{\tau } \left[ m {\overline{y}} -(2m+M) x_{\tau } \right] >0 \\&\quad \iff x_{\tau } < \frac{ {\overline{y}}}{ \left( \frac{M}{m} + 2 \right) } . \end{aligned}$$

Since \({\overline{y}}\), M and m are given constants, the above inequality is satisfied if \(x_{\tau }\) is small enough, that is, if \(\tau \) is sufficiently close to 0. \(\square \)

1.4 Proof of Theorem 2

Proof

Note that by translating the distribution (along the 45\(^{\circ }\) line) if necessary, we may assume without loss of generality that \({\underline{y}}=0\).

Let us fix any \(\tau \in (0,{\overline{\tau }}]\) and let \(x_{\tau }, y_{\tau }\) denote the \(\tau \)-quantiles of X and Y, respectively. Of course, \(x_{\tau }> {\underline{x}}\). Figure 12 below will be useful to illustrate the reasoning in this proof.

Let \( {\bar{M}}\) be the average \(f_{X}(x)\) density on the region \(x \in [{\underline{x}}, x_{\tau }]\), that is, \({\bar{M}} \equiv \frac{ \tau }{x_{\tau } - {\underline{x}} }\). Therefore, \(m \leqslant {\bar{M}} \leqslant M\) and \(m \leqslant f(x,y) \leqslant M\) for all \((x,y) \in [{\underline{x}}, x_{{\overline{\tau }}}]\times [ {\underline{y}},{\overline{y}}]\). We want to show that \(q(w,\tau ) < x_{\tau }\) for all \(w \in [0,1)\). To show this, it is enough to show that for any \(q\geqslant x_{\tau }\) and \( w \in [0,1)\), the probability under the line \(wx + (1-w) y \leqslant q\) is larger than \(\tau \), that is,

$$\begin{aligned} h(w,q)&= \int _{{\underline{x}}}^{ \frac{x_{\tau }}{w} } \left[ \int _{0}^{\frac{x_{\tau }-wx}{1-w}} f(x,y) dy \right] dx >\tau . \end{aligned}$$

For this, it is sufficient to show that the area in blue in Fig. 12 is larger than \(\tau \) already for \(q=x_{\tau }\), that is,

$$\begin{aligned} q \geqslant x_{\tau } \implies h(w,q) \geqslant h(w,x_{\tau })&= \int _{{\underline{x}}}^{ \frac{x_{\tau }}{w} } \left[ \int _{0}^{\frac{x_{\tau }-wx}{1-w}} f(x,y) dy \right] dx >\tau . \end{aligned}$$
(16)

Since \(f(x,y) >m\) for all \((x,y)\in [{\underline{x}},x_{{\overline{\tau }}}] \times [{\underline{y}},{\overline{y}}] \cup [{\underline{x}},{\overline{x}}] \times [{\underline{y}}, x_{{\overline{\tau }}}]\), we have

$$\begin{aligned} h(w,x_{\tau })&> \frac{m}{2} \left( \frac{x_{\tau }}{w} - {\underline{x}} \right) \left( \frac{x_{\tau }-w{\underline{x}}}{1-w} \right) = \frac{ m (x_{\tau }- w {\underline{x}} )^{2}}{ 2w(1-w)}. \end{aligned}$$

Thus, to establish (16), it is sufficient to show that

$$\begin{aligned} \frac{ m (x_{\tau }- w {\underline{x}} )^{2}}{ 2w(1-w)}> \tau&\iff m (x_{\tau }- w {\underline{x}} )^{2}> 2 \tau w(1-w) \\&\iff m \left[ (x_{\tau })^{2} - 2w {\underline{x}} x_{\tau }+ w^{2}{\underline{x}}^{2} \right] - 2 \tau w+ 2w^{2}\tau> 0 \\&\iff w^{2} \left( 2\tau + m {\underline{x}}^{2} \right) -2w \left( m{\underline{x}} x_{\tau } +\tau \right) + m x_{\tau }^{2} > 0 . \end{aligned}$$

Let us define the quadratic polynomial:

$$\begin{aligned} p(w) \equiv w^{2} \left( 2\tau + m {\underline{x}}^{2} \right) -2w \left( m{\underline{x}} x_{\tau } +\tau \right) + m x_{\tau }^{2} . \end{aligned}$$
(17)
Fig. 12
figure 12

Illustration of the proof

Notice that \(p(0)=mx_{\tau }^{2}>0\) and for \(w=1\):

$$\begin{aligned}&p(1)=\left( 2\tau + m {\underline{x}}^{2} \right) -2 \left( m{\underline{x}} x_{\tau } +\tau \right) + m x_{\tau }^{2}\\ {}&= m \left( {\underline{x}}^{2} -2 {\underline{x}} x_{\tau } + x_{\tau }^{2} \right) =m(x_{\tau }-{\underline{x}})^{2} > 0 . \end{aligned}$$

Let \(w_V\) denote the vertex of the quadratic p(w), which is given by

$$\begin{aligned} w_{V} = \frac{ m x_{\tau }{\underline{x}} +\tau }{ 2 \tau + m {\underline{x}}^{2} }. \end{aligned}$$

Note that \(w_V\) is always positive.

We can conclude that the quadratic form p(w) in (17) is positive if \(p(w_{V})>0\). To verify this, we can substitute \(w_{V}\) in the quadratic form to obtain:

$$\begin{aligned} p(w_{V})&=\left[ \frac{ m x_{\tau }{\underline{x}} +\tau }{ 2 \tau + m {\underline{x}}^{2} } \right] ^{2} \left( 2\tau + m {\underline{x}}^{2} \right) -2 \left[ \frac{ m x_{\tau }{\underline{x}} +\tau }{ 2 \tau + m {\underline{x}}^{2} } \right] \left( m{\underline{x}} x_{\tau } +\tau \right) + m x_{\tau }^{2} \\&=\frac{- \left( m x_{\tau }{\underline{x}} +\tau \right) ^{2}}{ 2 \tau + m {\underline{x}}^{2} } +m x_{\tau }^{2} = \frac{m x_{\tau }^{2} (2\tau + m{\underline{x}}^{2} ) - \left( m x_{\tau }{\underline{x}} +\tau \right) ^{2}}{ 2 \tau + m {\underline{x}}^{2} } , \end{aligned}$$

which is positive as long as

$$\begin{aligned}&2\tau m x_{\tau }^{2} + m^{2} x_{\tau }^{2}{\underline{x}}^{2}> \left( m x_{\tau }{\underline{x}} +\tau \right) ^{2} = m^{2}x_{\tau }^{2} {\underline{x}}^{2} + 2 \tau mx_{\tau } {\underline{x}} + \tau ^{2} \\&\quad \iff 2 \tau m x_{\tau } \left( x_{\tau } - {\underline{x}} \right)> \tau ^{2} \\&\quad \iff 2 m x_{\tau } \left( x_{\tau } - {\underline{x}} \right) > \tau . \end{aligned}$$

Since \(\tau = {\bar{M}} (x_{\tau } - {\underline{x}})\), the above condition is equivalent to

$$\begin{aligned} 2 m x_{\tau }> {\bar{M}} \iff x_{\tau } > \frac{ {\bar{M}}}{2m}. \end{aligned}$$

Since \({\bar{M}} \leqslant M\) and \(x_{\tau } > {\underline{x}} \geqslant \frac{M}{2m}\), we have \(x_{\tau } > \frac{{\bar{M}}}{2m}\), thus proving the result. \(\square \)

Remark A.1

One can also conclude that \(p(w)>0\) for all \(w \in [0,1]\) if \(w_V > 1\). Unfortunately, the condition for this is stronger than the one given in Theorem 2. Indeed, \(w_V > 1\) if

$$\begin{aligned} w_{v} =\frac{ m x_{\tau }{\underline{x}} +\tau }{ 2 \tau + m {\underline{x}}^{2} } \geqslant 1, \end{aligned}$$

which is equivalent to

$$\begin{aligned}&m x_{\tau }{\underline{x}} +\tau \geqslant 2 \tau + m {\underline{x}}^{2} \\&\quad \iff m{\underline{x}} (x_{\tau }-{\underline{x}}) \geqslant \tau . \end{aligned}$$

Since \(\tau = {\bar{M}} (x_{\tau } - {\underline{x}})\), the above condition simplifies to

$$\begin{aligned} m {\underline{x}} \geqslant {\bar{M}} \iff {\underline{x}} \geqslant \frac{{\bar{M}}}{m}. \end{aligned}$$

Since \({\bar{M}} \leqslant M\), if we have \({\underline{x}} \geqslant \frac{M}{m}\) then \({\underline{x}} \geqslant \frac{{\bar{M}}}{m}\). However, this assumption \({\underline{x}} \geqslant \frac{M}{m}\) is stronger than the condition in Theorem 2.

1.5 Proof of Theorem 3

Proof

Fix \(\tau \geqslant {\overline{\tau }}\) and assume that \({\underline{x}}={\underline{y}}\) and \({\overline{x}}={\overline{y}}\). We want to show that \(w^{*}(\tau )=1\) for sufficiently high \(\tau \), which is equivalent to show that \(q(w,\tau ) < x_{\tau }\) for all \(w \in (0,1)\) and \(\tau \geqslant {\overline{\tau }}\). To show this, it is enough to show that for any \(q> x_{\tau }\) the probability under the line \(wx + (1-w) y \leqslant q\) is larger than \(\tau \), that is,

$$\begin{aligned} q> x_{\tau } \implies h(w,q)> h(w,x_{\tau })&= \int _{A (w, x_{\tau })} f(x,y) dy dx \geqslant \tau , \end{aligned}$$

where \(A(w,q) \equiv \{(x,y) \in [ {\underline{x}}, {\overline{x}} ] \times [{\underline{x}},{\overline{x}}]: wx + (1-w) y \leqslant q \}\). This is equivalent to show that

$$\begin{aligned} 1- h(w,x_{\tau })&= \int _{ A^{c}(w,x_{\tau })} f(x,y) dy dx \leqslant 1-\tau . \end{aligned}$$

For estimating this integral, we have to consider three intervals for w: \((0,{\underline{w}})\), \([{\underline{w}}, {\overline{w}}]\) and \(( {\overline{w}}, 1)\), where \({\underline{w}}\) and \( {\overline{w}}\) correspond to the lines that pass by \(({\underline{x}},{\overline{x}})\) and \(({\overline{x}},{\underline{x}})\) respectively, that is,

$$\begin{aligned} {\underline{w}} \ {\underline{x}} + \left( 1-{\underline{w}} \right) {\overline{x}}&= x_{\tau } \iff {\underline{w}} = \frac{ {\overline{x}}- x_{\tau } }{{\overline{x}}-{\underline{x}}} ; \\ {\overline{w}} \ {\overline{x}} + \left( 1-{\overline{w}} \right) {\underline{x}}&= x_{\tau } \iff {\overline{w}} = \frac{x_{\tau } - {\underline{x}} }{{\overline{x}}-{\underline{x}}} . \end{aligned}$$

Note that \({\underline{w}}=1-{\overline{w}} < {\overline{w}}\) for large \(\tau \).

Fig. 13
figure 13

Illustration for the proof of Theorem 3: the case of \(\tau \) close to 1

Figure 13 shows the set \(A(w,x_{\tau })\) in blue, for \(w \in [{\underline{w}}, {\overline{w}}]\). Thus, if \(w \in [{\underline{w}}, {\overline{w}}]\),

$$\begin{aligned} 1- h(w,x_{\tau })&= \int _{ \frac{x_{\tau }-(1-w) {\underline{x}} }{w} } ^{{\overline{x}}} \left[ \int ^{{\overline{x}}}_{\frac{x_{\tau }-wx}{1-w}} f(x,y) dy \right] dx . \end{aligned}$$

For \(w \in (0, {\underline{w}})\),

$$\begin{aligned} 1-h(w,x_{\tau }) = \int _{A^{c} (w, x_{\tau })} f(x,y) dy dx = \int _{{\underline{x}}} ^{{\overline{x}}} \left[ \int ^{{\overline{x}}}_{\frac{x_{\tau }-wx}{1-w}} f(x,y) dy \right] dx , \end{aligned}$$

and, for \(w \in ({\overline{w}},1)\),

$$\begin{aligned} 1-h(w,x_{\tau }) = \int _{A^{c} (w, x_{\tau })} f(x,y) dy dx = \int _{ {\underline{x}}} ^{{\overline{x}}} \left[ \int ^{{\overline{x}}}_{\frac{x_{\tau }-(1-w)y}{w}} f(x,y) dx \right] dy. \end{aligned}$$

We will consider each of the above cases.

Case 1: \(w \in [{\underline{w}},{\overline{w}}]\).

Since \(m \leqslant f(x,y) \leqslant M\) for all \((x,y) \in [ x_{{\overline{\tau }}}, {\overline{x}} ] \times [{\underline{x}},{\overline{x}}]\), we have

$$\begin{aligned} 1- h(w,x_{\tau })&\leqslant \frac{M}{2} \left[ {\overline{x}} - \frac{x_{\tau }-(1-w) {\overline{x}} }{w} \right] \left( {\overline{x}} - \frac{x_{\tau }-w{\overline{x}}}{1-w} \right) . \end{aligned}$$

Using \({\overline{x}}={\overline{x}}\), the left hand side becomes \(\frac{M \left( {\overline{x}} - x_{\tau } \right) ^{2 }}{2w(1-w)}\). Thus, it is enough to prove that

$$\begin{aligned} \frac{M \left( {\overline{x}} - x_{\tau } \right) ^{2 }}{2w(1-w)} \leqslant 1- \tau&\iff w(1-w) \geqslant \frac{M \left( {\overline{x}} - x_{\tau } \right) ^{2 }}{2 (1-\tau )}. \end{aligned}$$

Since \(w\in [{\underline{w}},{\overline{w}}]\), we have

$$\begin{aligned} w(1-w) \geqslant \min \{ {\underline{w}}(1-{\underline{w}}), {\overline{w}}(1-{\overline{w}})\} = {\underline{w}}{\overline{w}} =\frac{ x_{\tau } -{\underline{x}}}{{\overline{x}}-{\underline{x}}} \cdot \frac{ {\overline{x}}-x_{\tau } }{{\overline{x}}-{\underline{x}}}. \end{aligned}$$

Thus, it is sufficient to show that:

$$\begin{aligned}&\frac{ (x_{\tau } -{\underline{x}})({\overline{x}}-x_{\tau } ) }{({\overline{x}}-{\underline{x}})^{2}} \geqslant \frac{M \left( {\overline{x}} - x_{\tau } \right) ^{2 }}{2(1-\tau )} \nonumber \\&\quad \iff 2(1-\tau ) (x_{\tau } -{\underline{x}}) \geqslant M ( {\overline{x}}-{\underline{x}})^{2} \left( {\overline{x}} - x_{\tau } \right) . \end{aligned}$$
(18)

We now define \({\bar{m}} = \frac{({\overline{x}} - x_{\tau })({\overline{x}}-{\underline{x}})}{1-\tau }\) so that \(1-\tau ={\bar{m}}({\overline{x}}-x_{\tau })({\overline{x}}-{\underline{x}})={\bar{m}}({\overline{x}}-x_{\tau })({\overline{x}}-{\underline{x}})\). Since

$$\begin{aligned} 1-\tau = \int _{x_{\tau }}^{{\overline{x}}} \left[ \int _{{\underline{x}}}^{{\overline{x}}}f(x,y)dy \right] dx \geqslant m ({\overline{x}}- x_{\tau }) ({\overline{x}}-{\underline{x}}), \end{aligned}$$

we have \({\bar{m}} \geqslant m \implies \frac{M}{{\bar{m}}} \leqslant \frac{M}{m}\). Therefore, the assumption gives

$$\begin{aligned} x_{\tau } -{\underline{x}} \geqslant x_{{\overline{\tau }}} - {\underline{x}} \geqslant \frac{ M ( {\overline{x}}-{\underline{x}}) }{ 2m} \geqslant \frac{ M ( {\overline{x}}-{\underline{x}}) }{ 2 {\bar{m}} }, \end{aligned}$$

as we wanted to show.

Case 2: \(w \in (0,{\underline{w}})\).

Since \(m \leqslant f(x,y) \leqslant M\) for all \((x,y) \in [ x_{{\overline{\tau }}}, {\overline{x}} ] \times [{\underline{x}},{\overline{x}}]\), we have

$$\begin{aligned} 1- h(w,x_{\tau })&\leqslant \frac{M}{2} \left( {\overline{x}} - {\underline{x}} \right) \left( {\overline{x}} - \frac{x_{\tau }-w{\overline{x}}}{1-w} \right) = \frac{M \left( {\overline{x}} - {\underline{x}} \right) \left( {\overline{x}} - x_{\tau } \right) }{2(1-w)}. \end{aligned}$$

Thus, it is enough to prove

$$\begin{aligned} \frac{M \left( {\overline{x}} - {\underline{x}} \right) \left( {\overline{x}} - x_{\tau } \right) }{2(1-w)} \leqslant 1- \tau&\iff 1-w \geqslant \frac{M \left( {\overline{x}} - {\underline{x}} \right) \left( {\overline{x}} - x_{\tau } \right) }{2 (1-\tau )}. \end{aligned}$$

Since \(w\in (0,{\underline{w}})\), we have \(1-w > 1- {\underline{w}} = \frac{x_{\tau } - {\underline{x}} }{{\overline{x}}-{\underline{x}}}\). Thus, it is sufficient to show that:

$$\begin{aligned} 2(1-\tau ) (x_{\tau } -{\underline{x}}) \geqslant M ( {\overline{x}}-{\underline{x}})^{2} \left( {\overline{x}} - x_{\tau } \right) . \end{aligned}$$

This is exactly condition (18) above, which we have shown to be implied by the assumption.

Case 3: \(w \in ({\overline{w}},1)\).

Since \(m \leqslant f(x,y) \leqslant M\) for all \((x,y) \in [ x_{{\overline{\tau }}}, {\overline{x}} ] \times [{\underline{x}},{\overline{x}}]\), we have

$$\begin{aligned} 1- h(w,x_{\tau })&\leqslant \frac{M}{2} \left( {\overline{x}} - {\underline{x}} \right) \left[ {\overline{x}} - \frac{x_{\tau }-(1-w) {\overline{x}} }{w} \right] = \frac{M \left( {\overline{x}} - {\underline{x}} \right) \left( {\overline{x}} - x_{\tau } \right) }{2w } . \end{aligned}$$

Thus, it is enough to prove

$$\begin{aligned} \frac{M \left( {\overline{x}} - {\underline{x}} \right) \left( {\overline{x}} - x_{\tau } \right) }{2w } \leqslant 1- \tau&\iff 2 (1-\tau ) w \geqslant M \left( {\overline{x}} - {\underline{x}} \right) \left( {\overline{x}} - x_{\tau } \right) . \end{aligned}$$

Since \(w\in ({\overline{w}},1)\), \(w > {\overline{w}}=\frac{ x_{\tau } - {\underline{x}} }{{\overline{x}}-{\underline{x}}} \), it is sufficient to show again condition (18) above, which we have shown to be implied by the assumption of Theorem 3. \(\square \)

1.6 Proof of Proposition 3.10

Proof

Assume first that q(w) is differentiable at \(w\in (0,1)\). Taking the total derivative with respect to w on the equation \( h(w,q) = \tau \), we obtain

$$\begin{aligned} \partial _{w}h(w,q) + \partial _{q}h(w,q) \cdot q'(w)= 0 \implies q'(w) = - \frac{ \partial _{w}h(w,q)}{ \partial _{q}h(w,q)}. \end{aligned}$$
(19)

From (8), it is clear that h is differentiable and that \(\partial _{q} h(w,q)>0\). Therefore, by the Implicit Function Theorem, q is differentiable and given by (19). Now, we will calculate \(\partial _{w}h(w,q)\) and \(\partial _{q}h(w,q)\). Define the function \(y(w,q) \equiv \frac{q-wx}{1-w}\). Then,

$$\begin{aligned} \partial _{q}y(w,q)&= \frac{1}{1-w}, \text { and } \\ \partial _{w}y(w,q)&= \frac{(-x)(1-w)-(q-wx)(-1)}{(1-w)^{2}} = \frac{q-x}{(1-w)^{2}}. \end{aligned}$$

Support on \({\mathbb {R}}\)

Here we consider first the case in which \({\mathscr {I}}_{X}={\mathscr {I}}_{Y} = {\mathbb {R}}\). The other two cases are similar and considered below.Footnote 15 It is clear from (8) and Assumption 1 that h is \(C^{1}\) and:

$$\begin{aligned} \partial _{w } h(w,q) = \int _{-\infty }^{\infty } f \left( x, \frac{q-wx}{1-w} \right) \left[ \frac{q-x}{(1-w)^{2}} \right] dx \end{aligned}$$
(20)

and

$$\begin{aligned} \partial _{q } h(w,q) = \int _{-\infty }^{\infty } f \left( x, \frac{q-wx}{1-w} \right) \frac{1}{1-w} dx. \end{aligned}$$
(21)

From this,

$$\begin{aligned} \partial _{w } h(w,q)&= \frac{q}{(1-w)^{2}} \int _{-\infty }^{\infty } f \left( x, \frac{q-wx}{1-w} \right) dx - \frac{1}{(1-w)^{2}} \int _{-\infty }^{\infty } x f \left( x, \frac{q-wx}{1-w} \right) dx \\&=\frac{ q}{1-w} \partial _{q } h(w,q) - \frac{1}{(1-w)} \text {E}[Z] \cdot \partial _{q } h(w,q) \\&= \frac{ 1}{1-w} \cdot \left( q - \text {E}[Z] \right) \cdot \partial _{q } h(w,q). \end{aligned}$$

Therefore,

$$\begin{aligned} q'(w) = - \frac{\partial _{w } h(w,q) }{\partial _{q } h(w,q)} = \frac{ 1}{1-w} \left( \text {E}[Z]-q \right) , \end{aligned}$$

as we wanted to show. \(\square \)

Support on \([0,\infty )\)

Now, we consider the case \({\mathscr {I}}_{X}={\mathscr {I}}_{Y}={\mathbb {R}}_{+}=[0, \infty )\). In this case, we have

$$\begin{aligned} h(w,q) = \int _{0}^{\frac{q}{w}} \int _{0}^{ \frac{q-wx}{1-w}} f(x,y) dy dx. \end{aligned}$$
(22)

Let us define:

$$\begin{aligned} g(w,q,x) \equiv \int _{0}^{ \frac{q-wx}{1-w}} f(x,y) dy , \end{aligned}$$

for \(x < \frac{q}{w}\) and \( g \left( w,q, \frac{q}{w} \right) = 0\) so that g is continuous and differentiable. Also,

$$\begin{aligned} h(w,q) = \int _{0}^{\frac{q}{w}} g(w,q,x) dx. \end{aligned}$$

Then,

$$\begin{aligned} \partial _{w } h(w,q)&= g \left( w,q, \frac{q}{w} \right) \left( \frac{ -q}{w^{2}} \right) + \int _{0}^{\frac{q}{w}} \partial _{w} g(w,q,x) dx \\&= \frac{1}{(1-w)^{2}} \int _{0}^{\frac{q}{w}} \left( q-x \right) f \left( x, \frac{q-wx}{1-w} \right) dx , \end{aligned}$$

and

$$\begin{aligned} \partial _{q } h(w,q)&= g \left( w,q, \frac{q}{w} \right) \left( \frac{ 1}{w } \right) + \int _{0}^{\frac{q}{w}} \partial _{q} g(w,q,x) dx \\&= \frac{1}{1-w} \int _{0}^{\frac{q}{x}} f \left( x, \frac{q-wx}{1-w} \right) dx. \end{aligned}$$

From this, we observe that the same expressions remain valid:

$$\begin{aligned} \partial _{w } h(w,q)&= \frac{q}{(1-w)^{2}} \int _{0}^{\frac{q}{w}} f \left( x, \frac{q-wx}{1-w} \right) dx \\&\quad - \frac{1}{(1-w)^{2}} \int _{0}^{\frac{q}{w}} x f \left( x, \frac{q-wx}{1-w} \right) dx \\&= \frac{ 1}{1-w} \cdot \left( q - \text {E}[Z] \right) \cdot \partial _{q } h(w,q) \end{aligned}$$

and

$$\begin{aligned} q'(w) = - \frac{\partial _{w } h(w,q) }{\partial _{q } h(w,q)} = \frac{ 1}{1-w} \left( \text {E}[Z]-q \right) , \end{aligned}$$
(23)

as before.

Support on [0, c]

We now consider the case \({\mathscr {I}}_{X}={\mathscr {I}}_{Y}=[0,c]\). There are two subcases to consider, as show in Fig. 14.

Fig. 14
figure 14

An illustration of Z for the case \({\mathscr {I}}_{X}={\mathscr {I}}_{Y}=[0,c]\). The blue area has probability \(\tau \)

Case (a)

In this case,

$$\begin{aligned} h(w,q) = \int _{0}^{\frac{q-(1-w)c}{w}} \int _{0}^{ c} f(x,y) dy dx + \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} \int _{0}^{ \frac{q-wx}{1-w}} f(x,y) dy dx . \end{aligned}$$

Let

$$\begin{aligned} g(\alpha ,\beta ,x) \equiv \int _{\alpha }^{ \beta } f(x,y) dy . \end{aligned}$$

Therefore,

$$\begin{aligned} h(w,q) = \int _{0}^{\frac{q-(1-w)c}{w}} g(0,c,x) dx + \int _{\frac{q-(1-w)c}{w}}^{ \frac{q}{w}} g(0, \frac{q-wx}{1-w},x) dx . \end{aligned}$$

Thus,

$$\begin{aligned} \partial _{w}h(w,q)&= g \left( 0,c, \frac{q-(1-w)c}{w} \right) \partial _{w}\left[ \frac{q - (1-w)c}{1-w}\right] \\&\quad + \int _{0}^{\frac{q-(1-w)c}{w}} \partial _{w} \left[ g(0,c,x) \right] dx \\&\quad + g \left( 0, 0, \frac{q}{w} \right) \left( \frac{-q }{w^{2}} \right) - g \left( 0, c, \frac{q-(1-w)c}{w} \right) \partial _{w}\left[ \frac{q - (1-w)c}{1-w}\right] \\&\quad + \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} \partial _{w} \left[ g(0, \frac{q-wx}{1-w},x) \right] dx . \end{aligned}$$

Since \(g(0,0, \cdot )=0\) and \(\partial _{w} g(0,c,x)=0\), and \(g \left( 0,c, \frac{q-(1-w)c}{w} \right) \partial _{w}\left[ \frac{q - (1-w)c}{1-w}\right] \) appear with + and - signs, we have

$$\begin{aligned} \partial _{w}h(w,q)&= \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} \partial _{w} \left[ g(0, \frac{q-wx}{1-w},x) \right] dx \\&= \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} f \left( x, \frac{q-wx}{1-w} \right) \partial _{w} \left[ \frac{q-wx}{1-w} \right] dx \\&= \frac{q-x}{(1-w)^{2}} \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} f \left( x, \frac{q-wx}{1-w} \right) dx . \end{aligned}$$

We have:

$$\begin{aligned} \partial _{q}h(w,q)&= g \left( 0,c, \frac{q-(1-w)c}{w} \right) \partial _{q}\left[ \frac{q - (1-w)c}{1-w}\right] \\&\quad + \int _{0}^{\frac{q-(1-w)c}{w}} \partial _{q} \left[ g(0,c,x) \right] dx \\&\quad + g \left( 0, 0, \frac{q}{w} \right) \left( \frac{1 }{w } \right) - g \left( 0, c, \frac{q-(1-w)c}{w} \right) \partial _{q}\left[ \frac{q - (1-w)c}{1-w}\right] \\&\quad + \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} \partial _{q} \left[ g(0, \frac{q-wx}{1-w},x) \right] dx . \end{aligned}$$

Anagously to the previous case,

$$\begin{aligned} \partial _{q}h(w,q)&= \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} \partial _{q} \left[ g(0, \frac{q-wx}{1-w},x) \right] dx \\&= \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} f \left( x, \frac{q-wx}{1-w} \right) \partial _{q} \left[ \frac{q-wx}{1-w} \right] dx \\&= \frac{1}{(1-w)^{2}} \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} (q-x) f \left( x, \frac{q-wx}{1-w} \right) dx . \end{aligned}$$

Note that in this case,

$$\begin{aligned} f_{Z}(z)= \frac{f \left( z, \frac{q-wz}{1-w} \right) }{ \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} f \left( x, \frac{q-wx}{1-w} \right) dx }, \end{aligned}$$

so that

$$\begin{aligned} \text {E}[Z] = \frac{ \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} x f \left( x, \frac{q-wx}{1-w} \right) dx }{\int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}}f \left( x, \frac{q-wx}{1-w} \right) dx}. \end{aligned}$$

From this, we observe that the same expressions remain valid:

$$\begin{aligned} \partial _{w } h(w,q)&= \frac{q}{(1-w)^{2}} \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} f \left( x, \frac{q-wx}{1-w} \right) \\ {}&dx - \frac{1}{(1-w)^{2}} \int _{\frac{q-(1-w)c}{w}}^{\frac{q}{w}} x f \left( x, \frac{q-wx}{1-w} \right) dx \\&=\frac{ q}{1-w} \partial _{q } h(w,q) - \frac{1}{(1-w)} \text {E}[Z] \cdot \partial _{q } h(w,q) \\&= \frac{ 1}{1-w} \cdot \left( q - \text {E}[Z] \right) \partial _{q } h(w,q), \end{aligned}$$

and

$$\begin{aligned} q'(w) = - \frac{\partial _{w } h(w,q) }{\partial _{q } h(w,q)} = \frac{ 1}{1-w} \left( \text {E}[Z]-q \right) , \end{aligned}$$

as before.

Case (b)

In this case,

$$\begin{aligned} h(w,q) = \int _{0}^{\frac{q-(1-w)c}{w}} \int _{0}^{ c} f(x,y) dy dx + \int _{\frac{q-(1-w)c}{w}}^{c} \int _{0}^{ \frac{q-wx}{1-w}} f(x,y) dy dx , \end{aligned}$$

and

$$\begin{aligned} h(w,q) = \int _{0}^{\frac{q-(1-w)c}{w}} g(0,c,x) dx + \int _{\frac{q-(1-w)c}{w}}^{c} g(0, \frac{q-wx}{1-w},x) dx . \end{aligned}$$

Thus,

$$\begin{aligned} \partial _{w}h(w,q)&= g \left( 0,c, \frac{q-(1-w)c}{w} \right) \partial _{w}\left[ \frac{q - (1-w)c}{1-w}\right] \\&\quad + \int _{0}^{\frac{q-(1-w)c}{w}} \partial _{w} \left[ g(0,c,x) \right] dx \\&\quad - g \left( 0, c, \frac{q-(1-w)c}{w} \right) \partial _{w}\left[ \frac{q - (1-w)c}{1-w}\right] \\&\quad + \int _{\frac{q-(1-w)c}{w}}^{c} \partial _{w} \left[ g(0, \frac{q-wx}{1-w},x) \right] dx . \end{aligned}$$

Since \(\partial _{w} g(0,c,x)=0\), and \(g \left( 0,c, \frac{q-(1-w)c}{w} \right) \partial _{w}\left[ \frac{q - (1-w)c}{1-w}\right] \) appear with + and - signs,

$$\begin{aligned} \partial _{w}h(w,q)&= \int _{\frac{q-(1-w)c}{w}}^{c} \partial _{w} \left[ g(0, \frac{q-wx}{1-w},x) \right] dx \\&= \int _{\frac{q-(1-w)c}{w}}^{c} f \left( x, \frac{q-wx}{1-w} \right) \partial _{w} \left[ \frac{q-wx}{1-w} \right] dx \\&= \frac{q-x}{(1-w)^{2}} \int _{\frac{q-(1-w)c}{w}}^{c} f \left( x, \frac{q-wx}{1-w} \right) dx . \end{aligned}$$

Also,

$$\begin{aligned} \partial _{q}h(w,q)&= g \left( 0,c, \frac{q-(1-w)c}{w} \right) \partial _{q}\left[ \frac{q - (1-w)c}{1-w}\right] \\&\quad + \int _{0}^{\frac{q-(1-w)c}{w}} \partial _{q} \left[ g(0,c,x) \right] dx \\&\quad - g \left( 0, c, \frac{q-(1-w)c}{w} \right) \partial _{q}\left[ \frac{q - (1-w)c}{1-w}\right] \\&\quad + \int _{\frac{q-(1-w)c}{w}}^{c} \partial _{q} \left[ g(0, \frac{q-wx}{1-w},x) \right] dx . \end{aligned}$$

Therefore,

$$\begin{aligned} \partial _{q}h(w,q)&= \int _{\frac{q-(1-w)c}{w}}^{c} \partial _{q} \left[ g(0, \frac{q-wx}{1-w},x) \right] dx \\&= \int _{\frac{q-(1-w)c}{w}}^{c} f \left( x, \frac{q-wx}{1-w} \right) \partial _{q} \left[ \frac{q-wx}{1-w} \right] dx \\&= \frac{1}{(1-w)^{2}} \int _{\frac{q-(1-w)c}{w}}^{c} (q-x) f \left( x, \frac{q-wx}{1-w} \right) dx . \end{aligned}$$

Note that in this case,

$$\begin{aligned} f_{Z}(z)= \frac{f \left( z, \frac{q-wz}{1-w} \right) }{ \int _{\frac{q-(1-w)c}{w}}^{c} f \left( x, \frac{q-wx}{1-w} \right) dx }, \end{aligned}$$

so that

$$\begin{aligned} \text {E}[Z] = \frac{ \int _{\frac{q-(1-w)c}{w}}^{c} x f \left( x, \frac{q-wx}{1-w} \right) dx }{\int _{\frac{q-(1-w)c}{w}}^{c}f \left( x, \frac{q-wx}{1-w} \right) dx}. \end{aligned}$$

From this, we observe that the same expressions remain valid:

$$\begin{aligned} \partial _{w } h(w,q)&= \frac{q}{(1-w)^{2}} \int _{\frac{q-(1-w)c}{w}}^{c} f \left( x, \frac{q-wx}{1-w} \right) dx \\&\quad - \frac{1}{(1-w)^{2}} \int _{\frac{q-(1-w)c}{w}}^{c} x f \left( x, \frac{q-wx}{1-w} \right) dx \\&=\frac{ q}{1-w} \partial _{q } h(w,q) - \frac{1}{(1-w)} \text {E}[Z] \cdot \partial _{q } h(w,q) \\&= \frac{ 1}{1-w} \cdot \left( q - \text {E}[Z] \right) \partial _{q } h(w,q) \end{aligned}$$

and

$$\begin{aligned} q'(w) = - \frac{\partial _{w } h(w,q) }{\partial _{q } h(w,q)} = \frac{ 1}{1-w} \left( \text {E}[Z]-q \right) , \end{aligned}$$

as before. \(\square \)

1.7 Proof of Corollary 3.11

Proof

The proof of this result is immediate from imposing \(q'(w^*)=0\) in Proposition 3.10, and noting that \(q''(w)=\frac{\partial }{\partial w} E[Z_{w,q}]\), with \(q''(\cdot )\) the second derivative with respect to w. To show this, we take the first derivative of \(q'(w)\) in (10). Then,

$$\begin{aligned} q''(w)= \frac{\left( \frac{\partial }{\partial w} E[Z] - q'(w) \right) (1-w) + E[Z]-q(w)}{(1-w)^2}. \end{aligned}$$
(24)

Then, noting that \(E[Z]=q(w^*)\) and \(q'(w^*)=0\), we have

$$\begin{aligned} q''(w)= \frac{1}{(1-w)^2} \frac{\partial }{\partial w} E[Z]. \end{aligned}$$
(25)

\(\square \)

1.8 Proof of Proposition 3.13

Proof

The proof of this result is shown in three different stages. First, we note from Corollary 3.11 that the first order condition characterizing an extremum of the optimization problem (7) is \(E[Z]=q(w^*)\), with \(w^* \in (0,1)\). Now, given that the density function \(f_Z(z)\) is evaluated over the line \(y=\frac{q(w^*)-w^* x}{1-w^*}\), it is not difficult to see that the projection of this line on the y-axis is \(y^*=E[Z]\) for \(x^*=E[Z]\). This result implies that for different quantile values \(q(w^*) \equiv q(w^*,\tau )\) indexed by \(\tau \in (0,1)\), the mean value of the random variable Z characterized by the density function \(f_Z(z)\) in (9) is in the \(45\%\) degree line such that \((x^*,y^*)=(E[Z],E[Z])\).

Second, we prove that condition (11) evaluated at \(w^*\) guarantees the symmetry around E[Z] of the density function \(f_{Z}\) for \(x \in {\mathscr {I}}_{Z}={\mathscr {I}}_{Z}^{w^*,q}= \{x \in {\mathbb {R}}: \left( x, \frac{q-w^*x}{1-w^*} \right) \in {\mathscr {I}}_{X} \times {\mathscr {I}}_{Y} \}\). More formally, condition (11) implies \(f_{Z}(E[Z]+\varepsilon )=f_{Z}(E[Z]-\varepsilon )\) for all \(\varepsilon >0\). This is so by noting that \(f_{Z}(E[Z]+\varepsilon )= \frac{f \left( E[Z]+\varepsilon , E[Z] - \frac{w^*}{1-w^*} \varepsilon \right) }{\int _{{\mathscr {I}}_{Z}} f \left( t, \frac{E[Z]-tw^*}{1-w^*} \right) dt}\) and \(f_{Z}(E[Z]-\varepsilon )= \frac{f \left( E[Z]-\varepsilon , E[Z] + \frac{w^*}{1-w^*} \varepsilon \right) }{\int _{{\mathscr {I}}_{Z}} f \left( t, \frac{E[Z]-tw^*}{1-w^*} \right) dt}\).

Now, we show that under assumption 2, the density function \(f_{Z}\) is unimodal. To show this, we note that \(f_Z(z)=\frac{f \left( x, \frac{E[Z] - w^* x}{1-w^*} \right) }{\int _{{\mathscr {I}}_{Z}} f \left( t, \frac{E[Z]-tw^*}{1-w^*} \right) dt}\), such that under the change of variable \(x=E[Z]+\varepsilon \), we obtain \(f_Z(z)=\frac{f \left( E[Z]+\varepsilon , E[Z] - \frac{w^*}{1-w^*} \varepsilon \right) }{\int _{{\mathscr {I}}_{Z}} f \left( t, \frac{E[Z]-tw^*}{1-w^*} \right) dt}\). Assumption 2 evaluated at \(\mu =E[Z]\) implies that the numerator of this expression increases with \(\varepsilon \) up to \(\varepsilon =0\) and then decreases. This condition is sufficient to show that \(f_Z(z)\) is unimodal with mode at E[Z].

These findings (unimodality and symmetry of \(f_Z(\cdot )\)) apply to every \(\tau \in (0,1)\) and imply that the point \((x^*,y^*)=(E[Z],E[Z])\) divides the line \(\frac{q(w^*)-w^* x}{1-w^*}\) in two equal segments for all quantile values \(q(w^*) \equiv q(w^*,\tau )\) indexed by \(\tau \in (0,1)\). This property also implies that E[Z] is the median of the distribution of Z with support the projection of the line on the x-axis. In this scenario no other combination \({\widetilde{w}}\), with \({\widetilde{w}} \ne w^*\), and such that \(q({\widetilde{w}},\tau )\) and \(q(w^*,\tau )\) defines two different lines for the same \(\tau \in (0,1)\), yields a line \(y=\frac{q({\widetilde{w}})-{\widetilde{w}} x}{1-{\widetilde{w}}}\) that intersects \(y=x\) at \((x^*,y^*)=(E[Z],E[Z])\). Let \(({\widetilde{x}},{\widetilde{x}})\) denote such intersection, and let \(\tau _0 \in (0,1)\) be defined by the condition \(\tau \le P(Z \le q)\) for \(\tau \le \tau _0\), and \(P(Z \le q) < \tau \) for \(\tau \in (\tau _0,1]\). Then, the condition \(\tau \le P(Z \le q)\), for all \(\tau \in (0,\tau _0]\), implies that the projection of the crossing point \(({\widetilde{x}},{\widetilde{x}})\) on the x-axis, given by \({\widetilde{x}}\), is smaller than the corresponding projection for \(w^*\), that is \(x^*=E[Z]\). Then, for all \({\widetilde{w}} \in (0,1)\) with \({\widetilde{w}} \ne w^*\), it follows that \({\widetilde{x}} < E[Z]\), that is equivalent to the condition \(q({\widetilde{w}})<q(w^*)\) since \(q({\widetilde{w}})={\widetilde{w}} x + (1-{\widetilde{w}}) y\). Then, for \(x=y={\widetilde{x}}\), we have \(q({\widetilde{w}})={\widetilde{x}}\) and for \(x=y= E[Z]\), we have \(q(w^*) =E[Z]\). Hence, the quantity \(w^*\) maximizes the quantile function for all \(\tau \in (0,\tau _0]\).

It remains to see that the condition \(q(w^{*})=E[Z]\) characterizes a minimum of the optimization problem (7) for \(\tau > \tau _0\). In this scenario, the solution to the optimization problem (7) is also in the \(45^o\) degree line, however, for \(\tau \in (\tau _{0},1]\), it follows that \(P(Z \le q) < \tau \) (see Fig. 5b). Then, for any \({\widetilde{w}}=w^{*} \pm \varepsilon \), with \(\varepsilon >0\), the projection of the crossing point \(({\widetilde{x}},{\widetilde{x}})\) on the x-axis is larger than the projection of the crossing point (E[Z], E[Z]) associated to \(w^{*}\). Thus, for any \({\widetilde{w}}=w^{*} \pm \varepsilon \) with \(\varepsilon >0\), it follows that \({\widetilde{x}}>E[Z]\) and, using the above arguments, \(q({\widetilde{w}})>q(w^*)\) for all \({\widetilde{w}} \ne w^*\), with \(w^*, {\widetilde{w}} \in (0,1)\).

The continuity of q(w) with respect to w implies that the solution to (7) is a corner solution. For \(w^*=\frac{1}{2}\) in the region \(\tau \in (0,\tau _0]\), the solution for \(\tau >\tau _0\) is indistinctively zero or one due to the symmetry of condition (11). Otherwise, if \(w^* \ne \frac{1}{2}\) in the region \(\tau \in (0,\tau _0]\), the solution to the maximization problem for \(\tau >\tau _0\) is given by the random variable with largest upside potential. More formally, it will be one if \(F_X(z) < F_Y(z)\) uniformly over z, for z sufficiently large. In contrast, the solution will be zero for \(\tau >\tau _0\) if \(F_X(z) > F_Y(z)\) uniformly over z, for z sufficiently large. This property is determined by whether \(w^*<\frac{1}{2}\) in condition (11) or not. \(\square \)

1.9 Proof of Corollary 3.14

Proof

To prove the result in the corollary, it is sufficient to show that the iid assumption implies assumptions 2 and 3 with \(w^*=\frac{1}{2}\). Then, applying Proposition 3.13 the result follows.

To prove assumption 2, we note that under the iid condition, for any \(\mu \in {\mathbb {R}}\), it holds that \(f(\mu +\varepsilon , \mu -\varepsilon ) = f_X(\mu +\varepsilon ) f_X(\mu -\varepsilon )\). Now, taking the first derivative with respect to \(\varepsilon \), we obtain

$$\begin{aligned} \frac{\partial f(\mu +\varepsilon , \mu -\varepsilon )}{\partial \varepsilon }= \frac{\partial f_X(\mu +\varepsilon )}{\partial \varepsilon }f_X(\mu -\varepsilon ) - \frac{\partial f_X(\mu -\varepsilon )}{\partial \varepsilon }f_X(\mu +\varepsilon ). \end{aligned}$$
(26)

Let \(m_X\) denote the mode of the random variable X. For \(\mu =m_X\), the unimodality of \(f_X(\cdot )\) implies that \(\frac{\partial f_X(\mu +\varepsilon )}{\partial \varepsilon }>0\) for \(\varepsilon <0\), \(\frac{\partial f_X(\mu +\varepsilon )}{\partial \varepsilon }=0\) at \(\varepsilon =0\) and \(\frac{\partial f_X(\mu +\varepsilon )}{\partial \varepsilon }<0\) for \(\varepsilon >0\). Under these conditions, expression (26) yields \(\frac{\partial f(\mu +\varepsilon , \mu -\varepsilon )}{\partial \varepsilon }>0\) for \(\varepsilon <0\); \(\frac{\partial f(\mu +\varepsilon , \mu -\varepsilon )}{\partial \varepsilon }=0\) at \(\varepsilon =0\) and \(\frac{\partial f(\mu +\varepsilon ,\mu -\varepsilon )}{\partial \varepsilon }<0\) for \(\varepsilon >0\), as stated in assumption 2.

The proof is a bit more complex for \(\mu \ne m_X\). In this case, for \(\mu <m_X\) there exists an interval \(|\varepsilon |< m_X - \mu \) that needs to be carefully evaluated. More specifically, for values of \(\varepsilon \) inside the interval, we note that expression (26) is positive if \(\varepsilon <0\) and negative if \(\varepsilon >0\). This follows from noting that for \(\varepsilon <0\) and \(\mu -m_X<\varepsilon \), the following condition is satisfied:

$$\begin{aligned} \frac{\partial f_X(\mu +\varepsilon )/\partial \varepsilon }{\partial f_X(\mu -\varepsilon )/\partial \varepsilon }>\frac{f_X(\mu +\varepsilon )}{f_X(\mu -\varepsilon )}. \end{aligned}$$
(27)

To show this condition it is sufficient to show that \(\frac{\partial f_X(\mu +\varepsilon )}{\partial \varepsilon }>\frac{\partial f_X(\mu -\varepsilon )}{\partial \varepsilon }\) given that \(f_X(\mu +\varepsilon )<f_X(\mu -\varepsilon )\) for \(\varepsilon <0\). This condition is, however, fulfilled for \(x \in (-\infty ,m_X)\) if the density function \(f_X\) is unimodal. Similarly, for \(\varepsilon >0\) and \(\varepsilon <m_X-\mu \), the following condition is satisfied:

$$\begin{aligned} \frac{\partial f_X(\mu +\varepsilon )/\partial \varepsilon }{\partial f_X(\mu -\varepsilon )/\partial \varepsilon }<\frac{f_X(\mu +\varepsilon )}{f_X(\mu -\varepsilon )}. \end{aligned}$$
(28)

This condition can be shown using the above arguments and noting that \(\varepsilon >0\).

For values of \(\varepsilon \) outside the interval \((0,m_x-\mu )\) the above proof follows straightforwardly.

For \(\mu > m_X\), there exists an interval \(|\varepsilon |< \mu - m_X\) that needs to be carefully evaluated. Nevertheless, the proof in this case follows as in the previous case. Also, for values of \(\varepsilon \) outside the interval the above proof follows similarly.

To verify Assumption 3, it is sufficient to show that for \(w^*=\frac{1}{2}\) and \(\varepsilon >0\), Eq. (11) becomes

$$\begin{aligned} f(\mu +\varepsilon , \mu -\varepsilon ) = f(\mu -\varepsilon , \mu +\varepsilon ), \end{aligned}$$
(29)

for all \(\mu \in {\mathbb {R}}\), with \(f(\cdot ,\cdot )\) the joint density function of the random variables X and Y. This condition is, however, satisfied by construction under the iid assumption. This is so because

$$\begin{aligned} f(\mu +\varepsilon , \mu -\varepsilon ) = f_X(\mu +\varepsilon ) f_X(\mu -\varepsilon )=f(\mu -\varepsilon , \mu +\varepsilon ), \end{aligned}$$
(30)

for all \(\mu \in {\mathbb {R}}\). \(\square \)

B Numerical simulation study

This Appendix presents additional results on the optimal asset allocation under quantile preferences (QP) for a portfolio \(S_{w}=w X+(1-w)Y\) given by the mixture of two continuous random variables X and Y, which are chosen from different distribution functions. For illustrative purposes, we consider pairs of Gaussian and Chi-squared random variables, under mutual independence and also with dependence. These results are also compared with the optimal asset allocation obtained under EU. Numerical computation of the portfolios are as described in Sect. 3.4.

1.1 B.1 Independent and identically distributed random variables

1.1.1 Two Chi-squared independent and identically distributed random variables

To illustrate the optimal portfolio allocation problem for asymmetric distributions, we investigate the case of two independent and identically distributed standard Chi-squared distributions, X and Y. Under risk aversion, the iid assumption is sufficient to guarantee that the optimal portfolio choice of a EU individual will be a fully diversified portfolio given by \(w^*=0.5\). For QP, the same result holds true for values of \(\tau \) identified with risk aversion. The left panel of Fig. 15 shows that the family of distribution functions \(F_{S_{w}}\) satisfies the single-crossing condition defined by Manski (1988); see also de Castro and Galvao (2021, Section 5). The right panel of Fig. 15 shows the optimal portfolio allocation is divided into two regions: a first region given by \({\widehat{w}}_n^{*}(\tau )=0.5\), for \(\tau \le \tau _0\approx 0.80\), and a second region given by \({\widehat{w}}_n^{*}(\tau )=\{0, 1\}\), for \(\tau >\tau _0\). The shift in the cut-off point \(\tau _0\) compared to \(\tau _0=0.5\) is due to the asymmetry of the Chi-squared distribution.

Fig. 15
figure 15

\(X,Y \sim \chi _{1}^{2}\). Left box plots the CDF of \(S_w\). Right box plots QP portfolio selection

1.2 B.2 Independently distributed random variables

1.2.1 Two Gaussian independent random variables with different variances

We extend the Gaussian case discussed above and consider the case of independent random variables X and Y with different variances. Let \(X\sim N(0,1)\) and \(Y\sim N(0,2)\), independent of each other. In this case Y is a mean-preserving spread of the random variable X. This observation implies that X second order stochastically dominates Y and, using Fishburn (1977), X is preferred to Y for EU individuals endowed with an increasing and concave utility function. We can extend this result to derive the optimal portfolio allocation of a EU risk averse individual. In this case, the combination \(S_{w^*}\) with \(w^*=2/3\) has zero mean and minimizes the variance of the family of random variables \(S_{w}\). This combination second order stochastically dominates any other convex combination including X and Y. Then, using Fishburn (1977), \(S_{w^*}\) is the optimal strategy for EU individuals endowed with an increasing and concave utility function.

For individuals endowed with QP, we note that condition (11) is satisfied for \(w^*=2/3\). Then, Proposition 3.13 implies that the optimal combination of a QP individual is \(w^*=\frac{2}{3}\) for \(\tau \le \tau _0\), with \(\tau _0=\frac{1}{2}\). This result is illustrated in the numerical exercise on the right panel of Fig. 16. Diversification takes place for \(\tau \le \frac{1}{2}\) and no diversfication is the optimal result for \(\tau >\frac{1}{2}\).

Fig. 16
figure 16

\(X \sim N(0, 1)\) and \(Y\sim N(0,2)\). Left panel plots the CDF of \(S_w\). Right panel plots QP portfolio selection

1.3 Two Gaussian independent random variables with different means

Consider now the case of two independent Gaussian random variables with different means. Let \(X\sim N(1,1)\) and \(Y\sim N(1.5,1)\), independent of each other. In this case the random variable Y first order stochastically dominates the random variable X. In this scenario the optimal portfolio decision of an EU individual has to be calculated for each utility function separately because it depends on how the individual faces the trade-off between risk and return.

The results for the QP case are provided in Fig. 17. The results show the presence of an interior solution for \(\tau \le \tau _0\), with \(\tau _0\) around 0.30, and a corner solution \({\widehat{w}}_n^{*}(\tau )=0\) for \(\tau >0.30\). The solution in the right tail of the distribution is rationalized by the first order stochastic dominance of Y over any other convex combination of X and Y for \(\tau \) sufficiently large.

Fig. 17
figure 17

\(X \sim N(1, 1)\) and \(Y\sim N(1.5,1)\). Left box plots the CDF of \(S_w\). Right box plots QP portfolio selection

1.4 B.3 Dependent random variables

We study now the case of two dependent random variables. We differentiate between symmetric and asymmetric random variables.

1.4.1 Two Gaussian dependent random variables with same mean

In this exercise, we consider a bivariate Normal random variable (XY) with covariance matrix \(\Sigma =[1 \ \rho , \rho \ 1]\), where \(\rho =0.5\) is the correlation parameter. This scenario is rationalized by Proposition 3.13. More specifically, condition (11) is satisfied for \({\widehat{w}}_n^{*}(\tau )=0.5\), such that an interior solution is obtained for \(\tau \le \tau _0=0.5\). This result is observed in Fig. 18. For values of \(\tau \) greater than 0.5 the optimal solution is, indistinctively, \({\widehat{w}}_n^{*}(\tau )=\{0,1\}\).

Fig. 18
figure 18

\((X,Y) \sim N(0,\Sigma )\) with \(\Sigma =[1 \ 0.5; 0.5 \ 1]\). Left box plots CDF of \(S_w\). Right box plots QP portfolio selection

1.4.2 Chi-squared dependent random variables

We complete the case of dependent assets with asymmetric distribution functions. Suppose that \(X\sim \chi ^2_1+1\) and \(Y\sim \chi ^2_2+X-2\) two dependent assets with expected value equal to two in both cases. For QP individuals, the optimal portfolio allocation has an interior solution for \(\tau \le \tau _0\), with \(\tau _0=0.60\). The presence of an interior solution for small values of \(\tau \) is rationalized by Theorem 1. However, in contrast to previous examples, assumptions 3 is not satisfied implying that the optimal portfolio allocation varies over \(\tau \) and there is no separation between risk aversion and risk loving behavior. Figure 19 reports the optimal portfolio allocation in this case.

Fig. 19
figure 19

\(X \sim \chi _{1}^{2}+1\) and \(Y \sim \chi _{2}^{2}+X-2\). Left box plots the CDF of \(S_w\). Right box plots QP portfolio selection

1.5 B.4 Optimal portfolio allocation when there is a risk-free asset

Building upon previous insights of the QP theory in Sect. 3.2, we can extend the mutual fund separation theorem to the case of one risk free asset with returns \({\bar{r}}\) and two risky assets \(R_1\), \(R_2\) with distribution functions \(F_{R_1}, F_{R_2}\), respectively, and such that \(F_{R_1}\) crosses \(F_{R_2}\) from below at point \(x_{12}\). This assumption implies that \(R_2\) is riskier than \(R_1\). The portfolio return is defined by the convex combination \(R_p=w_0 {\bar{r}} + w_1 R_1+ w_2 R_2\), with \(w_0+w_1+w_2=1\). The investor’s optimization problem is

$$\begin{aligned} \arg \max _{\{w_0, w_1, w_2\}} \text {Q}_{\tau }[ w_0 {\bar{r}} + w_1 R_1+ w_2 R_2]. \end{aligned}$$
(31)

In this case the individual’s optimal portfolio choice is given by the risk-free rate when \(\text {Q}_{\tau }[R_p(w)] < {\bar{r}}\), for any combination of weights \(w = \{w_0, w_1, w_2 \}\) that is different from \(w^* = \{w_0^*=1, w_1^*=w_2^*=0 \}\). For higher quantiles, the solution to the maximization problem (31) is, in principle, quantile-specific. We illustrate this scenario by simulating the returns on three assets with returns \({\bar{r}}\), \(R_1 \sim N(\mu _1,1)\) and \(R_2 \sim N(\mu _2,1)\). The solution for \(0 \le \tau \le 1\) is obtained by simulating \(n=10,000\) realizations of the random variables.

Fig. 20
figure 20

\(R_p=w_0 {\bar{r}} + w_1 R_1+ w_2 R_2\), with \({\bar{r}}\) constant, \(R_1 \sim N(\mu _1,1)\) and \(R_2 \sim N(\mu _2,1)\). Left box plots \({\bar{r}}=0.5\), \(\mu _1=\mu _2=0\). Right box plots \({\bar{r}}=0.25\), \(\mu _1=\mu _2=0\)

The left panel of Fig. 20 reports the case \({\bar{r}}=0.5\), \(\mu _1=\mu _2=0\). The right panel considers the case \({\bar{r}}=0.25\), \(\mu _1=\mu _2=0\). As discussed above, for \(\text {Q}_{\tau }[R_p(w)] < {\bar{r}}\), the optimal allocation to the portfolio is given by investing on the risk-free asset. For the parameterization \({\bar{r}}=0.5\), \(\mu _1=\mu _2=0\) - left panel - the value of \(\tau \) that yields the condition \(\text {Q}_{\tau }[R_p(w^*)] = {\bar{r}}\) is \(\tau _0 \approx 0.7\). For values of \(\tau >0.7\), the optimal allocation to the risk-free asset is zero, and the QP individual is indifferent between assets 1 and 2. The right panel characterized by a smaller risk-free return (\({\bar{r}}=0.25\)) presents a similar outcome. In this scenario the value of \(\tau \) that satisfies the condition \(\text {Q}_{\tau }[R_p(w^*)] = {\bar{r}}\) is \(\tau _0 \approx 0.6\). For higher quantiles, the optimal portfolio allocation is the same as for the left panel and entails a zero allocation to the risk-free rate.Footnote 16

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Castro, L.d., Galvao, A.F., Montes-Rojas, G. et al. Portfolio selection in quantile decision models. Ann Finance 18, 133–181 (2022). https://doi.org/10.1007/s10436-021-00405-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10436-021-00405-4

Keywords

JEL Classification

Navigation