1 Introduction

Over the last years, there was a steady increase in literature dealing with expectiles. These are measures of non-central location that have properties similar to quantiles. Therefore, expectiles can also be used as building blocks for measures of scale, skewness, etc. However, no such measure should be used without identification of the ordering it preserves.

Let us explain this in more detail. Since the seminal work of van Zwet (1964), Oja (1981), MacGillivray (1986), an axiomatic approach to measure statistical quantities is commonly accepted. It involves two main steps (Hürlimann 2002):

  • Define stochastic (partial) orders on sets of random variables or distribution functions that allow for comparisons of the given statistical quantity.

  • Identify measures of the statistical quantity by considering functionals of distributions that preserve the partial order, and use only such measures in practical work.

Given a dispersion or variability ordering \(\le _D\), the general axiomatic approach requires that a scale measure \(\delta \) satisfies

  1. D1.

    For \(c,d\in {\mathbb {R}}\), \(\delta (cX+d)=|c |\delta (X)\).

  2. D2.

    If \(X\le _D Y\), then \(\delta (X)\le \delta (Y)\).

Similarly, given a skewness order \(\le _S\), a skewness measure \(\gamma \) should satisfy

  1. S1.

    For \(c>0\) and \(d\in {\mathbb {R}}\), \(\gamma (cX+d) = \gamma (X)\).

  2. S2.

    The measure satisfies \(\gamma (-X)=-\gamma (X)\).

  3. S3.

    If \(X\le _S Y\), then \(\gamma (X)\le \gamma (Y)\).

The generally accepted strongest dispersion and skewness orders are the dispersive order (Bickel and Lehmann 1979) and the convex transformation order (van Zwet 1964), respectively.

Research which treats expectile-based measures and related stochastic orderings includes Bellini (2012), Bellini et al. (2014, 2018a), Klar and Müller (2019), Eberl and Klar (2020, 2021), Arab et al. (2022). In particular, Eberl and Klar (2020, 2021) introduced expectile-based measures of skewness which possess quite promising properties and have close connections to other skewness functionals. However, it remained unanswered whether these measures preserve the convex transformation order. It is one of the aims of the present work to answer this question in the affirmative.

As part of these measures of skewness, interexpectile distances (also called interexpectile ranges) appear quite naturally; they have also been used in a finance context by Bellini et al. (2018b, 2020, 2021). We introduce orders of variability based on slightly more general quantities and show that the so-called weak expectile dispersive order is equivalent to the dilation order. Hence, interexpectile ranges preserve the dispersive ordering.

Up to now, statistical properties of empirical interexpectile ranges do not seem to have been investigated. We show that they are a good compromise between the standard deviation on the one hand and robust measures as the interquartile range on the other.

This paper is organized as follows. In Sect. 2, we recall the definitions of expectiles and some of their properties. Section 3 discusses expectile skewness. It is shown that expectile-based skewness measures are consistent with the convex transformation order. In Sect. 4, we introduce strong and weak expectile dispersive orders and show that the latter is equivalent to the dilation order. It follows that the interexpectile range preserves the dispersive order. These concepts are illustrated in Sect. 5, using the Lomax or Pareto type II distribution. Empirical interexpectile ranges are analyzed and compared with other scale measures in Sect. 6, using standardized asymptotic relative efficiencies.

2 Preliminaries

Throughout the paper, we assume that all mentioned random variables X are non-degenerate, have a finite mean (denoted as \(X \in L^1\)), and are defined on a common probability space \((\Omega ,{\mathcal {A}}, P)\). Further, we assume that the supports of the underlying distributions are intervals and that these distributions have no atoms. Hence, the distribution function (cdf) \(F_X\) of X has a strictly increasing and continuous inverse on (0, 1).

Expectiles \(e_X(\alpha )\) of a random variable \(X\in L^{2}\) have been defined by Newey and Powell (1986) as the minimizers of an asymmetric quadratic loss:

$$\begin{aligned} e_X(\alpha )=\arg \min _{t\in {\mathbb {R}}}\left\{ E \ell _\alpha (X-t) \right\} \text {,} \end{aligned}$$
(1)

where

$$\begin{aligned} \ell _\alpha (x) = {\left\{ \begin{array}{ll} \alpha x^2, &{} \text{ if } x \ge 0, \\ (1-\alpha ) x^2, &{} \text{ if } x < 0, \end{array}\right. } \end{aligned}$$

and \(\alpha \in (0,1)\). For \(X\in L^{1}\) (but \(X\notin L^{2}\)), equation (1) has to be modified (Newey and Powell 1986) to

$$\begin{aligned} e_X(\alpha ) = \arg \min _{t\in {\mathbb {R}}} \left\{ E\left[ \ell _\alpha (X-t) - \ell _\alpha (X) \right] \right\} . \end{aligned}$$
(2)

The minimizer in (1) or (2) is always unique and is identified by the first order condition

$$\begin{aligned} \alpha E\left( X-e_{X}(\alpha )\right) _{+}=(1-\alpha )E\left( X-e_{X}(\alpha )\right) _{-}\text {,} \end{aligned}$$
(3)

where \(x_{+}=\max \{x,0\}\), \(x_{-}=\max \{-x,0\}\). This is equivalent to characterizing expectiles via an identification function, which, for any \(\alpha \in (0, 1)\) is defined by

$$\begin{aligned} I_\alpha (x, y) = \alpha (y - x) \mathbbm {1}_{\{y \ge x\}} - (1 - \alpha ) (x - y) \mathbbm {1}_{\{y < x\}} \end{aligned}$$

for \(x, y \in {\mathbb {R}}\). The \(\alpha \)-expectile of a random variable \(X \in L^1\) is then the unique solution of

$$\begin{aligned} E I_\alpha (t, X) = 0, \quad t \in {\mathbb {R}}. \end{aligned}$$

Similarly, the empirical \(\alpha \)-expectile \({\hat{e}}_n(\alpha )\) of a sample \(X_1,\ldots ,X_n\) is defined as solution of

$$\begin{aligned} I_\alpha (t, {\hat{F}}_n) = \frac{1}{n} \sum _{i=1}^{n} I_\alpha (t, X_i) = 0, \quad t \in {\mathbb {R}}. \end{aligned}$$

Like quantiles, expectiles are measures of non-central location and have similar properties [see, e.g., Newey and Powell (1986) and Bellini et al. (2014)]. Clearly, expectiles depend only on the distribution of the random variable X and can be seen as statistical functionals defined on the set of distribution functions with finite mean on \({\mathbb {R}}\).

Throughout the paper, we make use of the following stochastic orders:

Definition 1

  1. (i)

    \(F_X\) precedes \(F_Y\) in the usual stochastic order (denoted by \(F_X \le _{st} F_Y\)) if \(E\phi (X)\le E\phi (Y)\) for all increasing functions \(\phi \), for which both expectations exist.

  2. (ii)

    \(F_X\) precedes \(F_Y\) in the convex order (denoted by \(F_X \le _{cx} F_Y\)) if \(E\phi (X) \le E\phi (Y)\) for all convex functions \(\phi \), for which both expectations exist.

  3. (iii)

    \(F_X\) precedes \(F_Y\) in the convex transformation order (denoted by \(F_X\le _{c} F_Y\)) if \(F_Y^{-1}\circ F_X\) is convex.

  4. (iv)

    \(F_X\) precedes \(F_Y\) in the expectile (location) order (denoted by \(F_X\le _{e} F_Y\)) if

    $$\begin{aligned} e_X(\alpha ) \le e_Y(\alpha ) \text{ for } \text{ all } \alpha \in (0,1). \end{aligned}$$

The first three orders are well-known (Shaked and Shantikumar 2007; Müller and Stoyan 2002); the expectile (location) order was introduced in Bellini et al. (2018a). Since the usual stochastic order \(\le _{st}\) is equivalent to the pointwise ordering of the quantiles, the definition of the expectile ordering is quite natural, just replacing quantiles by expectiles.

As noted by Jones (1994), expectiles are the quantiles of a suitably transformed distribution. Indeed, the first order condition (3) can be written in the equivalent form

$$\begin{aligned} \alpha&=\frac{E\left( X-t\right) _{-}}{E\left| X-t\right| } =\frac{\pi _X(t)-\mu +t}{2\pi _X(t)-\mu +t} =:\breve{F}_X(t), \end{aligned}$$
(4)

where \(\mu =EX\) and \(\pi _X(t) = E(X-t)_+\) denotes the stop-loss transform of X. Hence, \(\breve{F}_X(t)=e_X^{-1}(t)\), and the expectile order could also be defined by \(\breve{F}_X(t)\ge \breve{F}_Y(t)\) for all t. The expectile order is weaker than the usual stochastic order, i.e. \(X\le _{st}Y\) implies \(X\le _{e}Y\) (Bellini 2012).

3 Expectile skewness

In the following, we summarize some results which are important with respect to expectile-based quantification of skewness. Bellini et al. (2018a), Theorem 12 and Corollary 13, show the following equivalence.

Proposition 2

Let \(X,Y\in L^{1}\) with \(EX=EY=\mu \), \(m=\mathrm{ess\,inf}(X), M=\mathrm{ess\,sup}(X)\). Then the following conditions are equivalent:

  1. (a)

    \(X\le _{e}Y\).

  2. (b)

    \(\pi _{X}(x)\ge \pi _{Y}(x)\), for each \(x\in (m,\mu )\) and \(\pi _{X}(x)\le \pi _{Y}(x)\), for each \(x\in (\mu ,M)\).

  3. (c)

    \((X-\mu )_{-}\ge _{cx}(Y-\mu )_{-}\) and \((X-\mu )_{+}\le _{cx}(Y-\mu )_{+}\).

Eberl and Klar (2021) introduced a family of scalar measures of skewness

$$\begin{aligned} {\tilde{s}}_2(\alpha ) = \frac{ e_X(1-\alpha )+e_X(\alpha )-2\mu }{ e_X(1-\alpha )-e_X(\alpha ) }, \quad \alpha \in (0,1/2), \end{aligned}$$

and called a distribution right-skewed (left-skewed) in the expectile sense if \({\tilde{s}}_2(\alpha )\ge (\le )\, 0\) for all \(\alpha \in (0,1/2)\) and equality does not hold for each \(\alpha \in (0,1/2)\). If equality holds for each \(\alpha \), the distribution is symmetric. They also defined a normalized version \(s_2(\alpha )={\tilde{s}}_2(\alpha )/(1-2\alpha )\), and proved that \(-1<s_2(\alpha )<1\), and both inequalities are sharp for any \(\alpha \in (0, 1/2)\).

Moreover, Eberl and Klar (2021) introduced a function

$$\begin{aligned} S_X(t) = \frac{\pi _X(\mu +t)-\pi _X(\mu -t)}{t} + 1 = \frac{1}{t} \int _{\mu -t}^{\mu + t} F_X(z) dz - 1, \quad t>0, \end{aligned}$$

which has been shown to be strongly related to \(s_2\): it holds that \(s_2(\alpha )\ge 0\) for \(\alpha \in (0,1/2)\) if and only if \(S_X(t)\ge 0\) for each \(t>0\) (Eberl and Klar 2021). Based on \(S_X\), the following two skewness orders have been defined, which are both weaker than van Zwet’s skewness order \(\le _c\) (Eberl and Klar 2021).

Definition 3

  1. (a)

    Let \({\tilde{S}}_X(t)=S_X(td_X)\), where \(d_X=E|X-EX|\) denotes the mean absolute deviation from the mean (MAD). Then, we write \(F_X\le _{sf} F_Y\) if \({\tilde{S}}_X(t) \le {\tilde{S}}_Y(t) \, \forall \, t>0\).

  2. (b)

    \(F_Y\) is more skew with respect to mean and MAD than \(F_X\) (\(F_X <_\mu ^{d} F_Y\)), if \(F_X(d_X x+EX)\) and \(F_Y(d_Y x+EY)\) cross each other exactly once on each side of \(x=0\), with \(F_X(EX)\le F_Y(EY)\).

Arab et al. (2022) introduced a skewness order as follows.

Definition 4

Let \(X,Y\in L^{1}\), and define \({\tilde{X}}=(X-EX)/E|X-EX|, {\tilde{Y}}=(Y-EY)/E|Y-EY|\), with cdf’s \(F_{{\tilde{X}}}\) and \(F_{{\tilde{Y}}}\). X is smaller than Y in the s-order, denoted by \(F_X\le _s F_Y\), if

$$\begin{aligned} \int _{-\infty }^x F_{{\tilde{X}}}(t)dt&\ge \int _{-\infty }^x F_{{\tilde{Y}}}(t)dt, \quad \forall x \le 0, \end{aligned}$$

and

$$\begin{aligned} \int _x^{\infty } {\bar{F}}_{{\tilde{X}}}(t)dt&\le \int _x^{\infty } {\bar{F}}_{{\tilde{Y}}}(t)dt, \quad \forall x\ge 0, \end{aligned}$$

where \({\bar{F}}=1-F\) denotes the survival function.

Using well-known conditions for the convex order [see, e.g., (3.A.7) and (3.A.8) in Shaked and Shantikumar (2007)], \(X\le _s Y\) holds if \({\tilde{X}}_{-}\ge _{cx}{\tilde{Y}}_{-}\) and \({\tilde{X}}_{+}\le _{cx}{\tilde{Y}}_{+}\). A comparison with Proposition 2c) yields that

$$\begin{aligned} X\le _s Y \quad \Leftrightarrow \quad {\tilde{X}} \le _{e} {\tilde{Y}} \end{aligned}$$
(5)

With regard to the standardized variables, (5) establishes that larger expectiles correspond to a more skewed distribution. The equivalence in (5) was derived in Arab et al. (2022), Theorem 14, using a different argument.

Definition 4 entails \(\int _{-x}^{x} (F_{{\tilde{X}}}(t)-F_{{\tilde{Y}}}(t)) dt \le 0\) for all \(x>0\). However, this is equivalent to \({\tilde{S}}_X(x) \le {\tilde{S}}_Y(x)\) for all \(x>0\). Hence, \(\le _{sf}\), the skewness order defined by \({\tilde{S}}_X\), is weaker than the \(\le _s\)-order. On the other hand, the proof of Theorem 13(2.) in Arab et al. (2022) shows that \(\le _s\) is weaker than the order \(\le _\mu ^d\). Taking into account Theorem 3 in Eberl and Klar (2021), we obtain the following chain of implications:

$$\begin{aligned} X\le _c Y \;\Rightarrow \; X\le _\mu ^d Y \;\Rightarrow \; X\le _s Y \;\Rightarrow \; X\le _{sf} Y. \end{aligned}$$
(6)

The following theorem is the main result in this section, showing that the expectile-based skewness measure \(s_2\) (and, hence, \({\tilde{s}}_2\)) preserves van Zwet’s skewness order.

Theorem 5

Let \(X,Y\in L^{1}\). Then, \(s_2\) is consistent with \(\le _c\), i.e. \(X\le _c Y\) implies \(s_{2,X}(\alpha )\le s_{2,Y}(\alpha )\) for each \(\alpha \in (0,1/2)\).

Proof

By (6) and (5), \(X\le _c Y\) implies \({\tilde{X}} \le _{e} {\tilde{Y}}\), i.e. \(e_{{\tilde{X}}}(\alpha ) \le e_{{\tilde{Y}}}(\alpha )\) for all \(\alpha \in (0,1)\). A straightforward computation shows that \(s_{2,{\tilde{X}}}(\alpha )\le s_{2,{\tilde{Y}}}(\alpha )\) is equivalent to

$$\begin{aligned} e_{{\tilde{Y}}}(1-\alpha ) e_{{\tilde{X}}}(\alpha ) \le e_{{\tilde{X}}}(1-\alpha ) e_{{\tilde{Y}}}(\alpha ), \quad \forall \alpha \in (0,1/2), \end{aligned}$$

which, in turn, is equivalent to

$$\begin{aligned} e_{{\tilde{Y}}}(1-\alpha )\, |e_{{\tilde{X}}}(\alpha )|\ge e_{{\tilde{X}}}(1-\alpha )\, |e_{{\tilde{Y}}}(\alpha )|, \quad \forall \alpha \in (0,1/2). \end{aligned}$$
(7)

Since, by assumption, \(e_{{\tilde{Y}}}(1-\alpha )\ge e_{{\tilde{X}}}(1-\alpha )\) and \(|e_{{\tilde{X}}}(\alpha )|\ge |e_{{\tilde{Y}}}(\alpha )|\) for \(\alpha \in (0,1/2)\), inequality (7) holds. Since \(s_2\) is invariant under location-scale transforms, \(s_{2,X}(\alpha )\le s_{2,Y}(\alpha )\) holds as well for all \(\alpha \in (0,1/2)\). \(\square \)

Remark 6

If we write \(X \le _{s_2} Y\) if \(s_{2,X}(\alpha )\le s_{2,Y}(\alpha )\) for all \(\alpha \in (0,1/2)\), the proof of Theorem 5 shows that \(\le _{s_2}\) is a weaker order of skewness than \(\le _{s}\). The question if \(\le _{s_2}\) implies \(\le _{sf}\) or vice versa is open. Another expectile based order, stronger than \(\le _{s_2}\), could be defined analogously to the quantile-based skewness order \(\le _c\), i.e. by

$$\begin{aligned} \frac{e_X(w)-2e_X(v)+e_X(u)}{ e_X(w)-e_X(u)}&\le \frac{e_Y(w)-2e_Y(v)+e_Y(u)}{ e_Y(w)-e_Y(u)} \end{aligned}$$
(8)

for all \(0<u<v<w<1\); this order is equivalent to the convexity of \(\breve{F}_Y^{-1} \circ \breve{F}_X\), where \(\breve{F}_X\) is defined in (4) [cp. Eberl and Klar (2021) for the quantile case]. However, the above results are not strong enough to show that this order is weaker than \(\le _c\).

For the suitability of \(s_2\) as a skewness measure, including its empirical counterpart, we refer to Eberl and Klar (2020, 2021). To gain an impression, consider a Bernoulli distribution with success probability p. Clearly, for \(p=1/2\), any skewness measure should be zero. For decreasing p, the distribution becomes more and more skewed to the right (note that the center decreases to 0), and a skewness measure should converge to its maximal possible value. Similarly, it should converge to its minimal possible value for \(p\rightarrow 1\). For the expectile skewness, we obtain \(s_2(\alpha )=1-2p\), independent of \(\alpha \). The moment skewness is \((1-2p)/(p(1-p))\); quantile-based skewness measures are not uniquely defined for discrete distributions.

4 Expectile dispersive order and related measures

Additionally to the stochastic orders in Sect. 2, we need the following dispersion or variability orders.

Definition 7

  1. (i)

    \(F_X\) precedes \(F_Y\) in the dispersive order (written \(F_X \le _{disp} F_Y\)) if

    $$\begin{aligned} F_X^{-1}(v)-F_X^{-1}(u) \le F_Y^{-1}(v)-F_Y^{-1}(u) \quad \forall \ 0<u<v<1. \end{aligned}$$
    (9)
  2. (ii)

    \(F_X\) precedes \(F_Y\) in the weak dispersive order (written \(F_X \le _{w-disp} F_Y\)) if (9) is fulfilled for all \(0<u \le 1/2 \le v<1\).

  3. (iii)

    \(F_X\) precedes \(F_Y\) in the expectile dispersive order (written \(F_X \le _{e-disp} F_Y\)) if

    $$\begin{aligned} e_X(v)-e_X(u) \le e_Y(v)-e_Y(u) \quad \forall \ 0<u<v<1. \end{aligned}$$
    (10)
  4. (iv)

    \(F_X\) precedes \(F_Y\) in the weak expectile dispersive order (written \(F_X \le _{we-disp} F_Y\)) if (10) is fulfilled for all \(0<u \le 1/2 \le v<1\).

  5. (v)

    \(F_X\) precedes \(F_Y\) in the dilation order (written \(F_X \le _{dil} F_Y\)) if

    $$\begin{aligned} X-EX \le _{cx} Y-EY. \end{aligned}$$
    (11)

Remark 8

The first ordering in Definition 7 is well-known [see, e.g., Bickel and Lehmann (1979), Oja (1981), Shaked (1982), Shaked and Shantikumar (2007)]. The defining condition of the weak dispersive order, which is obviously weaker than the dispersive order, can be equivalently written as

$$\begin{aligned} F_Y^{-1}(u)-F_X^{-1}(u)&\le (\ge ) \, F_Y^{-1}(1/2)-F_X^{-1}(1/2) \quad \text {for } u \le (\ge ) 1/2. \end{aligned}$$
(12)

This ordering was used in Bickel and Lehmann (1976) for symmetric distribution; for arbitrary distributions, it was introduced in MacGillivray (1986), Definition 2.6, and denoted by \(\le _1^m\). In introducing strong and weak expectile dispersive orders, we mimic these definitions. The following results show that the weak ordering defined in such a way has appealing theoretical properties; further, it corresponds to the expectile-based scale measures treated in Sect. 6. The expectile equivalent of the (strong) dispersive order from Def. 7(i) is shortly discussed in Example 12 and Remark 13.

We remark that Bellini et al. (2021) have introduced the following expectile based variability order which is even weaker than the weak expectile dispersive order:

$$\begin{aligned} X \le _{\Delta -ex} Y, \quad \text {if } e_X(1-\alpha )-e_X(\alpha ) \le e_Y(1-\alpha )-e_Y(\alpha ) \; \forall \ \alpha \in (0,1/2). \end{aligned}$$

This order can be seen as a dispersion counterpart to the skewness order \(\le _{s_2}\) in Remark 6. The dilation order in Definition 7(v) is less well-known compared to the dispersive order, see Belzunce et al. (1997) or Fagiuoli et al. (1999). However, many of its properties follow from properties of the convex order.

Our main result in this section shows that the weak expectile dispersive order and the dilation order coincide. The proof uses the idea in the proof of Theorem 14 in Arab et al. (2022).

Theorem 9

Let \(X,Y\in L^1\) with strictly increasing cdf’s \(F_X\) and \(F_Y\) which are continuous on their supports. Then, \(F_X \le _{we-disp} F_Y\) if, and only if, \(F_X \le _{dil} F_Y\).

Proof

Define \({\tilde{X}}=X-EX, {\tilde{Y}}=Y-EY\). Then, \(X \le _{dil} Y\) is equivalent to \({\tilde{X}} \le _{cx} {\tilde{Y}}\), and therefore to \(\pi _{{\tilde{X}}}(t)\le \pi _{{\tilde{Y}}}(t) \, \forall t\). By continuity of \(\pi _{{\tilde{X}}}\) and \(\pi _{{\tilde{Y}}}\), this is equivalent to \(\pi _{{\tilde{X}}}(t)/|t|\le \pi _{{\tilde{Y}}}(t)/|t|\, \forall t\ne 0\), and hence to

$$\begin{aligned} \frac{\pi _{{\tilde{X}}}(t)}{t} \ge \frac{\pi _{{\tilde{Y}}}(t)}{t} \; \forall t<0 \quad \text {and} \quad \frac{\pi _{{\tilde{X}}}(t)}{t} \le \frac{\pi _{{\tilde{Y}}}(t)}{t} \; \forall t>0. \end{aligned}$$
(13)

Note that, using the properties of the stop-loss transform (Müller and Stoyan 2002, Theorem 1.5.10), \(\pi _{{\tilde{X}}}(t)/t \le -1\) for \(t<0\), and \(\pi _{{\tilde{X}}}(t)/t \ge 0\) for \(t>0\). Now, applying to both sides of the inequalities the transformation \(h(x)=(x+1)/(2x+1), x\ne -1/2\), which is decreasing for \(x<-1/2\) as well as for \(x>-1/2\), shows that (13) is equivalent to

$$\begin{aligned} \breve{F}_{{\tilde{X}}}(t) \le \breve{F}_{{\tilde{Y}}}(t) \; \forall t<0 \quad \text {and} \quad \breve{F}_{{\tilde{X}}}(t) \ge \breve{F}_{{\tilde{Y}}}(t) \; \forall t>0, \end{aligned}$$
(14)

where \(\breve{F}_X\) is the expectile cdf defined in (4). In turn, (14) is equivalent to

$$\begin{aligned} e_{{\tilde{X}}}(p) \ge e_{{\tilde{Y}}}(p) \; \forall p<1/2 \quad \text {and} \quad e_{{\tilde{X}}}(p) \le e_{{\tilde{Y}}}(p) \; \forall p>1/2. \end{aligned}$$

This means that \(e_X(p)-EX \ge e_Y(p)-EY \, \forall p<0 \quad \text {and} \quad e_X(p)-EX \le e_Y(p)-EY \, \forall p>0\), which is equivalent to \(X \le _{we-disp} Y\) [cp. the representation of the weak dispersive order in (12)]. \(\square \)

Remark 10

Bellini (2012), Theorem 3(b) already proved that \(X\le _{cx} Y\) implies \(e_X(\alpha ) \ge e_Y(\alpha )\) for each \(\alpha \le 1/2\) and \(e_X(\alpha ) \le e_Y(\alpha )\) for each \(\alpha \ge 1/2\). From this, the implication \(F_X \le _{dil} F_Y \,\Rightarrow \, F_X \le _{we-disp} F_Y\) follows, see also Bellini et al. (2018b). However, Theorem 9 also yields the reverse direction. Moreover, its proof is rather elementary, whereas Bellini (2012) employed techniques from monotone comparative statics.

Since, for random variables with finite mean, the dispersive order implies the dilation order [Shaked and Shantikumar (2007), Theorem 3.B.16], the next result is a direct consequence of Theorem 9.

Corollary 11

Let \(F_X\) and \(F_Y\) be strictly increasing and continuous on their supports with finite expectations. Then, \(X\le _{disp}Y\) implies \(X\le _{we-disp}Y\).

In general, the (strong) expectile dispersive order in Definition 7(iii) is rather difficult to handle. However, the following example shows that this order does not imply the dispersive order [see also Remark 13c)].

It is an open question if the reverse direction holds, i.e. if the dispersive order implies the expectile dispersive order.

Example 12

Let \(p_X, p_Y \in (0, 1)\), \(p_X \ne p_Y\) and \(0< a_X < a_Y\). Furthermore, let \({\tilde{X}} \sim \mathrm {Bin}(1, p_X), {\tilde{Y}} \sim \mathrm {Bin}(1, p_Y)\) and \(X = a_X \cdot {\tilde{X}}\) and \(Y = a_Y \cdot {\tilde{Y}}\). It follows directly that X and Y are not comparable with respect to \(\le _{disp}\) since \(\mathrm {range}(F_X) \subseteq \mathrm {range}(F_Y)\) is a necessary condition for \(X \le _{disp} Y\) (Müller and Stoyan 2002, Theorem 1.7.3). Further, a simple calculation yields \(\pi _X(t) = p_X (a_X - t)\) for \(t \in [0, a_X]\). It follows that

$$\begin{aligned} \breve{F}_X(t) = \frac{t(1 - p_X)}{p_X a_X + t (1 - 2p_X)} \quad \text {and} \quad e_X(\alpha ) = \frac{\alpha p_X a_X}{(1-\alpha ) + p_X(2\alpha - 1)} \end{aligned}$$

for \(t \in [0, a_X]\) and \(\alpha \in (0, 1)\) with analogous results for Y. Overall,

$$\begin{aligned} (e_Y \circ \breve{F}_X)(t) = \frac{p_Y (1- p_X) a_Y t}{p_X (1 - p_Y) a_X + t (p_Y - p_X)} \end{aligned}$$

for \(t \in [0, a_X]\). Since \(X \le _{disp} Y\) is equivalent to \((F_Y^{-1} \circ F_X)' \ge 1\) [see, e.g., Oja 1981, p. 157], \(X \le _{e-disp} Y\) is equivalent to \((e_Y \circ \breve{F}_X)' \ge 1\). Because of \(\lim _{p_X \rightarrow p_Y} (e_Y \circ \breve{F}_X)(t) = \frac{a_Y}{a_X} t\) for all \(t \in [0, a_X]\), \(X \le _{e-disp} Y\) holds if the difference between \(p_X\) and \(p_Y\) is sufficiently small.

Since the involved distributions are discrete, this example does not fit in the general setting of this work. However, the statement of this example remains valid if the distributions of both X and Y are sufficiently closely approximated by continuous distributions (e.g. by linear interpolation). Overall, it is proved that \(X \le _{e-disp} Y \not \Rightarrow X \le _{disp} Y\) in general.

5 Dispersion orders for the Lomax distribution

To illustrate the various concepts, we consider the Lomax or Pareto type II distribution having density, distribution and quantile function

$$\begin{aligned} f(t)&= f(t;\alpha ,\lambda ) \ = \ \frac{\alpha }{\lambda } \left( 1+\frac{t}{\lambda }\right) ^{-(\alpha +1)}, \quad t\ge 0, \\ F(t)&= F(t;\alpha ,\lambda ) \ = \ 1 - \left( 1+t/\lambda \right) ^{-\alpha }, \quad t\ge 0, \\ q(p)&= q(p;\alpha ,\lambda ) \ = \ \lambda \left( (1-p)^{-1/\alpha }-1 \right) , \quad 0<p<1, \end{aligned}$$

where \(\alpha \) and \(\lambda \) are positive parameters. Accordingly, the stop-loss transform is given by

$$\begin{aligned} \pi (t)&= \frac{\lambda }{\alpha -1} \left( 1+\frac{t}{\lambda }\right) ^{1-\alpha }, \quad t\ge 0, \end{aligned}$$

and the expectile cdf \(\breve{F}\) can explicitly computed by (4). Further, if \(X\sim F(\cdot ;\alpha ,\lambda )\),

$$\begin{aligned} EX=\frac{\lambda }{\alpha -1} \; (\alpha>1),&\quad Var(X)= \frac{\lambda ^2\alpha }{(\alpha -1)^2(\alpha -2)} \; (\alpha >2). \end{aligned}$$

In the following, assume \(X\sim F(t;\alpha _1,\lambda _1)\) and \(Y\sim F(t;\alpha _2,\lambda _2)\).

5.1 Dispersive order

If X and Y have densities \(f_X\) and \(f_Y\), then \(X\le _{disp}Y\) if, and only if,

$$\begin{aligned} f_Y\left( F^{-1}_Y(p) \right)&\le f_X\left( F^{-1}_X(p) \right) \quad \forall p\in (0,1) \end{aligned}$$

[Shaked and Shantikumar 2007, (3.B.119)]. Applied to the Lomax distribution, \(X\le _{disp}Y\) iff

$$\begin{aligned} \frac{\alpha _2}{\lambda _2}(1-p)^{1/\alpha _2}&\le \frac{\alpha _1}{\lambda _1}(1-p)^{1/\alpha _1}, \quad 0<p<1. \end{aligned}$$
(15)

For p converging to 0, (15) is fulfilled if

$$\begin{aligned} \frac{\lambda _1}{\alpha _1} \le \frac{\lambda _2}{\alpha _2}. \end{aligned}$$
(16)

Looking at \(p\rightarrow 1\) shows that

$$\begin{aligned} \alpha _1 \ge \alpha _2 \end{aligned}$$
(17)

is a second necessary condition for \(X\le _{disp}Y\). However, (16) and (17) are also sufficient: Taking the reciprocal of (15) shows that the inequality is equivalent to

$$\begin{aligned} \frac{1}{\alpha _1} q(p;\alpha _1,\lambda _1) + \frac{\lambda _1}{\alpha _1} \; \le \; \frac{1}{\alpha _2} q(p;\alpha _2,\lambda _2)+\frac{\lambda _2}{\alpha _2}. \end{aligned}$$
(18)

Now, (16) and (17) are necessary and sufficient for \(X\le _{st} Y\) Bellini et al. (2018a). Hence, \(q(p;\alpha _1,\lambda _1)\le q(p;\alpha _2,\lambda _2)\) for all p, and (18) holds under (16) and (17). Overall, we have

$$\begin{aligned} X\le _{st} Y \text { as well as } X\le _{disp} Y&\; \text {if and only if} (16) \;\hbox {and}\; (17) \;\hbox {hold}. \end{aligned}$$

5.2 Weak expectile dispersive order

Here, we have to assume \(\alpha _1,\alpha _2>1\) for the expected values to exist. Define \(H(p)=\frac{1}{1-p}\int _p^1 F^{-1}(u)du\). Fagiuoli et al. (1999) showed that \(X\le _{dil}Y\) if, and only if,

$$\begin{aligned} H_X(p)-H_X(0)&\le H_Y(p)-H_Y(0), \quad 0<p<1. \end{aligned}$$
(19)

For the Lomax distribution, \(H(p)=(\alpha q(p)+\lambda )/(\alpha -1)\) and \(H(0)=\lambda /(\alpha -1)\). Hence, (19) holds iff

$$\begin{aligned} \frac{\alpha _1 \, q(p;\alpha _1,\lambda _1)}{\alpha _1-1}&\le \frac{\alpha _2 \, q(p;\alpha _2,\lambda _2)}{\alpha _2-1}, \quad 0<p<1. \end{aligned}$$
(20)

A discussion of the behavior of q(p) for \(p\rightarrow 1\) shows that (17) is necessary for (20). Further, a second order Taylor expansion around \(p=0\) yields \(q(p)=\lambda p/\alpha +O(p^2)\). Therefore, (20) can only be satisfied if

$$\begin{aligned} \frac{\lambda _1}{\alpha _1-1} \le \frac{\lambda _2}{\alpha _2-1}. \end{aligned}$$
(21)

Thus, (17) and (21) are necessary for \(X\le _{dil}Y\). We now show that they are also sufficient. Since \(cq_X(p)=q_{cX}(p)\), and \(\lambda \) is a scale parameter, (20) is equivalent to

$$\begin{aligned} q\left( p;\alpha _1,\frac{\alpha _1\lambda _1}{\alpha _1-1}\right)&\le q\left( p;\alpha _2,\frac{\alpha _2\lambda _2}{\alpha _2-1}\right) , \quad 0<p<1. \end{aligned}$$

This, in turn, is equivalent to

$$\begin{aligned} {\bar{F}}(p;\alpha _1,{\tilde{\lambda }}_1)&\le {\bar{F}}(p;\alpha _2,{\tilde{\lambda }}_2), \quad 0<p<1, \end{aligned}$$
(22)

where \({\tilde{\lambda }}_i=\alpha _i\lambda _i/(\alpha _i-1), i=1,2.\) The above results about the usual stochastic order yield that (22) is satisfied iff \(\alpha _1\ge \alpha _2\) and \({\tilde{\lambda }}_1/\alpha _1 \le {\tilde{\lambda }}_2/\alpha _2\), which coincide with (17) and (21).

Bellini et al. (2018a) have shown that these two conditions are also necessary and sufficient for the so-called increasing convex order (\(\le _{icx}\)). Hence, we have the following result:

$$\begin{aligned} X\le _{we-disp} Y \text { as well as } X\le _{icx} Y&\; \text {if and only if} (17) \;\hbox {and}\; (21) \;\hbox {hold}. \end{aligned}$$

Remark 13

  1. (a)

    Assume \(\alpha _1\ge \alpha _2>2\) and that (21) holds. Then, \(\alpha _1/(\alpha _1-2)\le \alpha _2/(\alpha _2-2)\), and \(Var(X)\le Var(Y)\) follows. Generally, the variance preserves the weak expectile dispersive order and therefore also the dilation order (see also the following section).

  2. (b)

    If \(\alpha _1\ge \alpha _2\) and \(\lambda _1/(\alpha _1-1)=\lambda _2/(\alpha _2-1)\), i.e. \(EX=EY\), then \(X\le _{icx} Y\) corresponds to \(X\le _{cx} Y\). According to the last section, \(X\le _{we-disp} Y\) holds as well in this case. On the other hand, if X and Y are random variables from different Lomax distributions, but with equal means, they can never be ordered in expectile order Bellini et al. (2018a). Since the dispersive order implies the stochastic order for distributions with the same finite left endpoint of their supports (Shaked and Shantikumar 2007, Theorem 3.B.13), the expectile dispersive order then also implies the expectile (location) order (by applying the cited result to the expectile cdf’s \(\breve{F}_X\) and \(\breve{F}_Y\)). Hence, X and Y can also not be ordered with respect to the expectile dispersive ordering. This example shows that \(\le _{we-disp}\) is strictly weaker than \(\le _{e-disp}\).

  3. (c)

    Let \(\alpha _1 = 3, \lambda _1 = \sqrt{3}, \alpha _2 = 2, \lambda _2 = 1\). Since \(\alpha _1 >\alpha _2\) and \(EX=\sqrt{3}/2 < EY=1\), one has \(X \le _{we-disp} Y\). On the other hand, (16) is not satisfied. Hence, \(X \le _{disp} Y\) does not hold. In this example, X also precedes Y in the expectile location order (Bellini et al. 2018a, Theorem 23). The left panel of Fig. 1 shows the interexpectile ranges of X and Y; clearly, \(e_X(1-p)-e_X(p)\le e_Y(1-p)-e_Y(p)\) for \(0<p<1\). The right panel shows a plot of \(e_Y(\breve{F}_X(x))-x\), which is increasing in x. This indicates that also \(X \le _{e-disp} Y\) holds [Shaked and Shantikumar 2007, (3.B.10)]. Hence, similarly to Example 12, this shows that \(X \le _{e-disp} Y\) does not imply \(X \le _{disp} Y\) in general.

Fig. 1
figure 1

Left panel: Interexpectile ranges of X (in black) and Y (in red) under Lomax distributions. Left panel: plot of \(e_Y(\breve{F}_X(x))-x\). (Color figure online)

6 Interexpectile ranges and their empirical counterparts

It is clear that any functional of the form \(E[\varphi (X-EX)]\), where \(\varphi \) is a convex function, preserves the dilation order. Then, by Theorem 9, this also holds for the weak expectile dispersive order. Examples are the standard deviation \(\sigma (F)=\{E(X-EX)^2\}^{1/2}\) and the mean absolute deviation around the mean \(E|X-EX|\). In this section, however, we want to have a closer look on the interexpectile range (IER) \({\mathcal {E}}_{\alpha }=e_X(1-\alpha )-e_X(\alpha ), \alpha \in (0,1/2)\). These scale measures obviously preserve the \(\le _{we-disp}\) ordering, and, hence, also the dilation order; they have already appeared as a scaling factor in the definition of \(s_2\). Moreover, (implicit) interexpectile differences have been used to extract information about the risk-neutral distribution of a financial index by Bellini et al. (2018b, 2020).

Using the properties of expectiles (see, e.g., Bellini et al. 2014), \({\mathcal {E}}_{\alpha }\) obviously satisfies property D1. in the introduction. Hence, it is a measure of variability with respect to the dispersive order, the most fundamental variability ordering, but also with respect to the dilation order. The latter is an order with respect to the mean, whereas the first order is location-free. Further properties of the theoretical IER can be found in Bellini et al. (2018b), Proposition 3.1.

The results in Sect. 3 show that the MAD arises quite naturally when dealing with expectile-based skewness orders. Our next result bounds \({\mathcal {E}}_{\alpha }\) in terms of the MAD.

Theorem 14

For \(\alpha \in (0,1/2)\),

$$\begin{aligned} \frac{1-2\alpha }{1-\alpha }E|X-\mu |&< {\mathcal {E}}_{\alpha } \, < \frac{1-2\alpha }{\alpha } E|X-\mu |. \end{aligned}$$

In particular, \(\frac{2}{3} E|X-\mu |< {\mathcal {E}}_{1/4} < 2 E|X-\mu |\).

Proof

For any \(\tau \in (0,1)\setminus \{1/2\}\), the first order condition (3) can be rewritten as

$$\begin{aligned} E(X - e_X(\tau ))_+&= \frac{1-\tau }{1-2\tau }(\mu -e_X(\tau )), \end{aligned}$$
(23)
$$\begin{aligned} E(X - e_X(\tau ))_-&= \frac{\tau }{1-2\tau }(\mu -e_X(\tau )). \end{aligned}$$
(24)

Since \(E(X - t)_+\) is strictly decreasing and \(E(X - t)_-\) is strictly increasing, and since \(e_X(\alpha )<\mu <e_X(1-\alpha )\), we further obtain

$$\begin{aligned} E(X-e_X(1-\alpha ))_+&< E(X-\mu )_+ < E(X-e_X(\alpha ))_+, \end{aligned}$$
(25)
$$\begin{aligned} E(X-e_X(\alpha ))_-&< E(X-\mu )_- < E(X - e_X(1 - \alpha ))_-. \end{aligned}$$
(26)

Adding the terms in (25) and (26), and using Eqs. (23) and (24) then yields

$$\begin{aligned} \frac{\alpha }{1-2\alpha } \left( e_X(1-\alpha )-e_X(\alpha )\right)&< E|X-\mu |\, < \frac{1-\alpha }{1-2\alpha } \left( e_X(1-\alpha )-e_X(\alpha )\right) , \end{aligned}$$

or

$$\begin{aligned} \frac{1-2\alpha }{1-\alpha }E|X-\mu |&< e_X(1-\alpha )-e_X(\alpha ) \, < \frac{1-2\alpha }{\alpha } E|X-\mu |. \end{aligned}$$

\(\square \)

For any \(\alpha \in (0,1/2)\), we define the population counterparts of \({\mathcal {E}}_{\alpha }\) by \(\widehat{{\mathcal {E}}}_{\alpha ,n} ={\hat{e}}_n(1-\alpha )-{\hat{e}}_n(\alpha )\). For empirical expectiles, a multivariate central limit theorem as well as strong consistency holds (see, e.g., Holzmann and Klar 2016); from these results, a central limit theorem and strong consistency of \(\widehat{{\mathcal {E}}}_{\alpha ,n}\) can be derived. Using the notations \(\eta (\tau _1,\tau _2)= E[I_{\tau _1}(e_X(\tau _1),X) I_{\tau _2}(e_X(\tau _2),X)]\) for \(\tau _1,\tau _2 \in (0,1)\) and \(a_\alpha =\alpha +(1-2\alpha )F(e_X(\alpha ))\), the following holds true.

Theorem 15

Let \(F_X\) be a cdf with \(E|X|<\infty \), and \(\alpha \in (0, 1/2)\).

  1. (a)

    If \(E X^2<\infty \) and \(F_X\) does not have a point mass at \(e_X(\alpha )\) or \(e_X(1-\alpha )\), then

    $$\begin{aligned} \sqrt{n} \left( \widehat{{\mathcal {E}}}_{\alpha ,n}-{\mathcal {E}}_\alpha \right) {\mathop {\longrightarrow }\limits ^{{\mathcal {D}}}}{\mathcal {N}}(0, \sigma ^2_{\mathcal {E}}(\alpha )), \end{aligned}$$

    where

    $$\begin{aligned} \sigma ^2_{\mathcal {E}}(\alpha )&=\, \frac{\eta (\alpha ,\alpha )}{a_\alpha ^2} - \frac{2\eta (\alpha ,1-\alpha )}{a_\alpha \cdot a_{1-\alpha }} + \frac{\eta (1-\alpha ,1-\alpha )}{a_{1-\alpha }^2}. \end{aligned}$$
  2. (b)

    \(\widehat{{\mathcal {E}}}_{\alpha ,n}\) is a strongly consistent estimator of \({\mathcal {E}}_\alpha \), i.e.

    $$\begin{aligned} \widehat{{\mathcal {E}}}_{\alpha ,n} {\mathop {\longrightarrow }\limits ^{a.s.}}{\mathcal {E}}_\alpha . \end{aligned}$$

A natural competitor to \({\mathcal {E}}_\alpha \) is the interquantile range (IQR) \({\mathcal {Q}}_{\alpha } = q_X(1-\alpha )-q_X(\alpha ), \alpha \in (0,1/2)\), where \(q_X(\alpha )\) denotes the \(\alpha \)-quantile of the distribution of X. By definition, the interquantile range preserves the dispersive order. However, it is not consistent with the dilation order, which may be seen as a disadvantage in specific applications (Bellini et al. 2020).

A general comparison result between the IQR and IER is not possible: \({\mathcal {Q}}_{\alpha }\) may be smaller than \({\mathcal {E}}_\alpha \) for some \(\alpha \), and larger for other ones. However, such a comparison is possible for symmetric log-concave distributions such as the normal, logistic, uniform or Laplace distribution. This follows directly from Corollary 7 and the preceding results in Arab et al. (2022):

Proposition 16

Let \(F_X\) be a symmetric cdf with finite mean and log-concave density. Then,

$$\begin{aligned} {\mathcal {E}}_\alpha&< {\mathcal {Q}}_{\alpha } \quad \text {for each } \alpha \in (0, 1/2). \end{aligned}$$

In the following, we compare the efficiency of the empirical IER as an estimator of the variability for specific distributions with other measures of dispersion, in particular with the IQR. Writing \(\widehat{{\mathcal {Q}}}_{\alpha ,n} ={\hat{q}}_n(1-\alpha )-{\hat{q}}_n(\alpha )\), where \({\hat{q}}_n(\alpha )\) is the sample quantile, one has

$$\begin{aligned} \sqrt{n}(\widehat{{\mathcal {Q}}}_{\alpha ,n}-{\mathcal {Q}}_\alpha ) {\mathop {\longrightarrow }\limits ^{{\mathcal {D}}}}{\mathcal {N}}\left( 0,\sigma ^2_{\mathcal {Q}}(\alpha )\right) , \end{aligned}$$

where

$$\begin{aligned} \sigma ^2_{\mathcal {Q}}(\alpha ) = \frac{\alpha (1-\alpha )}{f^2(q_X(1-\alpha ))} - \frac{2\alpha ^2}{f(q_X(\alpha ))f(q_X(1-\alpha ))} +\frac{\alpha (1-\alpha )}{f^2(q_X(\alpha ))}, \end{aligned}$$

and f denotes the density of \(F_X\).

If \(X\sim {\mathcal {N}}(\mu ,\sigma ^2)\), we obtain \(\sigma ^2_{\mathcal {Q}}(\alpha ) = 2\alpha (1-2\alpha )/f^2(q_X(\alpha ))\), and \({\mathcal {Q}}(\alpha )=2\sigma \Phi ^{-1}(1-\alpha )\). Therefore,

$$\begin{aligned} \sqrt{n}\left( \frac{\widehat{{\mathcal {Q}}}_{\alpha ,n}}{2\Phi ^{-1}(1-\alpha )} - \sigma \right) {\mathop {\longrightarrow }\limits ^{{\mathcal {D}}}}{\mathcal {N}}\left( 0,\tau ^2_{\mathcal {Q}}(\alpha )\sigma ^2\right) , \end{aligned}$$

where

$$\begin{aligned} \tau ^2_{\mathcal {Q}}(\alpha ) = \frac{\alpha (1-2\alpha )}{2\left( \Phi ^{-1}(1-\alpha )\right) ^2 \, \varphi ^2\left( \Phi ^{-1}(1-\alpha )\right) }. \end{aligned}$$

On the other hand, the sample standard deviation \({\widehat{\sigma }}_n=(1/(n-1)\sum _{i=1}^n(X_i-{\bar{X}}_n)^2)^{1/2}\) has asymptotic variance \((\mu _4-\sigma ^4)/(4\sigma ^2)\), where \(\mu _4=E(X-\mu )^4\), which simplifies to \(\sigma ^2/2\) under normality. Hence, the standardized asymptotic relative efficiency (ARE) is given by

$$\begin{aligned} sARE(\widehat{{\mathcal {Q}}}_{\alpha ,n},{\widehat{\sigma }}_n) = \frac{1/2}{\tau ^2_{\mathcal {Q}}(\alpha )}. \end{aligned}$$

We term \(\tau ^2_{\mathcal {Q}}(\alpha )\) the standardized asymptotic variance (standardized ASV). Proceeding in the same way with the IER leads to the corresponding quantity \(\tau ^2_{\mathcal {E}}(\alpha )=\sigma ^2_{\mathcal {E}}(\alpha )/(2e_X(1-\alpha ))^2\). Figure 2 shows the standardized ASV’s \(\tau ^2_{\mathcal {Q}}(\alpha )\) and \(\tau ^2_{\mathcal {E}}(\alpha )\) as functions of \(\alpha \).

Fig. 2
figure 2

Standardized asymptotic variances \(\tau ^2_{\mathcal {Q}}(\alpha )\) (in black) and \(\tau ^2_{\mathcal {E}}(\alpha )\) (in red) under normality. (Color figure online)

For the interquartile range, i.e. the choice \(\alpha =1/4\), one obtains the well known result \(sARE(\widehat{{\mathcal {Q}}}_{1/4,n},{\widehat{\sigma }}_n)=0.368\); the standardized ARE takes the maximal value 0.652 for \(\alpha =0.0692\) (Fisher 1920; David 1998). Whereas the latter estimator is more efficient, an advantage of using more central quantiles such as quartiles is their greater stability.

Proceeding to the IER, Fig. 2 shows that the standardized ASV’s are generally smaller and quite stable over a large range of \(\alpha \)-values. For \(\alpha =1/4\), we obtain \(sARE(\widehat{{\mathcal {E}}}_{1/4,n},{\widehat{\sigma }}_n)=0.934\); the standardized ARE takes the maximal value 0.967 for \(\alpha =0.118\). Hence, the IER is a quite efficient scale estimator under normality compared to the standard deviation.

To analyze the behaviour of the IER under distributions with longer tails than the normal, we consider in the following Student’s t-distribution with \(\nu \) degrees of freedom, denoted by \(t_{\nu }\), and having density function \(c_{\nu }(1+x^2/\nu )^{-(\nu +1)/2}, x\in {\mathbb {R}}\), for \(\nu >0\), where \(c_{\nu }=\Gamma ((\nu +1)/2)/(\sqrt{\nu \pi }\Gamma (\nu /2))\).

Additionally to the dispersion measures used so far, we consider the MAD \(d=E|X-EX|\), estimated by \({\widehat{d}}_n=1/n \sum _{i=1}^n|X_i-{\bar{X}}_n|\), and Gini’s mean difference \(g=E|X-Y|\), where Y is a independent copy of X. The usual estimator of g is the sample mean difference \({\widehat{g}}_n=\frac{2}{n(n-1)}\sum _{i<j}|X_i-X_j|\). The asymptotic variance of \({\widehat{d}}_n\) is given by

$$\begin{aligned} ASV({\widehat{d}}_n)&= Var\left( |X-\mu |+ (2F(\mu )-1)X\right) \end{aligned}$$

(Van der Vaart 1998, Example 19.25), which simplifies to \(ASV({\widehat{d}}_n)=\sigma ^2-d^2\) for symmetric distributions. In this case, the ASV is the same as for the sample MAD around the median \(1/n\sum _{i=1}^n|X_i-q_X(1/2)|\) (Gerstenberger and Vogel 2015). An explicit expression for the ASV of \({\widehat{g}}_n\) under the t-distribution can be found in Gerstenberger and Vogel (2015), Table 3.

Similarly as above, given a scale measure \(\delta \) and its estimator \({\hat{\delta }}\), we call the ratio \(ASV({\hat{\delta }})/\delta ^2\) standardized asymptotic variance. Table 1 shows the standardized ASV of the different estimators for the t-distribution with various degrees of freedom.

Table 1 Standardized ASV of different scale estimators for the \(t_{\nu }\)-distribution

Concerning the comparison between the interquantile and interexpectile range, we observe a similar behavior as for the normal distribution. Whereas the standardized ASV’s vary strongly with \(\alpha \) for the IQR, they are quite stable for the IER. To allow for a better comparison of the relative efficiencies, Table 2 shows the minimum of the standardized ASV’s in each line of Table 1, divided by the standardized ASV’s. Hence, for each distribution, the table shows the sARE with respect to the most efficient estimator. Whereas \(\widehat{{\mathcal {Q}}}_{0.15}\) and \({\widehat{d}}\) are most efficient for \(\nu =3\) and \(\nu =4,5\), respectively, the IER gets in the lead for degrees of freedom between 6 and 10. For higher degrees of freedom, Gini’s mean difference and the standard deviation are most efficient.

Table 2 sARE of different scale estimators with respect to the most efficient estimator for the \(t_{\nu }\)-distribution

As a final example, we use the normal inverse Gaussian (NIG) distribution (Barndorff-Nielsen 1997), which allows for skewed and heavy-tailed distributions; further, all moments exist and have simple explicit expressions. Table 3 shows the standardized ASV’s of the different estimators for the NIG distribution with shape parameters \(\alpha \) and \(\beta \). For \(\beta =0\), the distribution is symmetric. The two remaining parameters are chosen such that \(EX=0\) and \(Var(X)=1\); hence, the third and fourth moment in columns 3–4 corresponds to the moment skewness and kurtosis. Note that, for \(\beta =0\) and \(\alpha \rightarrow \infty \), the distribution converges to the standard normal.

Table 3 Standardized ASV of different scale estimators for the NIG-distribution

From Table 3 we see that \(\widehat{{\mathcal {E}}}_{0.25}, {\widehat{d}}\) and \({\widehat{g}}\) have comparable standardized ASV’s. They are all quite efficient compared to \({\widehat{\sigma }}\) for distributions near to the normal. They are considerably more efficient than \({\widehat{\sigma }}\) for skewed and long-tailed distributions as the NIG distribution with \(\alpha =1, \beta =0.5\), and compare well with the IQR in this case. All in all, these measures seem to be a reasonable compromise between the standard deviation on the one hand, and the interquartile range on the other.