Exercises

1.Academia das Ciências de Lisboa (Lisbon Academy of Sciences), Lisbon, Portugal.
1.Academia das Ciências de Lisboa (Lisbon Academy of Sciences), Lisbon, Portugal.
3.1 Obtain the inequality \(\mathrm{ \frac{max ( 1,e^{w} ) }{1+e^{w}} \leq k ( w ) \leq 1 }\) or \(\mathrm{max \left( u,1-u \right) \leq \Lambda \left( u \right) \leq 1 }\) from the Boole-Frechet inequality and a passage to the limit.
3.2 Show that the sets \(\mathrm{ \{ k \left( w \right) \} }\) or \(\mathrm{ \{ A \left( u \right) \} }\) are convex, closed and symmetrical (corresponding to exchangeability \(\mathrm{ k \left( w \right) =k \left( -w \right) ~or~A \left( u \right) =A \left( 1-u \right) ) }\).
3.3 Obtain the conditions that \(\mathrm{ k \left( w \right) }\) or \(\mathrm{ A \left( u \right) }\) must satisfy in the absolutely continuous case.
3.4 The equidistribution median curve for the case with reduced Gumbel margins is between the curves \(\mathrm{ e^{-x}+e^{-y}=log~2 }\) and \(\mathrm{ max(e^{-x},e^{-y})=log~2 }\); are these curves convex?: study the corresponding situation for the equisurvival median curve (case of \(\mathrm{ A \left( u \right) }\) ).
3.5 For bivariate distributions attracted for maxima to \(\mathrm{ \Lambda \left( x,y \right) }\), show that if there is negative dependence (i.e., \(\mathrm{ F \left( x,y \right) \leq F \left( x,+ \infty \right) F \left( + \infty,y \right)}\) ) then \(\mathrm{ \Lambda \left( x,y \right) = \Lambda \left( x \right) \Lambda \left( y \right) }\) (asymptotic attraction to independence).
3.6 We can expect that for bivariate samples \(\mathrm{ \begin{array}{c} { } \mathrm{n } \\ \mathrm{ (max} \\ \mathrm{1} \end{array}X_{i}, \begin{array}{c} { } \mathrm{n } \\ \mathrm{ max } \\ \mathrm{1} \end{array}Y_{i}) }\) are asymptotically independent. Obtain conditions for this, but show that in the singular case where \(\mathrm{ Prob\{ X+Y=a \} =1 }\) this is not the case. Recall that \(\mathrm{ \begin{array}{c} { } \mathrm{n } \\ \mathrm{ min } \\ \mathrm{1} \end{array} Y_{i} = \begin{array}{c} { } \mathrm{n } \\ \mathrm{ max } \\ \mathrm{1} \end{array}(-Y_{i} ) }\), and use Sibuya's necessary and sufficient condition or Geffroy's sufficient condition for independence in maxima.
3.7 In the characterization of \(\mathrm{ A \left( u \right) }\), show that if \(\mathrm{ A \left( 0 \right) =A \left( 1 \right) =~1 }\) and \(\mathrm{ A \left( u \right) }\) convex in \(\mathrm{ [0,1] }\) the conditions
a) \(\mathrm{ 0 \leq A^{'} ( 0 ) ,A'(1) \leq 1 }\) ,
b) \(\mathrm{ max ( u,1-u ) \leq A ( u ) \leq 1 }\) and
c) \(\mathrm{ A( u ) / ( 1-u ) }\) non-decreasing and \(\mathrm{ A( u ) / u }\) non-increasing are equivalent. Show also that \(\mathrm{ A^{'} ( 0 ) \leq A^{'} ( u ) \leq A^{'} ( 1 ) ,A( u ) \geq 1+A' ( 0 ) u }\), and \(\mathrm{ A ( u ) \geq 1-A' ( 1 ) ( 1-u ) }\).
3.8 Obtain the relation between the function \(\mathrm{ P \left( u,v \right) }\) defined by Sibuya and the structure function \(\mathrm{ \bar{S} \left( \xi , \eta \right) }\). Convert the results of Sibuya on \(\mathrm{ P \left( u,v \right) }\) concerning asymptotic independence and asymptotic complete dependence (diagonal case) to \(\mathrm{ \bar{S} \left( \xi , \eta \right) }\).
3.9 The condition of \(\mathrm{ k" ( w ) }\), for the bivariate distribution function \(\mathrm{ \Lambda \left( x,y \right) }\), implies that \(\mathrm{ \varphi \left( w \right) = \left( 1+e^{w} \right) k \left( w \right) }\) and \(\mathrm{ \Psi \left( w \right) = \left( 1+e^{-w} \right) k \left( w \right) }\) are such that \(\mathrm{ \varphi '' ( w ) \geq \varphi' ( w ) }\) and so \(\mathrm{ \varphi' \left( w \right) \geq 0 }\) and \(\mathrm{ \Psi " ( w ) + \Psi ' ( w ) \geq 0 }\) and so \(\mathrm{ \Psi '( w) \leq 0 }\), which eliminates the two monotonicity conditions.
3.10 Show that if \(\mathrm{ \left( X,Y \right) }\) has the distribution function \(\mathrm{ \Lambda \left( x,y \right) }\) then the pair \(\mathrm{ \left( V,W\right) }\) with \(\mathrm{ e^{-V}=e^{-X}+e^{-Y},W=Y-X }\) has the distribution function
\(\mathrm{ Prob \{ V \leq v,W \leq w\}= \Lambda \left( v \right) +( \frac{k ' \left( w \right) }{k \left( w \right) }+\frac{1-e^{-w}}{1+e^{-w}} ) \Lambda ( v-log\,k \left( w \right) ) }\)
\(\mathrm{ + \int _{- \infty}^{w}\frac{e^{- \rho }}{ \left( 1-e^{- \rho } \right) ^{2}} ( e^{-v}k \left( \rho \right) -1 ) \Lambda ( v-log\,k \left( p \right) ) dp }\).
3.11 For the distribution function \(\mathrm{ D \left( w \right) }\), show that \(\mathrm{ D^{a} \left( w+b \right) }\), if \(\mathrm{ a \geq 1 }\), is also a \(\mathrm{ D \left( w \right) }\) - function for convenient \(\mathrm{ b }\).
3.12 Show that if \(\mathrm{ \left( X,Y \right) }\) is a bivariate pair with reduced Gumbel margins, and so distribution function \(\mathrm{ \Lambda \left( x,y \right) }\) then \(\mathrm{ Prob \{ X \leq x,Y-X \leq w \} =D \left( w \right) \Lambda \left( x,x+w \right)} \).
3.13 Translate maxima results to minima results and vice-versa; use, in particular, the relation between the distribution function and the survival function.
3.14 Show that \(\mathrm{ \Lambda ( w+\gamma) }\) and \(\mathrm{ \frac{w+ \theta }{2~ \theta } }\) for \(\mathrm{ 0 \leq \theta \leq 2, \vert w \vert < \theta }\) are \(\mathrm{ D \left( w \right) }\) - functions.
3.15 Suppose that \(\mathrm{ \left( X,Y \right) }\) is an independent random pair with Gumbel margins whose parameters are \(\mathrm{ \left( \lambda _{X}, \delta \right) }\) and \(\mathrm{ \left( \lambda _{Y}, \delta \right) }\) (the same dispersion parameter). Compute \(\mathrm{ Prob\{ Y> X \} }\). Solve the corresponding question for minima with exponential margins and compare the results.
3.16 Compute \(\mathrm{ k(.) }\) for the use of max-technique; what is the corresponding technique if \(\mathrm{ S \left( x,y \right) }\) is used? Interpret the statistical implications for small samples of the fact that the index of dependence is \(\mathrm{ \leq 1/4 }\); what does this imply for the discrimination between models?
3.17 Show that for general bivariate distribution of maxima with reduced Gumbel margins \(\mathrm{ Prob \{ X \leq Y \} =\frac{1}{2}-\frac{k' \left( 0 \right) }{k \left( 0 \right) }\,so \vert \frac{k' \left( 0 \right) }{k \left( 0 \right) } \vert\leq 1/2 }\) and that for minima with standard exponential margins we have, correspondingly, \(\mathrm{ Prob \{ X^{'} \leq Y^{'}\}=\frac{1}{2}-\frac{A' \left( 1/2 \right) }{4~A \left( 1/2 \right) } }\)and so \(\mathrm{ \vert \frac{A' \left( 1/2 \right) }{4~A \left( 1/2 \right) } \vert \leq 2 }\).
3.18 Let \(\mathrm{ G_{1} \left( x \right) }\) and \(\mathrm{ G_{2} \left( y \right) }\) be distribution function such that \(\mathrm{ G_{1}^{k} \left( \lambda _{k}+ \delta _{k}x \right) \rightarrow \Lambda \left( x \right) }\) and \(\mathrm{ G_{1}^{k} \left( \lambda _{k}^{'}+ \delta _{k}^{'}~y \right) \rightarrow \Lambda \left( y \right) }\). As stated, we know by Fréchet inequalities, that any bivariate distribution function \(\mathrm{ F \left( x,y \right) }\) such that \(\mathrm{ F \left( x,+ \infty \right) =G_{1} \left( x \right) ,F \left( + \infty,y \right) =G_{2} \left( y \right) }\)satisfies the inequalities \(\mathrm{ \underline{D} \left( x,y \right) =max \left( 0,G_{1} \left( x \right) +G_{2} \left( y \right) -1 \right) F \left( x,y \right) \leq min \left( G_{1} \left( x \right) ,G_{2} \left( y \right) \right) =\bar{D} \left( x,y \right) }\). Show that the family of distribution functions \(\mathrm{ \{ \theta\, \underline{D} \left( x,y \right) + \left( 1- \theta \right) \bar{D} \left( x,y \right) \} }\) is attracted, with the same attraction coefficients for the margins, to \(\mathrm{ exp \{ - [ \theta \left( e^{-x}+e^{-y} \right) + \left( 1- \theta \right) max \left( e^{-x},e^{-y} \right) ]\} }\), which is the Gumbel model.
3.19 Show that the bivariate model of maxima with Weibull margins and shape parameter \(\mathrm{ \alpha =1 \left( \xi =-e^{-x}, \eta =-e^{-y} \right) \,is~ \Psi _{1} \left( \xi , \eta \right) =exp \{ \left( \xi + \eta \right) k ( exp\frac{ \xi }{ \eta } ) \} }\)and the corresponding logistic is \(\mathrm{ \Psi _{1L} \left( \xi , \eta \right) =exp \{ ( \left( - \xi \right) ^{{1}/{ \left( 1- \theta \right) }}+ \left( - \eta \right) ^{{1}/{ \left( 1- \theta \right) }} ) ^{1- \theta } \} }\).
3.20 Show that if we have a reduced sample \(\mathrm{ \{ \left( x_{i},y_{i} \right) ,i=1,2,\dots,n \} }\), and we expect to observe \(\mathrm{ x }\), the MSE linear predictor of \(\mathrm{ y \left( y^{*}= \alpha + \beta ~x+ \sum _{1}^{n} \,\varphi _{i}\,x_{i}+\sum _{1}^{n} \Psi _{i}~y_{i} \right) }\)is given by the usual linear regression \(\mathrm{ L_{y} \left( x \right) = \gamma +\rho \left( x-\gamma \right) }\), the sample having an intervention only through the computation of \(\mathrm{ { \rho}^{*}=r }\), as usual.
3.21 For bivariate extremes with reduced Gumbel margins, the linear regression line of \(\mathrm{ y }\) in \(\mathrm{ x }\) is \(\mathrm{ L_{y} \left( x \right) = \gamma +\rho \left( x-\gamma \right) }\). Show that all regression lines are in the area: if \(\mathrm{ x \leq \gamma }\) then \(\mathrm{ x \leq L_{y} \left( x \right) \leq \gamma }\) and if \(\mathrm{ x \geq \gamma }\) then \(\mathrm{ \gamma \leq L_{y} \left( x \right) \leq x }\).
3.22 Show that, as all regression lines are in the area described above, statistical choice of bivariate model through linear regression is difficult even for moderate sample sizes.
3.23 Draw the graphs of \(\mathrm{ A ( u \vert \theta) }\) for different models (and values of \(\mathrm{ \theta =0.,.5 ~and ~1.0 }\) ). They are all in in the triangle bounded by \(\mathrm{ max \left( u,1-u \right) \leq A \left( u \vert \theta \right) \leq 1 }\). If they are not close to independence or to the diagonal case it is difficult to distinguish them.
3.24 Using the explicit expression of the correlation coefficients, show that:
for the logistic model: \(\mathrm{ \rho\left( \theta \right) = \theta ( 2- \theta ),v \left( \theta \right) =2^{-2^{1- \theta }} -1}\);
for th e emixed model: \(\mathrm{ \rho \left( \theta \right) =\frac{6}{ \pi ^{2}} \left( arccos \left( 1- \theta ∕2 \right) \right) ^{2}=\frac{24}{ \pi ^{2}} ( arctg\sqrt[]{\frac{ \theta }{4- \theta }} ) ^{2},v \left( \theta \right) =2^{ \theta /2}-1 }\);
for the Gumbel model : \(\mathrm{ \rho \left( \theta \right) \frac{12}{ \pi ^{2}}= \int _{0}^{ \theta }\frac{log \left( 2-t \right) }{1-t}dt,v \left( \theta \right) =2^{ \theta } -1 }\);
for the biextremal model : \(\mathrm{ \rho \left( \theta \right) =-\frac{6}{ \pi ^{2}} \int _{0}^{ \theta }\frac{log\,t}{1-t}dt,v ( \theta ) ={2}^{ \theta }-1 }\).
3.25 Analyse the data of Table 1, contained in Gumbel and Goldstein (1964):
Year |
Men |
Women |
Year |
Men |
Women |
1905 |
100.88 |
102.54 |
1932 |
102.55 |
104.87 |
1906 |
101.17 |
106.13 |
1933 |
103.17 |
105.98 |
1907 |
104.65 |
103.46 |
1934 |
103.98 |
103.38 |
1908 |
105.12 |
102.12 |
1935 |
106.09 |
105.32 |
1909 |
102.57 |
101.69 |
1936 |
103.43 |
103.77 |
1910 |
101.70 |
102.92 |
1937 |
105.72 |
105.86 |
1911 |
100.49 |
102.78 |
1938 |
103.24 |
104.27 |
1912 |
100.90 |
106.15 |
1939 |
103.25 |
105.45 |
1913 |
103.06 |
105.02 |
1940 |
103.40 |
105.71 |
1914 |
102.63 |
103.56 |
1941 |
101.66 |
106.15 |
1915 |
102.69 |
106.52 |
1942 |
106.48 |
104.71 |
1916 |
100.82 |
101.50 |
1943 |
101.26 |
103.83 |
1917 |
102.52 |
104.01 |
1944 |
105.12 |
105.19 |
1918 |
100.08 |
105.01 |
1945 |
104.88 |
105.03 |
1919 |
101.67 |
104.52 |
1946 |
102.41 |
105.88 |
1920 |
101.41 |
103.94 |
1947 |
104.22 |
107.49 |
1921 |
101.76 |
103.14 |
1948 |
102.88 |
105.83 |
1922 |
102.57 |
104.33 |
1949 |
103.57 |
103.41 |
1923 |
101.63 |
102.32 |
1950 |
105.12 |
105.64 |
1924 |
103.47 |
103.56 |
1951 |
103.80 |
103.53 |
1925 |
105.48 |
103.86 |
1952 |
102.94 |
107.90 |
1926 |
104.01 |
105.87 |
1953 |
103.00 |
104.42 |
1927 |
105.83 |
103.31 |
1954 |
106.50 |
104.85 |
1928 |
105.00 |
104.37 |
1955 |
103.36 |
103.97 |
1929 |
102.78 |
102.72 |
1956 |
103.15 |
107.89 |
1930 |
102.61 |
105.01 |
1957 |
102.54 |
104.46 |
1931 |
105.55 |
104.40 |
1958 |
104.92 |
104.12 |
Test the fit of the Gumbel distribution to each set of oldest ages of death (men, women) and, after that, test for (the expected) independence.
Add to Table 1 the data of Table 2, given in Fransén and Tiago de Oliveira (1984).
Year |
Men |
Women |
Year |
Men |
Women |
1959 |
104.23 |
104.77 |
1965 |
105.28 |
104.31 |
1960 |
103.59 |
106.13 |
1966 |
104.93 |
104.98 |
1961 |
103.74 |
107.10 |
1967 |
105.27 |
105.83 |
1962 |
103.00 |
104.56 |
1968 |
105.92 |
106.35 |
1963 |
104.25 |
110.07 |
1969 |
101.81 |
107.58 |
1964 |
104.12 |
106.15 |
1970 |
104.02 |
105.42 |
Do the same tests as before and compare the estimators. Also compare the estimators of the margins for Table 1and Tables 1 and 2. Use the previous result to estimate the probability that women have greater ages of death than men (take \(\mathrm{ \delta ^{*} }\) as the average of the \(\mathrm{ \delta' s }\) for Tables 1 and 2) and test if women “live longer” than men.
3.26 Consider any river at two points with a sample of \(\mathrm{ k \geq 50 }\) yearly floods. Check if the margins are Weibull, Gumbel or Frechet distributed. If they are not Gumbel distributed, make the necessary transformations to be so, and then:
3.27 Obtain subroutines for statistical choice of bivariate maxima models with Gumbel margins or for bivariate minima with exponential margins.
3.28 Show that the “estimated” best linear predictor \(\mathrm{ \frac{y^{*}- \hat{\lambda} _{y}}{ \hat{\delta} _{y}}= \gamma +r\cdot ( \frac{x^{*}- \hat{\lambda} _{x}}{ \hat{\delta }_{x}}- \gamma ) }\)is to be very be expected very close to \(\mathrm{ \frac{y^{*}-\bar{y}}{s_{y}}=r\cdot ( \frac{x-\bar{x}}{s_{x}} ) }\)as
\(\mathrm{ \bar{ x}- \hat{\lambda} _{x}- \gamma \,\hat{ \delta} _{x}\approx0,\bar{ y}- \hat{\lambda} _{y}- \gamma \,\hat{ \delta} _{y}\approx0 }\)and \(\mathrm{ {\hat{ \delta} _{y}}/{ \hat{\delta} _{x}}\approx s_{y}/s_{x} }\).
3.29 Study the estimator \(\mathrm{ \tilde{~ \theta} }\) of \(\mathrm{ ~ \theta }\) in the biextremal case with Gumbel margins (moments, convergence, asymptotic distribution, etc.) .
3.30 Show that from the sample distribution function of the \(\mathrm{ w_{i,}D^{*} \left( w \right) =\frac{1}{n} \sum _{1}^{n}H \left( w-w_{i}\right) \stackrel {p}\rightarrow D(w) }\)we can obtain estimators \(\mathrm{ {k}^*(w) }\) of \(\mathrm{ {k}(w) }\) and \(\mathrm{ {D}^*(w) }\) of \(\mathrm{ {D}(w) }\), both being non-parametric but not intrinsic, which is a more restrictive condition. It is easy to prove this by showing that \(\mathrm{ {A} (u) }\) is non-convex (using exponential margins).
3.31 Consider the trivariate model with Gumbel margins\(\mathrm{ \Lambda \left( x_{1},x_{2},x_{3} \right) =exp \{ - \left( e^{-x_1}+e^{-x2}+e^{-x_3} \right) +\theta _{3}\,min \left( e^{-x_1},e^{-x_2} \right) + \theta _{2}\,min \left( e^{-x_1},e^{-x_3} \right) + }\)\(\mathrm{ \theta _{1}\,min ( e^{-x_2},e^{-x_3} ) - \tau\,min ( e^{-x_1},e^{-x_2},e^{-x_3} ) \} }\).
Using the inequalities for the margins, show that
\(\mathrm{ 0 \leq \theta _{1}, \theta _{2}, \theta _{3} \leq 1 }\) and \(\mathrm{ max \left( 0, \theta _{1}+ \theta _{2}+ \theta _{3}-1 \right) \leq \tau \leq {( \theta _{1}+ \theta _{2}+ \theta _{3}})/{2} }\)and so \(\mathrm{ \theta _{1}+ \theta _{2}+ \theta _{3} \leq 2 }\).