key: cord-0477843-cq8g50ek authors: McNeil, Alexander J.; Neslehova, Johanna G.; Smith, Andrew D. title: On attainability of Kendall's tau matrices and concordance signatures date: 2020-09-17 journal: nan DOI: nan sha: ff2b8b1665ca68725ae979736a83373191302290 doc_id: 477843 cord_uid: cq8g50ek Methods are developed for checking and completing systems of bivariate and multivariate Kendall's tau concordance measures in applications where only partial information about dependencies between variables is available. The concept of a concordance signature of a multivariate continuous distribution is introduced; this is the vector of concordance probabilities for margins of all orders. It is shown that every attainable concordance signature is equal to the concordance signature of a unique mixture of the extremal copulas, that is the copulas with extremal correlation matrices consisting exclusively of 1's and -1's. A method of estimating an attainable concordance signature from data is derived and shown to correspond to using standard estimates of Kendall's tau in the absence of ties. The set of attainable Kendall rank correlation matrices of multivariate continuous distributions is proved to be identical to the set of convex combinations of extremal correlation matrices, a set known as the cut polytope. A methodology for testing the attainability of concordance signatures using linear optimization and convex analysis is provided. The elliptical copulas are shown to yield a strict subset of the attainable concordance signatures as well as a strict subset of the attainable Kendall rank correlation matrices; the Student t copula is seen to converge, as the degrees of freedom tend to zero, to a mixture of extremal copulas sharing its concordance signature with all elliptical distributions that have the same correlation matrix. A characterization of the attainable signatures of equiconcordant copulas is given. In many real-world applications of statistics a modeler is required to impute missing information on the dependencies between variables, typically in the form of correlations. This problem is particularly common in finance and insurance where data on certain risks are often sparse or non-existent. Some financial institutions use copulas parameterized in part by expert-elicited correlations to build joint models of key risks affecting their solvency and profitability (Embrechts et al., 2002; Shaw et al., 2011) . For example, a large insurer analyzing excess risk due to the Covid-19 pandemic might consider the interplay between mortality risk, business interruption risk, financial investment risk and trade credit insurance risk. While dependencies between these risks may be significant and non-negligible, they are also difficult to quantify. In the related area of asset management a model for the dependencies between asset returns is essential for optimizing a portfolio. However many assets have no track record and plausible values must be entered if they are to be included in the analysis. Assessing dependence in the absence of data is also relevant in causal inference when unmeasured confounders are present (Stokes et al., 2020) . Imputing missing information on dependence is a challenging problem because of the complex relationships between different pairs or subgroups of variables. The general problem of determining the compatibility of lower-dimensional margins of higher-dimensional distributions has only been partially resolved (see Joe, 1996 Joe, , 1997 . In this paper, we investigate the related problem of compatibility of correlation measures for subgroups of variables. Even for the classical linear correlation of Pearson, the set of attainable correlation matrices has not been fully characterized. Because Pearson's correlation depends on the marginal distributions, the set of attainable matrices is generally much more complicated than the elliptope of positive semi-definite correlation matrices investigated, for example, by Marić (2015, 2019) , unless attention is restricted to special choices of margins such as the normal (Embrechts et al., 2002) . Recently, Hofert and Koike (2019) described the set of attainable Blomqvist's beta matrices, while Embrechts et al. (2016) found conditions characterizing matrices of tail dependence coefficients. Devroye and Letac (2015) and Wang et al. (2019) consider the set of attainable Spearman rank correlation matrices; this coincides with the elliptope of linear d × d correlation matrices when d 9 but not when d 12, while the case where d ∈ {10, 11} remains to be settled. In this paper, we provide a complete solution to the attainability and compatibility problem when dependence is measured by the widely-used Kendall's tau rank correlation (Kendall, 1938; Kruskal, 1958; Joe, 1990) which parametrizes many popular dependence models. In order to take into account bivariate and higher-order associations between subsets of variables, we quantify the dependence of a multivariate random vector using a vector-valued measure that we call the concordance signature, which underlies all bivariate and multivariate Kendall's tau coefficients; this concept is made precise in Section 2. A note on our use of the terms attainability and compatibility may be helpful at this point. We use attainable when we talk about a logically coherent collective entity that could belong to a probability distribution, such as a concordance signature or a Kendall's tau matrix. We use compatible when we talk about the relationship between sub-components of an attainable entity, such as Kendall's tau rank correlations for different pairs of variables. We also refer to attainable signatures as being compatible with probability distributions. Our main result is to fully characterize the set of attainable concordance signatures of continuous multivariate distributions. As a by-product, we prove the conjecture of Hofert and Koike (2019) that the set of attainable Kendall rank correlation matrices is identical to the set of convex combinations of the extremal correlation matrices, i.e., the correlation matrices consisting exclusively of 1's and −1's; this set is also known as the cut polytope (Laurent and Poljak, 1995) . As we show, this characterization follows from the links between concordance signatures, multivariate Bernoulli distributions, and extremal mixture copulas, i.e., mixtures of the 2 d−1 possible copulas with extremal correlation matrices (Tiit, 1996) . We explain why the set of attainable Kendall correlation matrices is identical to the set of attainable correlation matrices for multivariate Bernoulli random vectors with symmetric margins as derived by Marić (2015, 2019) . The methodology we propose has a number of important applications. First, it allows us to determine whether a set of estimated or expert-elicited Kendall correlations (bivariate or multivariate) is compatible with a valid multivariate distribution. If it is, we then propose a method of determining the set in which any remaining unmeasured Kendall correlations must lie using stan- dard techniques from linear optimization and convex analysis. To illustrate, suppose, for example, we have Bitcoin, Etherium and Litecoin in our portfolio and we invest in a new cryptocurrency 'X4coin'. Based on return data for Bitcoin, Etherium and Litecoin in 2017, the attainable Kendall rank correlations between the as yet unobserved X4coin and the other three currencies can be shown to lie in the three-dimensional polytope in the left panel of Figure 1 . As soon as we form an opinion on the rank correlation between X4coin and one of the cryptocurrencies, this further restricts the set of possible rank correlations between X4coin and the other two currencies, as shown in the right panel of Figure 1 ; full details are given later in Example 3. We also prove that sample concordance signatures based on classical estimators of Kendall's tau are themselves concordance signatures of valid multivariate distributions. Our finding that the concordance signature of a multivariate distribution with continuous margins is always equal to the concordance signature of a unique mixture of the extremal copulas then offers a powerful technique in Monte Carlo simulation or risk analyses. It allows us to draw realizations from a model with identical concordance signature to that estimated from the data but with an extreme form of tail dependence. Finally, our methodology allows us to take a closer look at the dependence structure inherent in different classes of distributions. We show that the concordance signatures of the family of elliptical copulas form a strict subset of the set of all possible attainable concordance signatures. The surprising consequence is the existence of Kendall correlation matrices that do not correspond to any elliptically distributed random vector; this behavior provides a contrast with Pearson correlation matrices, each of which is the correlation matrix of at least one elliptical distribution. Moreover, we show that the d-dimensional Student t copula with correlation matrix P ∈ R d×d and degreeof-freedom parameter ν > 0 converges pointwise to a mixture of extremal copulas which shares its concordance signature as ν → 0. We also investigate the concordance signatures of exchangeable and equiconconcordant distributions. Throughout this paper, we use bold symbols such as x = (x 1 , . . . , x d ) to denote vectors in R d and understand expressions such as x + y as componentwise operations; similarly x y implies that all components are ordered. Let X = (X 1 , . . . , X d ) denote a generic random vector with continuous margins F 1 , . . . , F d . As is well known, the unique copula C of X is the distribution function of the random vector (Joe, 1997; Nelsen, 2006) . For any subset I ⊆ D = {1, . . . , d} with I = ∅ we write X I and U I to denote sub-vectors of X and U , and C I for the copula of X I , i.e., the distribution function of U I . The multivariate probability of concordance of X I is defined as where X * is a random vector independent of X but with the same distribution. Note that (1) implies κ {i} = 1 for a singleton and we adopt the convention κ ∅ = 1. For subsets with cardinality |I| 2, the concordance probabilities quantify the association between the components of X I . Indeed κ I is related to Kendall's tau τ I (Kendall, 1938; Kruskal, 1958) when |I| = 2 and to its multivariate analogue (Joe, 1990; Genest et al., 2011) Moreover, following Nelsen (2006) , κ I depends only on the copula C I of X I according to Without loss of generality, we can thus refer to κ I as the concordance probability of the copula C I of X I and use the notation κ I = κ(C I ) = κ(X I ) = κ(U I ) interchangeably. To summarize the various dependencies inherent in the vector X, we can collect the values κ I for all subsets I ∈ P(D), where P(D) denotes the power set of D. This motivates calling the vector (κ I : I ∈ P(D)) the full concordance signature of X. Sub-vectors of the full concordance signature are known as partial concordance signatures. From Proposition 1 in Genest et al. (2011) , which is based on the exclusion-inclusion principle, we can deduce that for any set I ⊆ D of odd cardinality, Thus the full concordance signature can be deduced from the concordance probabilities for subsets of D of even cardinality. We formalize ideas in the following definition. Definition 1. A label set is any collection S of subsets of D such that ∅ ∈ S. The vector κ S (C) = (κ(C I ) : I ∈ S) is called the partial concordance signature of the copula C for the label set S. By convention, the elements of S are taken in lexicographical order. When the label set is the even power set E(D) = {I : I ⊆ D, |I| even} (containing the empty set ∅ by convention) the partial concordance signature is called the even concordance signature of C and is denoted κ(C). When the label set is P(D) the partial concordance signature is called the full concordance signature of C and is denotedκ(C). Our main tools for the study of concordance signatures are mixtures of so-called extremal copulas and their relationship with multivariate Bernoulli distributions. In this section we develop the necessary notation and theory. An extremal copula C with index set J ⊆ D is the distribution function of the random vector where for any x ∈ R, x + = max(x, 0) denotes the positive part of x and we employ the convention min j∈∅ u j = 1. In dimension d 2 there are 2 d−1 extremal copulas and we enumerate them in the following way. For k ∈ {1, . . . , 2 d−1 } let s k = (s k,1 , . . . , s k,d ) be the vector consisting of the digits of k − 1 when represented as a d-digit binary number. For example, when d = 4 we have exactly eight extremal copulas corresponding to the vectors s 1 = (0, 0, 0, 0), s 2 = (0, 0, 0, 1), s 3 = (0, 0, 1, 0), s 4 = (0, 0, 1, 1), s 5 = (0, 1, 0, 0), s 6 = (0, 1, 0, 1), s 7 = (0, 1, 1, 0), s 8 = (0, 1, 1, 1). For each k ∈ {1, . . . , 2 d−1 }, s k = (s k,1 , . . . , s k,d ) and 1 − s k = (1 − s k,1 , . . . , 1 − s k,d ) are opposite vertices of the unit hypercube connected by one of its 2 d−1 main diagonals, 1 being the vector of ones. The kth extremal copula in dimension d, denoted C (k) , spreads its probability mass uniformly along the latter diagonal joining s k and 1 − s k . Its index set J k ⊆ D = {1, . . . , d} is defined as the set of indices corresponding to zeros in s k ; that is, j ∈ J k if s k,j = 0 and j ∈ J k if s k,j = 1. For the 4-dimensional example above we have J 1 = {1, 2, 3, 4}, J 1 = ∅, J 2 = {1, 2, 3}, J 2 = {4} and so on. Note also that J 1 = D so that C (1) (u) = min(u 1 , . . . , u d ) is the comonotonicity or Fréchet-Hoeffding upper bound copula. Extremal copulas owe their name to the fact that their correlation matrices are extremal correlation matrices, that is, matrices consisting exclusively of 1's and −1's. Indeed, for any k ∈ {1, . . . , 2 d−1 }, we find that the matrix of pairwise Kendall's tau values for the kth extremal copula C (k) is P (k) = (2s k − 1)(2s k − 1) . This matrix is simultaneously the matrix of pairwise Pearson and Spearman correlations of C (k) . Definition 2. An extremal mixture copula is a copula of the form C * = 2 d−1 k=1 w k C (k) , where for all k ∈ {1, . . . , 2 d−1 }, C (k) is the kth extremal copula, w k 0, and 2 d−1 k=1 w k = 1. The following proposition shows how the extremal mixture copulas are related to multivariate Bernoulli distributions; the proof is given in A. Proposition 1. Let U be a standard uniform random variable and B a d-dimensional multivariate Bernoulli vector independent of U . Then the distribution function of the vector is an extremal mixture with weights given, for each k ∈ {1, . . . , 2 d−1 }, by where s k is the vector consisting of the digits of k − 1 when represented as a d-digit binary number. Conversely, any extremal mixture copula C * = The class of Bernoulli distributions satisfying (7) is infinite since the mass w k can be split between the events {B = s k } and {B = 1 − s k } in an arbitrary way. It is important for the arguments used in this paper to single out a representative and we do this by setting B = Y , where Y is radially symmetric about 0.5. This means that Y d = 1 − Y and implies that, for each k ∈ {1, . . . , 2 d−1 }, Such distributions are also known as palindromic Bernoulli distributions (Marchetti and Wermuth, 2016) and they are fully parameterized by the 2 d−1 probabilities P(Y = s k ) = 0.5w k , k ∈ {1, . . . , 2 d−1 }. This means that there is a bijective mapping between the extremal mixture copulas and the radially symmetric multivariate Bernoulli distributions in dimension d. A number of constraints apply to radially symmetric Bernoulli random vectors. Apart from the fact that P(Y i = 1) = 0.5 for all i, we will make use of the following insight, proved in A. Proposition 2. (i) The radially symmetric Bernoulli distribution of a vector Y in dimension d is uniquely determined by the vector of probabilities p Y = (p I : I ∈ E(D) \ ∅), where p I = P(Y I = 1) and E(D) is the even power set as in Definition 1. (ii) The vector p Y in (i) is the shortest vector of the form (p I : I ∈ S \ ∅) for a label set S ⊂ P(D) which uniquely determines the distribution of Y for all radially symmetric Bernoulli random vectors Y . We close this section by noting that the independence between U and Y in the stochastic representation U of an extremal mixture copula cannot be dispensed with. An example of a copula that violates this condition is provided in A and reveals an interesting contrast between the extremal copulas and the mixtures of extremal copulas. While a necessary and sufficient condition for a vector U to be distributed according to an extremal copula is that its bivariate margins should be extremal copulas (Tiit, 1996) , the analogous statement does not hold for extremal mixture copulas. Although it is necessary that the bivariate margins of an extremal mixture copula are extremal mixture copulas, the example shows that this is not sufficient. An additional condition is required, as detailed in the next proposition which is proved in A. Proposition 3. The distribution of a random vector U = (U 1 , . . . , U d ) is a mixture of extremal copulas if and only if its bivariate marginal distributions are mixtures of extremal copulas and for all u ∈ [0, 1], In this section we characterize concordance signatures of arbitrary copulas in dimension d. To do so, we first calculate concordance signatures of extremal mixture copulas and investigate their properties. For any k ∈ {1, . . . , 2 d−1 } let C (k) be an extremal copula with index set J k as specified in Section 3. For I ⊆ D we introduce the notation noting that a {i},k = 1 and a ∅,k = 1 for all k and i. Then we have the following result. (ii) If C is an extremal mixture copula of the form 2 d−1 k=1 w k C (k) then, for any I ⊆ D, Proof. To show (i), it suffices to consider sets with |I| 2. For k ∈ {1, . . . , 2 d−1 }, let X (k) = (X i , X (k) j ) such that i ∈ J k and j ∈ J k , or vice versa, is a pair of countermonotonic random variables and hence discordant with probability 1. To establish (ii), note again that (11) holds trivially for sets I which are singletons or the empty set. For sets such that |I| 2 we can use (3) to write if j = 1 or k = 1 but j = k, 0 if j = 1 and k = 1. Hence we can verify that κ(C) = 2(w 2 1 /2) + 4w 1 {(w 2 /4) + · · · + (w 2 d−1 /4)} = w 2 1 + w 1 (1 − w 1 ) = w 1 , which is the weight on C (1) = M , the d-dimensional comonotonicity copula. If U is distributed according to C then the vector U I is distributed according to a mixture of extremal copulas in dimension |I| and it follows that κ(U I ) = κ(C I ) =w 1 wherew 1 is the weight attached to the case where U I is a comonotonic random vector. This is given bỹ where the second equality follows from part (i). Remark 1. It may be noted that the extremal copulas C (k) are the only copulas that give rise to extremal concordance signatures, that is concordance signatures consisting only of zeros and ones as in part (i) of Proposition 4. This follows from the fact that two variables are comonotonic if and only if the bivariate concordance probability is one and countermonotonic if and only if the bivariate concordance probability is zero (Embrechts et al., 2002) ; thus a copula with an extremal signature must have bivariate margins that are extremal copulas. Tiit (1996) showed that, if the bivariate margins of a copula are extremal copulas, then the copula is an extremal copula. From Proposition 4 we see that for any label set S and any k ∈ {1, . . . , 2 d−1 }, the partial concordance signature of C (k) is κ S (C (k) ) = (a I,k : I ∈ S). We also see that the partial concordance signature of an extremal mixture copula C = d k=1 w k C (k) is a convex combination of partial signatures of extremal copulas with the same weights, so that κ S (C) = d k=1 w k κ S (C (k) ). The following two key properties of the even concordance signature of an extremal mixture link back to the Bernoulli representation in Proposition 1; the proof relying on Proposition 2 is given in A. Proposition 5. Let a k = (a I,k : I ∈ E(D)), for k ∈ {1, . . . , 2 d−1 }, and let κ(C) = 2 d−1 k=1 w k a k be the even concordance signature of the extremal mixture C = d k=1 w k C (k) . (i) κ(C) uniquely determines C. Moreover, it is the minimal partial concordance signature of C which uniquely determines C in all cases. (ii) The vectors a k , k ∈ {1, . . . , 2 d−1 } are linearly independent. Proof. For part (i) recall that a random vector U ∼ C has the stochastic representation U It follows from the radial symmetry property that κ I (C) = 2P(Y I = 1). By Proposition 2 the vector (P(Y I = 1) : I ∈ E(D) \ ∅) is the minimal vector of event probabilities that pins down the law of Y in all cases and hence, by equation (8), the weights . If another extremal mixture copulaC shares the same values for the concordance probabilities of even order, then the weights must be identical. For part (ii) suppose, on the contrary, that the vectors a k are linearly dependent, that is, there exist scalars µ k , k ∈ {1, . . . , 2 d−1 } such that µ k = 0 for at least one k and To prove the result, we need to show that this assumption leads to a contradiction. We select a copula C such that C = 2 d−1 k=1 w k C (k) and w k > 0 for all k. Because the first component of a k is 1 for all k, we have that 2 d−1 k=1 µ k = 0. Hence, there exists at least one k * such that µ k * > 0. From (12) we have that for each α ∈ R, 2 d−1 k=1 (w k − αµ k )a k = κ(C). Let α * = min((w k /µ k ) : µ k > 0). Clearly, α * > 0 because the set over which the minimum is taken contains at least w k * /µ k * and because w k > 0 for all k. Now define w * = w − α * µ. These weights are non-negative and sum up to one, while w * = w. This implies that κ(C) = 2 d−1 k=1 w k a k = 2 d−1 k=1 w * k a k but this is not possible because the mixture weights are unique by part (i). Proposition 5 shows that the set of attainable even concordance signatures of d-dimensional extremal mixture copulas is the convex hull and implies in particular that K is a convex polytope with vertices a k , k ∈ {1, . . . , 2 d−1 }. Similarly the set of full concordance signatures of extremal mixtures is the convex hull ofã k = (a I,k : I ∈ P(D)), k ∈ {1, . . . , 2 d−1 }. We are now ready for the main result of this paper which shows that these sets are also the sets of attainable even and full concordance signatures for any d-dimensional copula, and hence of any random vector with continuous margins. Theorem 1. Let C be a d-dimensional copula and κ(C) = (κ I : I ∈ E(D)) its even concordance signature. Then there exists a unique extremal mixture copula C * = 2 d−1 k=1 w k C (k) such that κ(C) = κ(C * ) . The weights w k , which are non-negative and add up to 1, are the unique solution to the system of 2 d−1 linear equations given by Proof. For a vector U ∼ C we can write, for any I ∈ P(D) with |I| 1, where U * is an independent copy of U . If we define the random vectors V = sign(U * − U ) and Y = (1/2)(V + 1), then the concordance probabilities of C are given by so that the concordance signature of C is determined by the distribution of Y , which is a radially symmetric Bernoulli distribution. The radial symmetry follows from the fact that V From (14) we can conclude that, for every I ∈ P(D), where we have used the fact that the union of all the disjoint events {Y = s k } and {Y = 1 − s k } forms a partition of {0, 1} d in the first equality, the notation (10) and (8) in the second and (11) in the final equality. Thus the weights w k specifying the extremal mixture copula are precisely the probabilities that specify the law of the radially symmetric Bernoulli vector Y through w k = 2P(Y = s k ) and the uniqueness of the set of weights follows from the uniqueness of the law of Y . The sufficiency of solving the linear equation system (13) to determine the weights w k and the existence of a unique solution follow from Proposition 5, in particular the fact that (13) is a system of 2 d−1 equations with 2 d−1 unknowns and the vectors a k are linearly independent. The implication of Theorem 1 is that we can find the mixture weights w for a given concordance signature using simple linear algebra. To see this, let κ = κ(C) denote the even concordance signature of a d-dimensional copula C and let A d be the 2 d−1 × 2 d−1 matrix with columns a k . Proposition 5 implies that A d is of full rank, and hence invertible. The linear equation system (13) can be written in the form κ = A d w and must have a unique solution which can be found by calculating w = A −1 d κ. Theorem 1 further guarantees that w 0 and that its components sum up to 1. For example, when d = 4 we would have The fact that the even concordance signature is required to determine the weights of the extremal mixture copula in all possible cases allows us to view the even concordance signature of a copula as the minimal complete concordance signature. Indeed, the fact that A d is of full rank has the following interesting corollary, namely that a formula such as (4) can only hold for sets of odd cardinality. Corollary 1. When I is a set of even cardinality there is no linear formula valid for all copulas relating κ I to concordance probabilities of order |I | |I|. Proof. If the statement were not true, the rows of the matrix A d would be linearly dependent, contradicting the assertion that A d is of full rank. Remark 2. The implications of Corollary 1 can be extended. In view of (2) an analogous result could be stated for the multivariate Kendall's tau coefficients: when |I| is odd, a linear formula relating τ I to the values for lower-dimensional subsets exists (see Proposition 1 in Genest et al. (2011) ) but no such formula exists when |I| is even. Moreover, as we will discuss in Section 7, the concordance probabilities of elliptical copulas are equal to twice the orthant probabilities of Gaussian distributions centred at the origin: thus if Z ∼ N (0, Σ) is a Gaussian random vector, recursive linear formulas exist for the orthant probabilities P(Z I 0) when |I| is odd but not when |I| is even. Consider a random sample X 1 , . . . , X n from a distribution with copula C and continuous margins F 1 , . . . , F d . In this section, we explain how the full concordance signatureκ(C) can be estimated intrinsically, i.e., in such a way that the estimated signature is attainable. We do this under the assumption that there are no ties in the data; this is not restrictive because the continuity of the margins ensures the absence of ties with probability 1. For any I with |I| 2, empirical estimators of κ I = κ(C I ) can be derived from empirical estimators of τ I = τ (C I ) using (2). When d = 2 and I = {k, } for some distinct k, ∈ {1, . . . , d}, the classical estimator of τ I going back to Kendall (1938) and Hoeffding (1947) is This is a special case of the estimator of τ I for |I| 2 proposed and investigated by Genest et al. (2011) , which is given by Plugging this estimator into (2) yields an empirical estimator of κ I of the form From the theory of U-statistics (Hoeffding, 1948) , we know that the empirical concordance signaturẽ κ n = (κ I,n : I ∈ P(D)) satisfies √ n(κ n −κ) N (0, Σ) as n → ∞, whereκ =κ(C), Σ is the covariance matrix of the random vector with components C I (U I ) +C I (U I ) andC I is the survival function of C I . The following result shows that the empirical concordance signatureκ n is in fact the concordance signature of a d-dimensional copula. Theorem 2. Assuming that n 2 and there are no ties in the sample, there exists a unique d-dimensional extremal mixture copula C n such thatκ n =κ(C n ). Proof. Let Y ij = sign(X i − X j ) + 1 /2 for i = j and set These empirical weights are estimators of w k = P(Y = s k ) + P(Y = 1 − s k ), the weights of the extremal mixture which has the same concordance signature as C. Because the sample is assumed to have no ties, the weights w k are clearly positive and sum to 1. Thus they describe a mixture of extremal copulas, say C n . To establish the result, it suffices to show that for any I ∈ P(D) with |I| 2, κ I,n = κ(C n,I ), where C n,I is the margin of C n corresponding to the index set I. This can be seen as follows. First, where the last expression is κ I,n × n(n − 1)/2 since there are no ties in the sample. Remark 3. While the probability of ties in a sample from a distribution with continuous margins is zero, rounding effects may lead to occasional ties in practice. In this case it may be that some of the vectors Y ij have components equal to 0.5. Let us suppose that Y ij has k such values for k ∈ {1, . . . , d}. A possible approach to incorporating this information in the estimator is to replace Y ij by the 2 k vectors that have zeros and one in the same positions as Y ij , each weighted by 2 −k , and to generalize (16) to be a weighted sum of indicators. For example, the observation Y ij = (1, 0, 0.5, 0.5) would be replaced by (1, 0, 1, 1), (1, 0, 1, 0), (1, 0, 0, 1) and (1, 0, 0, 0), each weighted by 1/4. This would still deliver estimates w k that are positive and sum to one and thus yield a proper concordance signature. Theorem 1 allows us to test whether a vector of putative concordance probabilities κ S = (κ I : I ∈ S) with label set S could be a partial concordance signature of some copula C. We now present a number of applications of this result. Let us first say that the vector κ S is attainable if there exists a d-dimensional copula C such that κ S = κ S (C). We first turn to the characterization of matrices of pairwise Kendall's taus, also termed Kendall rank correlation matrices, which are widely used to summarize pairwise associations in a random vector. In view of (2), this can be achieved using Theorem 1 by considering the label set S = {∅, {1, 2}, . . . , {1, d}, . . . , {d − 1, d}}. For a random vector X we denote the Kendall rank correlation matrix and the linear correlation matrix by P τ (X) and P ρ (X) respectively. First note that Kendall rank correlation matrices are correlation matrices, i.e., positive semi-definite matrices with ones on the diagonal and all elements in [−1, 1] . This is because if U is a random vector distributed as a copula C, then P τ (U ) = P ρ (sign(U − U * )), where U * is an independent copy U * . In Theorem 3 below we provide an answer to the attainability question for Kendall rank correlation matrices using Theorem 1 and the following corollary. Recall from Section 3 that the Kendall rank correlation matrix, which is also the linear correlation matrix of the kth extremal copula, is given by P (k) = (2s k − 1)(2s k − 1) . Corollary 2. Let U be distributed according to the mixture of extremal copulas given by C Proof. Proposition 4 and equation (2) imply P τ (U ) = 2 d−1 k=1 w k P τ (U (k) ); the equality P ρ (U ) = 2 d−1 k=1 w k P ρ (U (k) ) was proved by Tiit (1996) . Hence (17) follows from the equality of linear and Kendall correlations for extremal copulas. Theorem 3. The d × d correlation matrix P is a Kendall rank correlation matrix if and only if P can be represented as a convex combination of the extremal correlation matrices in dimension d, that is, Proof. If P is of the form (18) then it is the Kendall's tau matrix of the extremal mixture copula C * = 2 d−1 k=1 w k C (k) by Corollary 2. If P is the Kendall's τ matrix of an arbitrary copula C then, by Theorem 1, it is also the Kendall's tau matrix of the extremal mixture copula with the same concordance signature and must take the form (17) by Corollary 2. The set of convex combinations of extremal correlation matrices is known as the cut polytope. Laurent and Poljak (1995) showed that the cut polytope is a strict subset of the so-called elliptope of correlation matrices in dimensions d 3; see also Section 3.3 of Hofert and Koike (2019) . For example, the positive-definite correlation matrix 1 12   12 −5 −5 −5 12 −5 −5 −5 12   is in the elliptope but not in the cut polytope. Therefore it cannot be a matrix of Kendall's tau values. If τ = −5/12 and we set κ = (1 + τ )/2 and κ = (1, κ, κ, κ), there is no solution to the equation A 3 w = κ on the 4-dimensional unit simplex and a representation of the form (18) is impossible. Huber and Marić (2019) have shown that the cut polytope is also the set of attainable linear correlation matrices for multivariate distributions with symmetric Bernoulli margins. This can be deduced from Theorem 1 and Proposition 1 using the following lemma. , where B is a Bernoulli vector with symmetric margins and U is a uniform random variable independent of B. Then P τ (U ) = P ρ (B). Proof. First note that P τ (U ) = P ρ (U ) by Proposition 1 and Corollary 2. By writing and using the fact that E(B i ) = 0.5 for all i ∈ {1, . . . , d}, we find that Since var(U i ) = 1/12 and var(B i ) = 1/4 for all i ∈ {1, . . . , d} we conclude that P ρ (U ) = 4E(BB ) − 11 = P ρ (B). This completes the proof. Proposition 6. The set of Kendall rank correlation matrices of copulas is identical to the set of linear correlation matrices of Bernoulli random vectors with symmetric margins. Proof. If P τ (U ) is a Kendall rank correlation matrix then, by Theorem 1, it is identical to the Kendall rank correlation matrix of a random vector U * with distribution given by the extremal mixture copula with the same concordance signature as U . It follows from Proposition 1 that U * has the stochastic representation where, without loss of generality, Y has a radially symmetric Bernoulli distribution (with symmetric margins). Lemma 1 gives P τ (U ) = P τ (U * ) = P ρ (Y ). Conversely if P ρ (B) is the correlation matrix of a Bernoulli random vector B with symmetric margins (not necessarily radially symmetric) then, by Proposition 1, we can take an independent uniform random variable U and construct a random vector U * = U B + (1 − U )(1 − B) with an extremal mixture copula. Lemma 1 gives P ρ (B) = P τ (U * ). Remark 4. It is clear from the proof that the set of linear correlation matrices of Bernoulli random vectors with symmetric margins is equal to the set of linear correlation matrices of radially symmetric Bernoulli random vectors. This insight also appears in Theorem 1 of Huber and Marić (2019). We now turn to the problem of determining whether a putative partial concordance signature is attainable, and, if it is, of calculating the set in which the remaining unmeasured concordance probabilities must lie. We begin with a simple example that illustrates what happens in four dimensions when all pairwise Kendall rank correlations are equal and form a so-called equicorrelation matrix. Example 1. Let C be a copula in dimension d = 4 with Kendall rank correlation matrix P τ (C) equal to the equicorrelation matrix with off-diagonal element 2κ 2 − 1; in other words the bivariate concordance probabilities κ {i,j} for all pairs of random variables are equal to κ 2 . The even concordance signature is then κ(C) = (1, κ 2 , κ 2 , κ 2 , κ 2 , κ 2 , κ 2 , κ 4 ) where κ 4 = κ {1,2,3,4} . It is easy to verify that the linear system given in (15) is solved by the weight vector w = (κ 4 , w 1 (κ 2 , κ 4 ), w 1 (κ 2 , κ 4 ), w 2 (κ 2 , κ 4 ), w 1 (κ 2 , κ 4 ), w 2 (κ 2 , κ 4 ), w 2 (κ 2 , κ 4 ), w 1 (κ 2 , κ 4 )) , where w 1 (κ 2 , κ 4 ) = (3κ 2 − 1)/2 − κ 4 and w 2 (κ 2 , κ 4 ) = 1 − 2κ 2 + κ 4 . Since the weights must satisfy 0 w i (κ 2 , κ 4 ) 1, the concordance probabilities must satisfy κ 2 ∈ [1/3, 1] and κ 4 ∈ [max(2κ 2 − 1, 0), (3κ 2 − 1)/2]. The left panel of Figure 2 shows the set of attainable values for τ 2 = 2κ 2 − 1 and τ 4 = (8κ 4 − 1)/7. It is notable that the mere fact that the bivariate concordance probabilities are equal limits the range of attainable correlations significantly. In principle, for any copula, the pair (τ 2 , τ 4 ) must always lie in the rectangle [−1, 1] × [−1/7, 1] shown in the plot. However, in the case of equicorrelation, the attainable set is considerably smaller. We provide some more general results on equiconcordant copulas (copulas for which κ I = κĨ when |I| = |Ĩ|) in Section 8. In general, attainability and compatibility problems can be solved by linear programming. To see how this is achieved, consider a label set S ⊂ E(D) strictly contained in the even power set and suppose we want to test whether the vector κ S = (κ I : I ∈ S) with label set S is attainable. Let (1) d w = κ S , w 0} and note that every element of the set satisfies the sum condition on the weights, since S is a label set containing ∅ and the first row of A (1) d consists of ones. If κ S is an attainable partial concordance signature then this set of weight vectors is non-empty and forms a convex polytope, that is, a set of the form In this case it is possible to use the method of Avis and Fukuda (1992) to find the vertices w 1 , . . . , w m of the polytope. The set of attainable even concordance signatures containing the partial signature κ S is then given by the convex hull of the points κ i = A d w i , i ∈ {1, . . . , m}, while the set of attainable unspecified elements of the concordance signature is given by the convex hull of the points {κ As the dimension increases and the discrepancy between the length of κ S and the length of an even signature increases, the vertex enumeration algorithm can become computationally infeasible. In such cases we may be content to simply test κ S for attainability by finding a single even concordance signature that contains κ S . To do so, we can attempt to solve the optimization problem min A (2) where · denotes the Euclidean norm. This is a standard minimization problem with both equality and inequality constraints. The putative partial concordance signature κ S is attainable if a solution exists and the solution to the optimization problem will be the weight vector which gives the (collectively) smallest values for the missing concordance probabilities. To get the (collectively) largest values we could solve min A (2) Note also that if we are only concerned with finding the smallest or largest missing values for certain missing concordance probabilities, then we can drop rows of A (2) d . Example 2. For d = 5 let a partial concordance signature be given by κ {i,j} = 2/3 for all pairs of variables and κ {1,2,3,4} = κ {1,2,3,5} = 0.4. To complete the concordance signature, three further concordance probabilities are required: κ {1,2,4,5} , κ {1,3,4,5} and κ {2,3,4,5} . Using the method of Avis and Fukuda (1992) the set of possible weight vectors is non-empty and has nine vertices; thus the specified partial signature is attainable. The set of attainable values for the missing values forms a polytope in 3d which is shown in Figure 2 . We now return to the motivating example at the beginning of the paper. Using real data on cryptocurrency returns we illustrate the use of the signature estimation method and we show how missing rank correlation values may be inferred from existing values. Example 3. We take the multivariate time series of cryptocurrency prices (in US dollars) for Bitcoin, Ethereum, Litecoin and Ripple. From these data we compute the daily log-returns for the calendar year 2017, giving us 365 4-dimensional data points. The estimated even concordance signature is where all figures are given to 3 decimal places. The Kendall rank correlations of the copula C will exactly match the estimated values from the data. Suppose we are not interested in Ripple but rather in a new cryptocurrency X4coin for which we have no data. We want to estimate the correlations between X4coin and the first three cryptocurrencies. By computing the convex hull of the set of points in (20), and converting the concordance probabilities to Kendall's tau values, we obtain the polytope shown in the left panel of Figure 1 . Suppose that we now decide that a plausible value for τ {1,4} is 0.598, which is actually the estimated rank correlation between Bitcoin and Ripple implied by the estimated concordance signature above. Then the remaining two values must lie in the convex set shown in the right panel of Figure 1 , which is a section of the 3-dimensional set, shown as a cut in the left panel. This is certainly the case for the estimated values of the Etherium-Ripple and Litecoin-Ripple rank correlations which are shown as a point within the set. The concordance signature is identical for the copulas of all continuous elliptical distributions with the same correlation matrix P . This follows because the individual probabilities of concordance are identical for all such copulas. This is proved in Genest et al. (2011, Section 2 .1) in the context of an analysis of multivariate Kendall's tau coefficients. Let X have an elliptical distribution centred at the origin with dispersion matrix equal to the correlation matrix P and assume that P(X = 0) = 0. If X * is an independent copy of X, then the concordance probabilities (1) are given by κ I = 2P (X I < X * I ) = 2P (X I − X * I < 0). The random vector X I − X * I also has an elliptical distribution centred at the origin. Using the stochastic representation for elliptical distributions (Fang et al., 1990; McNeil et al., 2015) we can write X I − X * I d = R 1 AS and X I d = R 2 AS where S is a random vector uniformly distributed on the unit sphere, A is a matrix such that AA = P and R 1 and R 2 are positive scalar random variables, both independent of S. It follows that From this calculation, we see that the positive scalar random variable R 2 plays no role, so that the concordance probabilities κ I are the same for any elliptical random vector with the same dispersion matrix P = AA . Moreover, they are equal to twice the orthant probabilities for a centred elliptical distribution with dispersion matrix P . In practice it is easiest to calculate the orthant probabilities of a multivariate normal distribution X ∼ N (0, P ) and this is the approach we take in our examples; see B for an illustration involving a 6 × 6 correlation matrix. Every linear correlation matrix can be the correlation matrix of a multivariate elliptical (or multivariate normal) distribution. We now show by means of a counterexample that an analogous statement is not true of Kendall rank correlation matrices. Example 4. The positive-definite correlation matrix is a Kendall rank correlation matrix but is not the Kendall rank correlation matrix of an elliptical distribution. corresponding to the maximum attainable fourth order concordance probability of 0.0425. In this case w 1 and w 2 are precisely the two vertices of the polytope of attainable weights given by the set (19), which takes the form of a line segment connecting w 1 and w 2 . Any weight vector in this set will give the Kendall rank correlation matrix P τ . Now let us assume that P τ in (24) corresponds to an elliptical copula. Lindskog et al. (2003) and Fang and Fang (2002) have shown that the Kendall rank correlation matrix of an elliptical copula with correlation matrix P is given by the componentwise transformation P τ = 2π −1 arcsin(P ). It must be the case that P = sin(πP τ /2) is the correlation matrix of the elliptical copula. However, by calculating the eigenvalues we find that P is not positive semi-definite, which is a contradiction. We now turn to the copula of the multivariate Student t distribution C t ν,P with degree of freedom parameter ν and correlation matrix parameter P . It is unusual to consider this copula for degrees of freedom ν < 1, but as it turns out, the Student t copula converges to an extremal mixture as ν → 0. The proof of the following remarkable property relies on some limiting results for the univariate and multivariate t distribution which are collected in C. Theorem 4. As ν → 0 the d-dimensional t copula C t ν,P converges pointwise to the unique extremal mixture copula that shares its concordance signature. Proof. Let the function h ν (w, s) be defined by where F ν is the distribution function of a t distribution with ν degrees of freedom, G d,ν is the distribution function of the radial component of a d-dimensional multivariate t distribution with ν degrees of freedom and A is a d × d matrix such that AA = P ; such a matrix can be constructed for any positive semi-definite P . Let S = (S 1 , . . . , S d ) be uniformly distributed on the unit sphere and let W be an independent uniform random variable. Then X = G −1 d,ν (W )AS has a multivariate t distribution and U = h ν (W, S) has joint distribution function C t ν.P . We want to show that the joint distribution function of h ν (W, S) converges to the joint distribution function of an extremal mixture as ν → 0. We first argue that the random vector given by AS satisfies P (AS) j = 0 = 0. Let a j denote the jth row of A. If a j = 0 then the jth row and column of P would consist of zeros implying that the jth margin of the multivariate t distribution of X was degenerate; this case can be discounted because the t copula with such a matrix P is not defined. Suppose therefore that P(a j S = 0) > 0 for a j = 0. If R is the radial random variable corresponding to the multivariate normal, then P(a j RS = 0) = P(a j Z = 0) > 0, where Z is a vector of d independent standard normal variables. However a j Z is univariate normal with variance a j a j > 0 and cannot have an atom of mass at zero. We can therefore define the set A = {s : s s = 1, (As) j = 0, j = 1, . . . , d} such that P(S ∈ A) = 1. Given that S = s ∈ A, then and hence the conditional distribution function of U given S = s has the form where I As is the set of indices j for which (As) j > 0. Writing, for any u ∈ (0, 1) d , we can use Proposition 8 in C and Lebesgue's Dominated Convergence Theorem to conclude that C t ν.P (u) converges, as ν → 0, to We now show that this limit is a mixture of extremal copulas. To this end, consider the random vector sign(AS) + 1 /2. This has the same distribution as the multivariate Bernoulli random vector Y whose distribution is defined by the probabilities p I = P(Y I = 1) = P (AS) j > 0, j ∈ I for I ⊆ D; the random vectors sign(AS)+1 /2 and Y differ only on the null set where components of AS are zero. Moreover, the distribution of Y is radially symmetric since the spherical symmetry of S implies AS The limiting distribution (25) may be written in the form (1 − 2u j ) + + P(Y = y) and using the index set notation defined in Section 3 this may also be written as and we need to check that C k is in fact the kth extremal copula C (k) given by (5). To do so, we have to distinguish the four cases below: (i) Suppose that there is at least one j ∈ J k and at least one j ∈ J k such that u j 0.5. Then C k (u) = 0 = C (k) (u). Table 1 : Probabilities associated with diagonals of unit cube for copula in Figure 3 . (0, 0, 0) ↔ (1, 1, 1) 51.3% (0, 0, 1) ↔ (1, 1, 0) 5.1% (0, 1, 0) ↔ (1, 0, 1) 15.4% (0, 1, 1) ↔ (1, 0, 0) 28.2% (ii) Suppose that for all j ∈ J k , u j > 0.5 and there exists at least one j ∈ J k such that u j 0.5. (iii) The case when for all j ∈ J k , u j > 0.5 and there exists at least one j ∈ J k such that u j 0.5 is analogous to case (ii) and is omitted. (iv) Suppose that for all j ∈ D, u j > 0.5. Then Finally, we need to verify that the concordance signature of the limiting mixture of extremal copulas is the same as the concordance signature of C t ν,P for any ν > 0. If κ I denotes a concordance probability for the t copula, we need to show that κ I = 2 d−1 k=1 w k a I,k , which is the corresponding concordance probability for the limit. Recall that the vector X = G −1 d,ν (W )AS has a multivariate t distribution. Equation (23) implies that where the final step uses the reasoning employed in the proof of Theorem 1. Figure 3 shows a scatterplot of the copula C t ν,P when ν = 0.03 and P = (ρ ij ) is the 3 × 3 matrix with elements ρ 12 = 0.2, ρ 13 = 0.5 and ρ 23 = 0.8. Clearly the points are distributed very close to the four diagonals of the unit cube. The limiting weights attached to extremal copulas associated with each diagonal are given in Table 1 . Remark 5. The proof holds even when the matrix P is not of full rank. However, because in such cases the copula is distributed on a strict subspace of the unit hypercube [0, 1] d , the limiting extremal mixture copula has zero mass on certain diagonals of the hypercube. Suppose, for example, that rows i and j of the matrix A satisfying AA = P are identical. Then the components Y i and Y j of the vector Y defined in the proof are identical, i.e., they are both 0 or both 1. For any vector s k such that s k,i = s k,j it must be the case that w k = P(Y = s k ) + P(Y = 1 − s k ) = 0 and so the kth diagonal would have zero mass. : Scatterplot of data with distribution C t ν,P when ν = 0.03 and P is the 3 × 3 matrix with elements ρ 12 = 0.2, ρ 13 = 0.5 and ρ 23 = 0.8. As we saw in Example 1, the constraint that certain concordance probabilities are equal further restricts the set of attainable signatures. We close this paper by investigating this phenomenon in greater detail. To this end, let Π(u) represent a permutation of the vector u = (u 1 , . . . , u d ) and recall that a copula C is said to be exchangeable if C(Π(u)) = C(u) for all u ∈ [0, 1] d and any permutation Π. A weaker notion of symmetry based on concordance signatures is the following: Definition 3. A copula C is said to be equiconcordant if its even concordance signature κ(C) = {κ I : I ∈ E(D)} has the property that κ I = κĨ whenever |I| = |Ĩ|. Every exchangeable copula C is equiconcordant but not every equiconcordant copula is exchangeable. However, as we will show, the notions are equivalent for the class of extremal mixture copulas. To develop the necessary arguments we introduce some further notation for extremal copulas building on the notation of Section 3. Let η k = max(|J k |, |J k |) define the comonotonic number of the extremal copula C (k) or, in other words, the size of the larger of the two groups of comonotonic random variables described by C (k) . If we order the distinct values of η k from largest to smallest we obtain a vector (h 1 , . . . , h m(d) ) = (d, d − 1, . . . , d/2 ) with m(d) = 1 + d/2 elements, where · and · are the floor and ceiling functions respectively. Each value h i can be associated with a multiplicity, which is the number of extremal copulas with comonotonic number h i . Provided d = 2h i , the multiplicity is the number of ways of choosing the h i members of the larger comonotonic group from a set of d variables and is therefore µ i = d h i . When d = 2h i the binomial coefficient double counts the number of extremal copulas with comonotonic number h i ; for example the extremal copula for which the first h i variables are comonotonic is identical to the extremal copula for which the last h i variables are comonotonic. Thus, in this case, the mutiplicity is given by µ i = 1 To understand the notation it is helpful to consider a concrete case. When d = 4 we have η 1 = 4, η 2 = η 3 = η 5 = η 8 = 3 and η 4 = η 6 = η 7 = 2; thus in this case the 8 extremal copulas can be split into m(4) = 3 groups according to comonotonic number and we find (h 1 , h 2 , h 3 ) = (4, 3, 2) and (µ 1 , µ 2 , µ 3 ) = (1, 4, 3) . We also need to consider the behaviour of extremal copulas under permutations of the arguments. Clearly, the first extremal copula C (1) is exchangeable, but the others are not. For example, under the permutation Π(u) = (u 3 , u 4 , u 1 , u 2 ) we find that C (2) (Π(u)) = (min(u 3 , u 4 , u 1 ) + u 2 − 1) + = (min j∈J 5 u j + min j∈J 5 u j − 1) + = C (5) (u). and, more generally, , C (4) → C (4) , C (6) → C (6) and C (7) → C (7) . It is clear that every extremal copula is mapped either to itself or to another extremal copula with the same comonotonic number. For other permutations the mapping may differ and extremal copulas may form cycles of length greater than two, but they will always be mapped to copulas with the same comonotonic number. Proposition 7. Let C * = 2 d−1 k=1 w k C (k) be an extremal mixture copula. The following statements are equivalent: 3. For all k, l ∈ {1, . . . , 2 d−1 } the weights satisfy w k = w l whenever η k = η l . Proof. We show that 1 ⇒ 2 ⇒ 3 ⇒ 1. The first of these implications is immediate. If I andĨ are two subsets of D with the same cardinality then exchangeability implies C * I = C * I and therefore κ I = κ(C * I ) = κ(C * I ) = κĨ . 2 ⇒ 3. We use an inductive argument based on comonotonic number h i for i ∈ {1, . . . , m(d)}. Observe first that for i = 1 the set I = D is the single subset of D with cardinality |I| = d = h 1 and in this case κ I = w 1 , the weight on the first extremal copula C (1) , which is the only extremal copula with comonotonic number h 1 . Now suppose that w k = w l whenever η k = η l = h j for all j ∈ {1, . . . , i} and i < m(d). We want to show that w k = w l whenever η k = η l = h i+1 . Consider the subsets I such that |I| = h i+1 . For each of these subsets we have a formula for the concordance probability of the form κ(C * I ) = 2 d−1 k=1 w k a I,k by Proposition 4 and if a I,k = 1 it must be the case that η k h i+1 by (10). For each distinct subset I there is an equal number of copulas C (k) with a I,k = 1 and η k = h j for each j ∈ {1, . . . , i} and, by assumption, the weights on copulas with the same comonotonic number are equal. If h i+1 = d/2 there are µ i+1 extremal copulas C (k) with η k = h i+1 and µ i+1 sets I with |I| = h i+1 and for each distinct set I there is a distinct extremal copula C (k) such that a I,k = 1. If h i+1 = d/2 there are µ i+1 extremal copulas with η k = h i+1 and 2µ i+1 sets I with |I| = h i+1 ; in this case each distinct set I is weighted on a single extremal copula but for each C (k) there are two sets I with a I,k = 1. In either case, the equality of the concordance probabilities for the sets I implies that the weights w k are identical for all copulas with η k = h i+1 . 3 ⇒ 1. For a permutation Π let the function l Π (k) give the identity of the extremal copula to which C (k) is mapped under Π and recall that this copula has the same comonotonic number as C (k) . Then, for all u ∈ [0, 1] d , where the final step follows because w k = w l whenever η k = η l . To complete the proof we need to show that the function l Π is simply a permutation of the indices k = 1, . . . , 2 d−1 implying that C * (Π(u)) = C * (u). To see this assume that l π (k 1 ) = l Π (k 2 ) for k 1 = k 2 so that, for any u ∈ [0, 1] d , C (k 1 ) (Π(u)) = C l(k 1 ) (u) = C l(k 2 ) (u) = C (k 2 ) (Π(u)). If Π −1 denotes the inverse permutation satisfying Π(Π −1 (u)) = u, then this would imply that contradicting the assumption that k 1 = k 2 . This result allows us to conclude that for any equiconcordant copula C, the extremal mixture copula C * sharing its signature with C is exchangeable. Moreover, we can infer that the set of attainable even equiconcordance signatures takes the form and we can write this in a simplifed way as where we note that the vectors b i are obtained by averaging the vectors a k corresponding to extremal copulas with the same comonotonic number and are linearly independent (because the vectors a k are linearly independent). The weights v i now express the total weight assigned to all copulas with the same comonotonic number. The set of attainable equiconcordance signatures is clearly also a convex polytope with vertices given by the vectors b 1 , . . . , b m(d) . The 2 d−1 × m(d) matrix with columns b i has identical rows corresponding to sets I andĨ with identical cardinality. Duplicate rows can simply be dropped to retain exactly m(d) rows corresponding to each distinct even cardinality in the set E = {0, 2, . . . , 2 d/2 }. This results in an invertible square matrix B d . Let us write the elements of the even concordance signature corresponding to each distinct cardinality as k = (κ i : i ∈ E) and refer to k as the skeletal signature of an equiconcordant copula. We can then consider the linear equation system k = B d v to solve attainability and compatibility problems for skeletal signatures. For example, the equation system for d = 7 is and the columns of B 7 form the 4 vertices of the convex polytope of attainable vectors for the skeletal signature; upon removal of the 1's in the first component these describe an irregular tetrahedron inside the unit cube [0, 1] 3 as shown in Figure 4 . In higher dimensions it quickly becomes computationally infeasible to calculate the matrix B d by collapsing the 2 d−1 × 2 d−1 matrix A d . We end by giving a general formula for the elements of the matrix B d using combinatorial arguments. where l i = 2(i − 1), h j is the jth distinct value of the comonotonic number in reverse order from largest to smallest and, by convention, n p = 0 if p < 0. Proof. Obviously the first row of B d consists of ones corresponding to the sum constraint on the weights so let i > 1. To begin with, we exclude the case where d is even and h j = d/2. In all other cases the denominator of the fraction above is the multiplicity µ j of the comonotonic number h j . Let I ⊆ D be any subset with cardinality |I| = l i . The numerator is the number of columns of the matrix A d that correspond to extremal copulas with comonotonic number h j and which have a one in the row corresponding to I. Without loss of generality let us consider the set I consisting of the first l i elements of D. In this case we need to simply count the number of extremal copulas C (k) with comonotonic number h j for which the vectors s k that we use to code the extremal copulas have zeros in the first l i entries. In the case where d is even and h j = d/2 the denominator is twice the multiplicity µ j of the comonotonic number h j . In this case there are exactly the same number of zeros and ones in the µ j vectors s k coding extremal copulas with comonotonic number h j and so there are d−l i h j −l i ways of assigning the remaining zeros to the other positions. However, since d−l i h j −l i = d−l i d−h j −l i when h j = d/2, we can use the same general formula. The methods and examples in this paper are documented in the R package KendallSignature at https://github.com/ajmcneil/KendallSignature. Proof of Proposition 1. Let U be of the form (6) where U and B are independent. For any u ∈ [0, 1], where in the final step we have used the fact that the set of possible outcomes of the Bernoulli vector B can be written as the disjoint union From (5), one can see that the distribution function of U is indeed an extremal mixture with weights as given in (7). Conversely, given an extremal mixture copula of the form C * = 2 d−1 k=1 w k C (k) , it suffices to consider any Bernoulli vector B independent of U with the property that w k = P(B = s k )+P(B = 1 − s k ); this is possible because the events {B = s k } and {B = 1 − s k } are disjoint and their union forms a partition of {0, 1} d by (26). We can then retrace the steps of the argument in reverse to establish the representation (6). Proof of Proposition 2. For part (i), note that all probabilities of the form p I = P(Y I = 1) for sets I with odd cardinality are fully determined by the equivalent probabilities for lower-dimensional sets of even cardinality. This follows from the fact that When I is odd, we then have where the last equality follows from the fact that p A = 0.5 whenever |A| = 1. For example, The conclusion follows since the vector of joint event probabilities (p I : I ∈ P(D) \ ∅) uniquely specifies the distribution of a Bernoulli random vector. For part (ii) observe that the vector p Y has length equal to Radially symmetric multivariate Bernoulli distributions in dimension d have 2 d−1 − 1 free parameters, with one being deducted for the sum constraint 2 2 d−1 k=1 P(Y = s k ) = 1. Thus the vector p Y is the minimal vector of its kind that is required to fully specify the distribution of Y for all radially symmetric Bernoulli random vectors Y . Example 5. Let the joint distribution of (U, Y 1 , Y 2 , Y 3 ) be specified by P(U u, Y 1 = y 1 , Y 2 = y 2 , Y 3 = y 3 ) = 1 8 u + (−1) (y 1 +y 2 +y 3 ) θu(1 − u) 4 for u ∈ [0, 1] and y i ∈ {0, 1}, i ∈ {1, 2, 3}. Note that this is an increasing function in u for any fixed (y 1 , y 2 , y 3 ) and defines a valid distribution. It may be easily verified, by summing over the outcomes for the Y i variables, that the marginal distribution of U is standard uniform, while letting u → 1 shows that the random vector Y = (Y 1 , Y 2 , Y 3 ) consists of iid Bernoulli variables with success probability 0.5 (and is radially symmetric). It may also be verified that the marginal distributions P (U u, Y i = y i ) = 0.5u so that the pairs (U, Y i ) are independent for all i, while the marginal Since the pairs (U, Y i ) are independent it is easy to see that the components U i = U Y i + (1 − U )(1 − Y i ) are uniform, implying that the distribution of U is a copula. Since the triples (U, Y i , Y j ) are mutually independent, the bivariate margins of U are extremal mixtures by Proposition 1. To calculate the copula C of U we observe that, for u = (u 1 , u 2 , u 3 ), C(u 1 , u 2 , u 3 ) = 4 i=1 P(U u, Y = s k ) + P(U u, Y = 1 − s k ) = 2 4 i=1 P(U u, Y = s k ). The final equality follows because our model has the property that (U, Y 1 , Y 2 , Y 3 ) . The copula C has 4 distinct terms, each associated with a diagonal of the unit cube: 4C(u 1 , u 2 , u 3 ) = min(u 1 , u 2 , u 3 ) + θ 4 min(u 1 , u 2 , u 3 ) max(1 − u 1 , 1 − u 2 , 1 − u 3 ) + 1 {min(u 1 ,u 2 )+u 3 −1 0} min(u 1 , u 2 ) − θ 4 min(u 1 , u 2 ) max(1 − u 1 , 1 − u 2 ) − 1 + u 3 + θ 4 u 3 (1 − u 3 ) + 1 {min(u 1 ,u 3 )+u 2 −1 0} min(u 1 , u 3 ) − θ 4 min(u 1 , u 3 ) max(1 − u 1 , 1 − u 3 ) − 1 + u 2 + θ 4 u 2 (1 − u 2 ) + 1 {min(u 2 ,u 3 )+u 1 −1 0} min(u 2 , u 3 ) − θ 4 min(u 2 , u 3 ) max(1 − u 2 , 1 − u 3 ) − 1 + u 1 + θ 4 u 1 (1 − u 1 ) . However, unless θ = 0, C is not of the form (5) and is not an extremal mixture copula. Proof of Proposition 3. If U follows an extremal mixture copula then, by Proposition 1, it has the stochastic representation U d = U B + (1 − U )(1 − B) for some Bernoulli random vector B and it is clear that all the bivariate margins have the same structure. From the independence of U and B we obtain Conversely, let us suppose that all the bivariate margins of U are mixtures of extremal copulas. This implies that almost surely, U takes values in the set (1 − u 1 ) s k,j , j = 1}. Let E k = {U j = U (1−s k,j ) 1 (1 − U 1 ) s k,j , j = 1} represent the event that U lies on the kth diagonal and set w k = P(E k ). By the law of total probability, we get On the diagonals of the hypercube, {U ∈ E} ∩ {1 {U j =U 1 } = 1 − s k,j , j = 1} = E k , so that w k = P{1 {U j =U 1 } = 1 − s k,j , j = 1} and conditioning on the event E k is identical to conditioning on the event {1 {U j =U 1 } = 1 − s k,j , j = 1}. We thus obtain that where the last equality follows from (9) and (5). This shows that C is indeed a mixture of extremal copulas as claimed. The concordance probabilities for a continuous elliptical distribution the 6 × 6 correlation matrix P = 1 16         16 1 2 3 4 5 1 16 6 7 8 9 2 6 16 10 11 12 3 7 10 16 13 14 4 8 11 13 16 15 5 9 12 14 15 16 are given in Table 2 , along with the weights of the unique mixture of extremal copulas sharing the same concordance signature. C Some limiting properties of the univariate and multivariate Student distribution as ν → 0 Let F ν , F −1 ν and f ν denote the df, inverse df and density of a univariate t distribution with ν degrees of freedom. The df satisfies where 2 F 1 denotes the hypergeometric function and Γ the gamma function. We will show that, for fixed u = 0.5, the quantile function F −1 ν (u) is unbounded as a function of ν as ν → 0. To that end we first prove the following lemma. Lemma 2. lim ν→0 F ν (x) = 0.5 for all x ∈ R. Proof. The lemma is trivially true for x = 0 so we consider x = 0. Making the substitution y = x/ √ ν in (29) to conclude that lim ν→0 2F ν ( √ νy) − 1 ν = ln y + 1 + y 2 = sign(y) ln |y| + 1 + y 2 . Reversing the earlier substitution and setting x = √ νy now gives lim ν→0 2F ν (x) − 1 ν sign(x) ln |x| + √ ν + x 2 − 0.5 ln ν = 1. The result follows from the fact that the denominator tends to 0 as ν → 0. (28). if u > 0.5. Proof. The case u = 0 is obvious, since F −1 ν (0.5) = 0 for all ν > 0. To show that lim ν→0 F −1 ν (u) = −∞ for u < 0.5, we need to show that, for all k < 0, there exists δ such that ν < δ ⇒ F −1 ν (u) k. Suppose we fix an arbitrary k < 0. Since F ν (k) → 0.5 as ν → 0, there exists δ > 0 such that ν < δ implies that F ν (k) > u, since u < 0.5. But then, for any ν < δ, it follows that k ∈ {x : F ν (x) u} and hence F −1 ν (u) = inf{x : F ν (x) u} k. Analogously, to show that lim ν→0 F −1 ν (u) = ∞ for u > 0.5, we need to show that, for all k > 0, there exists δ such that ν < δ ⇒ F −1 ν (u) k. Suppose we fix an arbitrary k > 0. Since F ν (k) → 0.5 as ν → 0, there exists δ > 0 such that ν < δ implies that F ν (k) < u, since u > 0.5. But then, for any ν < δ, it follows that k / ∈ {x : F ν (x) u}. Since F ν (x) is a non-decreasing function of x, any y < k also satisfies y / ∈ {x : F ν (x) u}. Hence F −1 ν (u) = inf{x : F ν (x) u} k. If the random vector X has a d-dimensional multivariate t distribution with ν degrees of freedom, then it has the stochastic representation X Note that, when u = 0.5, F −1 ν (u) = 0 for all ν and (31) clearly holds; we thus restrict our analysis of the density to the case where u = 0.5. Using the notation x ν,u = F −1 ν (u) and the expression (30) we have that Handbook of Mathematical Functions A pivoting algorithm for convex hulls and vertex enumeration of arrangements and polyhedra Copulas with prescribed correlation matrix Bernoulli and tail-dependence compatibility Correlation and dependency in risk management: properties and pitfalls The meta-elliptical distributions with given marginals Symmetric Multivariate and Related Distributions Estimators based on Kendall's tau in multivariate copula models On the distribution of the rank correlation coefficient τ when the variates are not independent A class of statistics with asymptotically normal distribution Compatibility and attainability of matrices of correlation-based measures of concordance Multivariate distributions with fixed marginals and correlations Admissible Bernoulli correlations Multivariate concordance Families of m-variate distributions with given margins and m(m − 1)/2 bivariate dependence parameters Multivariate Models and Dependence Concepts A new measure of rank correlation Ordinal measures of association On a positive semidefinite relaxation of the cut polytope Credit Risk: Measurement, Evaluation and Management Palindromic Bernoulli distributions Quantitative Risk Management: Concepts, Techniques and Tools An Introduction to Copulas Measurement and modelling of dependencies in economic capital Causal simulation experiments: Lessons from bias amplification Mixtures of multivariate quasi-extremal distributions having given marginals Compatible matrices of Spearman's rank correlation The authors are grateful to the associate editor and two anonymous referees for their feedback and insightful comments that improved this paper and spurred further research that led to the material in Section 8. This work was partially supported by a Discovery Grant from the Natural Sciences and Engineering Research Council (NSERC) of Canada awarded to J.G. Nešlehová (RGPIN-2015-06801).