Symmetry | 您所在的位置:网站首页 › wjhxxb/2022/6/PIC/wjhxxb-38-6-1059-13_mini.jpg › Symmetry |
Next Article in Journal / Special Issue
Skewness-Based Projection Pursuit as an Eigenvector Problem in Scale Mixtures of Skew-Normal Distributions
Previous Article in Journal
Initial Value Problems of Linear Equations with the Dzhrbashyan–Nersesyan Derivative in Banach Spaces
Previous Article in Special Issue
Distribution-Based Entropy Weighting Clustering of Skewed and Heavy Tailed Time Series
Journals
Active Journals
Find a Journal
Proceedings Series
Topics
Information
For Authors
For Reviewers
For Editors
For Librarians
For Publishers
For Societies
For Conference Organizers
Open Access Policy
Institutional Open Access Program
Special Issues Guidelines
Editorial Process
Research and Publication Ethics
Article Processing Charges
Awards
Testimonials
Author Services
Initiatives
Sciforum
MDPI Books
Preprints.org
Scilit
SciProfiles
Encyclopedia
JAMS
Proceedings Series
About
Overview
Contact
Careers
News
Blog
Sign In / Sign Up
Notice
clear
Notice
You are accessing a machine-readable page. In order to be human-readable, please install an RSS reader. Continue Cancel clearAll articles published by MDPI are made immediately available worldwide under an open access license. No special permission is required to reuse all or part of the article published by MDPI, including figures and tables. For articles published under an open access Creative Common CC BY license, any part of the article may be reused without permission provided that the original article is clearly cited. For more information, please refer to https://www.mdpi.com/openaccess. Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications. Feature papers are submitted upon individual invitation or recommendation by the scientific editors and must receive positive feedback from the reviewers. Editor’s Choice articles are based on recommendations by the scientific editors of MDPI journals from around the world. Editors select a small number of articles recently published in the journal that they believe will be particularly interesting to readers, or important in the respective research area. The aim is to provide a snapshot of some of the most exciting work published in the various research areas of the journal. ![]() ![]() ![]() Find support for a specific problem in the support section of our website. Get Support FeedbackPlease let us know what you think of our products and services. Give Feedback InformationVisit our dedicated information section to learn more about MDPI. Get Information clear JSmol Viewer clear first_page settings Order Article Reprints Font Type: Arial Georgia Verdana Font Size: Aa Aa Aa Line Spacing: Column Width: Background: Open AccessArticle Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula by![]() ![]() ![]() ![]() ![]() Abstract: Symmetric elliptical distributions have been intensively used in data modeling and robustness studies. The area of applications was considerably widened after transforming elliptical distributions into the skew elliptical ones that preserve several good properties of the corresponding symmetric distributions and increase possibilities of data modeling. We consider three-parameter p-variate skew t-distribution where p-vector μ is the location parameter, Σ : p × p is the positive definite scale parameter, p-vector α is the skewness or shape parameter, and the number of degrees of freedom ν is fixed. Special attention is paid to the two-parameter distribution when μ = 0 that is useful for construction of the skew t-copula. Expressions of the parameters are presented through the moments and parameter estimates are found by the method of moments. Asymptotic normality is established for the estimators of Σ and α . Convergence to the asymptotic distributions is examined in simulation experiments. Keywords: asymptotic normality; inverse chi-distribution; multivariate cumulants; multivariate moments; skew normal distribution; skew t-copula; skew t-distribution MSC: 62E20; 62H10; 62H12 1. IntroductionThe idea of constructing multivariate skewed distributions from symmetric elliptical distributions has attracted many statisticians over last 25 years. This generalization preserves several good properties of corresponding symmetric distributions and in the same time increases possibilities of data modeling. In 1996, Azzalini and Dalla Valle introduced multivariate skew normal distribution [1]. Their idea of transforming multivariate normal distribution into a skewed distribution has been carried over to several other distributions and many generalizations have been made since 1996 (see for instance, Theodossiou [2], Arellano-Valle and Genton [3]). From these developments multivariate skew t-distribution has become most frequently used [4]. This distribution became an important tool in many data modeling applications because of heavier tail area than the normal and skew normal distributions (see, for instance, Lee and McLachan [5], Marchenko [6], Parisi and Liseo [7], Fernandez and Steel [8], Jones [9]). A review of applications in finance and insurance mathematics is given in Adcock et al. [10]. Multivariate skew t-distribution has also been successfully used as a data model in environmental studies and modelling coastal floodings (Ghizzolini et al. [11], Thompson and Shen [12]). Several other applications of skew elliptical distributions can be found in the collective monograph edited by Genton [13] and in the book by Azzalini and Capitanio [14] which includes a comprehensive list of references on the topic. Another useful property that makes multivariate t-distribution and skew t-distribution attractive in financial and actuarial applications is the possibility to take tail dependence between marginals into account. Tail dependence of multivariate t- and skew t-distributions has been intensively studied in recent years, see Fung and Seneta [15] or Joe and Li [16], for example. In Kollo et al. [17] it is shown that the upper tail dependence coefficient of the skew t-distribution can be several times bigger than the corresponding value of multivariate t-distribution. Moments and inferential aspects related to the multivariate skew t-distribution have been also actively studied in Padoan [18], Galarza et al. [19], Zhou and He [20], Lee and McLachlan [5], DiCiccio and Monti [21], for instance. A construction based on the skew t-distribution—skew t-copula—has made this distribution even more attractive in applications where correlated differently distributed marginals are joined into a multivariate distribution. Demarta and McNeil introduced skew t-copula based on multivariate generalized hyperbolic distribution [22], Kollo and Pettere [23] defined skew t-copula on the basis of the multivariate skew t-distribution following Azzalini and Capitanio [4].The paper is organized as follows. In Section 2, we introduce necessary notation and notions from matrix algebra. In Section 3, skew t-distribution is defined through skew normal and inverse χ -distribution and the first four central moments are found via the moments of skew normal and inverse χ -distribution. The asymptotic normality for parameter estimators in the two-parameter case is also established in this section. In Section 4, obtained analytical results are illustrated by simulations for different scenarios and the speed of convergence to asymptotic distribution is also addressed. In Section 5, multivariate skew t-copula is defined and some properties are discussed. Technical proofs are presented in Appendix A. A comparison of density contours of Gaussian, skew normal and skew t-copula is presented in Supplementary Materials. 2. Notation and PreliminariesTo derive asymptotic normal distribution for the parameter estimators of skew t-distribution, we use matrix technique. Necessary notions and notation are presented in what follows.Let us first denote the moments of a random vector x in the following matrix representation: m k ( x ) = E x ⊗ x ′ ⊗ x ⊗ x ′ ⊗ … ︸ k times = E x x ′ ⊗ x x ′ ⊗ … . The corresponding central moment is m ¯ k ( x ) = m k ( x − E x ) . We use notions vec -operator, commutation matrix, Kronecker product, matrix derivative and their properties. For deeper insight into this technique we refer to Magnus and Neudecker [24], or Kollo and von Rosen [25]. If A is a p × q matrix, A = ( a 1 , … , a q ) , the p q column vector vec A is formed by stacking the columns a i under each other; that is vec A = a 1 … a q . We denote a p × p identity matrix by I p and a p q × p q commutation matrix by K p , q . Commutation matrix K p , q is an orthogonal p q × p q partitioned matrix consisting of q × p blocks with the property K p , q vec A = vec A ′ where A is a p × q -matrix. The Kronecker product A ⊗ B of a p × q matrix A and an r × s matrix B is a p r × q s partitioned matrix consisting of the r × s blocks A ⊗ B = [ a i j B ] , i = 1 , … , p ; j = 1 , … , q . The following properties of the notions above are repeatedly used. It is assumed that the dimensions of the matrices match with the operations.1. ( A ⊗ B ) ( C ⊗ D ) = ( A C ) ⊗ ( B D ) ;2. a b ′ = a ⊗ b ′ = b ′ ⊗ a ;3. vec ( A B C ) = ( C ′ ⊗ A ) vec B ;4. K p , 1 = I p , K 1 , q = I q ; 5.for A : p × q and B : r × s , A ⊗ B = K p , r ( B ⊗ A ) K s , q . Further, when deriving asymptotic distributions of the estimators of Σ and α in Section 3, we use the matrix derivative where partial derivatives ∂ y k l ∂ x i j are ordered by the Kronecker product (see Magnus and Neudecker, [24]) d Y d X = d d vec ′ X ⊗ vec Y , with X : p × q and Y = Y ( X ) : r × s . It follows from the definition that d Y d X = d vec Y d vec ′ X = d vec Y d X = d Y d vec ′ X . The main properties of the matrix derivative are listed below.(1) d X d X = I p q ;(2) d ( Y + Z ) d X = d Y d X + d Z d X , Z = Z ( X ) : r × s ;(3) d A X B d X = B ′ ⊗ A ;(4)If Z = Z ( Y ) , Y = Y ( X ) then d Z d X = d Z d Y d Y d X ;(5)If W = Y Z , Z = Z ( X ) : s × t then d W d X = ( Z ′ ⊗ I r ) d Y d X + ( I t ⊗ Y ) d Z d X ;(6)If X : p × p then d X − 1 d X = − ( X ′ ) − 1 ⊗ X − 1 ;(7)If X ∈ R p × p then d | X | d X = | X | vec ( X − 1 ) ′ , where | · | denotes the determinant;(8)If W = Y ⊗ Z , Z = Z ( X ) : m × n , then d W d X = ( I s ⊗ K r , n ⊗ I m ) ( I r s ⊗ vec Z ) d Y d X + ( vec Y ⊗ I m n ) d Z d X .(9)If y = y ( X ) , Z = Z ( X ) : m × n , then d ( y Z ) d X = vec Z d y d X + y d Z d X . 3. Multivariate Skew t ν -Distribution 3.1. DefinitionMultivariate skew t-distribution has been defined in different ways, see Kotz and Nadarajah [26] (Chapter 5). Kshirsagar [27] was the first to introduce a noncentral multivariate t-distribution, but the density expression was complicated and the distribution was not used in applications. From more recent definitions we refer to Fang et al. [28], but as the marginals in their definition have different degrees of freedom, the distribution had problems with practical use. Gupta [29] defined the multivariate skew t-distribution in a similar way like Azzalini and Capitanio [4], but their definition had several advantages and has become widely used in applications (for example, it is possible to use explicit expression of the moment generating function of skew normal distribution for calculating moments of skew t-distribution). We follow the approach by Azzalini and Capitanio [4].A p-vector x has multivariate skew t-distribution with ν degrees of freedom if it is the ratio of a p-variate skew normal vector and a chi-distributed random variable with ν degrees of freedom. We have slightly modified their definition of the skew normal vector and use the covariance matrix as a parameter instead of the correlation matrix. Let y be a p-variate skew normal vector with density f ( y ) = 2 f N p ( 0 , Σ ) ( y ) Φ ( α ′ y ) , where f N p ( 0 , Σ ) ( y ) is the density function (density) of the p-variate normal distribution N p ( 0 , Σ ) and Φ ( · ) is the cumulative distribution function (distribution function) of N ( 0 , 1 ) . Denote the distribution of vector y with density (1) by y ∼ S N p ( Σ , α ) , where Σ : p × p is the scale parameter and α ∈ R p is the shape or skewing parameter.Definition 1. Let y be skew normally distributed with density (1) and let Z be χ-distributed with ν degrees of freedom, independent from y . Then x = μ + ν y Z is skew t ν -distributed, x ∼ S t ν ( μ , Σ , α ) .The density function of x ∼ S t ν ( μ , Σ , α ) is of the form (a modification of Azzalini and Capitanio [4]) f x ( x ) = 2 f ν , p ( x ) F p + ν α ′ ( x − μ ) p + ν Q + ν , where Q = ( x − μ ) ′ Σ − 1 ( x − μ ) , F k is the distribution function of the univariate t-distribution with k degrees of freedom and f ν , p is the density function of the p-variate t-distribution with ν degrees of freedom and scale matrix Σ f ν , p ( x ) = Γ ν + p 2 ( ν π ) p / 2 Γ ν 2 | Σ | 1 / 2 1 + ( x − μ ) ′ Σ − 1 ( x − μ ) ν − ν + p 2 . 3.2. MomentsBy Formula (2), a skew t-distributed x is constructed using independent skew normal random vector y and χ -distributed Z. Because of the independence of y and Z, the moments of x can be found as products of the moments of the skew normal and inverse χ -distribution. The k-th moment, central moment and cumulant of a skew-normal random vector y are denoted by m k ( y ) , m ¯ k ( y ) and c k ( y ) , respectively. Explicit expressions of the raw moments, central moments and cumulants of y can be obtained by differentiating the moment generating function M ( t ) = 2 e 1 2 t ′ Σ t Φ α ′ Σ t 1 + α ′ Σ α and the cumulant generating function C ( t ) = ln M ( t ) . The first four cumulants for a skew-normal random vector y are (Kollo et al. [30]) c 1 ( y ) = E y = 2 π δ , c 2 ( y ) = D y = Σ − 2 π δ δ ′ , c 3 ( y ) = m ¯ 3 ( y ) = 2 π 4 π − 1 δ ⊗ δ ′ ⊗ δ , c 4 ( y ) = 8 π 1 − 3 π δ ⊗ δ ′ ⊗ δ ⊗ δ ′ , where δ equals δ = Σ α 1 + α ′ Σ α . From general relation between the fourth order cumulants and central moments we have (see, e.g., Kollo and von Rosen [25] (p. 187)) m ¯ 4 ( y ) = c 4 ( y ) + ( I p 2 + K p , p ) ( D y ⊗ D y ) + vec ( D y ) vec ′ ( D y ) . After some algebra, we get the expression for α from (5): α = Σ − 1 δ 1 − δ ′ Σ − 1 δ . The quantity under square root in (7) influences the stability of parameter estimates and will be used later in a related discussion. Thus, we add a notation for it: η = 1 − δ ′ Σ − 1 δ . Note also that each component δ i in vector δ = ( δ 1 , … , δ p ) ′ is directly connected to the corresponding marginal distribution, whereas this property does not hold for vector α = ( α 1 , … , α p ) ′ . For a more comprehensive discussion on parametrizations of skew-normal vector see Käärik et al. [31].To derive the central moments of the multivariate skew t ν -distribution, we also need the moments of the skew normally distributed vector y m 2 ( y ) = Σ ; m 3 ( y ) = 2 π ( vec Σ δ ′ + δ ⊗ Σ + Σ ⊗ δ − δ δ ′ ⊗ δ ) . Recall that a random variable V ν = 1 Z is inverse χ ν -distributed if its probability density function is (Lee [32] (p. 378)) f ( x ) = 1 2 ν / 2 − 1 Γ ν 2 x − ν − 1 exp − 1 2 x 2 . The moments of the inverse χ ν -distribution can be obtained by integration.Lemma 1. The moments of the inverse χ ν -distribution are: E V ν = Γ ν − 1 2 2 Γ ν 2 , ν > 1 ; E V ν k = Γ ν − k 2 2 k / 2 Γ ν 2 , ν > k . Proof. By standard integration technique, we can derive the expectation: E V ν = ∫ 0 ∞ x 1 2 ν / 2 − 1 Γ ν 2 x − ν − 1 exp − 1 2 x 2 d x = 1 Γ ν 2 ∫ 0 ∞ Γ ν − 1 2 2 ν − 1 2 − 1 2 Γ ν − 1 2 x − ν exp − 1 2 x 2 d x = Γ ν − 1 2 Γ ν 2 2 . Following the same approach, the k-th order moment can be derived as follows E V ν k = ∫ 0 ∞ x k 1 2 ν / 2 − 1 Γ ν 2 x − ν − 1 exp − 1 2 x 2 d x = 2 ν − 1 2 − 1 Γ ν − 1 2 2 ν 2 − 1 Γ ν 2 ∫ 0 ∞ 1 2 ν − 1 2 − 1 Γ ν − 1 2 x − ( ν − 1 ) − 1 x k − 1 exp − 1 2 x 2 d x = ∏ i = 0 k − 1 E V ν − i = Γ ν − k 2 2 k 2 Γ ν 2 , ν > k . □The expressions of the second, third and fourth moment are special cases of this result: E V ν 2 = 1 ν − 2 , ν > 2 ; E V ν 3 = E V ν ν − 3 , ν > 3 ; E V ν 4 = 1 ( ν − 2 ) ( ν − 4 ) , ν > 4 . After some elementary algebra we get expressions of the corresponding central moments m ¯ 3 ( V ν ) = E V ν 7 − 2 ν ( ν − 3 ) ( ν − 2 ) + 2 ( E V ν ) 2 ; m ¯ 4 ( V ν ) = 1 ( ν − 2 ) ( ν − 4 ) + 2 ν − 5 ( ν − 2 ) ( ν − 3 ) ( E V ν ) 2 − 3 ( E V ν ) 4 . To keep the notation simpler, we denote V = V ν .In order to derive the asymptotic normality for parameter estimators of S t ν , we need the first 4 central moments of the distribution. Let us start with general expressions for central moments m ¯ 3 ( x ) and m ¯ 4 ( x ) .Lemma 2. Let x be a p-dimensional random vector with finite fourth order moments. Then the third and fourth central moment of x can be expressed as follows: m ¯ 3 ( x ) = m 3 ( x ) − ( I p 2 + K p , p ) ( m 2 ( x ) ⊗ E x ) − vec m 2 ( x ) E x ′ + 2 E x E x ′ ⊗ E x , m ¯ 4 ( x ) = m 4 ( x ) − C ( x ) ( I p 2 + K p , p ) − ( I p 2 + K p , p ) ( C ( x ) ) ′ − 3 E x E x ′ ⊗ E x E x ′ + ( I p 2 + K p , p ) A ( x ) + A ( x ) K p , p + B ( x ) + ( A ( x ) + B ( x ) ) ′ , where A ( x ) = E x ′ ⊗ m 2 ( x ) ⊗ E x ; B ( x ) = E x ′ ⊗ vec m 2 ( x ) ⊗ E x ′ ; C ( x ) = m 3 ( x ) ⊗ E x ′ . Proof of the lemma is presented in Appendix A.1. Applying this lemma to skew t-distributed x , the next result follows.Lemma 3. Let x ∼ S t ν ( μ , Σ , α ) . Then its mean E x , covariance matrix D x , and central moments m ¯ 3 ( x ) and m ¯ 4 ( x ) can be expressed through the moments of inverse χ ν -distributed V and y ∼ S N p ( Σ , α ) in the following form: E x = μ + ν E V E y ; D x = ν m 2 ( V ) m 2 ( y ) − E x E x ′ = ν ( m 2 ( V ) D y + D V E y E y ′ ) ; m ¯ 3 ( x ) = ν 3 / 2 { E V 3 m ¯ 3 ( y ) − 2 ( E V 3 − ( E V ) 3 ) ( E y E y ′ ⊗ E y ) + ( E V 3 − E V E V 2 ) [ ( I p 2 + K p , p ) ( m 2 ( y ) ⊗ E y ) + vec m 2 ( y ) E y ′ ] } ; m ¯ 4 ( x ) = ν 2 { E V 4 m ¯ 4 ( y ) + ( E V 4 − E V E V 3 ) ( C ( y ) ( I p 2 + K p , p ) + ( I p 2 + K p , p ) C ( y ) ′ ) + 3 ( E V 4 − ( E V ) 4 ) E y E y ′ ⊗ E y E y ′ − ( E V 4 − E V 2 ( E V ) 2 ) [ ( I p 2 + K p , p ) A ( y ) + A ( y ) K p , p + B ( y ) + ( A ( y ) + B ( y ) ) ′ ] } . Proof is straightforward and follows from Lemma 2 after substituting in the required moments of skew normal distribution from Formulas (9) and (10) and the moments of inverse χ -distribution given in Lemma 1. 3.3. Point EstimatesLet us now focus on the case when μ = 0 and estimate the parameters of the two-parameter distribution S t ν ( Σ , α ) using the method of moments.Let us have a sample of size n, n > p , from S t ν ( Σ , α ) with sample mean x ¯ and unbiased sample covariance matrix S . Now, the estimates for parameters Σ and α through the moments of V and y can be obtained from Expressions (11) and (12): E x = Γ ν − 1 2 Γ ν 2 ν π δ , D x = ν 1 ν − 2 Σ − Γ ν − 1 2 Γ ν 2 2 1 π δ δ ′ , where δ is defined in (5). From here we get the following estimates Σ ^ = ν − 2 ν S + x ¯ x ¯ ′ ; α ^ = ν ν − 2 ( S + x ¯ x ¯ ′ ) − 1 x ¯ ν π Γ ν − 1 2 Γ ν 2 2 − ν ν − 2 x ¯ ′ ( S + x ¯ x ¯ ) − 1 x ¯ . 3.4. Asymptotic NormalityAs the estimators depend on both, the sample mean and the sample covariance matrix, we need their joint distribution to find the asymptotic distributions of the parameter estimators. In general, the following convergence in distribution takes place for x ¯ vec S if m ¯ 4 ( x ) exists (Parring [33]): n x ¯ vec S − E x ¯ vec S ⟶ D N ( 0 , D a s ) , where D a s = D x ⋮ m ¯ 3 ( x ) ′ m ¯ 3 ( x ) ⋮ m ¯ 4 ( x ) − vec ( D x ) vec ′ ( D x ) . For the asymptotic covariance matrix of the joint distribution of sample mean and covariance matrix from skew t-distribution, D a s S t , we get the expressions of necessary central moments from Equations (12)–(14). Derivation of asymptotic normality for estimators of parameters is based on Theorem 1 by Anderson [34] (pp. 132–133).Theorem 1. Assume that for { x n } n ( x n − a ) ⟶ D N p ( 0 , Σ ) and x n ⟶ P a , if n → ∞ . Let the function g ( x ) : R p → R q have continuous partial derivatives in a neighborhood of a . Then, if n → ∞ n ( g ( x n ) − g ( a ) ) ⟶ D N q ( 0 , ξ Σ ξ ′ ) , where ξ = d g ( x ) d x | x = a ≠ 0 is the matrix derivative.The vector ( x ¯ ′ , vec ′ S ) ′ satisfies assumptions of x n and we consider Σ ^ and α ^ as the functions of ( x ¯ ′ , vec ′ S ) ′ and derive asymptotic normality for the estimators. For Σ ^ the following statement holds.Theorem 2. Let us have a p-vector x ∼ S t ν ( Σ , α ) , where ν is fixed. Then, for the estimator Σ ^ in (15), the following convergence takes place, if n → ∞ n vec ( Σ ^ − Σ ) ⟶ D N p 2 ( 0 , D a s Σ ) , where D a s Σ = d ( ν − 2 ν ( S + x ¯ x ¯ ′ ) ) d ( x ¯ ′ , vec ′ S ) E x ¯ vec S D a s S t d ( ν − 2 ν ( S + x ¯ x ¯ ′ ) ) d ( x ¯ ′ , vec ′ S ) ′ E x ¯ vec S , the matrix D a s S t is the asymptotic covariance matrix of x ¯ vec S and the derivative d ( ν − 2 ν ( S + x ¯ x ¯ ′ ) ) d ( x ¯ ′ , vec ′ S ) E x ¯ vec S = ν − 2 ν ( I p 2 + K p , p ) I p ⊗ 2 π δ ⋮ I p 2 . Proof. To find the matrix derivative, we can rely on the corresponding expression for the skew normal distribution, derived by Kollo et al. [30] (Proposition 2) as the expression of the derivative for skew t-distribution differs only by a constant.For y ∼ S N p ( Σ , α ) , the derivative of Σ ^ is: ( I p 2 + K p , p ) I p ⊗ 2 π δ ⋮ I p 2 . From here, the derivative of Σ ^ for x ∼ S t ν ( Σ , α ) is: d Σ ^ d ( x ¯ ′ , vec ′ S ) E x ¯ vec S = ν − 2 ν ( I p 2 + K p , p ) I p ⊗ 2 π δ ⋮ I p 2 . This completes the proof. □The proof of asymptotic normality of estimator for parameter α , n ( α ^ − α ) ⟶ D N p ( 0 , D a s α ) , is similar. In order to shorten the notation from (16), we define M = ν π Γ ν − 1 2 Γ ν 2 2 − ν ν − 2 x ¯ ′ ( S + x ¯ x ¯ ′ ) − 1 x ¯ . This allows us to express the estimator α ^ as α ^ = Σ ^ − 1 x ¯ M . Theorem 3. Let us have a p-vector x ∼ S t ν ( Σ , α ) , where ν is fixed. Then, for the estimator α ^ in (18), the following convergence takes place, if n → ∞ n ( α ^ − α ) ⟶ D N p ( 0 , D a s α ) , where D a s α = d α ^ d ( x ¯ ′ , vec ′ S ) E x ¯ vec S D a s S t d α ^ d ( x ¯ ′ , vec ′ S ) ′ E x ¯ vec S , matrix D a s S t is the asymptotic covariance matrix and the derivative d α ^ d ( x ¯ ′ , vec ′ S ) = d α ^ d x ¯ ⋮ d α ^ d vec S , where d α ^ d x ¯ = M − 1 2 Σ ^ − 1 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ I p + M − 1 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ + M x ¯ x ¯ ′ Σ ^ − 1 . and d α ^ d vec S = − ν − 2 ν M − 1 2 ( x ¯ ′ Σ ^ − 1 ⊗ Σ ^ − 1 ) + 1 2 M − 1 ( Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 ⊗ x ¯ ′ Σ ^ − 1 ) . The derivatives in (19) for the asymptotic covariance matrix D a s α are obtained when the estimators x ¯ and Σ ^ are replaced by their expected values.Derivations in the proof of Theorem 3 are more technical than in the proof of Theorem 2 and therefore the proof is presented in Appendix A.2. 4. Simulations and DiscussionTo illustrate theoretical results, we consider a bivariate skew t-distribution x ∼ S t ν ( Σ , α ) and study the convergence speed of the estimates using several parameter sets. First we fix two values of η in (8), since, by the construction, this quantity has effect on the stability of the moments-based estimate for α . Namely, values of η close to zero make the moment-based estimates unstable (see, e.g., Formula (7)). Taking this into account, we choose one “problematic” value of η , η ≈ 0.03 , and one “safe” value, η ≈ 0.3 . Different values for Σ and α are chosen to obtain given values of η . In both cases, two α values of different magnitude are considered: small value is set to ( 0.5 , 1 ) ′ and moderate to ( 3 , 2 ) ′ . Furthermore, the off-diagonal values in matrix parameter Σ are pre-set to achieve low or high absolute value of σ 12 / σ 11 σ 22 with both signs. Thus, altogether 16 sets of parameters are used, the exact values can be found in Appendix A.3, Table A1. The number of degrees of freedom ν is taken equal to 4 in all cases considered.With all parameter sets, the data is simulated using sample sizes n from 100 to 10 6 (using sequence for the powers of 10 from 2 to 6 with 50 equally spaced points) and k = 3000 repetitions for each sample size. The mean absolute percentage deviation between the estimate and the theoretical asymptotic value (MAPD) is used as an indicator of convergence. Given k generated samples, MAPD is calculated as MAPD ( θ ) = 1 k ∑ j = 1 k θ ^ j − θ θ , where θ is the theoretical asymptotic value of estimate and θ ^ j is the parameter estimate from the jth sample.It must be noted that depending on the sample size and the value of η there are occasions when it is not possible to find an eligible estimate for α since the sample value for η is negative. To achieve the planned 3000 estimates for parameter sets with η ≈ 0.03 and small sample sizes there is up to 2.75 times more samples needed altogether. For parameter sets with η ≈ 0.3 the number of failed estimates is considerably smaller. Figure A1 in Appendix A.4 shows the ratio of the number of failed samples to the number of planned estimates from our simulations.Figure 1 shows resulting MAPD values for parameters depending on the logarithm of sample size. Red tones represent η ≈ 0.03 group and blue tones η ≈ 0.3 . Lighter tone is for smaller α and dark for moderate size α . Line types represent different sign and magnitude combinations of σ 12 / σ 11 σ 22 .As one can see, the convergence for diagonal elements of Σ is fastest and shows no remarkable differences for different parameter sets. However, MAPD for σ ^ 11 and σ ^ 22 is slightly bigger for sample sizes between 100 to 1000 if η is small.For the off-diagonal element of Σ , the convergence speed varies with different parameter sets. If the absolute value of σ 12 / σ 11 σ 22 is big, the pattern is actually similar to the convergence of diagonal elements, but for small values, the convergence is slower and MAPD can be close to or exceed value 1 even with sample sizes near 1000.Most recognizable distinction is seen with the MAPD values of the estimates of α and it is due to the η value. If η is small ( η ≈ 0.03 ), the MAPD value retains almost constantly high level with sample sizes from 100 up to 10,000, then starts to decrease with larger samples, but is still bigger than the MAPD value for the group with η ≈ 0.3 with sample sizes near 10 6 . For η ≈ 0.3 , the convergence is faster and reaches smaller values.Another aspect is the stability of estimates: with smaller η ( η ≈ 0.03 ), the estimates are also more unstable for sample sizes under 10 4 than the estimates corresponding to η ≈ 0.3 . Other variations (low and moderate α elements and different σ 12 / σ 11 σ 22 values) do not seem to have an effect on stability of the estimates.Next we compare asymptotic covariance matrices of the estimators found in Theorems 2 and 3 with their estimates from simulated samples. Parameter sets 1 and 9 from Table A1 in Appendix A.3 were used. The former (set 1) corresponds to safe η ≈ 0.3 value and the latter (set 9) to problematic η ≈ 0.03 value and these parameter sets have the same value of α . For both cases α ^ and Σ ^ were calculated 1000 times from samples of size 500. These parameter estimates were used to estimate D a s α and D a s Σ . For the simulations, the number of degrees of freedom ν = 5 was used, as it is the minimal value for calculating the asymptotic covariance matrices.For the parameter set 9 with η ≈ 0.03 , the theoretical covariance matrix for n α ^ is D a s α = 280.18 503.84 503.84 993.65 and its estimate D a s α ^ = 230.93 372.54 372.54 756.02 . As the theoretical expected value is n α ≈ ( 11.2 , 22.4 ) ′ we can say that the variance of estimates is high in this case, but the theoretical and estimated covariance matrices are in agreement, although the theoretical values are higher. For the estimator n vec Σ ^ , the results are similar. In this case, the theoretical covariance matrix is D a s Σ = 200.00 − 460.00 − 460.00 1213.50 − 460.00 1291.25 1291.25 − 3864.00 − 460.00 1291.25 1291.25 − 3864.00 1213.50 − 3864.00 − 3864.00 14,112.00 and the corresponding estimate: D a s Σ ^ = 182.98 − 425.81 − 425.81 1044.01 − 425.81 1191.52 1191.52 − 3293.59 − 425.81 1191.52 1191.52 − 3293.59 1044.01 − 3293.59 − 3293.59 11,819.03 . For the parameter set with η ≈ 0.3 , the same applies: theoretical and estimated covariance matrices are close while theoretical values are higher. Here the theoretical covariance matrix for n α ^ is D a s α = 12.11 13.01 13.01 21.38 and the estimate D a s α ^ = 9.59 9.23 9.23 14.71 . As these two parameter sets have the same value of α , we can see that for the second case the estimates vary much less. For η ≈ 0.3 , the theoretical covariance matrix for n vec Σ ^ is D a s Σ = 32.00 − 35.20 − 35.20 45.04 − 35.20 48.20 48.20 − 70.40 − 35.20 48.20 48.20 − 70.40 45.04 − 70.40 − 70.40 128.00 and its estimate D a s Σ ^ = 29.14 − 30.29 − 30.29 36.03 − 30.29 39.30 39.30 − 54.62 − 30.29 39.30 39.30 − 54.62 36.03 − 54.62 − 54.62 97.27 . Figure 2 presents scatterplots for estimates of α 1 and α 2 with theoretical 95% confidence ellipses. On the left panel ( η ≈ 0.03 ), the estimates of α are more scattered as compared to the right panel, where η ≈ 0.3 , as expected. In addition, the placing of points does not correspond to an ellipse on the left panel, indicating the asymptotics does not work with sample size n = 500 for distribution with η ≈ 0.03 . At the same time the result for η ≈ 0.3 can be considered good.Similar plots for the elements of Σ ^ are presented in Figure 3. Here the estimates and ellipses again fit well for η ≈ 0.3 case. For η ≈ 0.03 the fit is also better compared to estimates of α in the previous figure. 5. Skew t-Copula 5.1. DefinitionThe skew t-copula is introduced in Kollo and Pettere [23] (Chapter 15). When the location parameter μ = 0 , the copula has the form C p , ν ( u 1 , … , u p ; Σ , α ) = F p , ν ( F 1 , ν − 1 ( u 1 ; σ 11 , α 1 * ) , … , F 1 , ν − 1 ( u p ; σ p p , α p * ) ; Σ , α ) , where F 1 , ν − 1 ( u i ; σ i i , α i * ) denotes the inverse of the univariate skew t 1 , ν -distribution function and F p , ν is the distribution function of the p-variate skew t p , ν -distribution. The arguments σ i i are the main diagonal elements of the covariance matrix Σ and α i * can be derived by the following simplification of (7): α i * = σ i i − 1 δ i 1 − σ i i − 1 δ i 2 , i = 1 , … , p . Note that α i * -s are not equal to the components α i of the parameter vector α of the multivariate distribution. The corresponding copula density is c p , ν ( u ; Σ , α ) = f p , ν ( F 1 , ν − 1 ( u 1 ; σ 11 , α 1 * ) , … , F 1 , ν − 1 ( u p ; σ p p , α p * ) ) ′ ; Σ , α ∏ i = 1 p f 1 , ν F 1 , ν − 1 ( u i ; σ i i , α i * ) ; σ i i , α i * , where f p , ν ( u ; Σ , α ) is the density function of S t p , ν and the inverse functions F 1 , ν − 1 ( u i ; σ i i , α i * ) are as before.In the standard case when μ = 0 and the covariance matrix Σ is equal to the correlation matrix R , the density of the skew t-copula is of the form c p , ν ( u ; R , α ) = f p , ν ( F 1 , ν − 1 ( u 1 ; α 1 * ) , … , F 1 , ν − 1 ( u p ; α p * ) ) ′ ; R , α ∏ i = 1 p f 1 , ν F 1 , ν − 1 ( u i ; α i * ) ; α i * , where F 1 , ν ( u i , α i * ) , i = 1 , … , p are the distribution functions of the univariate standard skew t 1 , ν -distribution. Density (21) is the most often used version of the skew t-copula density in data analysis.As the numerator of the copula density consists of the density of multivariate skew t-distribution, the resulting copula is distribution function of a skewed multivariate distribution. When we want to apply skew t-copula in data analysis to join (possibly) differently distributed marginals into a multivariate distribution, we have to take into account the correlations between marginals. Pearson correlation matrix can easily be expressed through the parameter Σ . Unfortunately Pearson correlations are not invariant under copula transformations (inverse distribution functions). We have to switch over to rank correlations. Fortunately for continuous elliptical distributions, Kendall’s τ can be expressed through the linear correlation (see Lindskog et al. [35]) τ ( X i , X j ) = 2 π arcsin ( r i j ) , so we can switch over from linear correlations to rank correlations and vice versa. 5.2. Examples of Bivariate Skew t-Copula DensitiesIn the following we present some examples of two-dimensional copula densities to compare Gaussian, skew-normal and skew t-copulas using contour plots. First we present the formulas for the before mentioned copulas in bivariate case. The bivariate Gaussian copula has one parameter ρ and using the general formula for multivariate Gaussian copula from Cherubini et al. [36] (p. 148) we can express bivariate density as c N ( u 1 , u 2 ; ρ ) = 1 1 − ρ 2 exp − ρ 2 ( z 1 2 + z 2 2 ) − 2 ρ z 1 z 2 2 ( 1 − ρ 2 ) , where z i = Φ − 1 ( u i ) and u i ∈ [ 0 , 1 ] , i = 1 , 2 . Skew-normal copula has in addition to ρ parameter vector α and following Käärik et al. [37] its density in bivaraite case is of form c S N ( u 1 , u 2 ; ρ , α ) = 1 2 1 − ρ 2 exp − ρ 2 ( w 1 2 + w 2 2 ) − 2 ρ w 1 w 2 2 ( 1 − ρ 2 ) Φ ( α ′ ( w 1 , w 2 ) ′ ) Φ ( α 1 * w 1 ) Φ ( α 2 * w 2 ) , where α i * are given by (20) and w i = F − 1 ( u i ; α i * ) denotes inverse of the univariate skew-normal distribution function with parameter α i * , as u i ∈ [ 0 , 1 ] , i = 1 , 2 . The bivariate skew t-copula has parameters ρ , α and ν and using the general Formula (21) we get the bivariate density as c 2 , ν ( u 1 , u 2 ; ρ ; α ) = ν 4 1 − ρ 2 Γ ν 2 Γ ν + 1 2 2 1 + Q ( 1 − ρ 2 ) ν − ν + 2 2 ∏ i = 1 2 ( 1 + q i 2 ν ) − ν + 1 2 F 2 + ν ( α ′ ( q 1 , q 2 ) ′ 2 + ν Q + ν ) ∏ i = 1 2 F 1 + ν ( α i * q i 1 + ν q i 2 + ν ) , where q i = F 1 , ν − 1 ( u i ; α i * ) , u i ∈ [ 0 , 1 ] , i = 1 , 2 , denotes inverse of the univariate skew t distribution function with parameter α i * , Q = q 1 2 − 2 q 1 q 2 ρ + q 2 2 and F k is the distribution function of the univariate t k -distribution.For a first example, we consider here the case were parameter ρ is zero. Gaussian copula reduces to the product copula when ρ = 0 . If the corresponding parameter for skew-normal or skew t-copula equals zero, the resulting copula is not a product copula, this can be seen in Figure 4.In the figures, contours on the left correspond to skew-normal and on the right to skew t-copulas with different α values (values are given in the header of the figure, note that the same number of degrees of freedom ν = 4 is used for all examples here). Contours are given with step size of 0.2 starting from 0 and ending with 3. Dark blue, yellow and dark red colors represent the lowest, middle and highest contour levels for chosen scale.If ρ = 0 and both components of α are positive, the densities of skew-normal and skew t-copulas tend to show negative dependence that grows stronger if α i are bigger (see first and second row of Figure 4). If one α i is negative, the copula represents positive dependence, but the direction of dependence also depends on the magnitude of α i , i.e., the direction is not so clear for α i close to zero. When comparing contours in the second and third row we see that adding negative sign to the second component of α results in rotation of the density by 90o to the right. Rotation to the left would be the result if first element of α is negative and the other positive. We note that the two skew copulas give the product copula if in addition to ρ = 0 one element in the α vector is also zero.As a second example, we consider the case when ρ ≠ 0 . For Gaussian copula, the parameter ρ gives the direction and strength of dependence between marginals. For the two skewed copulas, the additional parameter α adds complexity to this relation. We can conclude from the Figures S1–S4. that if both α i are positive and ρ is negative as given here, the dependence for skewed copulas can be stronger as compared to Gaussian copula. For positive ρ , the opposite may happen: the skew-normal and skew t represent weaker dependence as compared to Gaussian copula or the dependence changes direction, see Figure S2 for ρ = 0.2 , for example. The effect of parameter α is more notable if its elements have high absolute values. 6. ConclusionsOur main theoretical results, presented in Theorems 2 and 3, assure the asymptotic normality for the moment based parameter estimators of the two-parameter multivariate skew t-distribution. We can make the following observations based on the simulations used to illustrate the analytical results. First, the behaviour of the estimates of α depends on the value of η : if the value of η is close to zero (we used value 0.03 ), the estimates of α are unstable and the convergence is really slow, MAPD values are greater than 0.5 with sample sizes up to 10 4 . For parameter sets with η value 0.3 , the results are more stable and, for the aforementioned sample size, the MAPD values were already less than 0.1 . Second, it occurs that the combination of small sample size and low η value can result in ineligible estimates of α . As a third point, we note that the estimates of Σ are stable and the convergence is faster than the convergence of the estimates of α . Furthermore, the value of η does not affect the estimates of Σ . As the two considered values of η were chosen arbitrarily, its effect needs further investigation. Finally we also define skew t-copula, a powerful and flexible tool that can be used to model dependent data with differently distributed marginals in fields like finance, actuarial science and more. Supplementary MaterialsThe following are available online at https://www.mdpi.com/article/10.3390/sym13061059/s1, Figure S1: Gaussian copula and examples of skew-normal and skew t-copulas when the components of α have small values, Figure S2: Gaussian copula and examples of skew-normal and skew t-copulas when the components of α have relatively high values, Figure S3: Gaussian copula and examples of skew-normal and skew t-copulas when the components of α have different magnitude, Figure S4: Gaussian copula and examples of skew-normal and skew t-copulas when components of α are equal.Author ContributionsConceptualization, T.K. and M.K.; methodology, T.K., M.K. and A.S.; simulations and data analysis, A.S.; writing—original draft preparation, T.K.; writing—review and editing, T.K., M.K. and A.S.; visualization, A.S. All authors have read and agreed to the published version of the manuscript.FundingThis work was supported by the Estonian Research Council (grant PRG 1197).AcknowledgmentsThe authors are thankful to the editor and three anonymous referees for their constructive feedback that certainly improved the quality of the paper.Conflicts of InterestThe authors declare no conflict of interest. Appendix A Appendix A.1. Proof of Lemma 2Proof. We have to express the third and fourth central moments through raw moments. m ¯ 3 ( x ) = E [ ( x − E x ) ( x − E x ) ′ ⊗ ( x − E x ) ] = E [ ( x x ′ − E x x ′ − x E x + E x E x ) ⊗ ( x − E x ) ] = E [ x x ′ ⊗ x − x x ′ ⊗ E x − E x x ′ ⊗ x + E x x ′ ⊗ E x − x E x ′ ⊗ x + x E x ′ ⊗ E x + E x E x ′ ⊗ x − E x E x ′ ⊗ E x ] = E ( x x ′ ⊗ x ) − E ( x x ′ ) ⊗ E x − E x E ( x ′ ⊗ x ) + E x E x ′ ⊗ E x − E ( x E x ′ ⊗ x ) + E x E x ′ ⊗ E x = m 3 ( x ) − m 2 ( x ) ⊗ E x − E x ⊗ m 2 ( x ) + 2 E x E x ′ ⊗ E x − E x ′ ⊗ E ( x ⊗ x ) = m 3 ( x ) − m 2 ( x ) ⊗ E x − E x ⊗ m 2 ( x ) + 2 E x E x ′ ⊗ E x − E x ′ ⊗ vec m 2 ( x ) . Finally m ¯ 3 ( x ) = m 3 ( x ) − ( I p 2 + K p , p ) ( m 2 ( x ) ⊗ E x ) − vec m 2 ( x ) E x ′ + 2 E x E x ′ ⊗ E x . Derivation of the fourth central moment is longer. We can use the intermediate expression from derivation of m ¯ 3 ( x ) after the third equality sign above m ¯ 4 ( x ) = E [ ( x − E x ) ( x − E x ) ′ ⊗ ( x − E x ) ( x − E x ) ′ ] = E { [ x x ′ ⊗ x − x x ′ ⊗ E x − E x x ′ ⊗ x + E x x ′ ⊗ E x − x E x ′ ⊗ x + x E x ′ ⊗ E x + E x E x ′ ⊗ x − E x E x ′ ⊗ E x ] ⊗ x ′ } − E { [ x x ′ ⊗ x − x x ′ ⊗ E x − E x x ′ ⊗ x + E x x ′ ⊗ E x − x E x ′ ⊗ x + x E x ′ ⊗ E x + E x E x ′ ⊗ x − E x E x ′ ⊗ E x ] ⊗ E x ′ } = m 4 ( x ) − m ¯ 3 ( x ) ⊗ E x ′ − E ( x x ′ ⊗ E x ⊗ x ′ ) − E ( E x x ′ ⊗ x ⊗ x ′ ) + E ( E x x ′ ⊗ E x ⊗ x ′ ) − E ( x E x ′ ⊗ x ⊗ x ′ ) + E ( x E x ′ ⊗ E x ⊗ x ′ ) + E x E x ′ ⊗ m 2 ( x ) − E x E x ′ ⊗ E x ⊗ E x ′ . The expectations in the last equality can be presented through moments when we use the property of the Kronecker product for vectors a b ′ = a ⊗ b ′ = b ′ ⊗ a E ( x x ′ ⊗ E x x ′ ) = E ( x ⊗ x ′ ⊗ E x ⊗ x ′ ) = E ( x ′ ⊗ x ⊗ x ′ ⊗ E x ′ ) = m 3 ′ ( x ) ⊗ E x ; E ( E x x ′ ⊗ x x ′ ) = E x ⊗ m 3 ′ ( x ) ; E ( E x x ′ ⊗ E x x ′ ) = E ( E x ⊗ x ′ ⊗ x ′ ⊗ E x ) = E x ⊗ vec ′ m 2 ( x ) ⊗ E x ; E ( x E x ′ ⊗ x x ′ ) = E ( x ⊗ E x ′ ⊗ x ⊗ x ′ ) = E ( E x ′ ⊗ x ⊗ x ′ ⊗ x ) = E x ′ ⊗ m 3 ( x ) ; E ( x E x ′ ⊗ E x x ′ ) = E ( x ⊗ E x ′ ⊗ E x ⊗ x ′ ) = E ( E x ′ ⊗ x ⊗ x ′ ⊗ E x ′ ) = E x ′ ⊗ m 2 ( x ) ⊗ E x . After plugging in the expression of m ¯ 3 ( x ) we have m ¯ 4 ( x ) = m 4 ( x ) − m 3 ( x ) ⊗ E x ′ + m 2 ( x ) ⊗ E x E x ′ + E x ⊗ m 2 ( x ) ⊗ E x ′ + E x ′ ⊗ vec m 2 ( x ) ⊗ E x ′ − 3 E x E x ′ ⊗ E x E x ′ − m 3 ′ ( x ) ⊗ E x − E x ⊗ m 3 ′ ( x ) + E x ⊗ vec ′ m 2 ( x ) ⊗ E x − E x ′ ⊗ m 3 ( x ) + E x ′ ⊗ m 2 ( x ) ⊗ E x + E x E x ′ ⊗ m 2 ( x ) . Applying notation in Lemma 2 A ( x ) = E x ′ ⊗ m 2 ( x ) ⊗ E x ; B ( x ) = E x ′ ⊗ vec m 2 ( x ) ⊗ E x ′ ; C ( x ) = m 3 ( x ) ⊗ E x ′ we get the statement after using property 5 of the Kronecker product m ¯ 4 ( x ) = m 4 ( x ) − C ( x ) ( I p 2 + K p , p ) − ( I p 2 + K p , p ) ( C ( x ) ) ′ − 3 E x E x ′ ⊗ E x E x ′ + A ( x ) + B ( x ) + ( A ( x ) + B ( x ) ) ′ + m 2 ( x ) ⊗ E x E x ′ + E x E x ′ ⊗ m 2 ( x ) . □ Appendix A.2. Proof of Theorem 3Proof. Let us start from Expression (18) for α ^ : α ^ = Σ ^ − 1 x ¯ M , where M is given in (17), which, taking (15) into account, can also be expressed as follows: M = ν π Γ ν − 1 2 Γ ν 2 2 − x ¯ ′ Σ ^ − 1 x ¯ . Let us now derive the needed derivatives step by step. d α ^ d x ¯ = d ( Σ ^ − 1 x ¯ M − 1 2 ) d x ¯ = M − 1 2 d ( Σ ^ − 1 x ¯ ) d x ¯ + Σ ^ − 1 x ¯ d M − 1 2 d x ¯ . Let us find the two derivatives in (A1). (a)The first derivative can be expressed as d ( Σ ^ − 1 x ¯ ) d x ¯ = Σ ^ − 1 + ( x ¯ ′ ⊗ I p ) d Σ ^ − 1 d x ¯ . Now the derivative d Σ ^ − 1 d x ¯ can be found as d Σ ^ − 1 d x ¯ = d Σ ^ − 1 d Σ ^ d Σ ^ d x ¯ = − ( Σ ^ − 1 ⊗ Σ ^ − 1 ) d ( ν − 2 ν ( S + x ¯ x ¯ ′ ) ) d x ¯ = − ν − 2 ν Σ ^ − 1 ⊗ Σ ^ − 1 x ¯ ⊗ I p + I p ⊗ x ¯ = − ν − 2 ν Σ ^ − 1 x ¯ ⊗ Σ ^ − 1 + Σ ^ − 1 ⊗ Σ ^ − 1 x ¯ . The Equations (A2) and (A3) imply that d ( Σ ^ − 1 x ¯ ) d x ¯ = Σ ^ − 1 − ( x ¯ ′ ⊗ I p ) ν − 2 ν Σ ^ − 1 x ¯ ⊗ Σ ^ − 1 + Σ ^ − 1 ⊗ Σ ^ − 1 x ¯ = Σ ^ − 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ Σ ^ − 1 + x ¯ ′ Σ ^ − 1 ⊗ Σ ^ − 1 x ¯ = Σ ^ − 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ Σ ^ − 1 + Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 . (b)The second derivative in (A1) is d M − 1 2 d x ¯ = d M − 1 2 d M d M d x ¯ = − 1 2 M − 3 2 d ( − x ¯ ′ Σ ^ − 1 x ¯ ) x ¯ = 1 2 M − 3 2 x ¯ ′ Σ ^ − 1 + x ¯ d ( Σ ^ − 1 x ¯ ) d x ¯ = 1 2 M − 3 2 x ¯ ′ Σ ^ − 1 + x ¯ ′ Σ ^ − 1 − x ¯ ′ ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ Σ ^ − 1 + Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 = 1 2 M − 3 2 2 x ¯ ′ Σ ^ − 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 + x ¯ ′ Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 = M − 3 2 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 . Now, applying the results of parts (a) and (b) into Equation (A1), yields d α ^ d x ¯ = M − 1 2 Σ ^ − 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ Σ ^ − 1 + Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 + Σ ^ − 1 x ¯ M − 3 2 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 = M − 1 2 [ Σ ^ − 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ Σ ^ − 1 + Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 + M − 1 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 ] = M − 1 2 [ 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ Σ ^ − 1 + M − 1 1 − ν − 2 ν x ¯ ′ Σ ^ − 1 x ¯ + M Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 ] . Taking the derivative of α ^ by vec S is similar, but simpler. First of all, following Equality (A1), we can write d α ^ d vec S = M − 1 2 d ( Σ ^ − 1 x ¯ ) d vec S + Σ ^ − 1 x ¯ d M − 1 2 d vec S . Let us find the two derivatives in the last sum. (a)The first derivative can be expressed as d ( Σ ^ − 1 x ¯ ) d vec S = ( x ¯ ′ ⊗ I p ) d Σ ^ − 1 d Σ ^ d Σ ^ d vec S = − ν − 2 ν ( x ¯ ′ Σ ^ − 1 ⊗ Σ ^ − 1 ) , (b)The second derivative is d M − 1 2 d vec S = d M − 1 2 d M d M d vec S = − 1 2 M − 3 2 d ( − x ¯ ′ Σ ^ − 1 x ¯ ) d vec S = − 1 2 M − 3 2 ( x ¯ ′ ⊗ x ¯ ′ ) ( Σ ^ − 1 ⊗ Σ ^ − 1 ) ν − 2 ν = − ν − 2 2 ν M − 3 2 ( x ¯ ′ Σ ^ − 1 ⊗ x ¯ ′ Σ ^ − 1 ) . The full expression of the derivative d α ^ d vec S is d α ^ d vec S = − ν − 2 ν M − 1 2 ( x ¯ ′ Σ ^ − 1 ⊗ Σ ^ − 1 ) + 1 2 M − 1 ( Σ ^ − 1 x ¯ x ¯ ′ Σ ^ − 1 ⊗ x ¯ ′ Σ ^ − 1 ) . □ Appendix A.3. Parameter Sets Kollo, T.; Käärik, M.; Selart, A. Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula. Symmetry 2021, 13, 1059. https://doi.org/10.3390/sym13061059 AMA StyleKollo T, Käärik M, Selart A. Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula. Symmetry. 2021; 13(6):1059. https://doi.org/10.3390/sym13061059 Chicago/Turabian StyleKollo, Tõnu, Meelis Käärik, and Anne Selart. 2021. "Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula" Symmetry 13, no. 6: 1059. https://doi.org/10.3390/sym13061059 Find Other Styles Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here. Article Metrics No No Article Access Statistics For more information on the journal statistics, click here. Multiple requests from the same IP address are counted as one view. Supplementary Material Supplementary File 1:ZIP-Document (ZIP, 3211 KiB) clear Zoom | Orient | As Lines | As Sticks | As Cartoon | As Surface | Previous Scene | Next Scene Cite Export citation file: BibTeX | EndNote | RIS MDPI and ACS StyleKollo, T.; Käärik, M.; Selart, A. Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula. Symmetry 2021, 13, 1059. https://doi.org/10.3390/sym13061059 AMA StyleKollo T, Käärik M, Selart A. Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula. Symmetry. 2021; 13(6):1059. https://doi.org/10.3390/sym13061059 Chicago/Turabian StyleKollo, Tõnu, Meelis Käärik, and Anne Selart. 2021. "Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula" Symmetry 13, no. 6: 1059. https://doi.org/10.3390/sym13061059 Find Other Styles Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here. clear Symmetry, EISSN 2073-8994, Published by MDPI RSS Content Alert Further Information Article Processing Charges Pay an Invoice Open Access Policy Contact MDPI Jobs at MDPI Guidelines For Authors For Reviewers For Editors For Librarians For Publishers For Societies For Conference Organizers MDPI Initiatives Sciforum MDPI Books Preprints.org Scilit SciProfiles Encyclopedia JAMS Proceedings Series Follow MDPI LinkedIn Facebook Twitter![]() |
CopyRight 2018-2019 实验室设备网 版权所有 |