Processing math: 86%
Research article

The generalized scale mixtures of asymmetric generalized normal distributions with application to stock data

  • Received: 11 August 2023 Revised: 23 November 2023 Accepted: 30 November 2023 Published: 07 December 2023
  • MSC : 62E15, 62F10, 62G30, 62P05

  • In this paper, we introduced a family of distributions with a very flexible shape named generalized scale mixtures of generalized asymmetric normal distributions (GSMAGN). We investigated the main properties of the new family including moments, skewness, kurtosis coefficients and order statistics. A variant of the expectation maximization (EM)-type algorithm was established by combining the proflie likihood approach (PLA) with the classical expectation conditional maximization (ECM) algorithm for parameter estimation of this model. This approach with analytical expressions in the E-step and tractable M-step can greatly improve the computational speed and efficiency of the algorithm. The performance of the proposed algorithm was assessed by some simulation studies. The feasibility of the proposed methodology was illustrated through two real datasets.

    Citation: Ruijie Guan, Aidi Liu, Weihu Cheng. The generalized scale mixtures of asymmetric generalized normal distributions with application to stock data[J]. AIMS Mathematics, 2024, 9(1): 1291-1322. doi: 10.3934/math.2024064

    Related Papers:

    [1] Pingyun Li, Chuancun Yin . Tail risk measures with application for mixtures of elliptical distributions. AIMS Mathematics, 2022, 7(5): 8802-8821. doi: 10.3934/math.2022491
    [2] Essam A. Ahmed, Laila A. Al-Essa . Inference of stress-strength reliability based on adaptive progressive type-Ⅱ censing from Chen distribution with application to carbon fiber data. AIMS Mathematics, 2024, 9(8): 20482-20515. doi: 10.3934/math.2024996
    [3] Guangshuai Zhou, Chuancun Yin . Family of extended mean mixtures of multivariate normal distributions: Properties, inference and applications. AIMS Mathematics, 2022, 7(7): 12390-12414. doi: 10.3934/math.2022688
    [4] Mashael A. Alshehri, Mohamed Kayid . Copulas generated by mixtures of weighted distributions. AIMS Mathematics, 2022, 7(5): 8953-8974. doi: 10.3934/math.2022499
    [5] Alanazi Talal Abdulrahman, Khudhayr A. Rashedi, Tariq S. Alshammari, Eslam Hussam, Amirah Saeed Alharthi, Ramlah H Albayyat . A new extension of the Rayleigh distribution: Methodology, classical, and Bayes estimation, with application to industrial data. AIMS Mathematics, 2025, 10(2): 3710-3733. doi: 10.3934/math.2025172
    [6] Muhammad Aslam, Mohammed Albassam . Neutrosophic geometric distribution: Data generation under uncertainty and practical applications. AIMS Mathematics, 2024, 9(6): 16436-16452. doi: 10.3934/math.2024796
    [7] Mohamed S. Eliwa, Essam A. Ahmed . Reliability analysis of constant partially accelerated life tests under progressive first failure type-II censored data from Lomax model: EM and MCMC algorithms. AIMS Mathematics, 2023, 8(1): 29-60. doi: 10.3934/math.2023002
    [8] Tabassum Naz Sindhu, Zawar Hussain, Naif Alotaibi, Taseer Muhammad . Estimation method of mixture distribution and modeling of COVID-19 pandemic. AIMS Mathematics, 2022, 7(6): 9926-9956. doi: 10.3934/math.2022554
    [9] H. M. Barakat, M. A. Alawady, I. A. Husseiny, M. Nagy, A. H. Mansi, M. O. Mohamed . Bivariate Epanechnikov-exponential distribution: statistical properties, reliability measures, and applications to computer science data. AIMS Mathematics, 2024, 9(11): 32299-32327. doi: 10.3934/math.20241550
    [10] Areej M. AL-Zaydi . On concomitants of generalized order statistics arising from bivariate generalized Weibull distribution and its application in estimation. AIMS Mathematics, 2024, 9(8): 22002-22021. doi: 10.3934/math.20241069
  • In this paper, we introduced a family of distributions with a very flexible shape named generalized scale mixtures of generalized asymmetric normal distributions (GSMAGN). We investigated the main properties of the new family including moments, skewness, kurtosis coefficients and order statistics. A variant of the expectation maximization (EM)-type algorithm was established by combining the proflie likihood approach (PLA) with the classical expectation conditional maximization (ECM) algorithm for parameter estimation of this model. This approach with analytical expressions in the E-step and tractable M-step can greatly improve the computational speed and efficiency of the algorithm. The performance of the proposed algorithm was assessed by some simulation studies. The feasibility of the proposed methodology was illustrated through two real datasets.



    In the wake of the swift progression of technology over recent decades, a noteworthy proliferation of diverse data types has ensued. A salient feature inherent in many of these datasets is their proclivity toward heavy tails and marked skewness, thereby rendering conventional univariate distributions inadequately applicable. Consequently, considerable scholarly focus has been directed toward the formulation of robust distributions adept at effectively modeling such intricately structured data.

    Andrews and Mallows (1974) [1] and West (1987) [2] proposed a scale mixtures of normal (SMN) distributions with the following stochastic representation

    Y=μ+U12Z, (1.1)

    where μR is the location parameter, ZN(0,σ2), and U is a positive random variable independent with Z. The SMN distributions provide great flexibility in modeling not only asymmetric but also heavy-tailed datasets. Although SMN distributions are attractive, when facing asymmetric population with strong skewness and heavy-tailed behavior simultaneously in practice, SMN distribution appears overwhelming. To cope with this problem, Branco and Dey (2001) [3] constructed the scale mixtures of skew-normal (SMSN) distribution by replacing the normal distribution of Eq (1.1) to the skew-normal distribution. This class of distributions not only contains the entire family of SMN distributions, but also accounts for asymmetry and heavy tails simultaneously. Ferreira, Lachos, and Bolfarine (2020)[4] pointed out that scale mixtures of the skew normal distribution can offer much-needed flexibility by combining both skewness and heavy tails. Basso (2010) [5] developed an EM-type algorithm to obtain the maximum likelihood estimation for the SMSN family and derived the observed information matrix. Kim and Genton (2011) [6] obtained the characteristic functions for the SMSN family by means of the simple stochastic representation. Lin et al. contributed a great deal of important work for scale mixture models based on the skew-normal distribution, which has been found to be effective in the treatment of heavy-tailed data involving asymmetric behaviors [7,8,9]. Mahdavi et al. [10] obtained the maximum likelihood estimation for scale-shape mixtures of flexible generalized skew normal distributions via EM-type algorithms.

    The utilization of the skew normal distribution in the aforementioned study is rooted in the seminal work of Azzalini (1985) [11]. However, it is essential to recognize the limitations of this distribution when confronted with highly skewed and leptokurtic data, as its ability to accurately fit such data may be constrained by the range of the skewness and kurtosis coefficients. To explore more recent advancements in the field of skew normal distribution, we recommend consulting the works of [12,13,14,15,16]. Acknowledging these limitations, it becomes imperative to develop alternative forms of skew normal distributions that can accommodate a broader spectrum of diverse data. One approach in this direction was introduced by Fernandez and Steel (1988) [17], who incorporated a skew parameter into a symmetric distribution. This innovation enabled the regulation of the probability density functions in both the positive and negative orthants, offering enhanced flexibility. Another notable contribution is the work of Zhu (2009) [18], who proposed the asymmetric exponential power distribution by introducing a skew parameter into the exponential power distribution. Zhu further demonstrated the consistency, asymptotic normality, and efficiency of the maximum likelihood estimation for this distribution. These advancements signify important strides in the development of alternative skew normal distributions that can better accommodate complex data patterns.

    In this work, we have introduced an asymmetric generalized normal (AGN) distribution by incorporating a skew parameter into the generalized normal (GN) distribution [19]. This extension provides a flexible framework for modeling various types of data exhibiting asymmetric behavior and heavy-tailed characteristics. However, in many practical scenarios, the observed data often exhibits strongly heavy tails and significant skewness. To address this challenge, we turn our attention to the family of generalized scale mixtures of the asymmetric generalized normal (GSMAGN) distributions. This family represents a generalization of traditional scale mixtures, wherein we replace the power 12 of the scale factor in Eq (1.1) with any power 1β, with β>0. By employing this generalization, we gain greater flexibility in capturing the complexities present in the data and better accommodate the wide range of tail and skewness behaviors encountered in practical applications.

    The rest of this paper will be organized as follows. In section two, some definitions and essential properties of the GSMAGN family are summarized. In section three, the explicit expression for moments of order statistics from the GSMAGN family are obtained under independent identically distributed (IID) cases. In section four, an efficient and tractable EM-type algorithm to obtain the maximum likelihood estimation of parameters is established. In section five, the consistency properties of the estimates and the estimation of standard errors are demonstrated. In section six, a simulation study is implemented to assess the performance of the proposed algorithm. In section seven, two real dataset analyses are presented to demonstrate the feasibility of the proposed model. Finally, conclusions are noted in section eight.

    Definition 2.1. Let the random variable X with the following probability density function (pdf) be

    fAGN(x;μ,σ,κ,β)=κβ21/β(1+κ2)Γ(1/β)σexp{12(κsign(xμ)|xμ|σ)β},xR, (2.1)

    where Γ() denotes the gamma function, μR is a location parameter, σ>0 is a scale parameter, κ>0 is the skewness parameter and β>0 is a shape parameter. We say that the random variable X follows an AGN distribution, denoted by XAGN(μ,σ,κ,β). In particular, if κ=1, the AGN distribution becomes a GN distribution [19] denoted by GN(μ,σ,β).

    Definition 2.2. A random variable Y is in the GSMAGN family if its stochastic representation is given by

    Y=μ+σ[k(U)]1/βX, (2.2)

    where XAGN(0,1,κ,β), μ is the location parameter, k() is a weight function, and U is a random variable.

    If U=u, then YU=uAGN(μ,[k(U)]1/βσ,κ,β) and the pdf of Y is given by

    fGSMAGN(y;μ,σ,κ,α,β)=+0fAGN(y;μ,[k(u)]1/βσ,κ,β)dH(u;α), (2.3)

    where H(;α) is the cumulative distribution function (CDF) of U indexed by parameter α. To imbue the proposed distributions that have some compelling mathematical properties to facilitate statistical inference, usually the weight function will be chosen for k(u)=1u.

    Proposition 2.1. If YGSMAGN(μ,σ,κ,α,β), then the stochastic representation of Y can be written as follows

    Y=μ+σU1βI(2Z)1β, (2.4)

    where UH(;α), ZGamma(1/β,1), I is a discrete random variable with probability mass function fI=κ21+κ21{I=κ}+11+κ21{I=1/κ}, and U,I,Z are independent of each other.

    Proposition 2.2. If YGSMAGN(μ,σ,κ,α,β), then

    E(Ym)=mi=0(mi)σiμmi2iβΓ[(i+1)/β]Γ(1/β)11+κ2[1κi+(1)iκ2+i]E(Uiβ),mN+.

    In particular, the expectation and variance of the random variable Y are respectively given by

    μY=E(Y)=μ+σ21βΓ(2/β)Γ(1/β)11+κ2(1κκ3)E(U1β),Var(Y)=σ222β{1+κ6κ2(1+κ2)Γ(3/β)Γ(1/β)E(U2β)(1κ2)2κ2[Γ(2/β)Γ(1/β)E(U1β)]2}.

    Proof. See Appendix Part 1.

    (ⅰ) AGN distribution

    In this case U=1, the CDF of Y is

    FAGN(y;μ,σ,κ,α,β)={κ2(1+κ2)[1γ(1β,12(μyκσ)β)],ify<μ,κ21+κ2+1(1+κ2)γ(1β,12(κ(yμ)σ)β),ifyμ,

    where γ(α;x)=1Γ(α)x0tα1etdt denotes the lower incomplete gamma function.

    (ⅱ) Asymmetric generalized t distribution (AGT)

    In this case we consider UGamma(α,β) (α=(α,β) for H(u;α)) and the pdf of Y is given by

    fAGT(y;μ,σ,κ,α,β)=2κ(2/β)1+1β(1+κ2)B(α,1/β)σ[1+β2(κsign(yμ)|yμ|σ)β](α+1/β),yR,

    and the CDF of Y can be represented as

    FAGT(y;μ,σ,κ,α,β)={κ2(1+κ2)[1I(1β,α;h1(y))],ify<μ,κ21+κ2+1(1+κ2)I(1β,α;h2(y)),ifyμ,

    where

    h1(y)=β(μy)β2κβσβ+β(μy)β,h2(y)=βκβ(yμ)β2σβ+βκβ(yμ)β.

    B(,) denotes the beta function, I(a,b;x)=1B(a,b)x0ta1(1t)b1dt denotes the incomplete beta function, and α,β>0 are two shape parameters, then we say that the random variable Y follows AGT distribution, denoted by YAGT(μ,σ,κ,α,β).

    Corollary 2.1. If YAGT(μ,σ,κ,α,β), then

    E(Ym)=mi=0(mi)σiμmi(2β)iβΓ(αi/β)Γ[(i+1)/β]Γ(α)Γ(1/β)11+κ2[1κi+(1)iκ2+i],αβ>m.

    In particular, the expectation and variance of Y are respectively given by

    μY=E(Y)=μ+σ(2β)1βΓ(α1/β)Γ(2/β)Γ(α)Γ(1/β),Var(Y)=σ2(2β)2β{1+κ6κ2(1+κ2)Γ(α2/β)Γ(3/β)Γ(α)Γ(1/β)(1κ2)2κ2[Γ(α1/β)Γ(2/β)Γ(α)Γ(1/β)]2}.

    Proposition 2.3. If YAGT(μ,σ,κ,α,β), then U|Y=yGamma(α+1β,[1β+12(κsign(yμ)|yμ|σ)β]1).

    Proof. See Appendix Part 1.

    (ⅲ) Asymmetric generalized slash distribution (AGSL)

    In this case, we have UBeta(α,1) and the pdf of Y is given by

    fAGSL(y;μ,σ,κ,α,β)=καβΓ(α+1/β)21/β(1+κ2)Γ(1/β)σ[12(κsign(yμ)|yμ|σ)β](α+1/β)×γ[α+1β;12(κsign(yμ)|yμ|σ)β],yR.

    All parameters play the same role as in AGT distribution, then we say that the random variable Y follows an AGSL distribution, denoted by YAGSL(μ,σ,κ,α,β).

    Proposition 2.4. If YAGSL(μ,σ,κ,α,β), then the pdf of UY=y is given by

    fUY(uy)=[12(κsign(yμ)|yμ|σ)β](α+1/β)Γ(α+1/β)γ[α+1β;12(κsign(yμ)|yμ|σ)β]uα+1/β1exp[u2(κsign(yμ)|yμ|σ)β],0<u<1.

    Proof. See Appendix Part 1.

    (ⅳ) Contaminated asymmetric generalized normal distribution (CAGN)

    In this case, U is a discrete random variable taking one of two states with the following probability mass function

    h(u;α)=α1(u=λ)+(1α)1(u=1),0<α<1,0<λ1,i.e.,α=(α,λ). (2.5)

    It follows immediately that the pdf of the random variable Y is given by

    fCAGN(y;μ,σ,κ,λ,α,β)=αfAGN(y;μ,λ1/βσ,κ,β)+(1α)fAGN(y;μ,σ,κ,β),

    and the CDF of Y is

    FCAGN(y;μ,σ,κ,λ,α,β)={κ2(1+κ2)κ2(1+κ2)[αγ(1β,h1(y))+(1α)γ(1β,h2(y))],ify<μ,κ2(1+κ2)+1(1+κ2)[αγ(1β,h1(y))+(1α)γ(1β,h2(y))],ify>μ,

    where

    h1(y)=λ2(κsign(yμ)|yμ|σ)β,h2(y)=12(κsign(yμ)|yμ|σ)β.

    Proposition 2.5. If YCAGN(μ,σ,κ,λ,α,β), the pdf of UY=y for 0<λ<1 is given by

    P(U=λY=y)=αfAGN(y;μ,λ1/βσ,κ,β)fCAGN(y;μ,σ,κ,λ,α,β),P(U=1Y=y)=(1α)fAGN(y;μ,σ,κ,β)fCAGN(y;μ,σ,κ,λ,α,β).

    Proof. See Appendix Part 1.

    Figure 1 illustrates the pdf of the GSMAGN family under three distinct scenarios, providing valuable insights into the impact of the shape parameters (α and β) on the pdf's shape. The chosen values of α and β exert a decisive influence on the characteristics of the density. When α and β assume smaller values (Figure 1a and 1c), the resulting densities exhibit longer tails and greater leptokurtosis. Conversely, larger values of α and β (Figure 1b) lead to contrasting effects. Notably, the shape parameter β has a predominant impact on the distribution's shape, while the parameter α governs the behavior of the tail. Moreover, all three subfigures demonstrate that the AGSL distribution possesses thicker tails compared to the other three distributions. Consequently, the AGSL distribution exhibits a favorable advantage in modeling observations characterized by skewness and heavy tails.

    Figure 1.  PDF of GSMAGN family for different parameter values.

    In the following, we provide the moments of order statistics generating from GSMAGN(0,1,κ,α,β) distribution under IID case.

    Proposition 3.1. Let X1,X2,,Xn be a random sample of size n from AGN(0,1,κ,β), and X1:n,X2:n,,Xn:n are corresponding order statistics. The mth moments of the largest order statistic is given by

    E(Xmn:n)=nn1i=0j=0(n1i)2mβ(1+κ2)n[Γ(1/β)]i+1[(1)i+mκ2n+m+κ2(ni1)m]b(i)jΓ(m+i+1β+j),

    where

    bj=(1)jj!(1β+j),b(s)j=jh=0b(s1)hbjh,b(1)j=bj,b(0)j={1,j=0,0,j>0,,j=0,1,.

    Proof. See Appendix Part 2.

    The Lauricella function of type A (see, for example, Exton(1978)[20]) is defined as

    F(n)A(a,b1,;c1,,cn;x1,,xn)=m1=0mn=0am1++mn(b1)m1(bn)mn(c1)m1(cn)mnxm11xmnnm1!mn!,

    where (f)k=f(f+1)(f+k1) denotes the ascending factorial.

    We also acquire the moments of the largest order statistic by means of the Lauricella function of type A.

    Proposition 3.2. If XAGN(0,1,κ,β), and X1:n,X2:n,,Xn:n are the order statistics, then the mth moments of the largest order statistic is given by

    E(Xmn:n)=nn1i=0(n1i)βiΓ(m+i+1β)2mβ(1+κ2)n[Γ(1/β)]i+1[(1)i+mκ2n+m+κ2(ni1)m]×F(i)A(m+i+1β,1β,,1β;1β+1,,1β+1;1,,1).

    Proof. See Appendix Part 2.

    Proposition 3.3. If XAGT(0,1,κ,α,β), 1β is an integer, and X1:n,X2:n,,Xn:n are the order statistics, then the mth moments of the largest order statistic Xn:n is given by

    E(Xmn:n)=n(2/β)mβ(1+κ2)n[(1)mκ2n+m[B(α,1/β)]n(1/β1)(n1)v=0Tv(1/β,n1)×B(m+1β,αn+vmβ)+n1j=0(1)j(n1j)1κm[B(α,1/β)]j(1/β1)jv=0Tv(1/β,j)×B(m+1β,α(j+1)+vmβ)],

    where Tv(1/β,0)=1, Tv(1/β,1)=Cv,v=0,1,,(1/β), and Cv=(1)v(1/β1v)1α+v. For n>2, Tv(1β,n1)=vi=0CiTvi(1β,n2),v=0,1,,(1/β)(n1).

    Proof. See Appendix Part 2.

    The formula for computing the m-th moments of the order statistics given in Proposition 3.3 is only applicable to the case that 1/β is an integer. We derive a general representation of the m-th moments of order statistics without restrictions on β.

    Introduced by Exton (1978) [20], the generalized Kampe de Feriet function is defined by

    FA:BC:D((a):(b1);;(bn);(c):(d1);,(dn);x1,,xn)=m1=0mn=0((a))m1++mn((b1))m1((bn))mn((c))m1++mn((d1))m1((dn))mn×xm11xmnnm1!mn!,

    where a=(a1,a2,,aA), bi=(bi,1,bi,2,,bi,B), c=(c1,,cC), di=(di,1,,di,D), for i=1,2,,n and ((f))k=((f1,f2,,fp))k=(f1)k(fp)k, (fi)k=fi(fi+1)(fi+k1).

    By using the the generalized Kampe de Feriet function, we derive the following proposition:

    Proposition 3.4. If 1/β is a real non-integer, the m-th moments of order statistics Xn:n from AGT(0,1,κ,α,β) can be computed by the following convergent expression

    E(Xmn:n)=[nJ(m,n1,κ)+n1j=0(1)m+j(n1j)nJ(m,(j+1),1/κ)],

    where

    J(m,n1,κ)=(1)mk2n+m(2/β)mβB(m+1β,αnmβ)(1+κ2)n[B(α,1/β)]nan1F1:21:1((αnmβ):(11β,α);;(11β,α);(αn+1β):(α+1);,(α+1);1,,1).

    Proof. See Appendix Part 2.

    The relationship between order statistics from CAGN distribution and AGN distribution can be stated as follows

    Proposition 3.5. If YCAGN(0,1,κ,λ,α,β), XAGN(0,1,κ,α,β), then we have

    E(Yr:n)=P(U=1)E(Xr:n)+P(U=λ)λ1βE(Xr:n)=[α+λ1β(1α)]E(Xr:n).

    Proof. See Appendix Part 2.

    In this section, an EM-type algorithm to determine the MLEs for the parameters of the GSMAGN(μ,σ,κ,α,β) family is established.

    Originally proposed by Dempster et al. (1977) [21], the EM algorithm has garnered significant attention as a powerful tool for handling estimation challenges in situations involving incomplete data. This algorithm offers numerous exceptional advantages. For a more comprehensive understanding of this subject, we highly recommend referring to the monograph by Lange (2013) [22]. A notable extension of the EM algorithm is the ECM algorithm introduced by Meng and Rubin (1993) [23]. The ECM algorithm not only preserves the desirable property of monotonic convergence exhibited by the Q-function in the classical EM algorithm but also facilitates faster convergence rates.

    Invoking Proposition 2.1, the hierarchical representation for the GSMAGN model can be described as

    YjUj=ujAGN(μ,u1/βjσ,κ,β),UjH(uj;α).

    Let Y=(y1,,yn)T denote the observed data, U=(u1,,un)T on behalf of the corresponding latent variable. Combining Y, U together we obtain the complete data denoted by W=(Y,U). Let Θ=(μ,σ,κ,α)T represent the vector of parameters, then the complete log-likelihood function can be expressed as follows

    lnL(ΘW)=n[lnκln(1+κ2)+lnβ(1/β)ln2lnΓ(1/β)lnσ]+1βnj=1lnuj12nj=1(κsign(yjμ)|yjμ|σ)βuj+nj=1lnh(uj;α). (4.1)

    Prior to outlining the precise steps of the estimation algorithm, it is imperative to highlight a crucial aspect. We adopt the assumption that the shape parameter β remains constant, and the optimal value for this parameter can be determined through the PLA. By integrating the ECM algorithm with the PLA, the computational speed and overall efficiency of the algorithm are significantly enhanced. Consequently, under this assumption, the parameter vector θi is transformed into θi=(μi,σi,αi) throughout the subsequent sections of this paper.

    It is well known that the EM framework is an iterative algorithm consisting of two steps, namely, the expectation step (E-step) and the maximization step (M-step).

    E-step: Given the observed data set Y and parameter estimation values ˆΘ(h) of the h-th iteration, the aim of this step involves calculation of several necessary conditional expectations. The so-called Q-function is expressed by

    Q(ΘˆΘ(h))=E[lnL(ΘW)Y,ˆΘ(h)]=n[lnκln(1+κ2)+lnβ(1/β)ln2lnΓ(1/β)lnσ]+1βnj=1A2j(Y,ˆΘ(h))12nj=1(κsign(yjμ)|yjμ|σ)βA1j(Y,ˆΘ(h))+nj=1Bj(Y,ˆΘ(h)), (4.2)

    where A1j(Y,ˆΘ(h))=E[UjY,ˆΘ(h)], A2j(Y,ˆΘ(h))=E[lnUjY,ˆΘ(h)], and Bj(Y,ˆΘ(h))=E[lnh(Uj;α)Y,ˆΘ(h)].

    For different members of the GSMAGN family, the crucial conditional expectations are calculated as follows

    ● For AGT

    A1j(Y,ˆΘ(h))=(ˆα(h)+1/β)1/β+ˆc(h)j,A2j(Y,ˆΘ(h))=φ(ˆα(h)+1/β)ln(1/β+ˆc(h)j),Bj(Y,ˆΘ(h))=[αlnβ+lnΓ(α)]1βA1j(Y,ˆΘ(h))+(α1)A2j(Y,ˆΘ(h)),

    where ˆc(h)j=12(κsign(yjˆμ(h))|yjˆμ(h)|ˆσ(h))β, and φ() is the Digamma function.

    ● For AGSL

    A1j(Y,ˆΘ(h))=γ(ˆα(h)+1/β+1;ˆc(h)j)(ˆα(h)+1/β)γ(ˆα(h)+1/β;ˆc(h)j)ˆc(h)j,A2j(Y,ˆΘ(h))={ˆc(h)j0t(ˆα(h)+1/β1)etlntdtΓ(ˆα(h)+1/β)γ(ˆα(h)+1/β;ˆc(h)j)lnˆc(h)j},Bj(Y,ˆΘ(h))=lnα(α1)A2j(Y,ˆΘ(h)).

    ● For CAGN

    A1j(Y,ˆΘ(h))=1ˆα(h)+ˆα(h)[ˆλ(h)]1+1/βexp((1ˆλ(h))2ˆc(h)j)1ˆα(h)+ˆα(h)ˆλ(h)exp((1ˆλ(h)2)ˆc(h)j),A2j(Y,ˆΘ(h))=ˆα(h)[ˆλ(h)]1/βlnˆλ(h)ˆα(h)[ˆλ(h)]1/β+(1ˆα(h))exp((ˆλ(h)1)2ˆc(h)j),Bj(Y,ˆΘ(h))=lnαˆα(h)fAGN(yj;ˆμ(h),[ˆλ(h)]1/βˆσ(h),β)fCAGN(yj;ˆμ(h),ˆσ(h),ˆλ(h),ˆα(h),β)+ln(1α)(1ˆα(h))fAGN(yj;ˆμ(h),ˆσ(h),β)fCAGN(yj;ˆμ(h),ˆσ(h),,ˆλ(h),ˆα(h),β).

    In the EM algorithm, the M-step needs to maximize the conditional expectation obtained by the E-step and the suggested framework can be briefly described as follows

    (1) According to the theory of robust statistics [24], ˆμ(h+1) can be updated through the following expression

    ˆμ(h+1)=nj=1ωj(Y,ˆΘ(h))yjnl=1ωl(Y,ˆΘ(h)), (4.3)

    where ωj(Y,Θ)=κβsign(yjμ)|yjμ|β2A1j(Y,Θ),

    (2) Update ˆσ(h+1) by the following equation

    ˆσ(h+1)=(β2n)1β[nj=1ωj(Y,ˆΘ(h))|yjˆμ(h+1)|2]1β, (4.4)

    where ˆΘ(h)=(ˆμ(h+1),ˆσ(h),ˆκ(h),ˆα(h)).

    (3) Update ˆκ(h+1) by the following equation

    ˆκ(h+1)={nj=1([yjˆμ(h+1)])βA1j(Y,ˆΘ(h))nj=1([yjˆμ(h+1)]+)βA1j(Y,ˆΘ(h)))}1/2(β+1), (4.5)

    such that Θ(h)=(ˆμ(h+1),ˆσ(h+1),ˆκ(h),ˆα(h)) and [x]+=max{x,0}, [x]=[x]+.

    (4) Regarding ˆα(h+1)

    ● For AGT: Update ˆα(h+1) as the root of

    [φ(α)+lnβ]1nnj=1A2j(Y,ˆΘ(h))=0, (4.6)

    where ˆΘ(h)=(ˆμ(h+1),ˆσ(h+1),ˆκ(h+1),ˆα(h)).

    ● For AGSL: Update ˆα(h+1) by

    ˆα(h+1)=nnj=1A2j(Y,ˆΘ(h)). (4.7)

    ● For CAGN: Update ˆα(h+1) by

    ˆα(h+1)=nj=1ˆα(h)[ˆλ(h)]1/βˆα(h)[ˆλ(h)]1/β+(1ˆα(h))exp((ˆλ(h)1)2ˆc(h)j). (4.8)

    Remark 4.1. For AGN, an interative procedure presented below to update ˆβ(h+1) is also recommended

    ˆβ(h+1)=ˆβ(h)Q(ΘˆΘ(h))β(2Q(ΘˆΘ(h))β2)1, (4.9)

    where

    Q(ΘˆΘ(h))β=n[1β+ln2β2+φ(1/β)β2]nˆc(h)ij(ln2+1βlnˆc(h)ij),2Q(ΘˆΘ(h))β2=n[1β+2ln2β2+2φ(1/β)β3+2φ(1/β)β4]nˆc(h)j(ln2+2βlnˆc(h)j).

    For a prescribed value ε>0, if the value of the actual log-likelihood function L(YΘ) of two successive iterations statifies |L(YˆΘ(h+1))/L(YˆΘ(h))1|<ε, then the iterative process is broken up. In our study, the tolerance ε is set equal to 105.

    It is well known that the starting value plays a decisive role in the implementation of the EM-type algorithm. As a result, the following steps are recommended to find reasonable starting value.

    Notes on implementation

    ● Random sample generation: From the stochastic representation in Proposition 2.1, the following procedures are recommended to generate random samples from GSMAGN.

    Step 1: Generate UH(u;α) and set ZGamma(1/β,1).

    Step 2: Generate WU(0,1). If W11+κ2, set I=1κ; else set I=κ.

    Step 3: Set Y=μ+σU1βI(2Z)1β and return Y.

    ● Initialization: The initial value for ˆμ(0), ˆσ(0) and ˆκ(0) can be computed by

    ˆμ(0)=mode(Y),ˆσ(0)=nj=1(yjˆμ(0))2n,ˆκ(0)l=nj=1I(yj<ˆμ(0))(y)ni=1I(yj>ˆμ(0))(y).

    (1) For AGT and AGSL: Compute the initial guess ˆα(0) through the following equation

    ˆα(0)=argmaxαnj=1lnf(yj;ˆμ(0),ˆσ(0),ˆκ(0),α,β).

    (2) For CAGN: The initial value for α and λ are uniformly generated in (0,1).

    Choosing the value of β

    Step 1: Run the AGN model and obtain an estimate ˜β that can be referenced, then consider grids of values {ˇβ(1),,ˇβ(R)} that contain ˜β (the initial value for β can be obtained from the moment estimator).

    Step 2: Substitute each β(q) into the Q-function to acquire simultaneous estimation for the unknown parameters denoted by ˆθ(q)i via the ECM algorithm. After that, calculate the value of the corresponding log-likelihood function through Eq (4.1).

    Step 3: Repeat Step 2 and then interpolate on β values and derive ˘β which maximizes the trace of the value of the log-likelihood function.

    Lemma 5.1. If XdPθ(x)=f(x;θ)dμ(x),θ=(θ1,,θs),x=(x1,,xn) are a n identically distributed f(x;θ) random variable, then parameter space Θ is an open set in Rs. Assume that

    (1) If θ1θ2, then Pθ1Pθ2.

    (2) For x in the range of X, f(x,θ)>0, and has a continuous first partial derivative with respect to θ.

    (3) For θ0Θ, there exists a neighborhood Uθ0Θ and a positive function G(x) if θ,θ in Uθ0, we have lnf(x;θ)lnf(x;θ)∣≤G(x)θθ, and Eθ0[G(X)]<.

    For θ0Θ, when n, there exists a solution ˆθ to likelihood equation (a.s.)

    lnL(x;θ)θ:=ni=1lnf(xi;θ)θ,

    and the solution ˆθ is strongly consistent.

    Theorem 5.1. If YGSMAGN(μ,σ,κ,α,β), y=(y1,y2,,yn) is random sample from population Y, where θ˜Θ={(μ,σ,κ,α,β):μR,σR+,κR+,αR+,βR+}, the maximum likelihood estimator ˆθ of θ satisfies

    Pθ(limnˆθ=θ)=1,θ˜Θ.

    Proof. See Appendix Part 3.

    The observed information matrix of the GSMAGN distribution is computed by

    I0(Θy)=2L((Θy))/ΘΘT.

    It is well known that, under some regularity conditions, the covariance matrix of the maximum likelihood estimators ˆΘ can approximated by the inverse of I0(Θ|y). We evaluate

    I0(Θy)=ni=1ˆsiˆsTi. (5.1)

    Following the results of Basso[5], the individual score can be described as ˆsi=L(Θ|yi)Θ|Θ=ˆΘ, where ˆsi=(ˆsμ,ˆsσ,ˆsκ,ˆsα). The elements in ˆsi are given by

    ˆsi,μ=Dμ(f(yi;Θ))f(yi;Θ),ˆsi,σ=Dσ(f(yi;Θ))f(yi;Θ),ˆsi,κ=Dκ(f(yi;Θ))f(yi;Θ),ˆsi,α=Dν(f(yi;Θ))f(yi;Θ).

    Dμ(f(yi;Θ)) denotes the derivation of the density function with respect to μ, and Dσ(f(yi;Θ)), Dκ(f(yi;Θ)) and Dα(f(yi;Θ)) are defined by analogy. Let

    Mi,k=0u1β+kexp{12(κsign(yiμ)|yiμ|σ)βu}dH(u;α).

    After some algebraic computation, let Ai=κsign(yiμ)|yiμ|σ and we obtain

    si,μ=β2σsign(μyi)κsign(yiμ)Aiβ1Mi,1Mi,0,si,σ=1σ+β2σAiβMi,1Mi,0,si,κ=1κ2κ(1+κ2)β2κsign(yiμ)AiβMi,1Mi,0.

    Dα(f(yi;Θ)) depends on each particular case mentioned in section 2. The components of the score vector are given in Appendix Part 3.

    In this section, we conduct Monte Carlo simulations and the performance of the proposed algorithm is assessed via Bias and mean squared error (MSE) under different censoring schemes and parameter values. These two measures will be computed with 1500 replications at the same time and they are defned as follows:

    Bias(θi)=115001500j=1(ˆθiθi),MSE(θi)=115001500j=1(ˆθiθi)2.

    Without losing the generality, we consider sample sizes n=50,100,300,500,800,1000, and change the values of the shape parameter β. To be more specific, in order to study the influence of sample variability on the estimation effects of proposed method, we consider the cases where β=1.5 and β=2.5, and for each case of β, the remaining parameter (μ,σ,κ,α) is taken to be equal to (10, 2.5, 1.5, 3.5).

    Figures 2 and 3 exhibit the Bias and MSE plots for the parameter estimates of the AGN, AGT, AGSL, and CAGN distributions across varying sample sizes. Based on the observations drawn from these two pictures, the following findings can be elucidated:

    Figure 2.  Bias and MSE of (μ,σ,κ,β,α)(μ=10,σ=2.5,κ=1.5,β=1.5,α=3.5).
    Figure 3.  Bias and MSE of (μ,σ,κ,β,α)(μ=10,σ=2.5,κ=1.5,β=2.5,α=3.5).

    (1) It is evident that both the Bias and MSE of the parameters for the different sub-models (AGN, AGT, AGSL, CAGN) diminish and converge to zero as the sample sizes increase. This convergence implies that the estimates obtained through the proposed EM-type algorithm adhere to the desirable asymptotic properties, signifying consistency.

    (2) Notably, when employing the proposed EM-type algorithm, the parameters κ and scale demonstrate relatively minor Bias across all sample sizes. Additionally, the MSE of the κ parameter exhibits significant sensitivity to changes in sample size.

    (3) Furthermore, Figures 2 and 3 reveal that the estimated performance of the EM algorithm for the AGT and AGSL distributions closely approximates that of these two distributions.

    (4) Conversely, the results indicate that the estimation method faces the challenge of dependency on distribution complexity. Comparing the maximum likelihood estimates of the four distributions within the GSMAGN family, it becomes evident that the EM-type algorithm effectively estimates the parameters of the AGN distribution. However, for the CAGN distribution, which possesses a relatively larger number of parameters, the Bias and MSE of σ and β are relatively substantial, resulting in less stable estimates.

    Through a comparative analysis of Figures 2 and 3, it becomes apparent that as the value of β increases, the Bias and MSE of the parameter estimates decrease for all four distributions overall. This observation indicates that the EM-type algorithm for maximum likelihood estimation is more suitable for distributions with lower leptokurtosis.

    In order to facilitate the empirical investigation, datasets encompassing the S&P 500 Composite Index (S&P 500) and the Shanghai Stock Exchange Composite Index (SSEC) indices, spanning from January 2, 1998 to June 18, 2023 were diligently compiled. These datasets were acquired from the esteemed source, https://finance.yahoo.com. A total of 6,408 samples were drawn from the S&P 500 and SSEC indices at corresponding time points to enable a comprehensive comparative analysis. Within the scope of this study, the return (Rt) in period t is defined as Rt=100(lnPtlnPt1), where Pt (for t=1,2,,N) represents the level of the S&P 500 and SSEC indices at time t.

    As elucidated by Wen et al. (2022) [25], notwithstanding divergent summary statistics between the S&P 500 and SSEC indices, both datasets deviate from the assumptions associated with a normal distribution. Furthermore, both sets of data exhibit characteristics such as heavy-tailedness, leptokurtosis, and left-skewness.

    To assess the adequacy of the proposed GSMAGN family, fitting procedures were conducted for the AGN, AGT, AGSL, and CAGN distributions using the S&P 500 and SSEC datasets separately. The estimation of distribution parameters was executed using the EM-type algorithm, as expounded in section four. To evaluate the goodness of fit for the four distributions and various metrics, including observed log-likelihood (l), Akaike information criterion (AIC), Bayesian information criterion (BIC), and the Efficient determination criterion (EDC), were employed. The fitting results were consolidated and presented in Tables 1 (for the S&P 500 dataset) and 2 (for the SSEC dataset). The ultimate values of the shape parameter β of the GSMAGN family were determined by the trace of the PLA, as presented in Figures 4 and 5.

    Table 1.  Parameter estimates for the S&P 500 data set and the associated AIC, BIC, EDC, and log-likelihood (l) values for different distributions.
    Model μ σ κ α β γ λ ν AIC BIC EDC l
    AGN 0.113 0.755 1.056 –– –– 1.374 –– –– 19695.64 19722.7 19751.65 -9843.819
    AGT 0.105 1.592 1.049 3.212 1.319 –– –– –– 19215.9 19249.72 19285.91 -9602.951
    AGSL 0.0847 0.376 1.038 3.055 1.136 –– –– –– 19227.94 19261.77 19298.01 -9608.971
    CAGN 0.108 0.508 1.053 0.131 1.248 0.293 –– –– 19237.83 19278.43 19321.92 -9612.917
    SN 0.859 2.248 –– –– –– –– -0.999 –– 20908.13 20929.02 20950.77 -10451.36
    ST 0.258 0.577 –– –– –– –– -0.298 2.79 19275.44 19302.5 19331.49 -9633.719
    SCN 0.295 0.59 –– –– –– 0.215 -0.342 0.108 19432.05 19465.88 19502.12 -9711.05
    BST 0.139 0.737 –– 1.321 1.455 –– –– –– 19274.54 19301.6 19346.61 -9633.269
    Note: Bold font highlights the best model according to the model selection criteria.

     | Show Table
    DownLoad: CSV
    Table 2.  Parameter estimates for the SSEC data set and the associated AIC, BIC, EDC, and log-likelihood (l) values for different distributions.
    Model μ σ κ α β γ λ ν AIC BIC EDC l
    AGN 0.105 0.933 1.046 –– –– 1.391 –– –– 21538.4 21565.31 21593.23 -10765.2
    AGT 0.0814 1.973 1.028 3.053 1.368 –– –– –– 21171.13 21204.76 21239.76 -10580.56
    AGSL 0.0779 0.46 1.027 2.932 1.138 –– –– –– 21169.24 21202.87 21237.78 -10579.62
    CAGN 0.074 0.569 1.024 0.159 1.167 0.375 –– –– 21161.66 21202.02 21243.9 -10574.83
    SN 1.042 3.314 –– –– –– –– -1.024 –– 22450.91 22471.09 22492.03 -11222.46
    ST 0.206 0.889 –– –– –– –– -0.19 2.868 21215.86 21242.77 21270.69 -10604.93
    SCN 0.239 0.821 –– –– –– 0.254 -0.227 0.122 21291.59 21328.22 21360.12 -10642.8
    BST 0.099 0.928 –– 1.39 1.466 –– –– –– 21232.15 21259.06 21318.45 -10612.07
    Note: Bold font highlights the best model according to the model selection criteria.

     | Show Table
    DownLoad: CSV
    Figure 4.  Trace of the PLA of the parameter β for fitting the S&P 500 dataset GSMAGN family.
    Figure 5.  Trace of the PLA of the parameter β for fitting the SSEC dataset GSMAGN family.

    Upon scrutinizing Table 1, with a focus on the AIC and BIC criteria, it is discerned that the AGSL distribution consistently outperforms the other three distributions in fitting the S&P 500 dataset. Subsequently, the AGT distribution emerges as the next best-fitting alternative, indicating its substantial advantage when fitting data characterized by skewness and leptokurtic behavior. Turning attention to Table 2, considering the AIC, BIC, EDC, and log-likelihood values, it is deduced that the GSMAGN family surpasses the alternative models in terms of goodness of fit for the SSEC dataset.

    To visually illustrate the fitting results, Figure 6 depicts histograms of the datasets alongside corresponding fitted densities. Notably, the fitted curves of AGT, AGSL and CAGN closely align with the contours of the histograms, attesting to the model's proficiency in capturing pertinent information inherent in the observations.

    Figure 6.  Histogram of S&P500 and SSEC data sets with fitted densities.

    This paper introduced a distribution family named GSMAGN, which holds significant relevance in the fitting of complex data characterized by skewness and heavy tails. The study focused on examining the properties of this family and deriving the explicit expression for the moment of order statistics. This expression serves as a valuable tool for parameter estimation based on the order statistics. To enhance the efficiency of parameter estimation for this model, an alternative variant of the EM-type algorithm was established. This variant combined the PLA with the classical ECM algorithm. The proposed method not only improves computational speed but also provides efficient estimations. Furthermore, closed-form expressions of the information matrix, along with the corresponding asymptotic properties of the MLE, were presented for this distribution family. To assess the performance of the proposed estimation method, Monte Carlo simulations were conducted under various scenarios. The results demonstrated that the proposed method satisfies the asymptotic property, and the estimation accuracy improves with larger sample sizes. The feasibility of the proposed methodology was illustrated through two real datasets. The encouraging results obtained in this study motivate us to further delve into the realm of multivariate GSMAGN family for effectively modeling datasets exhibiting asymmetry and heavy-tailedness in the future.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    All authors declare no conflicts of interest in this paper.

    Part 1: Proofs required for section two

    Proof of Proposition 2.2

    According to Proposition 2.1, Y=μ+σU1βI(2Z)1β, and U, I, Z are independent of each other, then using the binomial expansion, we can obtain

    E(Ym)=mi=0(mi)σiμmi2iβE(Ii)E(Ziβ)E(Uiβ),

    where ZGamma(1/β,1), then we can calculate that

    E(Ziβ)=1Γ(1/β)0zi+1βezdz=Γ((i+1)/β)Γ(1/β),

    and I is a discrete random variable with density function fI=κ21+κ21{I=κ}+11+κ21{I=1/κ}. Thus the ith moments of I are

    E(Ii)=(κ)iκ21+κ2+1κi11+κ2=11+κ2[1κi+(1)iκ2+i].

    By substituting E(Ii) and E(Ziβ) into the above expression, we can prove Proposition 2.2. In particular, for m=1 and m=2, we obtain E(Y) and E(Y2), respectively. This allows us to derive the expressions for E(Y) and Var(Y).

    Proof of Proposition 2.3

    For AGT distribution, invoking Eq (2.3), we have

    fAGT(y;μ,σ,κ,α,β)=1Γ(α)βα10uα1exp(uβ)fAGN(y;μ,u1/βσ,κ,β)du=κβ2α21/βΓ(1/β)Γ(α)(1+κ2)σ+0uαβexp{[12(κsign(yμ)|yμ|σ)β+1β]uβ}du.

    Let z=[12(κsign(yμ)|yμ|σ)β+1β]uβ, and we obtain

    +0uαβexp{[12(κsign(yμ)|yμ|σ)β+1β]uβ}du=[12(κsign(yμ)|yμ|σ)β+1β](α+1β)1β×+0zα+1β1ezdz=βα+1β1Γ(α+1β)[1+β2(κsign(tμ)|yμ|σ)β](α+1β).

    Hence, we can derive the pdf of AGT distribution.

    As a result, by using fU|Y(uy)=fAGN(y;μ,u1/βσ,κ,β)f(u)fAGT(y) and after some algebric manipulation, we can obtain Proposition 2.3.

    Proof of Proposition 2.4

    For AGSL distribution, we have

    fAGSL(y;μ,σ,κ,α,β)=α10uα1fAGN(y;μ,u1/βσ,κ,β)du=καβ21/β(1+κ2)Γ(1/β)σ10uα+1/β1exp[u2(κsign(yμ)|yμ|σ)β]du,

    let z=[12(κsign(yμ)|yμ|σ)β], and after some calculation we will derive the pdf of AGSL distribution.

    By using fU|Y(uy)=fAGN(y;μ,u1/βσ,κ,β)f(u)fAGSL(y) and after some algebric manipulation, we can obtain Proposition 2.4.

    Proof of Proposition 2.5

    If γi=1, then CAGN becomes AGN.

    If 0<γi<1, then the pdf of Uj can be written as

    H(uj;α)=α1uj1γ(1α)ujγ1γ,

    and the conditional pdf of Uj|Y=yj can be written as

    PUjYj(ujyj)=[αifAGN(yj;μi,γ1/βσ,α,β)]1uj1γ[(1α)fAGN(yj;μ,σ,κ,β)]ujγ1γfACGN(yj;μ,σ,κ,γ,α,β).

    Part 2: Proofs required for Section 3

    Proof of Proposition 3.1

    Here, we assume that X1,X2,,Xn are identically distributed AGN(0,1,κ,β) samples of size n and X1:n,X2:n,,Xn:n are corresponding order statistics. In the case of X<0, the cdf of X can be evaluated using the power series expansion

    FAGN(x,κ,β)=κ2(1+κ2)Γ(1/β)[Γ(1β)j=0(1)jj!(1β+j)[12(xκ)β]1β+j].

    Let bj=(1)jj!(1β+j), for j=0,1,, we have that b0=β0. Consider an integer s order expansion of a power series

    [j=0bjyj]s=j=0b(s)jyj,y>0,

    where

    b(s)j=jh=0b(s1)hbjh,b(1)j=bj,b(0)j={1,j=0,0,j>0.

    After that, we can obtain

    [F(x)]n1=κ2(n1)(1+κ2)n1[Γ(1/β)]n1n1i=0(1)i(n1i)[Γ(1β)]ni1{j=0bj[12(xκ)β]1β+j}i,

    such that

    I1=n0xm[F(x)]n1f(x)dx=nκ2n1β21/β(1+κ2)n1n1i=0(1)i(n1i)[Γ(1β)]i1j=0b(i)j0xm[12(xκ)β]iβ+jexp{12(xκ)β}dx.

    Let u=12(xκ)β and we can easily get

    I1=n(1)mκ2n+m2mβ(1+κ2)nn1i=0(1)i(n1i)[Γ(1β)]i1j=0b(i)jΓ(m+i+1β+j).

    Similarly,

    I2=n0xm[F(x)]n1f(x)dx=nn1i=0(n1i)κ2(ni1)m2mβ(1+κ2)n[Γ(1β)]i1j=0b(i)jΓ(m+i+1β+j).

    Consequently,

    E(Xmn:n)=nn1i=0j=0(n1i)2mβ(1+κ2)n[Γ(1/β)]i+1[(1)i+mκ2n+m+κ2(ni1)m]b(i)jΓ(m+i+1β+j).

    Proof of Proposition 3.2

    If XAGN(0,1,κ,β) when X0,

    FAGN(x,κ,β)=κ2(1+κ2)Γ(1/β)[Γ(1β)12(xκ)β0t1β1etdt].

    Since

    12(xκ)β0t1β1etdt=m=0(1)mm!(1β+m)[12(xκ)β]1β+m,

    we obtain

    [F(x)]n1=κ2(n1)(1+κ2)n1[Γ(1β)]n1n1i=0(1)i(n1i)[Γ(1β)]ni1{12(xκ)β0t1β1etdt}i=κ2(n1)(1+κ2)n1[Γ(1β)]n1n1i=0(1)i(n1i)[Γ(1β)]ni1×m1=0mi=0(1)m1++mim1!mi!(1β+m1)(1β+mi)[12(xκ)β]iβ+it=1mt.

    Denote H1(m,n1,κ)=0xm[F(x)]n1f(x)dx,

    H1(m,n1,κ)=κ2n1β21/β(1+κ2)nn1i=0(1)i(n1i)[Γ(1β)]i1m1=0mi=0(1)m1++mim1!mi!(1β+m1)(1β+mi)×0xm[12(xκ)β]iβ+it=1mtexp{12(xκ)β}dx,

    and let u=12(xκ)β, then, x=κ(2u)1β, dxdu=κβ21βu1β1

    H1(m,n1,κ)=(1)mκ2n+m2mβ(1+κ2)nn1i=0(1)i(n1i)[Γ(1β)]i1m1=0mi=0(1)m1++mim1!mi!(1β+m1)(1β+mi)Γ(m+i+1β+it=1mt).

    Using some algebraic operations, we can obtain

    H1(m,n1,κ)=(1)mκ2n+m2mβ(1+κ2)nn1i=0(β)i(n1i)[Γ(1β)]i1Γ(m+i+1β)×F(i)A(m+i+1β,1β,,1β;1β+1,,1β+1;1,,1).

    Denote H2(m,n1,κ)=0xm[F(x)]n1f(x)dx.

    In the same way, we can obtain

    H2(m,n1,κ)=n1i=0(n1i)κ2(ni1)m2mβ(1+κ2)n[Γ(1/β)]i+1Γ(m+i+1β)βi×F(i)A(m+i+1β,1β,,1β;1β+1,,1β+1;1,,1).

    Proof of Proposition 3.3

    If 1β is an integer, when x<0 we have

    F(x)=κ2(1+κ2)B(α,1/β)1[1+2β(κx)β]1t1/β1(1t)α1dt=κ2(1+κ2)B(α,1/β)(1/β1)i=0(1/β1i)1[1+2β(κx)β]1(1t)α+i1dt,

    and by binomial expansion we obtain

    F(x)=κ2(1+κ2)B(α,1/β)(1/β1)i=0Ci1[1+β2(xκ)β]α+i,

    where Ci=(1)i(1/β1i)1α+i.

    Let Tv(1/β,n1) be the coefficient of [1+β2(xκ)β]v in the expansion of {(1/β1)i=0Ci1[1+β2(xκ)β]i}n1. Clearly, we have Tv(1/β,0)=1.

    For v=0,1,,(1/β),

    Tv(1/β,1)=Cv,

    and for n>2, v=0,1,,(1/β)(n1),

    Tv(1β,n1)=vi=0CiTvi(1β,n2).

    We can calculate every coefficient directly, so [F(x)]n1 can be expressed as

    [F(x)]n1=κ2(n1)(1+κ2)n1B(α,1/β)(1/β1)(n1)v=0Tv(1/β,n1)1[1+β2(xκ)β]α(n1)+v,

    then, we have

    nI1(n1,κ)=n0xm[F(x)]n1f(x)dx=n(1)m(2/β)mβκ2n+m(1+κ2)n[B(α,1/β)]n(1/β1)(n1)v=0Tv(1/β,n1)×B(m+1β,αn+vmβ),

    and the above expression exists for αβ>mn. It is not hard to verify the following statement

    nI2(n1,κ)=n0xm[F(x)]n1f(x)dx=nn1j=0(1)m+j(n1j)I1(j,1/κ).

    Consequently,

    E(Xmn:n)=n(2/β)mβ(1+κ2)n[(1)mκ2n+m[B(α,1/β)]n(1/β1)(n1)v=0Tv(1/β,n1)×B(m+1β,αn+vmβ)+n1j=0(1)j(n1j)1κm[B(α,1/β)]j(1/β1)jv=0Tv(1/β,j)×B(m+1β,α(j+1)+vmβ)].

    Proof of Proposition 3.4

    Denote J(m,n1,κ)=0xm[F(x)]n1f(x)dx. When x<0, 1β is not an integer and we have

    F(x)=κ2(1+κ2)B(α,1/β)i=0(1)i(1/β1i)1α+i1[1+β2(xκ)β]α+i,

    where

    i=0(1)i(1/β1i)1α+i=i=0(11β)ii!(α+i).

    Let C=2κ2n1(2/β)1+1β(1+κ2)n[B(α,1/β)]n, such that, J(m,n1,κ) can be represented as

    J(m,n1,κ)=C0xm{i=0(11β)ii!(α+i)1[1+β2(xκ)β]α+i}n11[1+β2(xκ)β](α+1/β)=Cm1=0mn1=0Πn1i=1(11β)miΠn1i=1(α+mi)mi!0xm1[1+β2(xκ)β]αn+n1i=1mi+1/βdx.

    Let β2(xκ)β1+β2(xκ)β=u, then, the above integral can be converted to

    0xm1[1+β2(xκ)β]αn+n1i=1mi+1/βdx=(1)mκm+12m+1ββm+1β+110um+1β(1u)αn+n1i=1mimβ1du=(1)mκm+12m+1ββm+1β+1B(m+1β,αn+n1i=1mimβ).

    Substituting the above equation into J(m,n1,κ), and using some algebraic operations, we can obtain

    J(m,n1,κ)=(1)mk2n+m(2/β)mβB(m+1β,αnmβ)(1+κ2)n[B(α,1/β)]nan1×F1:21:1((αnmβ):(11β,α);;(11β,α);(αn+1β):(α+1);,(α+1);1,,1).

    It is worth noting that, when αβ>mn, B(m+1β,αn+n1i=1mimβ)<Γ(m+1β), On the other hand, for a sufficiently large N, we can bound that

    |max(m1,,mn1)>N(1b)m1(1b)m2(1b)mn1(a+m1)(a+m2)(a+mn1)m1!mn1!×B(m+1β,αn+n1i=1mimβ)|<Γ(m+1β)max(m1,,mn1)>N|(1b)m1(1b)m2(1b)mn1|(a+m1)(a+m2)(a+mn1)m1!mn1!=Γ(m+1β)[m1=0mn1=0|(1b)m1(1b)m2(1b)mn1|(a+m1)(a+m2)(a+mn1)m1!mn1!Nm1=0Nmn1=0|(1b)m1(1b)m2(1b)mn1|(a+m1)(a+m2)(a+mn1)m1!mn1!]=Γ(m+1β)[{m=0|(1b)m|(a+m)m!}n1{Nm=0|(1b)m|(a+m)m!}n1]<.

    This means that for a maximum order statistic from AGT(0,1,κ,α,β), it's m-th moments always exists.

    Proof of Proposition 3.5

    Suppose that YACGN(0,1,κ,λ,α,β), XAGN(0,1,κ,α,β), U is a discrete random variable with pdf defined in (2.5), then we have

    E(Yr:n)=P(U=1)E[Yr:nU=1]+P(U=λ)E[Yr:nU=λ]=P(U=1)E[[U1βX]r:nU=1]+P(U=λ)E[[U1βX]r:nU=λ]=P(U=1)E[Xr:nU=1]+P(U=λ)E[λ1βXr:nU=λ].

    In view of the independence between U and X,

    E(Yr:n)=P(U=1)E(Xr:n)+P(U=λ)λ1βE(Xr:n)=[α+λ1β(1α)]E(Xr:n).

    Part 3: Proofs in section five and details to obtain the information matrix

    Proof of Theorem 5.1

    It is easy to prove that the distributions of the GSMAGN family satisfy item one and two of Lemma 5.1. As a result, our main task is to verify the GSMAGN family satisfies item three under different circumstances.

    Let v=12(κsign(yμ)|yμ|σ)β. According to Proposition 2.1, we can obtain that vd=U1Z, which illustrates that E(vk)=E(UkZk). Taking the first partial derivation with respect to μ,σ,β,κ,

    vμ=β2σsign(μy)κsign(yμ)(κsign(yμ)|yμ|σ)β1,vσ=β2σ(κsign(yμ)|yμ|σ)β,vβ=12(κsign(yμ)|yμ|σ)βln(κsign(yμ)|yμ|σ),vκ=β2κsign(yμ)(κsign(yμ)|yμ|σ)β.

    For AGN

    lnfAGN=lnκ+lnβ1βln2ln(1+κ2)lnΓ(1/β)lnσv.

    Take sufficiently small ϵ>0, Uθ0={θ:∥θ0θ∥<ϵ}˜Θ,θUθ0, then

    |lnfAGNμ|=|vμ|β21/βσκsign(yμ)v11/β=g1(v),|lnfAGNσ|=|1σvσ|1σ+βσv1σ0ϵ+βσv=G1+g2(v),|lnfAGNβ|=|1β+1β2ln2+φ(1/β)1β2vβ|1β+1β2ln2+φ(1/β)β2+vβ1β0ϵ+1(β0ϵ)2ln2+φ(1/(β0ϵ))(β0ϵ)2+2v2β=G2+g3(v),|lnfAGNκ|=|1κ+2κ1+κ2vκ|1κ+2κ1+κ2+βκv1+3(κ0+ϵ)2(κ0+ϵ)[1+(κ0+ϵ)2]+βκv=G3+g4(v).

    Let G(y)=max{G1,G2,G3}+g1(v)+g2(v)+g3(v)+g4(v), for AGN, where U=1, and we can get E(vk)=E(Zk),ZGamma(1/β,1), for β>0. E(vk)<,k>1β, then we have E[G(y)]<.

    For AGT

    lnfAGTlnκ(1+1β)ln2βln(1+κ2)lnB(α,1/β)lnσ(α+1β)ln(1+βv),=lnκ(1+1β)ln2βln(1+κ2)lnΓ(α)lnΓ(1β)+lnΓ(α+1β)lnσ(α+1β)ln(1+βv).

    Take sufficiently small ϵ>0, Uθ0={θ:∥θ0θ∥<ϵ}˜Θ,θUθ0,

    |lnfAGTμ|=|(α+1β)β1+βvvμ|αβ+11+βvβ21/βσκsign(yμ)v11/βαβ2+β21/βσκsign(yμ)v11β=g1(v),|lnfAGTσ|=|1σ(α+1β)β1+βvvσ|1σ+(α+1β)β1+βvβσv1σ+αβ+1σ1σ0ϵ+(α0+ϵ)(β0+ϵ)σ0ϵ=G1,|lnfAGTα|=|φ(α)+φ(α+1β)ln(1+βv)|φ(α)+φ(α+1β)+ln(1+βv)φ(α0+ϵ)+φ(α0+ϵ+1β0ϵ)+βv=G2+g2(v),|lnfAGTβ|=|1β2ln2β+1β+1β2+1β2φ(1/β)1β2φ(α+1β)+1β2ln(1+βv)(α+1β)11+βv(v+βvβ)|1β2ln2β+1β+1β2+1β2φ(1/β)+1β2φ(α+1β)+1β2ln(1+βv)+(α+1β)11+βv(v+β2v2β)1β2[ln2β+(α+1)β+φ(1β)+φ(α+1β)+2]+(2α+2β+1)vβ1(β0ϵ)2[ln2β0ϵ+(α0+ϵ+1)(β0+ϵ)+φ(1β0ϵ)+φ(α0+ϵ+1β0ϵ)+2]+(2α+2β+1)vβ=G3+g3(v),|lnfAGTκ|=|1κ2κ1+κ2(α+1β)β1+βvvκ|1κ+2κ1+κ2+αβ+11+βvβκv1+3(κ0+ϵ)2(κ0+ϵ)[1+(κ0+ϵ)2]+(α0+ϵ)(β0+ϵ)+1κ0ϵ=G4.

    Let G(y)=max{G1,G2,G3,G4}+g1(v)+g2(v)+g3(v), for AGT, E(vk)=E(Uk)E(Zk), where UGamma(α,β),ZGamma(1/β,1). Invoking Propostion 2, if α>1, E(v)<, and if α+1β>1, E(v11β)<, such that, when α>1, we have E[G(y)]<. For AGSL

    lnfAGSL=lnκ+lnα+lnβ+lnΓ(α+1/β)1βln2lnΓ(1/β)lnσ+lnγ(α+1β,v)(α+1β)lnv,

    where γ(α+1β;v)=1Γ(α+1/β)v0tα+1β1etdt. By taking sufficiently small ϵ>0, Uθ0={θ:∥θ0θ∥<ϵ}˜Θ,θUθ0, then

    |lnfAGSLμ|=|1γ(α+1β;v)1Γ(α+1β)vα+1β1evvμ(α+1β)1vvμ||vα+1β1evv0tα+1β1etdtα+1/βv||vμ|(α+1/βvα+1/βvev)|vμ|αβ+121/βσκsign(yμ)v11β=g1(v),|lnfAGSLσ|=|1σ+1γ(α+1β;v)1Γ(α+1β)vα+1β1evvσ(α+1β)1vvσ|1σ+vα+1β1evv0tα+1β1etdtβσv+(α+1β)1vβσv2αβ+3σ2(α0+ϵ)(β0+ϵ)+3σ0ϵ=G1,
    \begin{align*} \left|\frac{\partial \mathrm{ln}f_{AGSL}}{\partial\alpha}\right| = &\bigg|\frac{1}{\alpha}+\varphi(\alpha+1/\beta)+\frac{1}{\gamma(\alpha+\frac{1}{\beta};v)}\bigg(\frac{1}{\varGamma(\alpha+\frac{1}{\beta})}\int_0^{v}t^{\alpha+\frac{1}{\beta}-1} \mathrm{ln}te^{-t} \mathrm{d}t\\ &-\frac{\varphi(\alpha+\frac{1}{\beta})}{\varGamma^2(\alpha+\frac{1}{\beta})}\int_0^{v}t^{\alpha+\frac{1}{\beta}-1}e^{-t} \mathrm{d}t\bigg)-\ln v\bigg|\\ &\leq\frac{1}{\alpha}+\varphi(\alpha+1/\beta)+\frac{\int_0^{v}t^{\alpha+\frac{1}{\beta}-1}\ln te^{-t} \mathrm{d}t}{\int_0^{v}t^{\alpha+\frac{1}{\beta}-1}e^{-t} \mathrm{d}t}+\varphi(\alpha+1/\beta)+|\ln v|\\ &\leq\frac{1}{\alpha}+2\varphi(\alpha+1/\beta)+2|\ln v|\\ &\leq\frac{1}{\alpha_0-\epsilon}+2\varphi(\alpha_0+\epsilon+1/(\beta_0-\epsilon))+2v = G_2+g_2(v), \end{align*}
    \begin{align*} \left|\frac{\partial \mathrm{ln}f_{AGSL}}{\partial\kappa}\right|& = \left|\frac{1}{\kappa}-\frac{2\kappa}{1+\kappa^2}+\frac{1}{\gamma(\alpha+\frac{1}{\beta};v)}\frac{1}{\varGamma(\alpha+\frac{1}{\beta})}v^{\alpha+\frac{1}{\beta}-1}e^{-v}\frac{\partial v}{\partial \kappa}-(\alpha+\frac{1}{\beta})\frac{1}{v}\frac{\partial v}{\partial \kappa}\right|\\ &\leq\frac{1}{\kappa}+\frac{2\kappa}{1+\kappa^2}+\frac{v^{\alpha+\frac{1}{\beta}-1}e^{-v}}{\int_0^{v}t^{\alpha+\frac{1}{\beta}-1}e^{-t} \mathrm{d}t}\frac{\partial v}{\partial \kappa}+(\alpha+\frac{1}{\beta})\frac{1}{v}\frac{\partial v}{\partial \kappa}\\ &\leq\frac{1}{\kappa}+\frac{2\kappa}{1+\kappa^2}+\frac{\alpha+1/\beta}{v}\frac{\beta}{\kappa}v+(\alpha+\frac{1}{\beta})\frac{1}{v}\frac{\beta}{\kappa}v\\ &\leq\frac{1}{\kappa_0-\epsilon}+\frac{2(\kappa_0+\epsilon)}{1+(\kappa_0-\epsilon)^2}+\frac{2(\alpha_0+\epsilon)(\beta_0+\epsilon)+1}{\kappa_0-\epsilon} = G_3, \end{align*}
    \begin{align*} \left|\frac{\partial\ln f_{AGSL}}{\partial\beta}\right|& = \left|\frac{1}{\beta}-\frac{1}{\beta^2}\varphi(\alpha+\frac{1}{\beta})+\frac{1}{\beta^2}\ln 2+\frac{1}{\beta^2}\varphi(\frac{1}{\beta})+\frac{v^{\alpha+\frac{1}{\beta}-1}e^{-v}}{\int_0^{v}t^{\alpha+\frac{1}{\beta}-1}e^{-t} \mathrm{d}t}\frac{\partial v}{\partial \beta}-\frac{1}{\beta^2}\frac{\int_0^{v}t^{\alpha+\frac{1}{\beta}-1} \mathrm{ln}te^{-t} \mathrm{d}t}{\int_0^{v}t^{\alpha+\frac{1}{\beta}-1}e^{-t} \mathrm{d}t}\right|\\ &\leq\frac{1}{\beta}+\frac{1}{\beta^2}\varphi(\alpha+\frac{1}{\beta})+\frac{1}{\beta^2} \mathrm{ln}2+\frac{1}{\beta^2}\varphi(\frac{1}{\beta})+\frac{\alpha+\frac{1}{\beta}}{v}\frac{\partial v}{\partial \beta}+\frac{1}{\beta^2} \mathrm{ln}v\\ &\leq\frac{1}{\beta}+\frac{1}{\beta^2}\varphi(\alpha+\frac{1}{\beta})+\frac{\alpha\beta+2}{\beta^2} \mathrm{ln}2+\frac{\alpha\beta+2}{\beta^2} \mathrm{ln}v\\ &\leq\frac{1}{\beta_0-\epsilon}+\frac{1}{(\beta_0-\epsilon)^2}\varphi(\alpha_0+\epsilon+\frac{1}{\beta_0-\epsilon})+\frac{(\alpha_0+\epsilon)(\beta_0+\epsilon)+2}{(\beta_0-\epsilon)^2} \mathrm{ln}2+\frac{\alpha\beta+2}{\beta^2}v\\ & = G_4+g_3(v). \end{align*}

    Let G(y) = \mathrm{max}\{G_1, G_2, G_3, G_4\}+g_1(v)+g_2(v)+g_3(v) . For AGSL, \mathrm{E}(v^k) = \mathrm{E}(U^{-k}) \mathrm{E}(Z^k) , where U\sim Beta(\alpha, 1), Z\sim Gamma(1/\beta, 1) . Invoking Propostion 2, we can obtain that if \alpha > 1 , then \mathrm{E}(v) < \infty . For \alpha+\frac{1}{\beta} > 1 , \mathrm{E}(v^{1-\frac{1}{\beta}}) < \infty , such that \alpha > 1 , \mathrm{E}[G(y)] < \infty .

    \bullet For CAGN

    f_{CAGN}(y;\mu,\sigma,\kappa,\lambda,\alpha,\beta) = \alpha f_{AGN}(y;\mu,\lambda^{-1/\beta}\sigma,\kappa,\beta)+(1-\alpha)f_{AGN}(y;\mu,\sigma,\kappa,\beta).

    As for CAGN, it can be written as a linear combination of two AGN distributions. According to the above proof, it can be obtained that the MLE of AGN distribution exists and satisfies the strong consistency. Therefore, it can be easily concluded that CAGN satisfies the same property.

    Details to obtain information matrix:

    \bullet For AGT

    M_{i,k} = \frac{1}{\beta^{\alpha}\Gamma(\alpha)}\frac{\Gamma(\alpha+\frac{1}{\beta}+k)}{\left[\frac{1}{2}{A_i}^{\beta}+\frac{1}{\beta}\right]^{\alpha+\frac{1}{\beta}+k}},

    and,

    s_{i,\alpha} = -\ln\beta-\varphi(\alpha)+\frac{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}-1}\ln u\exp\bigg\{-\bigg(\frac{1}{2}{A_i}^\beta+\frac{1}{\beta} \bigg) u\bigg\} \mathrm{d}u}{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}-1}\exp\bigg\{-\bigg(\frac{1}{2}{A_i}^\beta+\frac{1}{\beta} \bigg) u\bigg\} \mathrm{d}u}
    \begin{align*} s_{i,\beta} = &\frac{1-\alpha}{\beta}+\frac{\varphi(1/\beta)+\ln 2}{\beta^2}-\frac{1}{\beta^2}\frac{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}-1}\ln u\exp\bigg\{-\bigg(\frac{1}{2}{A_i}^\beta+\frac{1}{\beta} \bigg) u\bigg\} \mathrm{d}u}{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}-1}\exp\bigg\{-\bigg(\frac{1}{2}{A_i}^\beta+\frac{1}{\beta} \bigg) u\bigg\} \mathrm{d}u}\\ &+\left[\frac{1}{\beta^2}-\frac{1}{2}{A_i}^\beta\ln\bigg(\kappa^{ \mathrm{sign}(y_i-\mu)}\frac{|y_i-\mu|}{\sigma}\bigg)\right] \times\frac{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}}\exp\bigg\{-\bigg(\frac{1}{2}{A_i}^\beta+\frac{1}{\beta} \bigg) u\bigg\} \mathrm{d}u}{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}-1}\exp\bigg\{-\bigg(\frac{1}{2}{A_i}^\beta+\frac{1}{\beta} \bigg) u\bigg\} \mathrm{d}u}. \end{align*}

    \bullet For AGSL

    M_{i,k} = \alpha2^{\alpha+\frac{1}{\beta}+\kappa}{A_i}^{-[(\alpha+\kappa)\beta+1]}\gamma\bigg(\alpha+\frac{1}{\beta}+\kappa,\frac{1}{2}{A_i}^{\beta}\bigg),

    and,

    s_{i,\alpha} = \frac{1}{\alpha}+\frac{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}-1}\ln u\exp\bigg\{-\frac{1}{2}{A_i}^\beta u\bigg\} \mathrm{d}u}{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}-1}\exp\bigg\{-\frac{1}{2}{A_i}^\beta u\bigg\} \mathrm{d}u}.
    \begin{align*} s_{i,\beta} = \frac{1}{\beta}+\frac{\varphi(1/\beta)+\ln 2}{\beta^2}-\left[\frac{1}{\beta^2}+\frac{1}{2}{A_i}^\beta\ln\bigg(\kappa^{ \mathrm{sign}(y-\mu)}\frac{|y-\mu|}{\sigma}\bigg)\right] \times\frac{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}-1}\ln u\exp\bigg\{-\frac{1}{2}{A_i}^\beta u\bigg\} \mathrm{d}u}{\int_{0}^{1}u^{\alpha+\frac{1}{\beta}-1}\exp\bigg\{-\frac{1}{2}{A_i}^\beta u\bigg\} \mathrm{d}u}. \end{align*}

    \bullet For CAGN

    M_{i,k} = \lambda^{\frac{1}{\beta}+\kappa}\exp\bigg\{-\frac{\lambda}{2}{A_i}^\beta\bigg\}+(1-\alpha)\exp\bigg\{-\frac{1}{2}{A_i}^\beta\bigg\},

    and,

    s_{i,\alpha} = \frac{f_{AGN}(y_i;\mu,\lambda^{-\frac{1}{\beta}}\sigma,\kappa,\beta)-f_{AGN}(y_i;\mu,\sigma,\kappa,\beta)}{\alpha f_{AGN}(y_i;\mu,\lambda^{-\frac{1}{\beta}}\sigma,\kappa,\beta)+(1-\alpha)f_{AGN}(y_i;\mu,\sigma,\kappa,\beta)}.
    s_{i,\lambda} = \frac{\alpha\left[\frac{1}{\beta\lambda}-\frac{1}{2}A_i^\beta\right]f_{AGN}(y_i;\mu,\lambda^{-\frac{1}{\beta}}\sigma,\kappa,\beta)}{\alpha f_{AGN}(y_i;\mu,\lambda^{-\frac{1}{\beta}}\sigma,\kappa,\beta)+(1-\alpha)f_{AGN}(y_i;\mu,\sigma,\kappa,\beta)}.

    The information-based approximation (11) is asymptotically applicable. The standard errors of \hat{\Theta} can be obtained through the inverse of the empirical information matrix \mathbf{I_0(\Theta|y)} , with a sufficiently large sample size. The accuracy of the estimation can be guaranteed and the complexity of calculation can be reduced.



    [1] D. F. Andrews, C. L. Mallows, Scale mixtures of normal distribution, J. R. Stat. Soc. Ser. B Methodol., 36 (1974), 99–102. https://doi.org/10.1111/j.2517-6161.1974.tb00989.x doi: 10.1111/j.2517-6161.1974.tb00989.x
    [2] M. West, On scale mixtures of normal distributions, Biometrika, 74 (1987), 646–648. https://doi.org/10.1093/biomet/74.3.646 doi: 10.1093/biomet/74.3.646
    [3] M. D. Branco, D. K. Dey, A general class of multivariate skew-elliptical distributions, J. Multivar. Anal., 79 (2001), 99–113. https://doi.org/10.1006/jmva.2000.1960 doi: 10.1006/jmva.2000.1960
    [4] C. S. Ferreira, H. Bolfarine, V. H. Lachos, Linear mixed models based on skew scale mixtures of normal distribution, Commun. Stat. Simul. Comput., 51 (2020), 7194–7214. https://doi.org/10.1080/03610918.2020.1827265 doi: 10.1080/03610918.2020.1827265
    [5] R. M. Basso, V. H. Lachos, C. R. B. Cabral, P. Ghosh, Robust mixture modeling based on scale mixtures of skew-normal distributions, Comput. Stat. Data Anal., 54 (2010), 2926–2941. https://doi.org/10.1016/j.csda.2009.09.031 doi: 10.1016/j.csda.2009.09.031
    [6] H. M. Kim, M. G. Genton, Characteristic functions of scale mixtures of multivariate skew-normal distributions, Comput. Stat. Data Anal., 102 (2011), 1105–1117. https://doi.org/10.1016/j.jmva.2011.03.004 doi: 10.1016/j.jmva.2011.03.004
    [7] T. I. Lin, J. C. Lee, W. J. Hsieh, Robust mixture modeling using the skew t distribution, Stat. Comput., 17 (2007), 81–92. https://doi.org/10.1007/s11222-006-9005-8 doi: 10.1007/s11222-006-9005-8
    [8] T. I. Lin, H. J. Ho, C. R. Lee, Flexible mixture modelling using the multivariate skew-t-normal distribution, Stat. Comput., 24 (2014), 531–546. https://doi.org/10.1007/s11222-013-9386-4 doi: 10.1007/s11222-013-9386-4
    [9] I. Lin, J. C. Lee, Y. Y. Shu, Finite mixture modelling using the skew normal distribution, Stat. Sin., 17 (2007), 909–927. https://doi.org/10.2307/24307705 doi: 10.2307/24307705
    [10] A. Mahdavi, V. Amirzadeh, A. Jamalizadeh, T. I. Lin, Maximum likelihood estimation for scale-shape mixtures of flexible generalized skew normal distributions via selection representation, Comput. Stat., 36 (2021), 2201–2230. https://doi.org/10.1007/s00180-021-01079-2 doi: 10.1007/s00180-021-01079-2
    [11] A. Azzalini, A class of distributions which includes the normal ones, Scand. J. Stat., 12 (1985), 171–178.
    [12] A. Azzalini, A. Dalla Valle, The multivariate skew-normal distribution, Biometrika, 83 (1996), 715–726. https://doi.org/10.1093/biomet/83.4.715 doi: 10.1093/biomet/83.4.715
    [13] A. Azzalini, A. Capitanio, Statistical applications of the multivariate skew normal distribution, J. R. Stat. Soc. Ser. B Methodol., 61 (1999), 579–602. https://doi.org/10.1111/1467-9868.00194 doi: 10.1111/1467-9868.00194
    [14] R. B. Arellano-Valle, A. Azzalini, On the unification of families of skew-normal distributions, Scand. J. Stat., 33 (2006), 561–574. https://doi.org/10.1111/j.1467-9469.2006.00503.x doi: 10.1111/j.1467-9469.2006.00503.x
    [15] R. B. Arellano-Valle, M. G. Genton, On fundamental skew distributions, J. Multivar. Anal., 96 (2005), 93–116. https://doi.org/10.1016/j.jmva.2004.10.002 doi: 10.1016/j.jmva.2004.10.002
    [16] A. Azzalini, The skew-normal distribution and related multivariate families, Scand. J. Stat., 32 (2005), 159–188. https://doi.org/10.1111/j.1467-9469.2005.00426.x doi: 10.1111/j.1467-9469.2005.00426.x
    [17] C. Fernandez, M. F. J. Steel, Reference priors for non-normal two-sample problems, Test, 7 (1988), 179–205. https://doi.org/10.1007/BF02565109 doi: 10.1007/BF02565109
    [18] D. M. Zhu, V. Zinde-Walsh, Properties and estimation of asymmetric exponential power distribution, J. Econom., 148 (2009), 86–99. https://doi.org/10.1016/j.jeconom.2008.09.038 doi: 10.1016/j.jeconom.2008.09.038
    [19] R. J. Guan, X. Zhao, C. H. Cheng, Y. H. Rong, A new generalized t distribution based on a distribution construction method, Mathematics, 9 (2021), 2413. https://doi.org/10.1016/10.3390/math9192413 doi: 10.1016/10.3390/math9192413
    [20] H. Exton, Handbook of hypergeometric integrals: Theory, applications, tables, computer programs, New York: Halsted Press, 1978.
    [21] A. P. Dempster, N. M. Laird, D. B. Rubin, Maximum likelihood from incomplete data via the em algorithm, J. R. Stat. Soc. Ser. B Methodol., 39 (1977), 1–22. https://doi.org/10.1111/j.2517-6161.1977.tb01600.x doi: 10.1111/j.2517-6161.1977.tb01600.x
    [22] K. Lange, The EM algorithm, New York: Springer, 2013.
    [23] X. L. Meng, D. B. Rubin, Maximum likelihood estimation via the ecm algorithm: A general framework, Biometrika, 80 (1993), 267–278. https://doi.org/10.2307/2337198 doi: 10.2307/2337198
    [24] P. Huber, Robust statistics, New York: Wiley, 1981.
    [25] L. L. Wen, Y. J. Qiu, M. H. Wang, J. L. Yin, P. Y. Chen, Numerical characteristics and parameter estimation of finite mixed generalized normal distribution, Commun. Stat. Simul. Comput., 51 (2022), 3596–3620. https://doi.org/10.1080/03610918.2020.1720733 doi: 10.1080/03610918.2020.1720733
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1429) PDF downloads(82) Cited by(0)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog