Research article Special Issues

The multi-parameter estimation of discrete distribution without closed-form solutions by the US algorithm

  • Strong and steady convergence characterizes the upper-crossing/solution (US) algorithm, which is an effective method for identifying roots of a complicated nonlinear equation h(θ)=0. Only the case where one parameter of a distribution function can be directly specified by another parameter is taken into account by the research that is currently available. However, whether this approach can be applied in multi-parameter scenarios where one parameter cannot be clearly represented by the other is an issue deserving of more investigation. In order to extend the applicability of the US algorithm, this article used the Type Ⅰ discrete Weibull distribution with two parameters as an example. It then combined the US algorithm with the first-derivative lower bound (FLB) function method to estimate the complex situation where two parameters cannot be expressed as each other. Simulation studies and empirical analysis demonstrated that the US algorithm performs more accurately and steadily than the traditional Newton method.

    Citation: Yuanhang Ouyang, Ruyun Yan, Jianhua Shi. The multi-parameter estimation of discrete distribution without closed-form solutions by the US algorithm[J]. AIMS Mathematics, 2024, 9(9): 24507-24524. doi: 10.3934/math.20241193

    Related Papers:

    [1] Abdulaziz S. Alghamdi, Muhammad Ahsan-ul-Haq, Ayesha Babar, Hassan M. Aljohani, Ahmed Z. Afify . The discrete power-Ailamujia distribution: properties, inference, and applications. AIMS Mathematics, 2022, 7(5): 8344-8360. doi: 10.3934/math.2022465
    [2] Monthira Duangsaphon, Sukit Sokampang, Kannat Na Bangchang . Bayesian estimation for median discrete Weibull regression model. AIMS Mathematics, 2024, 9(1): 270-288. doi: 10.3934/math.2024016
    [3] Saurabh L. Raikar, Dr. Rajesh S. Prabhu Gaonkar . Jaya algorithm in estimation of P[X > Y] for two parameter Weibull distribution. AIMS Mathematics, 2022, 7(2): 2820-2839. doi: 10.3934/math.2022156
    [4] Rasha Abd El-Wahab Attwa, Shimaa Wasfy Sadk, Hassan M. Aljohani . Investigation the generalized extreme value under liner distribution parameters for progressive type-Ⅱ censoring by using optimization algorithms. AIMS Mathematics, 2024, 9(6): 15276-15302. doi: 10.3934/math.2024742
    [5] Haiping Ren, Xue Hu . Estimation for inverse Weibull distribution under progressive type-Ⅱ censoring scheme. AIMS Mathematics, 2023, 8(10): 22808-22829. doi: 10.3934/math.20231162
    [6] Xue Hu, Haiping Ren . Statistical inference of the stress-strength reliability for inverse Weibull distribution under an adaptive progressive type-Ⅱ censored sample. AIMS Mathematics, 2023, 8(12): 28465-28487. doi: 10.3934/math.20231457
    [7] A. M. Abd El-Raheem, Ehab M. Almetwally, M. S. Mohamed, E. H. Hafez . Accelerated life tests for modified Kies exponential lifetime distribution: binomial removal, transformers turn insulation application and numerical results. AIMS Mathematics, 2021, 6(5): 5222-5255. doi: 10.3934/math.2021310
    [8] Heba S. Mohammed, Zubair Ahmad, Alanazi Talal Abdulrahman, Saima K. Khosa, E. H. Hafez, M. M. Abd El-Raouf, Marwa M. Mohie El-Din . Statistical modelling for Bladder cancer disease using the NLT-W distribution. AIMS Mathematics, 2021, 6(9): 9262-9276. doi: 10.3934/math.2021538
    [9] Qasim Ramzan, Muhammad Amin, Ahmed Elhassanein, Muhammad Ikram . The extended generalized inverted Kumaraswamy Weibull distribution: Properties and applications. AIMS Mathematics, 2021, 6(9): 9955-9980. doi: 10.3934/math.2021579
    [10] Nora Nader, Dina A. Ramadan, Hanan Haj Ahmad, M. A. El-Damcese, B. S. El-Desouky . Optimizing analgesic pain relief time analysis through Bayesian and non-Bayesian approaches to new right truncated Fréchet-inverted Weibull distribution. AIMS Mathematics, 2023, 8(12): 31217-31245. doi: 10.3934/math.20231598
  • Strong and steady convergence characterizes the upper-crossing/solution (US) algorithm, which is an effective method for identifying roots of a complicated nonlinear equation h(θ)=0. Only the case where one parameter of a distribution function can be directly specified by another parameter is taken into account by the research that is currently available. However, whether this approach can be applied in multi-parameter scenarios where one parameter cannot be clearly represented by the other is an issue deserving of more investigation. In order to extend the applicability of the US algorithm, this article used the Type Ⅰ discrete Weibull distribution with two parameters as an example. It then combined the US algorithm with the first-derivative lower bound (FLB) function method to estimate the complex situation where two parameters cannot be expressed as each other. Simulation studies and empirical analysis demonstrated that the US algorithm performs more accurately and steadily than the traditional Newton method.



    Discrete lifetime data—such as the number of appliance failures of a particular brand within a given time frame, the total number of machine operations prior to a failure, the number of bullets fired by a weapon before the first malfunction, and the anticipated lifespan of humans (in years)—are frequently handled in reliability lifetime studies. For more classic examples, see Szymkowiak and Iwinska [1]. Data scientists typically employ discrete models as analysis tools, such as the Poisson distribution, negative binomial distribution, and geometric distribution, in order to more correctly define, analyze, and model these data. But in many situations, these discrete distribution functions are not the best options. For instance, seasonal or periodic data cannot be handled by the Poisson distribution, while underdispersed data cannot be described by the negative binomial distribution. More suitable discrete lifetime distributions are required to explore many additional kinds of complex discrete lifespan data. Discretizing continuous random variables is a useful strategy that yields a discrete life model with characteristics that are comparable to the continuous model.

    The essential concept of discretizing continuous random variables was first presented by Roy [2]. Specifically, let Y be a continuous random variable with a survival function denoted by S(y). Define the random variable Z=[Y] as the maximum integer less than or equal to Y. The probability mass function (PMF) P(Z=z) of Z can be expressed as

    P(Z=z)=SY(z)SY(z+1).

    Many researchers have introduced various new models for discrete life distributions by the approach. For instance, the discrete normal distribution was first introduced by Roy [3]. Using the generic method of discretizing a continuous distribution, Krishna and Pundir [4] introduced the discrete Burr and Pareto distributions. In addition, Bracquemond and Gaudoin [5] provided an extensive overview of discrete distributions, such as the Weibull distribution, that are employed in reliability to describe discrete lifetimes of nonrepairable systems. It is well-known that the Weibull distribution has become the most commonly used distribution for analyzing continuous life data due to its ability to fit various types of data and relatively simple structure (Johnson et al. [6]). At least three cases exist for the corresponding discrete Weibull distribution: (a) the Type Ⅰ discrete Weibull distribution, which maintains the form of the continuous survival function (SF), as introduced by Nakagawa and Osaki [7]; (b) the Type Ⅱ discrete Weibull distribution, as suggested by Stein and Dattero [8]; and (c) the three-parameter discrete Weibull distribution, as introduced by Padgett and Spurrier [9]. The most popular of them is the Type Ⅰ discrete Weibull distribution, whose features have been extensively researched by numerous academics. Englehardt and Li [10] employed the discrete Weibull distribution to analyze pathogen counts in treated water over time. Barbiero [11,12] compared several parameter estimation methods of this distribution, and solved the minimum Chi-square and least squares estimation. Vila et al. [13] studied in detail the basic theoretical properties of the Type Ⅰ discrete Weibull and analyzed the censored data. Yoo [14] extended the application of the discrete Weibull regression model to accommodate missing data. In addition, El-Morshedy et al. [15] conducted a detailed study on a new bivariate exponential discrete Weibull distribution.

    The primary goal in this study is to enhance the current techniques for estimating the complex discrete probability distribution model. The probability distribution's score function typically lacks an explicit analytical solution, hence the Newton approach is usually used to estimate the numerical solution for parameter estimation. Nevertheless, the algorithm's low convergence and strong dependence on the initial value make it challenging to achieve the best estimation outcomes. Recently, Liu et al. [16] employed the majorize minimize (MM) algorithm to enhance the resolution of the maximum likelihood estimation for the simplex distribution. Li and Tian [17] introduced a novel root-finding method known as the upper-crossing/solution (US) algorithm. In contrast to conventional iterative algorithms (like Newton's algorithm), the US algorithm can lessen the influence of initial values and achieve a strong, stable convergence to the objective equation's real root at each iteration. The benefits of this technique have been illustrated through the use of a few classic models, such as the Weibull distribution, gamma distribution, zeta distribution, and generalized Poisson distribution. Cai [18] has improved the maximum likelihood estimation of generalized gamma distribution parameters by combining the US algorithm with the second-derivative lower-bound function (SeLF) algorithm.

    The essence of the US algorithm is to identify a U-function U(θ|θ(t)), which simplifies the solution of the complicated nonlinear equation h(θ)=0 to the solution of the equation U(θ|θ(t))=0 with an explicit solution. Li and Tian [17] presented a variety of approaches to discover the U-function, among which the first-derivative lower bound (FLB) function method requires only by using the first derivative of the objective function h(θ), thereby diminishing algorithmic complexity. In previous research on the US algorithm, it was generally used to solve the roots of univariate nonlinear equations or the maximum likelihood estimation problem of a multi-parameter probability distribution with an explicit partial score function. Specifically, for a probability distribution with two parameters (α,β), while solving for maximum likelihood estimation, the estimator of the parameter α can be explicitly expressed by the other estimator of the parameter β(α). However, there is no further discussion provided in Li and Tian [17] about whether this approach can be applicable in more complex multi-parameter distributions, where an estimator of one parameter cannot be clearly represented by the other, and it is an issue deserving of more investigation.

    The rest of the paper will proceed as follows. Section 2 provides a detailed introduction to the US algorithm and FLB function method. The application of the suggested approach to the Type Ⅰ discrete Weibull distribution's maximum likelihood parameter estimation is covered in Section 3. Numerical simulation experiments will be conducted in Section 4 in order to evaluate the performance of the employed methods and compare them with alternative estimation approaches. Section 5 will demonstrate the applicability of the US algorithm through the analysis of two real data sets. Conclusions and discussions will be provided in Section 6.

    One of the most frequent issues in numerical computations is figuring out the zero point of a function or an equation's root. In classical statistics, the maximum likelihood estimate (MLE) of parameters and the calculation of maximum a posteriori probability in Bayesian statistics may typically be turned into the problem of solving the zero point of a nonlinear function h(θ). In summary, since h(θ) is a nonlinear function of a single variable θ, we must identify the unique root θ such that

    h(θ)=0,θΘR. (2.1)

    The US algorithm is the most recent method for discovering roots. It has a similar procedure to the commonly used EM (expectation maximum) and MM (maximize minimize) algorithms[17]. There are two primary steps in this process: the upper-crossing step (U-step) and the solution step (S-step). The two primary advantages of this algorithm are as follows:

    (a) It converges strongly and stably to the root θ of the Eq (2.1) with each iteration, that is, for an iterative points set sequence {θ(t)}t=0, there is

    θ(0)<θ(1)<<θ(t)<θorθ<θ(t)<<θ(1)<θ(0).

    (b) The Newton algorithm's sensitivity to the initial value is decreased.

    Two new symbols, sgn(α) and sgn(α), are introduced regarding the changing direction (CD) inequalities in order to simplify the explanation of the US algorithm. The specific definition is presented as follows: for two functions f1(x) and f2(x) on the same domain Q,

    f1(x)sgn(α)f2(x){f1(x)f2(x),α>0,f1(x)=f2(x),α=0,f1(x)f2(x),α<0,

    and

    f1(x)sgn(α)f2(x){f1(x)f2(x),α>0,f1(x)=f2(x),α=0,f1(x)f2(x),α<0.

    It is typically challenging to locate the root θ of the nonlinear equation h(θ)=0 directly. The US algorithm aims to create an alternative function U(θ|θ(T)) to replace h(θ), transforming the challenge of solving complex nonlinear equations into solving the equation U(θ|θ(T)) with explicit solutions. First, we assume that

    h(θ)<0,θ>θ and h(θ)>0,θ<θ. (2.2)

    If h(θ)<0 when θ<θ, then both sides of the equation h(θ)=0 can be multiplied by -1, which can also obtain the same root θ satisfying the assumption (2.2). Let θ(t) represent the solution after the (t-1)-th iteration, and the function U(θ|θ(t)) satisfying the following criteria is designated as the U-function of h(θ) at θ=θ(t):

    h(θ)U(θ|θ(t)),θ<θ(t),h(θ(t))=U(θ(t)|θ(t)),θ=θ(t),h(θ)U(θ|θ(t)),θ>θ(t). (2.3)

    According to the definition of the CD inequalities symbol, the above condition may be represented as

    h(θ)sgn(θθ(t))U(θ|θ(t)),θ,θ(t)Θ. (2.4)

    As described above, the US algorithm is an iterative approach for solving nonlinear equations, with each iteration including a U-step and an S-step. The purpose of the U-step is to find a U-function that satisfies the condition (2.4), whereas the S-step involves solving the simplified U-equation: U(θ|θ(t))=0 to obtain its root θ(t+1),

    θ(t+1)=sol{U(θ|θ(t))=0,θ,θ(t)Θ}. (2.5)

    In typical scenarios, θ(t+1) can be explicitly expressed, even as a linear equation. Through the iterative execution of these two steps, {θ(t)}t=0 can gradually converge to the real root θ of the U-equation.

    There are numerous U-functions for a given objective function h(θ); as Eq (2.4) illustrates, distinct U-functions correlate to distinct US algorithms. We may express the U-function using the lower-order derivatives of the goal function h(θ). This can be accomplished by a variety of techniques, such as the first-derivative lower bound (FLB), second-derivative lower-upper bound (SLUB) constants method, and third-derivative lower bound (TLB) constant method [17]. These three methodologies enhance efficient solutions when the objective function is complex and the solution is not closed, each with a distinct convergence speed. In terms of maximizing the objective function, the US algorithm based on the FLB approach shares qualities with the EM algorithm and the MM algorithm, both of which exhibit linear convergence. The FLB function technique, which is dependent on the target function's first-order derivative, is mostly used in this article to generate the required U-function. First for parameter space Θ, we suppose that there exists a certain first-derivative lower bound function b(θ) for the first derivative of h(θ), i.e.,

    h(θ)b(θ),θΘ. (2.6)

    The U-function of h(θ) at θ=θ(t) can be formally defined as follows

    U(θ|θ(t))Δ=h(θ(t))+θθ(t)b(z)dz,θ,θ(t)Θ. (2.7)

    In fact,

    h(θ)U(θ|θ(t))=[h(θ)h(θ(t))]θθ(t)b(z)dz=θθ(t)h(z)dzθθ(t)b(z)dz=θθ(t)[h(z)b(z)]dzsgn(θθ(t))0,θ,θ(t)Θ.

    Let θ be the unique root of the equation h(θ)=0, and then the corresponding US iteration is as follows:

    θ(t+1)=sol{U(θ|θ(t))=h(θ(t))+θθ(t)b(z)dz=0,θ,θ(t)>0}Δ=g(θ(t)), (2.8)

    where g(θ(t))=g(θ)+(θ(t)θ)h(θ)+0.5(θ(t)θ)2h(ˆθ) is the first-order Taylor expansion around θ, and θ is a point between θ(t) and θ.

    Although Li and Tian [17] proposed the idea of the US algorithm, in practical applications, only the distribution of univariate and binary parameters were studied. In the case of binary parameter distribution, when discussing the solution of the scoring equation, only one parameter can be explicitly expressed with another parameter. However, when one parameter cannot be explicitly expressed by the other parameter for this more general and complex situation, whether the US algorithm can be effectively applied is not further discussed, which is the issue to be carried out in this article. For the parameters of interest, the new FLB functions are constructed in this article, then, starting with initial values, the iterative values are updated using the corresponding S-step until the convergence criteria are met.

    The discrete Weibull distribution's score function has a complex double exponential form, which makes it impossible to depict its solution and, thus, prevents its two parameters from being mutually expressed. For investigating the US algorithm's applicability in complicated models, we combine the US algorithm with the FLB method in this section to optimize the maximum likelihood estimation.

    Assuming a random variable following the Weibull distribution W(λ,β), where λ>0 and β>0, the cumulative distribution function (CDF) of the Weibull distribution is defined as H(t,λ,β)=1eλtβ, where t>0. Define α=eλ, and then 0<α<1. If the probability mass function (PMF) for a random variable X can be represented as

    P(X=x;α,β)=α([x]1)βα[x]β,(x1),

    then we say that X follows the Type Ⅰ discrete Weibull distribution, denoted as XDW(θ). Here, [x] represents the maximum integer less than or equal to x. When β = 1, the discrete Weibull distribution degenerates to the geometric distribution Geo(q) with q=1α.

    Naturally, the cumulative distribution function of X takes the following form:

    F(x;α,β)=1α[x]β.

    This section will go into detail on the application of the US algorithm for maximum likelihood estimation of the Type Ⅰ discrete Weibull distribution. Assume X is a random variable with the Type Ⅰ discrete Weibull distribution DW(θ), where the parameter vector θ=(α,β)T is in the parameter space ΘR2. Let x=(x1,...,xn) denote the observed values of the random sample (X1,...,Xn). Then, the log-likelihood function of the parameter vector θ is given by

    (θ|x)=ni=1log(α(xi1)βα(xi)β).

    First, the first-order partial derivative of (θ|x) with respect to α can be calculated as

    (θ|x)α=ni=1α1[α(xi1)β(xi1)βαxβixβiα(xi1)βαxβi]Δ=h1(α). (3.1)

    Next, we construct the FLB function regarding the parameter α:

    h1(α)=ni=1[[(xi1)β1](xi1)βα(xi1)β2(xβi1)xβiαxβi2](α(xi1)βαxβi)(α(xi1)βαxβi)2ni=1[α(xi1)β(xi1)βαxβi1xβi]((xi1)βα(xi1)β1xβiαxβi1)(α(xi1)βαxβi)2ni=1[α2(xi1)β2[(xi1)β+(xi1)2β]+αxβi+(xi1)β2[(xi1)2β+x2βi]+α2xβi2[x2β+xβii]](α(xi1)βαxβi)ni=1α2(xi1)β2[2(xi1)2β+2x2βi+(xi1)β+xβi](α(xi1)βαxβi)2ni=14x2βi+2xβi(αα2)2=ni=13[4x2βi+2xβi][1α2+1(1α)2]Δ=b1(α). (3.2)

    Then, the US iteration of α can be obtained as follows:

    α(t+1)=sol[Uα(α,β|α(t),β(t))=h1(α(t))+αα(t)b1(z)dz=0]=sol[C1+3C2[1α11α]3C2[1α(t)11α(t)]=0]]=sol[(C3C1)α2(6C2+C3C1)α+3C2=0], (3.3)

    where C1=h1(α(t)), C2=ni=1[4x2β(t)i+2xβ(t)i], and C3=3C2[1α(t)11α(t)]. Similarly, we can obtain the first-order partial derivative of (θ|x) with respect to β,

    (θ|x)β=ni=1logα[α(xi1)β(xi1)βlog(xi1)αxβixβilog(xi)α(xi1)βαxβi]Δ=h2(β). (3.4)

    In order to construct the FLB function and derive the US algorithm without explicit solutions for the two parameters, we need the following two lemmas.

    Lemma 1. [18] Given that θ>0, we have

    eθ4e2max(θ(t)1,0)(2θ(t)θ)2,0θ2θ(t)andθ(t)>0.

    Lemma 2. [18] For any θ0, we have

    eθ23θ2.

    We will then build the FLB function with respect to β. First, we calculate h2(β)'s first derivative.

    h2(β)=ni=1log(α)[log2(xi1)(xi1)2βα2(xi1)β(1logα)+log2(xi)x2βiα2(xi1)β(1log(α))](α(xi1)βαxβi)2+ni=1log(α)[2log(α)log(xi)log(xi1)xβi(xi1)βlog2(xi1)(xi1)βlog2(xi)xβi]α2xβi(α(xi1)βαxβi)2ni=1[2log2(xi)x2βiα2(xi1)β[1log(α)]2log2(xi1)(xi1)βα2xβi[1log(α)]](α(xi1)βαxβi)ni=12log(α)(1log(α))log2(xi)x2βiα2(xi1)β(α(xi1)βαxβi)22log(α)(1log(α))[1α]2[ni=1log2(xi)x2βi].

    It can be deduced from Lemma 1 and Lemma 2 that

    ni=1log2(xi)x2βi=ni=1e2βlog(xi)log2(xi)ni=1[4e2max(2β(t)log(xi)1,0)I(log(xi)>0)[4β(t)log(xi)2β]2+2I(log(xi)0)3[2βlog(xi)]2]log2(xi)=ni=1[4e2max(2β(t)log(xi)1,0)I(xi>1)(4β(t)2β)2+I(xi=1)6β2],

    where I() is the indicator function. Then, the corresponding FLB function is obtained as follows:

    2log(α)(1log(α))[1α]2ni=1[e2max(2β(t)log(xi)1,0)I(xi>1)(2β(t)β)2+I(xi=1)6β2]Δ=b2(β). (3.5)

    Therefore, the US iterative process for parameter β can be given by

    β(t+1)=sol[Uβ(α,β|α(t+1),β(t))=h2(β(t))+ββ(t)b2(z)dz=0]=sol[h2(β(t))+a1[a22β(t)βI(xi=1)6β]a1[a2β(t)I(xi=1)6β(t)]]=sol[6a3β2a4β+a5=0], (3.6)

    where

    a1=2log(α(t+1))(1log(α(t+1)))(1α(t+1))2,a2=e2max(2β(t)log(xi)1,0)I(xi>1),a3=h2(β(t))a1[a2β(t)I(xi=1)6β(t)],a4=12a3β(t)+6a1a2+a1I(xi=1),a5=2a1I(xi=1)β(t).

    The algorithm process for estimating two parameters can be described as follows. In the first stage, we determine the FLB functions for parameters α and β using Eqs (3.2) and (3.5), respectively. Subsequently, we set two initial values α(t) and β(t), calculate Eq (3.3) to get α(t+1), and then compute β(t+1) via Eq (3.6) using α(t+1) and β(t). If both of the estimates for the parameters satisfy the convergence criteria, then their corresponding values will be returned. Otherwise, we resume to update the iteration value and repeat the preceding steps until the two estimated parameters converge.

    Algorithm :Calculating the MLEs of α and β via the US algorithm.
    Input: The initial value α(0) and β(0); The observed data Xobs={xi}ni=0;
    Output: ˆα,ˆβ.
    1 Select FLB function for parameters α and β, respectively;
    2 Set initial values α(t),β(t), t = 0;
    3 repeat
    4   Using α(t) and β(t), calculate α(t+1) based on (3.3);
    5   Using α(t+1) and β(t), calculate β(t+1) based on (3.6), update t = t + 1;
    6 until convergence.

    In this section, we conduct simulation studies to confirm the applicability to complex nonlinear equations and compare its performance to that of the classic Newton algorithm. First, we provide the calculation steps for parameter estimation using the Newton algorithm as follows:

    Step 1 : α(t+1)=α(t)(θ|x)α(α(t),β(t))/2(θ|x)α2(α(t),β(t)),Step 2 : β(t+1)=β(t)(θ|x)β(α(t+1),β(t))/2(θ|x)β2(α(t+1),β(t)).

    The sample size of the studies is set as n = (50,100,200), and the parameters are set as α=(0.2,0.4,0.6,0.8) and β=(0.5,1.0,1.5,2.0), respectively. We independently generated X(k)1,,X(k)niidDW(α,β), where k=1,,K(K = 1000). The MLE of the parameters under the US algorithm were computed via Eqs (3.2) and (3.4). For every combination of parameters, we ran 1000 iterations of the experiments and evaluated the two methods' fitting performance using the convergence percentage and the mean squared error (MSE) of parameter estimation.

    Tables 14 display the outcomes of the two algorithms' simulations for each scenario. The MSE of the parameters under both algorithms progressively drops as the sample size rises, according to the statistics in the table, suggesting that both techniques are asymptotically unbiased. When the value of β is fixed, as the value of α increases, the MSE of β will steadily decrease. Overall, the MSE of both parameters under the US algorithm is smaller than that of the Newton algorithm, suggesting that the US algorithm performs better when it comes to estimation. Furthermore, it is evident from the table's convergence percentages that the US algorithm is more stable.

    Table 1.  The MSE and percentage from simulated data for β=0.5.
    Sample size 50
    Parameters (α,β)=(0.2,0.5) (α,β)=(0.4,0.5)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00312 0.00234 0.00429 0.00345
    MSE (ˆβ) 0.02222 0.02031 0.00931 0.00631
    Parameters (α,β)=(0.6,0.5) (α,β)=(0.8,0.5)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00422 0.00217 0.00233 0.00117
    MSE (ˆβ) 0.00535 0.00115 0.00547 0.00062
    Sample size 100
    Parameters (α,β)=(0.2,0.5) (α,β)=(0.4,0.5)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00149 0.00122 0.00253 0.00208
    MSE (ˆβ) 0.01294 0.01072 0.00451 0.00307
    Parameters (α,β)=(0.6,0.5) (α,β)=(0.8,0.5)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00193 0.00178 0.00106 0.00081
    MSE (ˆβ) 0.00215 0.00079 0.00207 0.00047
    Sample size 200
    Parameters (α,β)=(0.2,0.5) (α,β)=(0.4,0.5)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00082 0.00069 0.00128 0.00117
    MSE (ˆβ) 0.00647 0.00544 0.00244 0.00183
    Parameters (α,β)=(0.6,0.5) (α,β)=(0.8,0.5)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00106 0.00073 0.00064 0.00048
    MSE (ˆβ) 0.00112 0.00057 0.00102 0.00045

     | Show Table
    DownLoad: CSV
    Table 2.  The MSE and percentage from simulated data for β=1.0.
    Sample size 50
    Parameters (α,β)=(0.2,1.0) (α,β)=(0.4,1.0)
    Algorithms Newton US Newton US
    Percentage 87% 100% 100% 100%
    MSE (ˆα) 0.00273 0.00214 0.00542 0.00259
    MSE (ˆβ) 0.06586 0.05798 0.05064 0.02234
    Parameters (α,β)=(0.6,1.0) (α,β)=(0.8,1.0)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00492 0.00297 0.00269 0.00090
    MSE (ˆβ) 0.02579 0.01286 0.01075 0.00106
    Sample size 100
    Parameters (α,β)=(0.2,1.0) (α,β)=(0.4,1.0)
    Algorithms Newton US Newton US
    Percentage 97% 100% 100% 100%
    MSE (ˆα) 0.00160 0.00114 0.00260 0.00150
    MSE (ˆβ) 0.04316 0.02192 0.02013 0.01119
    Parameters (α,β)=(0.6,1.0) (α,β)=(0.8,1.0)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00196 0.00123 0.00132 0.00065
    MSE (ˆβ) 0.01236 0.00674 0.00922 0.00091
    Sample size 200
    Parameters (α,β)=(0.2,1.0) (α,β)=(0.4,1.0)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00078 0.00048 0.00114 0.00074
    MSE (ˆβ) 0.01835 0.00674 0.00866 0.00607
    Parameters (α,β)=(0.6,1.0) (α,β)=(0.8,1.0)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00109 0.00065 0.00051 0.00030
    MSE (ˆβ) 0.00541 0.00320 0.00319 0.00088

     | Show Table
    DownLoad: CSV
    Table 3.  The MSE and percentage from simulated data for β=1.5.
    Sample size 50
    Parameters (α,β)=(0.2,1.5) (α,β)=(0.4,1.5)
    Algorithms Newton US Newton US
    Percentage 39% 100% 98% 100%
    MSE (ˆα) 0.00529 0.00229 0.00520 0.00276
    MSE (ˆβ) 0.10368 0.10098 0.08433 0.05747
    Parameters (α,β)=(0.6,1.5) (α,β)=(0.8,1.5)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00375 0.00309 0.00334 0.00187
    MSE (ˆβ) 0.05022 0.03443 0.04337 0.00426
    Sample size 100
    Parameters (α,β)=(0.2,1.5) (α,β)=(0.4,1.5)
    Algorithms Newton US Newton US
    Percentage 59% 100% 100% 100%
    MSE (ˆα) 0.00141 0.00108 0.00183 0.00152
    MSE (ˆβ) 0.04498 0.04198 0.05027 0.02460
    Parameters (α,β)=(0.6,1.5) (α,β)=(0.8,1.5)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00193 0.00115 0.00123 0.00071
    MSE (ˆβ) 0.02331 0.00817 0.01853 0.00367
    Sample size 200
    Parameters (α,β)=(0.2,1.5) (α,β)=(0.4,1.5)
    Algorithms Newton US Newton US
    Percentage 85% 100% 100% 100%
    MSE (ˆα) 0.00074 0.00070 0.00121 0.00099
    MSE (ˆβ) 0.03839 0.02496 0.01860 0.00719
    Parameters (α,β)=(0.6,1.5) (α,β)=(0.8,1.5)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00067 0.00049 0.00065 0.00048
    MSE (ˆβ) 0.00921 0.00393 0.00960 0.00324

     | Show Table
    DownLoad: CSV
    Table 4.  The MSE and percentage from simulated data for β=2.0.
    Sample size 50
    Parameters (α,β)=(0.2,2.0) (α,β)=(0.4,2.0)
    Algorithms Newton US Newton US
    Percentage 5% 100% 73% 100%
    MSE (ˆα) 0.00589 0.00235 0.00445 0.00286
    MSE (ˆβ) 0.50894 0.15827 0.06508 0.05573
    Parameters (α,β)=(0.6,2.0) (α,β)=(0.8,2.0)
    Algorithms Newton US Newton US
    Percentage 99% 100% 100% 100%
    MSE (ˆα) 0.00478 0.00198 0.00205 0.00153
    MSE (ˆβ) 0.11310 0.01770 0.06167 0.01236
    Sample size 100
    Parameters (α,β)=(0.2,2.0) (α,β)=(0.4,2.0)
    Algorithms Newton US Newton US
    Percentage 7% 100% 89% 100%
    MSE (ˆα) 0.00258 0.00155 0.00213 0.00153
    MSE (ˆβ) 0.02659 0.07759 0.05018 0.03908
    Parameters (α,β)=(0.6,2.0) (α,β)=(0.8,2.0)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00201 0.00092 0.00121 0.00084
    MSE (ˆβ) 0.04086 0.00654 0.03288 0.00463
    Sample size 200
    Parameters (α,β)=(0.2,2.0) (α,β)=(0.4,2.0)
    Algorithms Newton US Newton US
    Percentage 27% 100% 99% 100%
    MSE (ˆα) 0.00136 0.00077 0.00129 0.00102
    MSE (ˆβ) 0.10796 0.05642 0.04357 0.02541
    Parameters (α,β)=(0.6,2.0) (α,β)=(0.8,2.0)
    Algorithms Newton US Newton US
    Percentage 100% 100% 100% 100%
    MSE (ˆα) 0.00116 0.00054 0.00066 0.00041
    MSE (ˆβ) 0.02345 0.00357 0.01886 0.00200

     | Show Table
    DownLoad: CSV

    The trend of the predicted values of α and β using the US algorithm and Newton technique as the number of iterations grows is shown in Figure 1. From a stability standpoint, the Newton method shows significant instability and often requires multiple twists to preserve the correct trend, while the US algorithm approaches the true values of parameters monotonically. From a convergence speed perspective, the FLB method exhibits linear convergence, while the Newton algorithm demonstrates quadratic convergence. Consequently, the FLB method has a comparatively slower convergence rate.

    Figure 1.  Simulation results of both algorithms.

    This section describes the analysis of two different real data sets to illustrate the applicability of the US algorithm. The first data set in Table 5 contains the remission times in weeks of 20 leukemia patients with treatment studied by Hassan et al. [19]. Another data set is from the National Highway Traffic Safety Administration (www-fars.nhtsa.dot.gov) of the United States, which reports the number of fatalities due to motor vehicle accidents among children under the age of 5 in 32 states during the year 2022.

    Table 5.  The leukaemia patients data and the vehicle fatalities data.
    Data 1 1 3 3 6 7 7 10 12 14 15 18 19 22 26 28 29 34 40 48 49
    Data 2 15 1 6 3 23 6 1 25 14 4 13 15 7 14 2 2 8 12 3 6 12 13 2 8 2 10 5 15 47 7 3 4

     | Show Table
    DownLoad: CSV

    We model the two data sets with the Type Ⅰ discrete Weibull distribution and the geometric distribution. The fitting results for the leukemia patients data by two distributions are provided in Table 6. The results of the Cramer-von Mises test, Anderson-Darling test, and Kolmogorov-Smirnov test show that the two distributions can successfully fit the data set. The p-values from the three tests of the Type Ⅰ discrete Weibull distribution employing the US algorithm demonstrate the best fitting effect. Moreover, the values of Akaike information criterion (AIC) [20] and Bayesian information criterion (BIC) [21] also show that the DW distribution based on the US algorithm has better estimation effect.

    Table 6.  Fitting for the leukaemia patients data by the DW distribution and the geometric distribution.
    Methods Estimates AIC BIC p-value(KS) p-value(AD) p-value(CVM)
    DW distribution (New) α=0.9282 β=0.9477 163.1752 165.1667 0.4207 0.1952 0.1533
    DW distribution (US) α=0.9513 β=1.0000 161.9296 163.9211 0.9722 0.8817 0.8463
    Geo distribution q=0.0512 161.9783 162.9741 0.7035 0.4456 0.3946

     | Show Table
    DownLoad: CSV

    Table 7 shows the results of fitting the vehicle fatality data with the Type Ⅰ discrete Weibull distribution and the geometric distribution. Comparing the p-values of the three tests at the significance level α=0.05 reveals that the Geo distribution and the Type Ⅰ discrete Weibull distribution estimated by the Newton algorithm are considered to be insufficient. The fitting effectiveness of the DW distribution using the US algorithm is significant, as indicated by the values of AIC and BIC. The histogram for two different data sets evaluated by the DW distribution and the Geo distribution is shown in Figure 2. Figures 3 and 4 present QQ plots for these two distributions. It is also evident that the US algorithm performs better.

    Table 7.  Fitting for the vehicle fatalities data by the DW distribution and the geometric distribution.
    Methods Estimates AIC BIC p-value(KS) p-value(AD) p-value(CVM)
    DW distribution (New) α=0.9122 β=1.1756 212.8066 215.2212 0.1822 0.1324 0.1755
    DW distribution (US) α=0.8852 β=0.9672 209.7482 212.6796 0.4707 0.3948 0.4147
    Geo distribution q=0.1039 214.4938 215.9596 0.0957 0.0857 0.0941

     | Show Table
    DownLoad: CSV
    Figure 2.  Histogram of leukaemia patients data (left panel) and vehicle fatalities data (right panel), and the correlation density curve fitted by the DW and Geo distributions.
    Figure 3.  QQ plots of the first data for the DW (US) (left panel), DW (Newton) (middle panel), and Geo distributions (right panel).
    Figure 4.  QQ plots of the second data for the DW (US) (left panel), DW (Newton) (middle panel), and Geo distributions (right panel).

    The US algorithm is a novel iterative method with high stability and convergence. The existing research only involves simple models such as univariate nonlinear equations or univariate functions. This paper extends the US algorithm to more complex cases of two parameter discrete distribution functions, where one parameter cannot be explicitly represented by the other parameter estimate. In order to successfully estimate the parameters, this paper combines the FLB method to perform optimization estimation of a distribution function. The simulation results for the Type Ⅰ discrete Weibull distribution demonstrate that the US algorithm has good accuracy and stability. Simultaneously, for the purpose of demonstrating the applicability of the algorithm in complex situations, this paper conducted empirical research on two real data sets that follow the Type Ⅰ discrete Weibull distribution, namely, the data from patients with leukemia and children who die from motor vehicle accidents. After comparing and analyzing the US method with the conventional Newton algorithm, the results show that the recommended strategy has an excellent fitting effect.

    Yuanhang Ouyang: Formal analysis, Writing original draft, Software, Investigation, Methodology, Data curation; Ruyun Yan: Validation, Software, Formal analysis; Jianhua Shi: Validation, Resources, Writing-review and editing, Methodology. All authors have read and approved the final version of the manuscript for publication.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This research was conducted under a project titled "The National Social Science Fund of China" (20XTJ003).

    The authors declare that there are no conflicts of interest regarding the publication of this paper.



    [1] M. Szymkowiak, M. Iwińska, Characterizations of discrete weibull related distributions, Stat. Probab. Lett., 111 (2016), 41–48. https://doi.org/10.1016/j.spl.2016.01.004 doi: 10.1016/j.spl.2016.01.004
    [2] D. Roy, Reliability measures in the discrete bivariate set-up and related characterization results for a bivariate geometric distribution, J. Multivar. Anal., 46 (1993), 362–373. https://doi.org/10.1006/jmva.1993.1065 doi: 10.1006/jmva.1993.1065
    [3] D. Roy, The discrete normal distribution, Commun. Stat.-Theory Methods, 32 (2003), 1871–1883.
    [4] H. Krishna, P. S. Pundir, Discrete burr and discrete pareto distributions, Stat. Methodol., 6 (2009), 177–188. https://doi.org/10.1016/j.stamet.2008.07.001 doi: 10.1016/j.stamet.2008.07.001
    [5] C. Bracquemond, O. Gaudoin, A survey on discrete lifetime distributions, Int. J. Reliab., Qual. Saf. Eng., 10 (2003), 69–98. https://doi.org/10.1142/S0218539303001007 doi: 10.1142/S0218539303001007
    [6] K. Samuel, N. Balakrishnan, N. L. Johnson, Continuous Multivariate Distributions: Models and Applications, New York: John Wiley & Sons, Inc, 2000. https://doi.org/10.1002/0471722065
    [7] T. Nakagawa, S. Osaki, The discrete weibull distribution, IEEE Trans. Reliab., R-24 (1975), 300–301. https://doi.org/10.1109/TR.1975.5214915 doi: 10.1109/TR.1975.5214915
    [8] W. E. Stein, R. Dattero, A new discrete weibull distribution, IEEE Trans. Reliab., R-33 (1984), 196–197. https://doi.org/10.1109/TR.1984.5221777 doi: 10.1109/TR.1984.5221777
    [9] W. J. Padgett, J. D. Spurrier, On discrete failure models, IEEE Trans. Reliab., R-34 (1985), 253–256.
    [10] J. D. Englehardt, R. Li, The discrete weibull distribution: An alternative for correlated counts with confirmation for microbial counts in water, Risk Anal.: Int. J., 31 (2011), 370–381. https://doi.org/10.1111/j.1539-6924.2010.01520.x doi: 10.1111/j.1539-6924.2010.01520.x
    [11] A. Barbiero, A comparison of methods for estimating parameters of the type i discrete weibull distribution, Stat. Interface, 9 (2016), 203–212. https://doi.org/10.4310/SII.2016.v9.n2.a7 doi: 10.4310/SII.2016.v9.n2.a7
    [12] A. Barbiero, Least-squares and minimum chi-square estimation in a discrete weibull model, Commun. Stat.-Simul. Comput., 46 (2017), 8028–8048. https://doi.org/10.1080/03610918.2016.1263733 doi: 10.1080/03610918.2016.1263733
    [13] R. Vila, E. Y. Nakano, H. Saulo, Theoretical results on the discrete weibull distribution of nakagawa and osaki, Statistics, 53 (2019), 339–363. https://doi.org/10.1080/02331888.2018.1550645 doi: 10.1080/02331888.2018.1550645
    [14] H. Yoo, Application of discrete weibull regression model with multiple imputation, Commun. Stat. Appl. Methods, 26 (2019), 325–336. https://doi.org/10.29220/CSAM.2019.26.3.325 doi: 10.29220/CSAM.2019.26.3.325
    [15] M. El-Morshedy, M. Eliwa, A. El-Gohary, A. Khalil, Bivariate exponentiated discrete weibull distribution: Statistical properties, estimation, simulation and applications, Math. Sci., 14 (2020), 29–42.
    [16] P. Liu, K. C. Yuen, L. C. Wu, G. L. Tian, T. Li, Zero-one-inflated simplex regression models for the analysis of continuous proportion data, Stat. Interface, 13 (2020), 193–208.
    [17] X. Li, G.-L. Tian, The upper-crossing/solution (US) algorithm for root-finding with strongly stable convergence, 2022. https://doi.org/10.48550/arXiv.2212.00797
    [18] Y. Cai, Maximum likelihood estimates of parameters in generalized gamma distribution with self algorithm, 2023. https://doi.org/10.48550/arXiv.2306.16419
    [19] H. S. Bakouch, M. A. Jazi, S. Nadarajah, A new discrete distribution, Statistics, 48 (2012), 200–240. https://doi.org/10.1080/02331888.2012.716677
    [20] H. Bozdogan, Model selection and Akaike's information criterion (AIC): The general theory and its analytical extensions, Psychometrika, 52 (1987), 345–370. https://doi.org/10.1007/BF02294361 doi: 10.1007/BF02294361
    [21] S. I. Vrieze, Model selection and psychological theory: A discussion of the differences between the akaike information criterion (AIC) and the bayesian information criterion (BIC), Psychol. Methods, 17 (2012), 228. https://doi.org/10.1037/a0027127 doi: 10.1037/a0027127
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(816) PDF downloads(45) Cited by(0)

Figures and Tables

Figures(4)  /  Tables(7)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog