Research article Special Issues

Solving ANOVA problem with restricted Type Ⅰ and Type Ⅱ error rates

  • Received: 16 November 2024 Revised: 25 December 2024 Accepted: 08 January 2025 Published: 10 February 2025
  • MSC : 62F15, 62F03

  • The problem of solving the ordered one-way analysis of variance (ANOVA) (which consists of comparing a set of normal means) with restricted Type Ⅰ and Type Ⅱ error rates is considered in this paper. This case is more complicated than unordered one-way ANOVA because the detection of the monotonicity of means restrictions is necessary. To solve this issue, one of the possible formulations of the constrained Bayesian method (CBM) is applied here using the concept of directional hypotheses. The cases of known and unknown variances are examined. For unknown variances, the maximum likelihood ratio and Stein's Methods were used to overcome the problem connected with the complexity of hypotheses. The correctness of the developed methods and high quality (in comparison with existing methods) of obtained results were demonstrated by computing results of the simulated concrete scenarios. Moreover, the offered method controlled not only one Type of error, as methods do, but both Type Ⅰ and Type Ⅱ methods.

    Citation: Kartlos J. Kachiashvili, Joseph K. Kachiashvili, Ashis SenGupta. Solving ANOVA problem with restricted Type Ⅰ and Type Ⅱ error rates[J]. AIMS Mathematics, 2025, 10(2): 2347-2374. doi: 10.3934/math.2025109

    Related Papers:

    [1] Ran Tamir . Testing for correlation in Gaussian databases via local decision making. AIMS Mathematics, 2025, 10(4): 7721-7766. doi: 10.3934/math.2025355
    [2] Hassan Okasha, Mazen Nassar, Saeed A. Dobbah . E-Bayesian estimation of Burr Type XII model based on adaptive Type-Ⅱ progressive hybrid censored data. AIMS Mathematics, 2021, 6(4): 4173-4196. doi: 10.3934/math.2021247
    [3] Haiping Ren, Xue Hu . Estimation for inverse Weibull distribution under progressive type-Ⅱ censoring scheme. AIMS Mathematics, 2023, 8(10): 22808-22829. doi: 10.3934/math.20231162
    [4] Bing Long, Zaifu Jiang . Estimation and prediction for two-parameter Pareto distribution based on progressively double Type-II hybrid censored data. AIMS Mathematics, 2023, 8(7): 15332-15351. doi: 10.3934/math.2023784
    [5] Li Dong, Jingyong Tang . New convergence analysis of a class of smoothing Newton-type methods for second-order cone complementarity problem. AIMS Mathematics, 2022, 7(9): 17612-17627. doi: 10.3934/math.2022970
    [6] Amany E. Aly, Magdy E. El-Adll, Haroon M. Barakat, Ramy Abdelhamid Aldallal . A new least squares method for estimation and prediction based on the cumulative Hazard function. AIMS Mathematics, 2023, 8(9): 21968-21992. doi: 10.3934/math.20231120
    [7] Xue Hu, Haiping Ren . Statistical inference of the stress-strength reliability for inverse Weibull distribution under an adaptive progressive type-Ⅱ censored sample. AIMS Mathematics, 2023, 8(12): 28465-28487. doi: 10.3934/math.20231457
    [8] Tahani A. Abushal . Parametric inference of Akash distribution for Type-Ⅱ censoring with analyzing of relief times of patients. AIMS Mathematics, 2021, 6(10): 10789-10801. doi: 10.3934/math.2021627
    [9] Ahmed R. El-Saeed, Ahmed T. Ramadan, Najwan Alsadat, Hanan Alohali, Ahlam H. Tolba . Analysis of progressive Type-Ⅱ censoring schemes for generalized power unit half-logistic geometric distribution. AIMS Mathematics, 2023, 8(12): 30846-30874. doi: 10.3934/math.20231577
    [10] Mazen Nassar, Refah Alotaibi, Ahmed Elshahhat . Reliability analysis at usual operating settings for Weibull Constant-stress model with improved adaptive Type-Ⅱ progressively censored samples. AIMS Mathematics, 2024, 9(7): 16931-16965. doi: 10.3934/math.2024823
  • The problem of solving the ordered one-way analysis of variance (ANOVA) (which consists of comparing a set of normal means) with restricted Type Ⅰ and Type Ⅱ error rates is considered in this paper. This case is more complicated than unordered one-way ANOVA because the detection of the monotonicity of means restrictions is necessary. To solve this issue, one of the possible formulations of the constrained Bayesian method (CBM) is applied here using the concept of directional hypotheses. The cases of known and unknown variances are examined. For unknown variances, the maximum likelihood ratio and Stein's Methods were used to overcome the problem connected with the complexity of hypotheses. The correctness of the developed methods and high quality (in comparison with existing methods) of obtained results were demonstrated by computing results of the simulated concrete scenarios. Moreover, the offered method controlled not only one Type of error, as methods do, but both Type Ⅰ and Type Ⅱ methods.



    The problem of comparing the means of several normally distributed random variables has attracted researches' attention from the middle of the last century. Such problems arise in many practical applications such as when comparing treatment effects in clinical trials, or experimental studies to compare the effects of different experiments, or when comparing the failure rate of a machine with its age. In the framework of ANOVA, the problem can be formulated as follows [1]. Let us consider k normal populations N(μi,σ2i), i=1,...,k. Let Xi1,...,Xini be a random sample from N(μi,σ2i), independent from each other. The problem is to test hypotheses

    H0:μ1=...=μk vs. Ha:μiμj for some ij. (1)

    In some studies, researchers may believe that the true means are simply ordered (μ1...μk) before observing the data. For example, the treatment means may be a sequence of increasing dose levels of a drug. In this case, ANOVA takes the form

    H0:μ1=...=μk vs. Ha:μ1...μk, (2)

    with at least one strict inequality. This statement has received considerable attention in the statistical literature (see, for example, [1,2,3,4,5]).

    The simplest case k=2, i.e., the case with only two normal populations, called the Behrens–Fisher problem, is well studied. For its solution, at statement (1), a simple and accurate method is introduced in [6,7], which gives good results in both cases whether variances are equal or not. For the arbitrary case; moreover, when k>2, a compact but quite good review of the methods used to solve problem (1) is given in [8], on the basis that there is a procedure that gives satisfactory results by the Type Ⅰ error rate for all sample sizes, where k and parameter configurations do not exist [8,9]. To improve the situation, a parametric bootstrap (PB) approach is offered in [8,9] and it is compared with the most promoted tests: The Welch test, the James ([10]) second-order test, and the generalized F (GF) test [11,12]. The comparison was realized by Type Ⅰ error rate and powers using Monte Carlo simulation, showing that the best, based on the sample sizes, values of the error variances, and the number of means being compared is a PB test. A little worse is the James second-order test. The Welch test and the GF test have fairly poor Type Ⅰ errors when the value of k is medium or large and/or the sample sizes are small. In computed examples of [8], the minimum value of Monte Carlo estimates of Type Ⅰ error rates is equal to 0.04 even when k=2, meaning it is quite high. For other k, as a rule, it is greater.

    For statement (2), a one-sided studentized range test is offered in Hayter [3]. It provides simultaneous one-sided lower confidence bounds for the ordered pairwise comparisons μiμj, 1j<ik. Similar inference procedures are discussed in [13,14,15,16,17]. In some, the strong order of arranging the mathematical expectations is considered, and in others, the assumption that μi's follow a simple ordering is not made. The familywise error rate (FWER) is used as an optimality criterion in these cases, the computation of which is complicated for large k. To overcome this problem, the empirical Bayesian approach of multiple testing is offered in [1]. Its essence consists of the consideration of the families of hypotheses

    Hi0:μi+1μi=0 vs. Hia:μi+1μi>0,i=1,...,k1, (3)

    from a Bayesian view point. To test (3), there is a new step-down multiple testing procedure, which is a step-down version of the single-step procedure [18]. It controls the familywise error rate better than a single-step procedure, which provides a simple comparison of each p-value with a single critical value. For this reason, the step-down version is more powerful than the single-step procedure.

    The heart of the single-step procedure consists of the rejection of the Hi0 hypothesis if piα, where pi is Hi0 hypothesis corresponding p-value, and α(0,1) is the significant level. In the step-down test, the differences Ti=μi+1μi and the appropriate hypotheses are in ascending order. Testing hypothesis (3) begins with the largest T-value, i.e., with T(k1). If for some T(ki) hypothesis H(ki)0 is accepted, then testing is stopped, and all the remaining hypotheses whose order number is less than (ki) are accepted; otherwise, H(ki)0 is rejected, and the (i+1) th step is transited. In [1], the critical constants for each stage chosen so that FWER is controlled at a pre-specified level α. Because the empirical Bayesian approach is used in this method, which is based on the resampling techniques, it asymptotically controls the FWER.

    An alternative method for testing hypothesis (2), similar to the parametric bootstrap method of [8] to improve the Type-Ⅰ error rate, is offered in [4]. In particular, this method is an extended Hayter's test [3] for heterogeneous sample sizes and variances, and uses the idea of the PB offered in [8], based on the simulation of the sample from the null distribution with unbiased estimators of the parameters. The non-significant improvement of Type Ⅰ error compared to existing methods of the Welch test (WT) [19]), the ˉE2 test (which is a counterpart of the F test in the ordered case [5]) and Chen's one-stage method ([20]) for large sample sizes is shown through the modeling for the introduced novelty.

    For justice, it should be said that all the considered tests have quite high Type Ⅰ and Type Ⅱ error rates, the values that depend on the parameters of the problem under consideration (the values of k, variances and sample sizes) and their control is impossible at a predetermined level. Therefore, to solve problems (1) and (2), we aim to develop a method that will enable us to make a decision by restricting Type Ⅰ and Type Ⅱ error rates to a predetermined level. To achieve this goal, the CBM is used below [21,22].

    The structure of the rest of the paper is as follows. The application of CBM to solve the stated problem is given in Section 2, where formulas and criteria for testing individual and multiple Hypotheses defined by (6) are given. Algorithms for computation for known and unknown variances are developed in Section 3. Computation results for practical examples at known and unknown variances when the maximum likelihood ratio and Stein's Methods are used for overcoming the problem caused by complexity of hypotheses, are presented in Section 4. A short discussion of the obtained results is offered in Section 5, and a conclusion is presented in Section 6.

    Without loss of generality, we can suppose that all μi (i=1,...,k) are positive, because if this condition is not fulfilled, we can always achieve it by moving the origin.

    Let us rewrite hypothesis (1) as follows

    H0:θ1=θ2=...=θk1=0 vs. Ha:θi0, for some i(1,...,k1). (4)

    where θi=μi+1μi, i=1,...,k1. In this case, hypothesis (2) takes the form

    H0:θ1=θ2=...=θk1=0 vs. Ha:θi0, with at least one strict inequality.  (5)

    Using the concept of directional hypotheses [22,23,24], similarly to (3), hypothesis (4) can be considered as a set of directional hypotheses

    Hi0:θi=0 vs. Hi:θi<0 or Hi+:θi>0,i=1,...,k1. (6)

    Statement (6) enables us to test hypotheses (1) and (2) (which are the same as hypotheses (4) and (5)) simultaneously, using techniques developed in [22,23]. In particular: After testing hypothesis (6), if for even one value of i, the left sided hypothesis is accepted among accepted alternatives, accept alternative hypothesis in (4); and if all the accepted alternative hypotheses are right sided, accept alternative hypothesis in (5); otherwise accept the null hypothesis.

    Let us recall that for testing hypothesis (6), we have the following information: Xi1,...,Xini - a random sample from N(μi,σ2i), i=1,...,k, on the basis of which we compute

    ˉXi=1ninii=1Xij and S2i=1ni1nij=1(XijˉXi)2,i=1,...,k, (7)

    where ˉXi and S2i are the i th sample mean and sample variance.

    To test hypotheses concerning parameter θi, sufficient statistics is ˉYi=ˉXi+1ˉXi, where the distribution is N(ˉyi|θi,σ2i/ni+σ2i+1/ni+1), i=1,...,k1, because of independence of ˉXi+1 and ˉXi. Parameter θi=0 at basic hypothesis Hi0, and θi<0 or θi>0, at alternative hypotheses Hi or Hi+, accordingly.

    For ease of presentation and understanding, let us first consider one subset of individual hypotheses of (6), and for simplicity, omit index i. Let us use one of the possible formulations of CBM, in particular, restrictions on posterior probabilities of rejected true hypotheses (Task 7) considered in [22,23]. In this formulation, the Type Ⅱ error rate is minimized (that means the power is maximized), and the Type Ⅰ error rate is restricted. This means, that the statement of the problem is the following [21,23]:

    Gδ=max{Γ,Γ0,Γ+}{K0[p(H)P(xΓ|H)+p(H0)P(xΓ0|H0)+p(H+)P(xΓ+|H+)]} (8)

    subject to

    K1[p(H0)P(xΓ|H0)+p(H+)P(xΓ|H+)]r7,K1[p(H)P(xΓ0|H)+p(H+)P(xΓ0|H+)]r07,K1[p(H)P(xΓ+|H)+p(H0)P(xΓ+|H0)]r+7. (9)

    Here, for each subset of individual hypotheses of (6), the following loss functions are used

    L1(Hi,δj(x)=1)={0ati=j,K1atij; and L2(Hi,δj(x)=0)={K0ati=j,0atij; (10)

    where K1 is the loss of incorrectly accepted Hj when Hi is true and K0 the loss of incorrectly rejected Hj at testing it versus Hi, i,j(,0,+).

    Using the undetermined Lagrange multiplier method and the concepts of posterior probabilities, to solve problems (8) and (9), we have [23]

    Γ={x:K1(p(H0|x)+p(H+|x))<1λ7K0p(H|x)},Γ0={x:K1(p(H|x)+p(H+|x))<1λ07K0p(H0|x)},Γ+={x:K1(p(H|x)+p(H0|x))<1λ+7K0p(H+|x)}, (11)

    where Lagrange multipliers λ7, λ07, and λ+7 are determined so that in (9), the equalities hold.

    The most common criteria of estimation of the quality of testing directional hypotheses are the mixed directional false discovery rate (mdFDR) and the false acceptance rate (FAR), which have the following forms [21,22]

    mdFDR=P(ˉyΓ|H+)+P(ˉyΓ|H0)+P(ˉyΓ+|H)+P(ˉyΓ+|H0), (12)

    and

    FAR=P(ˉyΓ0|H)+P(ˉyΓ0|H+), (13)

    respectively.

    According to the theorems proved in [23], if the condition r7+r+7=qK1Pmin is satisfied, where 0<q<1, Pmin=min{p(H,H0,H+)}, then the following condition mdFDRq is fulfilled, and if the condition r7+r07+r+7=qK1Pmin is satisfied, the following condition mdFDR+FARq is fulfilled. Also, when λ07 is chosen so that in the second condition of (9) equality takes place, the following is fulfilled:

    FAR1Pminr07K1, (14)

    where Pmin=min{p(H),p(H+)}.

    For the case of directional hypotheses, the Type Ⅰ error rate (TIER) can be determined as follows

    TIER=P(ˉyΓ|H0)+P(ˉyΓ+|H0). (15)

    Comparing (12) and (15), it is clear that mdFDRTIER.

    Let us define Type Ⅱ error rates (TIIER) as follows

    TIIER=P(ˉyΓ0|H)+P(ˉyΓ0|H+). (16)

    Comparing (13) and (16), we are convincing that FAR=TIIER. Thus, to fulfill the condition 1Pminr07K1=q, the following takes place FAR=TIIERq, where 0<q<1.

    Theorem 1. CBM 7 with restriction levels (9), to satisfy a condition (r7+r+7)/(K1p(H0))=q, where 0<q<1, ensures a decision rule with Type I error rate less or equal to q, i.e., with the condition TIERq.

    Proof. From the first and the third conditions of the restrictions of (9), we write the following

    P(ˉyΓ|H0)[r7K1p(H+)P(ˉyΓ|H+)]1p(H0)=
    =r7K1p(H0)p(H+)p(H0)P(ˉyΓ|H+),

    and

    P(ˉyΓ+|H0)[r+7K1p(H)P(ˉyΓ+|H)]1p(H0)=
    =r+7K1p(H0)p(H)p(H0)P(ˉyΓ+|H).

    From here, and considering (15), we are convincing in the correctness of theorem 1.

    Because of the specific nature of the acceptance regions of CBM [21,22,23], it can so happen that making a simple decision on the basis of existed information becomes impossible, e.g., when the test statistic belongs to the intersection areas of the acceptance regions or does not belong to any of these regions. In such a situation, it becomes impossible to make a simple decision with a specified confidence level on the basis of the information, and more information is required to achieve this. If acquiring more information is impossible, then the restriction levels in (9) must be changed until a simple decision can be made. When acquiring more information is possible, we appeal to the sequential experiment, i.e., to increase a sample size and apply Procedure A.

    Procedure A

    The statistics for testing a subset of individual hypotheses of (6) is ˉyi=ˉxi+1ˉxi, which depends on two initial samples xi1,...,xin,... and xi+1,1,...,xi+1,n,.... Let us denote the test statistics on the basis of n observations by ˉyi,n. Then the sequential procedure is as follows:

    Step 1

    - if ˉyi,n belongs to only region Γi, accept hypothesis Hi,

    - if ˉyi,n belongs to only region Γi0, accept hypothesis Hi0,

    - if ˉyi,n belongs to only region Γi+, accept hypothesis Hi+,

    - otherwise continue sampling; collect next observations and compute new test statistics ˉyi,n+1;

    Step 2

    - if ˉyi,n+1 belongs to only region Γi, accept hypothesis Hi,

    - if ˉyi,n+1 belongs to only region Γi0, accept hypothesis Hi0,

    - if ˉyi,n+1 belongs to only region Γi+, accept hypothesis Hi+,

    - otherwise, continue sampling; collect observations and compute test statistics etc.

    The sampling continues until the test statistics belong to only one hypothesis acceptance region.

    Note 1: Obviously, the sample size at the beginning of the sequential test is equal to any particular value, i.e., n=n (in particular, it can be n=1), and if the desired level of reliability of making a decision is achievable for this amount of information, this corresponds to the parallel experiment on which the testing process finishes. In case of continuing sampling, the parallel experiment generalizes to the sequential experiment naturally.

    For testing multiple directional hypotheses (6), the concept of the total mixed directional false discovery rate (tmdFDR) is introduced in [23,25]

    tmdFDR=k1i=1mdFDRi. (17)

    To guarantee level q, in testing hypothesis (6), we have to consider k1 subsets of directional hypotheses. Then, for each of them, we use procedure A described above to provide a level of qi for i th subset of hypotheses, so that k1i=1qi=q is achieved.

    We act similarly to provide a level q for the total Type Ⅰ error rate (TTIER) and for the total Type Ⅱ error rate (TTIIER). Namely, we provide qi, the level of the appropriate criteria for the i th subset of the individual directional hypotheses. As a result, we have

    TTIER=k1i=1TIERi, (18)
    TTIIER=k1i=1TIIERi, (19)

    where TIERi and TIIERi are the Type Ⅰ error rate and the Type Ⅱ error rate, respectively, of the i th subset of directional hypotheses [23,26].

    The values of qi in all three cases (for tmdFDR, for TTIER, and for TTIIER) can be chosen to be equal, i.e., qi=q/(k1) or different, e.g., inversely proportional to the informational distances between the tested hypotheses in the subsets of directional hypotheses [23].

    To test multiple hypotheses, two family-wise error rates, type Ⅰ (FWERI) and type Ⅱ (FWERII), are used (see, for example, [26,27,28]), which are the same as TTIER and TTIIER in the considered case, i.e., we have TTIER=FWERI and TTIIER=FWERII.

    In all three considered cases, with the restriction of tmdFDR, TTIER, and TTIIER on the desired levels, we use the above described sequential Procedure A, where the sampling continues until a simple decision is made for all the subsets of multiple hypotheses (6). The stopping rules remain the same as in [23]. Theorems 3.2 and 3.3 of the work [23], proving the appropriateness of the stopping rule, are in force for the considered directional hypotheses as well.

    As mentioned in [24], "Currently, in many real-life applications, we indeed encounter situations where the number of individual hypotheses in the set of multiple hypotheses (6) is very big, i.e., when data is big. In such a situation, determination of Lagrange multipliers for each subset of an individual hypothesis requires a long time for computation. Though the computation of Lagrange multipliers is completed in the preparatory stage before making a decision, still the reduction of computation time is important for many practical applications from the operational and cost considerations." For this purpose, Theorem 4 is proved in [24] in accordance with the testing of all of subsets of individual hypotheses of (6) and can be used once with the same Lagrange multipliers, determined for a subset of individual hypothesis with lowest divergence among directional hypotheses at the level qi=q/(k1), satisfying condition k1i=1qi=q provides the total mixed directional false discovery rate (17), as well as the total Type Ⅰ error rate (18) and the total Type Ⅱ error rate (19), restricting all of them with level q.

    Let us suppose that the variances of Xi that is σ2i, i=1,...,k, are known. To realize the above developed method, the concretization of distribution laws of the statistics ˉYi at hypotheses H(0)i, H()i and H(+)i is necessary, i.e., it is necessary to determine densities p(ˉyi|H(0)i), p(ˉyi|H()i), and p(ˉyi|H(+)i).

    As mentioned in the above ˉYi is normally distributed. Therefore, taking into account the intervals of definition of the mathematical expectation at different hypotheses, we can suppose that p(ˉyi|H(0)i)=N(ˉyi|0,ˉσ2i) at H(0)i and p(ˉyi|H()i) and p(ˉyi|H(+)i) are the truncated normal densities N(0,ω10ˉσ2i) (ω0 known) over (,0) and (0,+), respectively [23,29]. Here, ˉσ2i=(σ2i+σ2i+1)/n=ϑ2i/n, were ϑ2i=(σ2i+σ2i+1). Thus, we have [24]

    p(ˉyi|H0i)=n2πϑiexp{nˉy2i2ϑ2i},p(ˉyi|Hi)=0n2πϑiexp{n(ˉyiθi)22ϑ2i}2ω02πϑiexp{ω20θ2i2ϑ2i}dθi==2ω0n+ω0n2πϑi(1Φ(ui))exp{ω0u2i2n},p(ˉyi|H+i)=+0n2πϑiexp{n(ˉyiθi)22ϑ2i}2ω02πϑiexp{ω20θ2i2ϑ2i}dθi==2ω0n+ω0n2πϑiΦ(ui)exp{ω0u2i2n}. (20)

    For these distributions, hypotheses acceptance regions (11) for observations ˉyi are the following [23]

    Γ={ˉyi:p(H0)n+ω0exp{u2i/2}+p(H+)2ω0Φ(ui)p(H)2ω0(1Φ(ui))<1λ7K0K1},Γ0={ˉyi:p(H)2ω0(1Φ(ui))+p(H+)2ω0Φ(ui)p(H0)n+ω0exp{u2i/2}<1λ07K0K1},Γ+={ˉyi:p(H)2ω0(1Φ(ui))+p(H0)n+ω0exp{u2i/2}p(H+)2ω0Φ(ui)<1λ+7K0K1}. (21)

    Here, n is the quantity of ˉyi on the basis of which hypothesis (6) is tested using sequential Procedure A, ui=nˉyi/(ϑin+ω0), and Φ() is the standard normal distribution function.

    The Lagrange multipliers are determined so that in condition (9), the equalities hold. For the solution of the relevant equations, the suitable probability integrals are computed by the Monte-Carlo method (see, for example, [21,23]).

    Let us suppose that the variances of Xi that is σ2i, i=1,...,k, are not known. In this case, we can consider two methods based on the maximum likelihood ratio and Stein's method [30,31,32,33,34]. The essence of these methods consist of the following. In the maximum likelihood ratio approach, the density with unknown parameters' values is replaced with the maximum of the density over all possible values of these parameters at the truth of testing hypotheses, computed by existing samples. Stein's method integrates the density over unknown parameters using special measures to obtain the density of the maximum invariant statistic, which can then be used to analyze the problems; for example, to find the uniformly most powerful invariant test [32].

    The method based on the maximum likelihood ratio can be applied when the numbers of observations ni, i=1,...,k, are large, which enable us to divide every subset of the observations into two parts xi,1,...,xi,mi and xi,mi+1,...,xi,ni, where mi<ni. Using mi observations, ˉS2i=S2i+S2i+1, i=1,...,k1, estimators of the variance ˉσ2i of the statistics ˉYi=ˉXi+1ˉXi are computed and are used in formulae (20) and (21) instead of ˉσ2i. Statistics ui are computed using observations xi,mi+1,...,xi,ni, i=1,...,k, to make decisions based on hypotheses acceptance regions (21) when testing hypothesis (6).

    When applying Stein's method, to average the influence of ˉσ2i, it is necessary to use a priori distribution of ˉS2i, which is difficult to determine functionally with known parameters at the considering case. Therefore, we conduct the following: Construct a confidence interval of ˉS2i=S2i+S2i+1 with the given confidence probability 1α (0<α<1) and assume that ˉS2i takes values from this interval according to a uniform distribution.

    Finding confidence intervals of ˉS2i, i=1,...,k1 is not difficult. It looks like:

    ˉS2i,L<ˉσ2i<ˉS2i,U, (22)

    where

    ˉS2i,L=(mi1)S2iχ2mi1,1α+(mi+11)S2i+1χ2mi+11,1α,ˉS2i,U=(mi1)S2iχ2mi1,α+(mi+11)S2i+1χ2mi+11,α (23)

    where χ2mi1,α is a quantile of the α-th order of the chi-square distribution with degree of freedom mi1.

    To overcome the uncertainty in the densities (20) caused by not knowing the value of the variance, we are averaging the effect of the variance on these densities by integrating them over the confidence interval of the variance with uniform density. In this case, conditional distribution densities (20) take the form

    p(ˉyi|H0i)=kiˉS2i,uˉS2i,L1xexp{aix2}dx,p(ˉyi|Hi)=di0ˉS2i,uˉS2i,L1x2exp{aibiz+ciz2x2}dzdx,p(ˉyi|H+i)=di+0ˉS2i,uˉS2i,L1x2exp{aibiz+ciz2x2}dzdx, (24)

    where ki=ni2π(ˉS2i.uˉS2i.L), di=niω0π(ˉS2i.uˉS2i.L), ai=niˉy2i2, bi=niˉyi, ci=ni+ω02.

    Let us denote:

    p0(ˉyi|H0i)=ˉS2i,uˉS2i,L1xexp{aix2}dx,p(ˉyi|Hi)=0ˉS2i,uˉS2i,L1x2exp{aibiz+ciz2x2}dzdx,p+(ˉyi|H+i)=+0ˉS2i,uˉS2i,L1x2exp{aibiz+ciz2x2}dzdx, (25)

    Hypotheses acceptance regions (11) take the following form using these denotations:

    Γ0={ˉyi:p(Hi)p(ˉyi|Hi)+p(Hi+)p+(ˉyi|Hi+)p(Hi0)p0(ˉyi|Hi0)2ω0π<1λ07K0K1},Γ={ˉyi:π2ω0p(Hi0)p0(ˉyi|Hi0)+p(Hi+)p+(ˉyi|Hi+)p(Hi)p(ˉyi|Hi)<1λ7K0K1},Γ+={ˉyi:p(Hi)p(ˉyi|Hi)+π2ω0p(Hi0)p0(ˉyi|Hi0)p(Hi+)p+(ˉyi|Hi+)<1λ+7K0K1}, (26)

    where the Lagrange multipliers are determined, so that in conditions (9), the equalities hold.

    To compare the offered method with the known ones, we present computation outcomes of the examples considered in ([4], p. 26, Tables 2 and 6) in Tables below where the results are obtained by the method offered above (see Item 3.1) and the Hayter PB method. As a priori probabilities, we take p(H)=p(H0)=p(H+)=1/3, coefficient ω0=4 for the results given in Tables 15, and ω0=13 for Stein's method for true alternative hypotheses the results of which are given in Table 6. To keep mdFDR, TIER, and TIIER at q=0.05, to test individual hypotheses of (6), we have to choose the following restriction levels in (9) for each subset of the hypotheses: r7=r07=r+7=0.0083(3). To test multiple hypotheses with the k1 subset of individual hypotheses for ensuring the same restriction levels of the criteria tmdFDR, TTIER, and TTIIER, we have to choose the following restriction levels: r7=r07=r+7=0.0083(3)/(k1). To test of all subsets of individual hypotheses, we use one and the same Lagrange multipliers, which are determined for a subset of individual hypotheses with the lowest divergence among directional hypotheses [24]. Because of k1=3, the restriction levels in (9) are equal to 0.0027(7). For the solution of equations (9), the necessary probability integrals are computed by the Monte Carlo method with a sample size equal to 5,000 (see, for example, [21,35]). The obtained values of the Lagrange multipliers are: λ7=5.625, λ07=0.965194702148438, and λ+7=2.34375. The values of TTIER, TTIIER, and tmdFDR are computed by the Monte Carlo simulation of random sequences with 5,000 observations. The codes for all necessary computations are written in MATLAB R2021b.

    Table 1.  Comparison of type-Ⅰ errors at k=4 (variances are known, null hypotheses are true).
    σ n CBM at q=0.05 AN α HPB AH
    TTIER tmdFL Nominal α=0.05
    [1,2,2,3] [12,12,12,12] 0.0190 0.0190 12.04 at H0
    11.89 at H0
    11.93 at H0
    0.0508 H0(5)
    [2,2,2,2] [12,12,12,12] 0.0082 0.0082 11.93 at H0
    12.07 at H0
    11.96 at H0
    0.0528 H0(5)
    [1,2,2,3] [6,8,10,12] 0.0081 0.0081 11.88 at H0
    11.93 at H0
    12.00 at H0
    0.0444 H0(5)
    [2,2,2,2] [6,8,10,12] 0.0136 0.0136 11.98 at H0
    11.87 at H0
    11.91 at H0
    0.0476 H0(5)
    [3,2,2,1] [6,8,10,12] 0.0109 0.0109 12.10 at H0
    11.96 at H0
    11.95 at H0
    0.0484 H0(5)
    [1,2,2,3] [6,6,6,6] 0.0082 0.0082 11.94 at H0
    11.97 at H0
    11.96 at H0
    0.0516 H0(5)
    [2,2,2,2] [6,6,6,6] 0.0054 0.0054 12.07 at H0
    12.00 at H0
    11.97 at H0
    0.0484 H0(5)
    [1,2,2,3] [6,3,3,6] 0.0135 0.0135 11.86 at H0
    11.94 at H0
    11.91 at H0
    0.0460 H0(5)
    [2,2,2,2] [6,3,3,6] 0.0136 0.0136 11.96 at H0
    11.90 at H0
    12.07 at H0
    0.0428 H0(5)
    [1,2,2,3] [3,3,3,3] 0.0109 0.0109 11.97 at H0
    12.02 at H0
    11.91 at H0
    0.0424 H0(5)
    [2,2,2,2] [3,3,3,3] 0.0217 0.0217 11.92 at H0
    11.88 at H0
    11.99 at H0
    0.0344 H0(5)
    [1,2,2,3] [2,3,2,3] 0.0082 0.0082 11.96 at H0
    12.01 at H0
    12.00 at H0
    0.0368 H0(5)
    [2,2,2,2] [2,3,2,3] 0.0136 0.0136 11.88 at H0
    12.00 at H0
    11.97 at H0
    0.0336 H0(5)
    [3,2,2,1] [2,3,2,3] 0.0136 0.0136 11.89 at H0
    11.95 at H0
    11.98 at H0
    0.0424 H0(5)
    Note: AN: Average number of observations necessary for making a decision; α HPB: Type-Ⅰ Error Rate for Hayter PB Test; αTTIER: Type Ⅰ error rate for CBM; AH: Accepted hypotheses; H0(5): basic hypothesis of (5) is accepted.

     | Show Table
    DownLoad: CSV
    Table 2.  Powers comparison table at k=4 (variances are known, alternative hypotheses are true).
    n σ μ CBM at q=0.05 AN HayterPB Nominal α=0.05 AH
    TTIER Power tmdFL TTIIER Power
    [6,8,10,12] [2,2,2,2] [0,0,0,2] 0.0107 1 0.0107 0 11.99 at H0
    11.88 at H0
    1.64 at H+
    0.6388 Ha(5)
    [0,0,0.7,1.3] 0.0052 1 0.0052 0 11.99 at H0
    4.955 at H+
    5.4 at H+
    0.2976 Ha(5)
    [0,1,2,3] 0 1 0 0 3.92 at H+
    3.41 at H+
    3.01 at H+
    0.7700 Ha(5)
    [-2,0,0,0] 0.0188 1 0.0188 0 2.028 at H+
    11.88 at H0
    11.91 at H0
    0.4864 Ha(5)
    [-1.3,-0.7,0,0] 0.0078 1 0.0078 0 7.35 at H+
    4.95 at H+
    11.92 at H0
    0.2600 Ha(5)
    [-3,-2,-1,0] 0 1 0 0 3.957 at H+
    3.45 at H+
    3.05 at H+
    0.7700 Ha(5)
    [1,2,2,3] [0,0,0,2] 0.0107 1 0.0107 0 11.93 at H0
    11.98 at H0
    2.029 at H+
    0.8020 Ha(5)
    [0,0,0.7,1.3] 0.0052 1 0.0052 0 11.92 at H0
    4.93 at H+
    7.15 at H+
    0.3384 Ha(5)
    [0,1,2,3] 0 1 0 0 2.92 at H+
    3.37 at H+
    3.84 at H+
    0.7356 Ha(5)
    [-2,0,0,0] 0.0080 1 0.0080 0 1.63 at H+
    11.99 at H0
    11.95 at H0
    0.2836 Ha(5)
    [-1.3,-0.7,0,0] 0 1 0 0 4.98 at H+
    5.10 at H+
    12.03 at H0
    0.2080 Ha(5)
    [-3,-2,-1,0] 0 1 0 0 2.98 at H+
    3.39 at H+
    3.85 at H+
    0.7356 Ha(5)
    [3,2,2,1] [0,0,0,2] 0.0107 1 0.0107 0 11.94 at H0
    11.91 at H0
    1.40 at H+
    0.7574 Ha(5)
    [0,0,0.7,1.3] 0.0026 1 0.0026 0 12.07 at H0
    5.07 at H+
    4.23 at H+
    0.2210 Ha(5)
    [0,1,2,3] 0 1 0 0 5.29 at H+
    3.42 at H+
    2.52 at H+
    0.6724 Ha(5)
    [-2,0,0,0] 0.0027 1 0.0027 0 2.53 at H+
    12.04 at H0
    12.03 at H0
    0.1886 Ha(5)
    [-1.3,-0.7,0,0] 0 1 0 0 11.23 at H+
    5.05 at H+
    11.96 at H0
    0.1108 Ha(5)
    [-3,-2,-1,0] 0 1 0 0 5.35 at H+
    3.43 at H+
    2.48 at H+
    0.6724 Ha(5)
    [6,8,10,12] [2,2,2,2] [0,0,0,-2] 0.0054 1 0.0054 0 12.01 at H0
    11.96 at H0
    2.28 at H-
    - Ha(4)
    [0,0,-0.7,1.3] 0.0104 1 0.0130 0 11.87 at H0
    8.56 at H-
    1.70 at H+
    - Ha(4)
    [0,-1,2,3] 0 1 0 0 6.15 at H-
    1.33 at H+
    3.07 at H+
    - Ha(4)
    [-2,0,1,0] 0 1 0 0 2.01 at H+
    3.42 at H+
    4.68 at H-
    - Ha(4)
    [-1.3,0.7,0,0] 0.0052 1 0.0052 0 1.99 at H+
    8.41 at H-
    11.97 at H0
    - Ha(4)
    [-3,2,-1,0] 0 1 0 0 1.03 at H+
    1.77 at H-
    3.09 at H+
    - Ha(4)
    Note: AN: Average number of observations necessary for making a decision; α HPB: Type-Ⅰ Error Rate for Hayter PB Test; αTTIER: Type Ⅰ error rate for CBM; AH: Accepted hypotheses; Ha(5): Alternative hypothesis of (5) is accepted; Ha(4): Alternative hypothesis of (6) is accepted.

     | Show Table
    DownLoad: CSV
    Table 3.  Comparison of Type-Ⅰ errors, k=4 (unknown variances are estimated by maximum likelihood method, null hypotheses are true).
    σ n CBM at q=0.05 AN α HPB AH
    TTIER tmdFL Nominal α=0.05
    [1,2,2,3] [12,12,12,12] 0.0190 0.0190 11.90 at H0
    11.96 at H0
    11.93 at H0
    0.0508 H0(5)
    [2,2,2,2] [12,12,12,12] 0.0055 0.0055 12.03 at H0
    12.03 at H0
    12 at H0
    0.0528 H0(5)
    [1,2,2,3] [6,8,10,12] 0.0054 0.0054 11.99 at H0
    11.99 at H0
    11.99 at H0
    0.0444 H0(5)
    [2,2,2,2] [6,8,10,12] 0.0163 0.0163 11.91 at H0
    11.90 at H0
    11.95 at H0
    0.0476 H0(5)
    [3,2,2,1] [6,8,10,12] 0.0191 0.0191 12.01 at H0
    11.93 at H0
    11.95 at H0
    0.0484 H0(5)
    [1,2,2,3] [6,6,6,6] 0.0163 0.0163 11.99 at H0
    11.87 at H0
    11.95 at H0
    0.0516 H0(5)
    [2,2,2,2] [6,6,6,6] 0.0217 0.0217 11.79 at H0
    12.05 at H0
    11.89 at H0
    0.0484 H0(5)
    [1,2,2,3] [6,3,3,6] 0.0082 0.0082 11.99 at H0
    12.03 at H0
    11.95 at H0
    0.0460 H0(5)
    [2,2,2,2] [6,3,3,6] 0.0245 0.0245 12 at H0
    11.85 at H0
    11.88 at H0
    0.0428 H0(5)
    [1,2,2,3] [3,3,3,3] 0.0190 0.0190 12.02 at H0
    11.85 at H0
    12.02 at H0
    0.0424 H0(5)
    [2,2,2,2] [3,3,3,3] 0.0164 0.0164 11.94 at H0
    12.04 at H0
    11.98 at H0
    0.0344 H0(5)
    [1,2,2,3] [2,3,2,3] 0.0163 0.0163 11.93 at H0
    11.88 at H0
    11.89 at H0
    0.0368 H0(5)
    [2,2,2,2] [2,3,2,3] 0.0163 0.0163 11.96 at H0
    11.95 at H0
    11.96 at H0
    0.0336 H0(5)
    [3,2,2,1] [2,3,2,3] 0.0164 0.0164 11.94 at H0
    11.95 at H0
    11.97 at H0
    0.0424 H0(5)
    Note: AN: Average number of observations necessary for making a decision; α HPB: Type-Ⅰ Error Rate for Hayter PB Test; αTTIER: Type Ⅰ error rate for CBM; AH: Accepted hypotheses; H0(5): basic hypothesis of (5) is accepted.

     | Show Table
    DownLoad: CSV
    Table 4.  Power comparison table, k=4 (unknown variances are estimated by the maximum likelihood method, alternative hypotheses are true).
    n σ μ CBM at q=0.05 AN HayterPB Nominal α=0.05 AH
    TTIER Power Power TTIIER Power
    [6,8,10,12] [2,2,2,2] [0,0,0,2] 0.008 1 0.008 0 11.95 at H0
    11.91 at H0
    2.16 at H+
    0.6388 Ha(5)
    [0,0,0.7,1.3] 0.01 1 0.01 0 11.94 at H0
    7.79 at H+
    7.70 at H+
    0.2976 Ha(5)
    [0,1,2,3] 0 1 0 0 4.28 at H+
    2.89 at H+
    2.13 at H+
    0.7700 Ha(5)
    [-2,0,0,0] 0.0054 1 0.0054 0 4.28 at H+
    11.98 at H0
    11.96 at H0
    0.4864 Ha(5)
    [-1.3,-0.7,0,0] 0 0.9948 0 0.0052 6.10 at H+
    6.01 at H+
    11.94 at H0
    0.2600 Ha(5)
    [-3,-2,-1,0] 0 1 0 0 2.74 at H+
    3.73 at H+
    3.43 at H+
    0.7700 Ha(5)
    [1,2,2,3] [0,0,0,2] 0.0054 1 0.0054 0 11.89 at H0
    11.97 at H0
    8.63 at H+
    0.8020 Ha(5)
    [0,0,0.7,1.3] 0.0105 1 0.0105 0 11.87 at H0
    5.98 at H+
    8.78 at H+
    0.3384 Ha(5)
    [0,1,2,3] 0 1 0 0 1.93 at H+
    2.55 at H+
    2.92 at H+
    0.7356 Ha(5)
    [-2,0,0,0] 0.0054 1 0.0054 0 3.77 at H+
    11.97 at H0
    12.01 at H0
    0.2836 Ha(5)
    [-1.3,-0.7,0,0] 0.0078 1 0.0078 0 4.91 at H+
    3.69 at H+
    11.95 at H0
    0.2080 Ha(5)
    [-3,-2,-1,0] 0 1 0 0 2.56 at H+
    2.74 at H+
    4.37 at H+
    0.7356 Ha(5)
    [3,2,2,1] [0,0,0,2] 0 0.9919 0 0.0081 11.99 at H0
    12.05 at H0
    1.12 at H+
    0.7574 Ha(5)
    [0,0,0.7,1.3] 0.0026 1 0.0026 0 11.97 at H0
    5.16 at H+
    3.06 at H+
    0.2210 Ha(5)
    [0,1,2,3] 0 1 0 0 4.70 at H+
    2.17 at H+
    2.39 at H+
    0.6724 Ha(5)
    [-2,0,0,0] 0.0081 1 0.0081 0 2.71 at H+
    11.98 at H0
    11.97 at H0
    0.1886 Ha(5)
    [-1.3,-0.7,0,0] 0 1 0 0 9.09 at H+
    7.84 at H+
    12.01 at H0
    0.1108 Ha(5)
    [-3,-2,-1,0] 0 1 0 0 4.33 at H+
    2.04 at H+
    1.84 at H+
    0.6724 Ha(5)
    [6,8,10,12] [2,2,2,2] [0,0,0,-2] 0.0054 1 0.0054 0 11.99 at H0
    12.067 at H0
    2.097 at H-
    - Ha(5)
    [0,0,-0.7,1.3] 0.0027 0.9973 0.0027 0.0027 12.04 at H0
    11.38 at H-
    1.98 at H+
    - Ha(4)
    [0,-1,2,3] 0 1 0 0 4.06 at H-
    1.12 at H+
    2.49 at H+
    - Ha(4)
    [-2,0,1,0] 0 1 0 0 1.96 at H+
    4.47 at H+
    5.65 at H-
    - Ha(4)
    [-1.3,0.7,0,0] 0.0026 1 0 0.0026 2.29 at H+
    8.74 at H-
    11.97 at H0
    - Ha(4)
    [-3,2,-1,0] 0 1 0 0 1.003 at H+
    1.90 at H-
    3.81 at H+
    - Ha(4)
    Note: AN: Average number of observations necessary for making a decision; α HPB: Type-Ⅰ Error Rate for Hayter PB Test; αTTIER: Type Ⅰ error rate for CBM; AH: Accepted hypotheses; Ha(5): Alternative hypothesis of (5) is accepted; Ha(4): Alternative hypothesis of (6) is accepted.

     | Show Table
    DownLoad: CSV
    Table 5.  Comparison of type-Ⅰ errors, k=4 (Stein's method, null hypotheses are true).
    σ n CBM at q=0.05 AN α HPB AH
    TTIER tmdFL Nominal α=0.05
    [1,2,2,3] [12,12,12,12] 0 0 2 at H0
    2 at H0
    4.18 at H0
    0.0508 H0(5)
    [2,2,2,2] [12,12,12,12] 0 0 2 at H0
    2 at H0
    1 at H0
    0.0528 H0(5)
    [1,2,2,3] [6,8,10,12] 0 0 2 at H0
    1 at H0
    6.05 at H0
    0.0444 H0(5)
    [2,2,2,2] [6,8,10,12] 0 0 1.03 at H0
    1.54 at H0
    1 at H0
    0.0476 H0(5)
    [3,2,2,1] [6,8,10,12] 0 0 1 at H0
    2 at H0
    1 at H0
    0.0484 H0(5)
    [1,2,2,3] [6,6,6,6] 0 0 2 at H0
    2 at H0
    2.25 at H0
    0.0516 H0(5)
    [2,2,2,2] [6,6,6,6] 0 0 2 at H0
    2 at H0
    1 at H0
    0.0484 H0(5)
    [1,2,2,3] [6,3,3,6] 0 0 1 at H0
    1 at H0
    1 at H0
    0.0460 H0(5)
    [2,2,2,2] [6,3,3,6] 0 0 1 at H0
    1 at H0
    1 at H0
    0.0428 H0(5)
    [1,2,2,3] [3,3,3,3] 0 0 1 at H0
    1 at H0
    2 at H0
    0.0424 H0(5)
    [2,2,2,2] [3,3,3,3] 0 0 1 at H0
    1 at H0
    1 at H0
    0.0344 H0(5)
    [1,2,2,3] [2,3,2,3] 0 0 1 at H0
    1 at H0
    13.25 at H0
    0.0368 H0(5)
    [2,2,2,2] [2,3,2,3] 0 0 1 at H0
    1 at H0
    1 at H0
    0.0336 H0(5)
    [3,2,2,1] [2,3,2,3] 0 0 1 at H0
    1 at H0
    13.81 at H0
    0.0424 H0(5)
    Note: AN: Average number of observations necessary for making a decision; α HPB: Type-Ⅰ Error Rate for Hayter PB Test; αTTIER: Type Ⅰ error rate for CBM; AH: Accepted hypotheses; H0(5): basic hypothesis of (5) is accepted.

     | Show Table
    DownLoad: CSV
    Table 6.  Power comparison table, k=4 (Stein's method, alternative hypotheses are true).
    n σ μ CBM at q=0.05 AN HayterPB Nominal α=0.05 AH
    Power TTIER, tmdFL TTIIER Power
    [6,8,10,12] [2,2,2,2] [0,0,0,2] 1 0 0 390.42 at H0
    389.50 at H0
    169.46 at H+
    0.6388 Ha(5)
    [0,0,0.7,1.3] 1 0 0 390 at H0
    621.5 at H+
    3001 at H+
    0.2976 Ha(5)
    [0,1,2,3] 1 0 0 350.90 at H+
    242.40 at H+
    480.70 at H+
    0.7700 Ha(5)
    [-2,0,0,0] 1 0 0 79.62 at H+
    390 at H0
    121.54 at H0
    0.4864 Ha(5)
    [-1.3,-0.7,0,0] 1 0 0 137.50 at H+
    65.13 at H+
    650.86 at H0
    0.2600 Ha(5)
    [-3,-2,-1,0] 1 0 0 1085.50 at H+
    146.60 at H+
    1014.75 at H+
    0.7700 Ha(5)
    [1,2,2,3] [0,0,0,2] 1 0 0 390.33 at H0
    390.42 at H0
    100.38 at H+
    0.8020 Ha(5)
    [0,0,0.7,1.3] 1 0 0 390.25 at H0
    1074.33 at H+
    1500.67 at H+
    0.3384 Ha(5)
    [0,1,2,3] 1 0 0 140.80 at H+
    205.80 at H+
    522.33 at H+
    0.7356 Ha(5)
    [-2,0,0,0] 1 0 0 27.44 at H+
    390.22 at H+
    554.44 at H0
    0.2836 Ha(5)
    [-1.3,-0.7,0,0] 1 0 0 353.50 at H+
    3563 at H+
    2156 at H0
    0.2080 Ha(5)
    [-3,-2,-1,0] 1 0 0 573.50 at H+
    297.25 at H+
    580 at H+
    0.7356 Ha(5)
    [3,2,2,1] [0,0,0,2] 1 0 0 390.58 at H0
    389.83 at H0
    122.69 at H+
    0.7574 Ha(5)
    [0,0,0.7,1.3] 1 0 0 391.25 at H0
    275.69 at H+
    359.75 at H+
    0.2210 Ha(5)
    [0,1,2,3] 1 0 0 2070.50 at H+
    671.33 at H+
    600.67 at H+
    0.6724 Ha(5)
    [-2,0,0,0] 1 0 0 74.15 at H+
    390.33 at H0
    75.38 at H0
    0.1886 Ha(5)
    [-1.3,-0.7,0,0] 1 0 0 1567.67 at H+
    174.25 at H+
    8.50 at H0
    0.1108 Ha(5)
    [-3,-2,-1,0] 1 0 0 3582 at H+
    790.50 at H+
    1384 at H+
    0.6724 Ha(5)
    [6,8,10,12] [2,2,2,2] [0,0,0,-2] 1 0 0 390.5 at H0
    391.67 at H0
    127.92 at H-
    - Ha(4)
    [6,8,10,12] [2,2,2,2] [0,0,-0.7,1.3] 1 0 0 389.2 at H0
    959.4 at H-
    76.17 at H+
    - Ha(4)
    [0,-1,2,3] 1 0 0 2801 at H-
    26.5 at H+
    287.5 at H+
    - Ha(4)
    [-2,0,1,0] 1 0 0 46 at H+
    72.33 at H+
    868.2 at H-
    - Ha(4)
    [-1.3,0.7,0,0] 1 0 0 213 at H+
    4592 at H-
    65 at H0
    - Ha(4)
    [-3,2,-1,0] 1 0 0 32 at H+
    318.27 at H-
    452 at H0
    - Ha(4)
    Note: AN: Average number of observations necessary for making a decision; α HPB: Type-Ⅰ Error Rate for Hayter PB Test; αTTIER: Type Ⅰ error rate for CBM; AH: Accepted hypotheses; Ha(5): Alternative hypothesis of (5) is accepted; Ha(4): Alternative hypothesis of (6) is accepted.

     | Show Table
    DownLoad: CSV

    When using CBM, principle complexity does not arise when the problem dimension increases [21,22]. The computation time required to make a decision in CBM is equal to the time required for the classical Bayes method. The time required to determine the Lagrange multipliers certainly increases, but it is realized at the preparatory stage and does not directly affect the decision-making time.

    Computational results, when variances are known, are given in Table 1 (null hypotheses are true) and Table 2 (alternative hypotheses are true) for different ni and σi, i=1,2,3,4. From here, it is seen that for all computed combinations, TTIER=tmdFDR and they all are significantly less than the type Ⅰ error rate of the Hayter PB method given in [4]. The powers for all computed combinations are equal to 1 and significantly surpass the appropriate values of the Hayter PB method. The offered method perfectly distinguishes one- and two-sided alternatives whose possibility is not discussed in ([4] p. 26, Table 2).

    Computational results obtained by the method involving the maximum likelihood estimator of the unknown variances, are given in Table 3 (null hypotheses are true) and Table 4 (alternative hypotheses are true). The computations give us the same results as we had in the case of the known variances, that is, the reliabilities of the decisions in CBM are significantly higher than the results obtained by the Hayter PB method [4]. However, the number of observations required to make a decision is greater in CBM for this case than for the case when variances are known. One circumstance is noteworthy: Calculating the variance estimator with less than 5–6 observations is not recommended due to the low accuracy of the obtained results in statistics. Nevertheless, in the calculated examples, even when the number of observations is 2–3, the reliability of the decisions is quite high. In our opinion, the reason for this, in addition to the high reliability of the CBM, lies in the relatively large number of observations used to make a decision.

    Computational results obtained by the Stein's method when the uniform distributions are used as a priori distributions of the unknown variances, are given in Table 5 (null hypotheses are true) and Table 6 (alternative hypotheses are true). The obtained results are very reliable. The probabilities of incorrect decisions are equal to zero, which are provided by the big number of observations necessary for making decisions when alternative hypotheses are true.

    The offered method of solving the ANOVA problem, for known and unknown variances of the results, with restricted Type Ⅰ and Type Ⅱ error rates based on CBM gives very reliable results that surpass the methods of today. It is a sequential method that requires a set of observations to make a decision, the number of which is especially large in Stein's method in the case of true alternative hypotheses. Nevertheless, due to the large practical application of the ANOVA problem and the fact that, in many cases, the number of data is quite large, and the requirements for the accuracy of the decisions are high, we consider the results to be important and to deserve high attention.

    It is worth noting that the CBM method using the maximum likelihood method requires more observations than the same CBM method using Stein's method to test the null hypothesis; this situation is reversed when testing alternative hypotheses. In this case, using Stein's method requires more observations than using the maximum likelihood method. The reason lies in the use of different values of ω0. It should be noted that the increase of ω0 causes the increase in the quantity of observations necessary to make a decision and, accordingly, the increase of the accuracy of made decisions [24,33].

    The computing results obtained by CBM and the Hayter PB methods are given in Tables 16. In order to compare the results obtained by these methods, q=0.05 is taken in CBM, since in Hayter PB methods, the calculations have been realized for a Type Ⅰ error rate of α = 0.05 (see Tables 24, 6 in [4]). Since each scenario has four normal population correspondences, i.e., k=4, we have to test three hypotheses each time (see (6)). In each case under consideration, the true hypotheses are determined by the values of θi=μi+1μi, i=1,...,k1. Therefore, in each scenario, three hypotheses (from the set of hypotheses H0, H and H+) are accepted, and depending on θi, i=1,...,k1, the values of the criteria TTIER, TTIIER, tmdFDR, and Power=1TTIIER are computed. The values criteria TTIER and tmdFDR are given when null hypotheses are true in all three cases (see Tables 1, 3, and 5), and the values all four considered criteria TTIER, TTIIER, tmdFDR, and Power=1TTIIER are given when the alternative hypothesis is true even for one of the three cases. In the columns n, σ, and μ, the values of sample sizes for considered normal populations, standard deviations, and mathematical expectations, respectively, are given for each considered case. The second to last the final column contains the results obtained by Hayter PB methods from [4], and the last column contains each scenario in the CBM accepted hypotheses defined by (5) or (6). The column AN (average number) contains the average number of observations necessary to make a decision in CBM (averaged over the results of 5,000 experiments).

    From the results of these Tables, the superiority of CBM over Hayter PB methods is evident. Moreover, unlike the Hayter PB method, CBM enables us to distinguish (4) and (5), which are the same as (1) and (2).

    The problem of testing hypotheses concerning the means of a set of normally distributed random variables at the restriction of both Type Ⅰ and Type Ⅱ error rates are considered in this paper. To solve the stated problem, one of the possible statements of CBM is used. In order to overcome the difficulties related to the complexity of testing hypotheses caused by not knowing the values of the variances of random variables, the maximum likelihood and Stein's methods are used. The ability to limit Type Ⅰ and Type Ⅱ error rates and other relevant criteria to the required level is shown for the developed algorithms. The latter is confirmed by the calculation results of practical examples by simulating different scenarios.

    Kartlos J. Kachiashvili: Conceptualization, Formal analysis, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing–original draft, Writing–review & editing; Ashis SenGupta: Conceptualization, Formal analysis, Investigation, Resources, Validation, Writing–original draft, Writing–review & editing; Joseph K. Kachiashvili: Data curation, Formal analysis, Investigation, Resources, Software, Visualization, Writing–original draft, Writing–review & editing.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    All Authors declare that they have no financial interests.



    [1] A. Aghamohammadi, M. R. Meshkani, M. Mohammadzadeh, A Bayesian approach to successive comparisons, Journal of Data Science, 8 (2010), 541–553. https://doi.org/10.6339/JDS.2010.08(4).630 doi: 10.6339/JDS.2010.08(4).630
    [2] R. E. Barlow, D. J. Bartholomew, J. M. Bremner, H. D. Brunk, Statistical inference under order restrictions: the theory and application of isotonic regression, J. Wiley & Sons Ltd, 1972.
    [3] A. J. Hayter, An one-sided studentized range test for testing against a simple ordered alternative, J. Amer. Stat. Assoc., 85 (1990), 778–785. https://doi.org/10.1080/01621459.1990.10474940 doi: 10.1080/01621459.1990.10474940
    [4] S. Ichiba, An alternative method for ordered ANOVA under unequal variances, Thesis of the Degree of Bachelor of Arts with Honours in Mathematics and Statistics, Acadia University, 2016.
    [5] T. Robertson, F. T. Wright, R. L. Dykstra, Order restricted statistical inference, Chichester: John Wiley and Sons, 1988.
    [6] B. L. Welch, The generalization of Student's problem when several different population variances are involved, Biometrika, 34 (1947), 28–35. https://doi.org/10.1093/biomet/34.1-2.28 doi: 10.1093/biomet/34.1-2.28
    [7] B. L. Welch, On the comparison of several mean values: an alternative approach, Biometrika, 38 (1951), 330–336. https://doi.org/10.1093/biomet/38.3-4.330 doi: 10.1093/biomet/38.3-4.330
    [8] K. Krishnamoorthy, F. Lu, T. Mathew, A parametric bootstrap approach for ANOVA with unequal variances: fixed and random models, Comput. Stat. Data Anal., 51 (2007), 5731–5742. https://doi.org/10.1016/j.csda.2006.09.039 doi: 10.1016/j.csda.2006.09.039
    [9] A. N. Dajani, Contributions to statistical inference for some fixed and random models, Ph.D. Dissertation, Baltimore County: Department of Mathematics and Statistics, University of Maryland, 2002.
    [10] G. S. James, The comparison of several groups of observations when the ratios of population variances are unknown, Biometrika, 38 (1951), 324–329. https://doi.org/10.1093/biomet/38.3-4.324 doi: 10.1093/biomet/38.3-4.324
    [11] S. Weerahandi, ANOVA under unequal error variances, Biometrics, 51 (1995), 589–599.
    [12] S. Weerahandi, Exact statistical methods for data analysis, New York: Springer, 1995. https://doi.org/10.1007/978-1-4612-0825-9
    [13] J. C. Hsu, Multiple comparisons: theory and methods, New York: Chapman and Hall, 1996. https://doi.org/10.1201/b15074
    [14] W. Liu, T. Miwa, A. J. Hayter, Simultaneous confidence interval estimation for successive comparisons of ordered treatment effects, J. Stat. Plan. Infer., 88 (2000), 75–86. https://doi.org/10.1016/S0378-3758(99)00209-8 doi: 10.1016/S0378-3758(99)00209-8
    [15] W. Liu, P. N. Somerville, Stepwise multiple tests for successive comparisons of treatment effects, Comput. Stat. Data Anal., 46 (2004), 189–199. https://doi.org./10.1016/S0167-9473(03)00137-3 doi: 10.1016/S0167-9473(03)00137-3
    [16] Y. Hochberg, A. C. Tamhane, Multiple comparisons procedures, Wiley, 1987. https://doi.org./10.1002/9780470316672
    [17] R. G. Miller, Simultaneous statistical inference, 2 Eds., New York: Springer, 1981. https://doi.org./10.1007/978-1-4613-8122-8
    [18] R. N. Machekano, A. E. Hubbard, Empirical Bayes approach to controlling familywise error: an application to HIV resistance data, U.C. Berkeley Division of Biostatistics Working Paper Series, April 2006, Working Paper 205.
    [19] A. J. Roth, Robust trend tests derived and simulated: analogs of the Welch and Brown-Forsythe tests, J. Amer. Stat. Assoc., 78 (1983), 972–980. https://doi.org/10.1080/01621459.1983.10477048 doi: 10.1080/01621459.1983.10477048
    [20] S.-Y. Chen, One-sided range test for testing against an ordered alternative under heteroscedasticity, Commun. Stat.-Simul. Comput., 29 (2000), 1255–1272. https://doi.org/10.1080/03610910008813663 doi: 10.1080/03610910008813663
    [21] K. J. Kachiashvili, Constrained Bayesian methods of hypotheses testing: a new philosophy of hypotheses testing in parallel and sequential experiments, New York: Nova Science Publishers, Inc., 2018.
    [22] K. J. Kachiashvili, Testing statistical hypotheses with given reliability, UK: Cambridge Scholars Publishing, 2023.
    [23] K. J. Kachiashvili, J. K. Kachiashvili, I. A. Prangishvili, CBM for testing multiple hypotheses with directional alternatives in sequential experiments, Sequential Anal., 39 (2020), 115–131. https://doi.org/10.1080/07474946.2020.1727166 doi: 10.1080/07474946.2020.1727166
    [24] K. J. Kachiashvili, J. K. Kachiashvili, A. SenGupta, Consideration of directional hypotheses in pairs for making a decision with given reliability, Bulletin of TICMI, 28 (2024), 3–28.
    [25] N. K. Bansal, G. G. Hamedani, M. Maadooliat, Testing multiple hypotheses with skewed alternatives, Biometrics, 72 (2016), 494–502. https://doi.org/10.1111/biom.12430 doi: 10.1111/biom.12430
    [26] K. J. Kachiashvili, Constrained Bayesian method for testing multiple hypotheses in sequential experiments, Sequential Anal., 34 (2015), 171–186. https://doi.org/10.1080/07474946.2015.1030973 doi: 10.1080/07474946.2015.1030973
    [27] S. K. De, M. Baron, Sequential Bonferroni methods for multiple hypothesis testing with strong control of family-wise error rates Ⅰ and Ⅱ, Sequential Anal., 31 (2012), 238–262. https://doi.org/10.1080/07474946.2012.665730 doi: 10.1080/07474946.2012.665730
    [28] S. K. De, M. Baron, Step-up and Step-down methods for testing multiple hypotheses in sequential experiments, J. Stat. Plan. Infer., 142 (2012), 2059–2070. https://doi.org/10.1016/j.jspi.2012.02.005 doi: 10.1016/j.jspi.2012.02.005
    [29] N. K. Bansal, R. Sheng, Beyesian decision theoretic approach to hypothesis problems with skewed alternatives, J. Stat. Plan. Infer., 140 (2010), 2894–2903. https://doi.org/10.1016/j.jspi.2010.03.013 doi: 10.1016/j.jspi.2010.03.013
    [30] S. Andersson, Distributions of maximal invariants using quotient measures, Ann. Statist., 10 (1982), 955–961. https://doi.org/10.1214/aos/1176345885 doi: 10.1214/aos/1176345885
    [31] R. A. Wijsman, Cross-sections of orbits and their application to densities of maximal invariants, In: Proceedings of the fifth Berkeley symposium on mathematical statistics and probability, Volume 1: statistics, 1967,389–400.
    [32] K. J. Kachiashvili, Constrained Bayesian method of composite hypotheses testing: singularities and capabilities, International Journal of Statistics in Medical Research, 5 (2016), 135–167. http://doi.org/10.6000/1929-6029.2016.05.03.1 doi: 10.6000/1929-6029.2016.05.03.1
    [33] K. J. Kachiashvili, A. SenGupta, Constrained Bayesian methods for union-intersection and intersection-union hypotheses testing problems, Iran. J. Sci., 48 (2024), 1579–1594. https://doi.org/10.1007/s40995-024-01693-x doi: 10.1007/s40995-024-01693-x
    [34] K. J. Kachiashvili, N. Mukhopadhyay, J. K. Kachiashvili, Constrained Bayesian method for testing composite hypotheses concerning normal distribution with equal parameters, Sequential Anal., 43 (2024), 147–178. https://doi.org/10.1080/07474946.2024.2326222 doi: 10.1080/07474946.2024.2326222
    [35] K. J. Kachiashvili, N. K. Bansal, I. A. Prangishvili, Constrained Bayesian method for testing the directional hypotheses, Journal of Mathematics and System Science, 8 (2018), 96–118. https://doi.org/10.17265/2159-5291/2018.04.002 doi: 10.17265/2159-5291/2018.04.002
  • Reader Comments
  • © 2025 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(311) PDF downloads(49) Cited by(0)

Figures and Tables

Tables(6)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog