Processing math: 81%
Research article

Soil erosion estimation using Erosion Potential Method in the Vjosa River Basin, Albania

  • Soil erosion is a major environmental threat to soil sustainability and productivity with knock-on effects on agriculture, climate change, etc. Factors influencing soil erosion are many and usually divided into natural and human causes. Massive deforestation, intensive agriculture, temperature and wind, rainfall intensity, human activities and climate changes are listed as the main causes of soil erosion. Calculation of the coefficient of soil erosion is very important to prevent the event. One of the methods used worldwide to calculate soil loss and the erosion coefficient is the Erosion Potential Method. In this study, 49 sub-basins of the Vjosa River Basin in Albania were evaluated. Results showed that the phenomenon of erosion is present in all sub-basins, varying from 0.01 to 0.71. Thus, the categorization of soil erosion varies from heavy to very slight erosion. Moreover, the overall sediment yield calculated for the Vjosa River Basin was 2326917 m3/year. In conclusion, the application of the Erosion Potential Method is reliable for evaluating erosion and can further be applied in our country's conditions.

    Citation: Oltion Marko, Joana Gjipalaj, Dritan Profka, Neritan Shkodrani. Soil erosion estimation using Erosion Potential Method in the Vjosa River Basin, Albania[J]. AIMS Environmental Science, 2023, 10(1): 191-205. doi: 10.3934/environsci.2023011

    Related Papers:

    [1] Xiaoxia Wang, Jinping Jiang . The uniform asymptotic behavior of solutions for 2D g-Navier-Stokes equations with nonlinear dampness and its dimensions. Electronic Research Archive, 2023, 31(7): 3963-3979. doi: 10.3934/era.2023201
    [2] Jie Zhang, Gaoli Huang, Fan Wu . Energy equality in the isentropic compressible Navier-Stokes-Maxwell equations. Electronic Research Archive, 2023, 31(10): 6412-6424. doi: 10.3934/era.2023324
    [3] Guochun Wu, Han Wang, Yinghui Zhang . Optimal time-decay rates of the compressible Navier–Stokes–Poisson system in R3. Electronic Research Archive, 2021, 29(6): 3889-3908. doi: 10.3934/era.2021067
    [4] Wei Shi, Xinguang Yang, Xingjie Yan . Determination of the 3D Navier-Stokes equations with damping. Electronic Research Archive, 2022, 30(10): 3872-3886. doi: 10.3934/era.2022197
    [5] Jie Qi, Weike Wang . Global solutions to the Cauchy problem of BNSP equations in some classes of large data. Electronic Research Archive, 2024, 32(9): 5496-5541. doi: 10.3934/era.2024255
    [6] Jianxia He, Qingyan Li . On the global well-posedness and exponential stability of 3D heat conducting incompressible Navier-Stokes equations with temperature-dependent coefficients and vacuum. Electronic Research Archive, 2024, 32(9): 5451-5477. doi: 10.3934/era.2024253
    [7] Guoliang Ju, Can Chen, Rongliang Chen, Jingzhi Li, Kaitai Li, Shaohui Zhang . Numerical simulation for 3D flow in flow channel of aeroengine turbine fan based on dimension splitting method. Electronic Research Archive, 2020, 28(2): 837-851. doi: 10.3934/era.2020043
    [8] Jingjing Zhang, Ting Zhang . Local well-posedness of perturbed Navier-Stokes system around Landau solutions. Electronic Research Archive, 2021, 29(4): 2719-2739. doi: 10.3934/era.2021010
    [9] Meng-Xue Chang, Bang-Sheng Han, Xiao-Ming Fan . Global dynamics of the solution for a bistable reaction diffusion equation with nonlocal effect. Electronic Research Archive, 2021, 29(5): 3017-3030. doi: 10.3934/era.2021024
    [10] Cheng Wang . Convergence analysis of Fourier pseudo-spectral schemes for three-dimensional incompressible Navier-Stokes equations. Electronic Research Archive, 2021, 29(5): 2915-2944. doi: 10.3934/era.2021019
  • Soil erosion is a major environmental threat to soil sustainability and productivity with knock-on effects on agriculture, climate change, etc. Factors influencing soil erosion are many and usually divided into natural and human causes. Massive deforestation, intensive agriculture, temperature and wind, rainfall intensity, human activities and climate changes are listed as the main causes of soil erosion. Calculation of the coefficient of soil erosion is very important to prevent the event. One of the methods used worldwide to calculate soil loss and the erosion coefficient is the Erosion Potential Method. In this study, 49 sub-basins of the Vjosa River Basin in Albania were evaluated. Results showed that the phenomenon of erosion is present in all sub-basins, varying from 0.01 to 0.71. Thus, the categorization of soil erosion varies from heavy to very slight erosion. Moreover, the overall sediment yield calculated for the Vjosa River Basin was 2326917 m3/year. In conclusion, the application of the Erosion Potential Method is reliable for evaluating erosion and can further be applied in our country's conditions.



    Many complex problems in real life are composed of conflicting and influential objectives, they need to be optimized simultaneously as much as possible in given constraints, that is, multi-objective optimization. These problems are very complex, difficult, but also very important. Multi-objective optimization is common and significant in actual life, and it is widely used in production and engineering design [1], job scheduling [2], management and decision-making [3], etc. In the multi-objective optimization problem, each objective restricts each other, it is impossible to have a solution that can make all the objectives achieve the optimal performance. Therefore, for the multi-objective optimization problem, the optimal solution is usually a set of non-inferior solutions, namely Pareto optimal solution set [4]. Therefore, the main tasks of multi-objective optimization are as follows: 1) Finding a set of solutions as close as possible to the Pareto front; 2) Finding a set of solutions as different as possible [4,5].

    Meta-heuristic methods have powerful global search capabilities, which design iterative equations by simulating the behavioral characteristics of biological groups or the development and structural characteristics of physical things. Therefore, they are more appropriate for settling complex multi-objective optimization problems [6]. At present, a large number of Meta-heuristic methods have been proposed and improved to solve multi-objective optimization problems, such as, monarch butterfly optimization (MBO) which simplifies and simulates the migration process of monarch butterfly [7], slime mould algorithm (SMA) which is proposed based on the oscillation mode of slime mould in nature [8], Moth search algorithm (MSA) [9] which is a new kind of metaheuristic algorithm inspired by the phototaxis and Lévy flights of the moths, Harris hawks optimization (HHO) which simulates the predatory behavior of Harris hawk [10].

    Meta-heuristic methods include two contradictory processes: exploration and exploitation, which need to establish a balance between them [11,12]. When exploitation is enhanced, randomization of the search is increased, which is helpful to avert falling into the local optimal solution in the process of optimization and improve the convergence accuracy. Conversely, when exploration is enhanced, local search capability of the meta-heuristic algorithm is more powerful, and the convergence speed is faster, but the approach is easier to fall into local optimum, especially the diversity of the obtained solutions becomes worse [13,14]. Meta-heuristic methods are population-based search and optimization methods whose efficacy mainly depends on a fine balance between exploration and exploitation [15]. Therefore, the balance between exploration and exploitation has been widely studied in meta-heuristic optimization algorithms [15,16,17,18,19,20].

    The meta-heuristic optimization algorithms should be equipped with special control parameters to adjust exploration and exploitation in different stages of optimization. The control parameters should be designed and adjusted according to the specific optimization problems. Generally, there are two ways to design the control parameters. One is the constant control strategy [21], that is, to keep the control parameters unchanged in the whole iterative search process. Another way is to dynamically adjust the value of control parameters in the iterative search process to achieve the balance between exploration and exploitation [22]. The first way is simple and easy to implement, but it is very difficult to clearly identify the exploration and exploitation phases. It is generally considered that the strategy of more exploration in the early phase and more exploitation in the later phase is usually employed in the whole search process, such as linear control strategy [1,14], and nonlinear control strategy [22,23]. Therefore, the method of dynamically adjusting the value of control parameters is widely employed in meta-heuristic optimization algorithms. As discussed in [24], a proper balance between exploration and exploitation may drive the search process towards better performance. The balance between exploration and exploitation is critical to the performance of a meta-heuristic optimization method. However, a scheme of adjusting control parameters that performs better in problem A might not work effectively in problem B. How to balance the relationship between exploration and exploitation in complex optimization problems has become the goal of meta-heuristic algorithms design or improvement, which is still an open question. For this problem, the solution in this paper is that different values of the same parameters can be employed to different subpopulations of a meta-heuristic optimization method at the same time to ensure that a suitable value of parameters can be employed for the specific optimization problem.

    As one of the dominant modern meta-heuristic optimization algorithms, grasshopper optimization algorithm (GOA) has been successfully applied to various optimization problems in several fields. The good performance of GOA has been proved in [1]. It has one control parameter, which balances between exploration and exploitation. Therefore, its advantages are over the other optimization algorithms, including ease of implementation, speed in searching, and ease of modifying algorithm components [25]. However, it suffers from slow convergence and is prone to getting stuck in local optima [25,26]. To resolve the issues above, a multi-group and co-evolution framework was proposed to improve GOA and apply to multi-objective optimization problems, which draws on the idea of co-evolution strategy in [24]. In order to achieve population diversity and evolutionary adaptability, each subpopulation of GOA is assigned a different parameter c, including linear adaptation, cosine adaption, arc adaption, etc. At the same time, the subpopulation is dynamically updated in the optimization process.

    The main contributions of this research are as follows:

    1) A multi-group and co-evolution framework is proposed to archive a fine balance between exploration and exploitation of swarm intelligence algorithm

    2) Multi-objective grasshopper optimization algorithm base on the multi-group and co-evolution framework is developed to improve the convergence and diversity of optimal solutions.

    3) Detailed experiments are designed on several benchmark functions to demonstrate the effectiveness of the proposed methods.

    The rest of this paper is organized as follows. Section 2 introduces the related work about multi-objective optimization problems. Section 3 describes the definition of the multi-objective optimization problem to be solved in this paper. In Section 4, the multi-group and co-evolution framework, and the multi-objective grasshopper optimization algorithm based on this framework are illustrated in detail. Section 5 includes presentation and analysis of outcomes on test functions. Finally, Section 6 concludes the main work of this paper and suggests the next step of research work.

    Multi-objective optimization problems do not have a globally unique optimal solution, but to discover a cluster of optimal solutions. Therefore, the key to solving multi-objective optimization problems is to find a set of non-inferior solutions in the solution space, that is, the Pareto front. The approaches for settling multi-objective optimization problems can be divided into 5 categories [5]: decomposition methods [27,28], interactive methods [29,30], fuzzy methods [31,32], decision aid methods [33,34], and methods using meta-heuristics [35,36].

    The meta-heuristic is an iterative generation progress that combines different heuristic algorithms to explore the search space. These algorithms are applied to settle complex optimization problems where traditional heuristics and optimization methods are not capable of being effective and efficient [37]. Constructing a dynamic balance mechanism between diversification and intensification is a key step. Diversification can explore the broader search space, while intensification is an in-depth exploitation of specific areas by using the accumulated experience in the previous search process. Meta-heuristic methods can quickly explore the areas containing high-quality solutions in the search space, meanwhile, they do not waste time to exploit areas that have been exploit before or can not find high-quality solutions. There are three meta-heuristic strategies: a dominance-based approach, an indicator-based alternative, and an adaptive proposal that incorporates both multi-objective strategies (dynamically allocating more resources to the most successful strategy during the execution) [38]. Meta-heuristic methods have widespread applications. The Non-dominated Sorting Genetic Algorithm (NSGA-II) is applied to settle the multi-objective optimization problem of telecommunication network, and then use the Choquet integral measure based on utility function to make the final choice [39]. A meta-heuristic method for RNA inverse folding problem is designed by Tian et al. [40], in which RNA inverse folding problem is defined as a multi-objective optimization problem, and the similarity between the target structure and the predicted structure is employed as a constraint. In [41], a novel flexible job shop scheduling problem based on linear programming model is built to schedule the spool fabrication activities. Then priority dispatching rules based heuristic scheduling approach is utilized to settle the problem. In [42], an extended Multi-row facility layout problem (MRFLP) has been studied, and the genetic algorithm is applied to resolve the optimization problem.

    Recently, many meta-heuristic nature-inspired algorithms are proposed. They are nature-inspired and population-based optimization algorithms, which have been developing and adapting in varying ways, such as learner performance based behavior algorithm (LPB) [43], Dragonfly Algorithm (DA) [44], cat swarm optimization (CSO) [45], Backtracking search optimization algorithm (BSA) [46], Donkey and Smuggler Optimization Algorithm (DSO) [47], Fitness Dependent Optimizer (FDO) [48] and IFDO [49], Particle Swarm Optimization Algorithm(PSO) [50], Firefly Algorithm (FA) [51], Grasshopper Optimization Algorithm (GOA) [1], and several hybrid algorithm of different nature-inspired algorithms including A hybrid of Shuffled Frog Leaping Algorithm and Genetic Algorithm (SFGA) [52], WOAGWO [53] hybridizing Whale optimization algorithm (WOA) [54] with Grey wolf optimization (GWO) [55], and a many-objective multi-agent coordination optimization algorithm (MOMCO) [56] hybridizing with EA.

    Swarm intelligence is an important meta-heuristic optimization method, which has been widely used in different fields as a very important optimization technology. Swarm intelligence optimization algorithm is a combination of randomness and some rules to solve the optimization problem, through the simulation of natural phenomena. This kind of algorithm has a general optimization process: firstly, the population is initialized randomly, and then the search direction of every individual in the population is guided according to the rules. Then when termination conditions are met (such as, the maximum number of iterations is reached), the algorithm stops, and the final solution is the global optimal solution. Therefore, the composition of their computational complexity is similar, which is composed of three main parts: population initialization, population individual updating and fitness calculation. Generally, it is closely related to the number of iterations M, the size of the population N, the dimension of the decision space D, the complexity of updating function f1(x) of the individual and the complexity of the fitness function f2(x). Therefore, the computational complexity of swarm intelligence can be presented as O(N)+O(MND)×O(f1(x))+O(MN)×O(f2(x)), where the big oh notation is used to show the computation complexity. O(N) is computational complexity of population initializing, O(MND)×O(f1(x)) is the computational complexity of population updating, and O(MN)×O(f2(x)) is the computational complexity of fitness calculation. Such as the computational complexity of PSO [50] is O(N)+O(MND)×O(N)+O(MN)×O(N) ), which takes O(MN2D); the computational complexity of WOA [54] and GOA [1] is also takes O(MN2D) respectively.

    Among these swarm intelligence algorithms, GOA has been proved to have good performance in literature [1]. Therefore, there is a lot of literature on GOA based improvements and related applications. A comprehensive survey of the GOA is summarized in [25], which analyzes several modified versions of GOA and its applications in different fields in detail. For the purpose of ameliorating the optimization performance, GOA_jDE [57] which combines GOA and jDE [58] is proposed. GOA_jDE can greatly improve the convergence efficiency, convergence speed and calculation precision on the benchmark test in [57]. The chaos theory is also introduced into the optimization process of GOA, which employs chaotic maps to keep the exploration and exploitation progress balance of GOA and accelerate its global convergence speed [59]. A multi-objective grasshopper optimization algorithm for robot path planning in static environments was proposed to optimize several indexes such as cost, distance, energy or time [60].

    In these GOA based optimization methods, the control parameter c is a significant parameter to maintain a balance between exploration and exploitation. The adjustment of parameter c can be divided into linear control strategy (LCS) and nonlinear control strategy (NCS), according to the classification method in [22]. In the original GOA, the linear control strategy is adopted for the control parameter c, which linearly decreases from the maximum value to the minimum value with the increase of the number of iterations. LSC transits linearly from the exploration phase to the exploitation phase. However, in order to balance exploration and exploitation more reasonably, some nonlinear control strategies (NCS) are proposed, such as cosine adaption, arc adaption, etc. A comprehensive table of adjustment methods of parameter c in GOA is presented in Table 1.

    Table 1.  Literature review of adjustment methods of parameter c in GOA.
    Reference Control strategy Method
    [1,14,57,58,60,61,62,63] LCS c=cmaxmcmaxcminM
    [64] NCS c=(cos(πmM)+cmax)(cmax+cmax2)
    [64] NCS c=(cmaxmM)2
    [59] NCS xdi+1=c1(t){Nj=1jic2(t)ubdlbd2s(|xdjxdi|)xjxidij}+Td
    where c1(t) and c2(t) are the values generated from a wide variety of different chaotic maps in the t-th iteration.
    [65] NCS c=cmin+(1cmin)(1mw)1w
    where the values of w vary between 0 and 1, if w is greater than 0.5, exploration is more important, but if w is less than 0.5, exploitation is more important [65].
    [66] NCS c={δ[1+k1(fg/(fmaxfmin+favg))],fgfavgk2,fgfavg where, fmax and fmin determine the maximum/minimum value of all agents fitness when AGOA do a search operation, favg determines the average fitness, fg denotes the average fitness of the three parents in selection operation. δ, k1 and k2 notify the constant value in between 0 to 1[66].
    Formal notations used in this paper:c is a decreasing coefficient to shrink the comfort zone, repulsion zone, and attraction zone, which maintains the balance between exploration and exploitation of GOA; m and M indicate the current iteration and total number of iterations respectively; cmax and cmin represent the maximum and minimum values of parameter c; N is the number of grasshoppers; ubd and lbd are the upper and lower boundary of the D-dimensional search space separately; xi represents the position of the i-th grasshopper.

     | Show Table
    DownLoad: CSV

    In the optimization process of meta-heuristic methods, how to find a fine balance between exploration and development is critical. Each specific balance control strategy has a good performance in a certain kind of optimization problems, but how to select the appropriate balance control strategy for specific complex optimization problem is very difficult, and it needs a very deep understanding of the optimization problem. In order to solve the above issues, a framework is designed in this paper, which integrates a variety of balance control strategies to achieve adaptive selection of the appropriate balance and improve the optimization effect.

    The ordinary form of multi-objective optimization problem definition is as follows [4]:

    Minmize:F(x)={f1(x),f2(x),,fM(x)} (1)

    where x=[x1,x2,,xD] is a point in the D-dimensional decision space (search space), f=[f1,f2,,fM] is an objective space with M optimization objectives.

    The solutions of multi-objective optimization problems are usually a group of non-inferior solutions called Pareto optimal solutions. The Pareto front is the boundary defined by the set of all point mapped from Pareto optimal solutions. Equations (2) and (3) define the set of Pareto optimal solutions Ps and Pareto front Pf separately [4]:

    Ps={x:[(i{1,2,,k},x:fi(x)fi(x))(j{1,2,,k}:fj(x)<fi(x))]} (2)
    Pf={f(x):xPs} (3)

    In multi-objective optimization problems, the sub-objective functions may be uncorrelated or conflicting. It is not possible to have a certain solution that can make all the sub-objectives achieve optimal performance. Therefore, the primary task of all multi-objective optimization algorithms is to seek out as many Pareto optimal solutions as possible [5]. The proposed method of this paper has two objects: 1) finding a set of solutions f(x) which converge to the Pareto front Pf as close as possible, as shown in Eq (4); 2) finding a group of solutions as diverse as possible, as shown in Eq (5).

    |f(xPf)|t<ϵ (4)
    Difference(f(x))>σ (5)

    In this paper, the multi-group and co-evolution framework of swarm intelligence algorithm is designed to further achieve these two objects. By grouping the entire population into different subpopulations, the proposed framework increases the diversity of population and the randomness of search, to meet the requirements of Eq (5). In the iterative process, the co-evolution mechanism between subpopulations is established to ameliorate the convergence speed and accuracy of swarm intelligence algorithm, satisfying the requirements of Eq (4) as much as possible. The solution space of multi-objective optimization problem is transformed into the search space of the swarm intelligence algorithm, and the metrics of x is also transformed into the search conditions and evolution criterions of the swarm intelligence algorithm. So, in the multi-objective optimization problems, the convergence speed refers to the number of iterations t needed to find several solutions satisfying Eq (4). The convergence accuracy is determined by ϵ in Eq (4), the smaller ϵ is, the higher the convergence accuracy is. The diversity of Pareto optimal solutions is determined by Eq (5). Each solution should keep a certain difference or distance. In the experiment, this paper does not directly and independently determine the parameters ϵ and σ, but indirectly chooses the multi-objective optimal solutions according to the overall performance of their convergence and diversity on the real Pareto front.

    Firstly, the multi-group and co-evolution framework of swarm intelligence algorithm is presented in Section 4.1, including the design of the grouping strategy of the population, the co-evolution strategy among subpopulations, and the selection strategy of key parameters. In Section 4.2, we first briefly introduce GOA, and then illustrate how to integrate GOA into the framework in detail. MOGOA based on the multi-group and co-evolution framework is described in Section 4.3.

    The multi-group and co-evolution framework is an optimization mechanism of swarm intelligence algorithm. It can keep the exploration and exploitation balance in the search process and ameliorate the speed and accuracy of convergence by building variety of subpopulations and establishing information interaction mode between subpopulations. As shown in Figure 1, the multi-group and co-evolution framework of a swarm intelligence algorithm is divided into two key components: the grouping mechanism and the co-evolution mechanism. The grouping mechanism includes two steps: population division and parameter setting. The co-evolution mechanism includes two steps also: communication and feedback between subpopulations.

    Figure 1.  Multi-group co-evolution framework of swarm intelligence algorithm.

    When the optimization objective is determined, the swarm intelligence algorithm will select and determine the population size, control variables and other key parameters of the population to search the optimization space. Depending on the parameter change mechanism in the search process, the exploration and exploitation of the swarm intelligence algorithm are constantly compromised. A commonly used method is: in the incipient stage of search, the exploration is more important which make the search agent cover more search space; and then the exploitation is gradually emphasized with the iterative process of the algorithm, the optimal solution is searched in the local space [4]. After the population is determined, the search agents start to work according to the established search strategies in the whole search iteration process. Therefore, all search agents in the population and their search strategies lack difference. For the purpose of increasing the difference and diversity of search agents, this paper proposes a grouping mechanism, which divides the entire population into different subpopulations. Each subpopulation can be set different key parameters and search strategies.

    The entire population can be divided into different subpopulations in different ways. According to the size of subpopulation, there are average divisions, random divisions and so on; the setting of the initial parameters and search strategy of each subpopulation can also be achieved by random or fixed assignment. Based on GOA, this paper focuses on the research and implementation of two grouping mechanisms: fixed and random assignment of parameters under the condition of average population division.

    Different subpopulations have different optimization paths, and different search agents also produce different search information in each subpopulation. The proposed co-evolution mechanism is used to determine the interaction pattern of this information, including those within and between subpopulations. After different subpopulations have completed the search in parallel, they can share the information of the optimal solution and the worst solution found in the current iteration, and constantly adjust the search strategy. Meanwhile, the performance of different subpopulations is feed back to the grouping mechanism, and then the size and some key parameters of subpopulations are updated according to the feed information to further improve the efficiency of the swarm intelligence algorithm.

    The idea of grasshopper optimization algorithm is briefly described firstly, then the implementation of GOA with the grouping and co-evolution mechanism is described in detail, and the pseudo code of GOA based on the multi-group and co-evolution framework is also given.

    Following is presenting the mathematical model employed to simulate the swarming behavior of grasshoppers [1]:

    Xi=r1Si+r2Gi+r3Ai (6)

    where Xi represents the position of the i-th grasshopper, Si signifies social interaction, Gi defines the gravity force on the i-th grasshopper, Ai shows the wind advection. random numbers r1, r2 and r3 selected from [0, 1] are used to provide random behaviors.

    Si is an important component of the model [1], as follows:

    Si=Nj=1jis(|xjxi|)xjxidij (7)

    where dij represents the distance between i-th and j-th grasshopper, N is the number of grasshoppers.

    In Eq (7), the function s(r) signifies the social forces [1], as follows:

    s(r)=ferler (8)

    where f defines the intensity of attraction and l indicates the attractive length scale.

    Gi is calculated as follows [1]:

    Gi=geg (9)

    where g is a constant which is used to adjust the effect of gravity and eg is a unity vector towards the center of earth.

    Ai is calculated as follows [1]:

    Ai=uew (10)

    where u is a constant which is used to adjust the effect of drift and ew is a unity vector determining the direction of wind.

    So, Eq (6) can be rewritten as follows [1]:

    xdi=Nj=1jis(|xdjxdi|)xjxidijgeg+uew (11)

    To solve optimization problems, Gi is omitted and Ai is replaced by optimization target as the wind direction in Grasshopper Optimization Algorithm (GOA) [1]. The above model is adjusted as follows:

    xdi+1=c{Nj=1jicubdlbd2s(|xdjxdi|)xjxidij}+Td (12)

    where ubd and lbd are the upper and lower boundary of the D-dimensional search space separately. Td is the optimal solution of the objective function in the current iteration as the wind direction. c is a key parameter in GOA, which is proportional to the amount of iterations. The parameter c helps to balance the value of repulsion/gravitation between grasshoppers dynamically [1].

    If the parameter c decreases too fast in the inchoate stage of GOA, it will make an insufficient exploration. Conversely, if the parameter c decreases too slowly in the later stage of GOA, the exploitation of GOA will be insufficient. For the purpose of keeping the exploration and exploitation balance of GOA, this paper ingrates GOA into the multi-group and co-evolution framework to effectively balance local and global search ability of GOA.

    Grouping algorithm for GOA includes two key parameters setting: subpopulation number and the parameter c of GOA.

    ●  The initial subpopulation number and size

    There are several ways to divide population of GOA into different groups, such as average division, dynamic division and random division. This paper adopts the average division method, which facilitates the comparison of the final results during the iterative update process, such as Eq (13). The number and the size of subpopulation are related to the specific optimization problem.

    S=(S1,S2,,Sns) (13)

    where the whole population S are divided into ns subpopulations, each of which have the same size.

    ●  Settings of the parameter c of GOA

    In this paper, three forms are selected to set the parameter c of GOA [10,57], as follows:

    c1=cmaxmcmaxcminM (14)
    c2=(cos(πmM)+cmax)(cmax+cmax2) (15)
    c3=(cmaxmM)2 (16)

    where c1, c2 and c3 are three different forms of parameter c, c1is a linear adaptation form of LCS, c2 is a cosine adaption form of NCS, c3 is an arc adaption form of NCS; cmax means the maximum, cmin means the minimum, m is the current iteration number, M is the maximum iteration number, in this paper, cmax and cmin are set to 1 and 0.00001, respectively.

    We use a fixed assignment strategy and an equal probability random assignment strategy to allocate different parameter c to each subpopulation in this paper.

    The fixed assignment strategy adopts the following methods:

    cSi=cj,with cj(c1,c2,,cn) (17)

    where Si presents the i-th subpopulation, cSi is the form of parameter c of the i-th subpopulation, n is the number of parameters c, cj is one of n different forms of parameter c.

    Before the search process starting, each subpopulation Si fixedly chooses a corresponding parameter cj which remains unchanged in the whole optimization process.

    The random assignment strategy adopts the following methods:

    cSi=cr,with r=rand(1,n),cr(c1,c2,,cn) (18)

    where r is a random integer between 1 and n, n is the number of parameters c that can be chosen.

    In each iteration, every subpopulation can choose a different form of parameter c with equal probability according to the calculated random number in Eq (18). The Grouping algorithm not only enhances the diversity of GOA's population, but also avoids GOA from falling into a local optimum. The multiplicity of population also increases the adaptability of the algorithm to solve optimization problems with different characteristics.

    The co-evolutionary algorithm builds connections between the different subpopulations, realizes the information interaction between subpopulations, and then adjusts the evolution direction of GOA. Before running the next iteration, we compare the advantages and disadvantages of the optimal solution obtained from each subpopulation in the current iteration, in order that update the optimal solution of the entire population. The optimal solution is assigned to each subpopulation as the new evolution direction of GOA. Therefore, after each iteration, all subpopulations determine a new evolution direction again based on the results of the previous information interaction. In the frequent iterative process, subpopulations constantly exchange information to achieve the purpose of co-evolution. The co-evolutionary algorithm increases the probability that the population converges to the optimal solution. The optimal solution obtained is more intensification in each independent running.

    The pseudo code of GOA based on the multi-group and co-evolution framework (GOA-MC) is shown in Figure 2. According to grouping algorithm, the group operation first divides the entire population into ns subpopulations. In the co-evolutionary operation, the evolution direction of entire population is updated according to the optimal solutions of all subpopulations in each iteration. Then using Eq (12) update the position of all grasshoppers according to the direction of evolution. In each iteration, the parameter c of subpopulations can change using Eq (18). The update of positions of grasshoppers runs until the termination condition is met. Finally, the optimal fitness and position of the objective function are given.

    Figure 2.  The pseudo code of GOA-MC.

    Results of the experiment in Section 5.1 verify that there is a remarkable enhancement in convergence speed and accuracy of GOA-MC algorithm compared to the initial GOA algorithm. Therefore, a MOGOA-MC method is proposed which integrate MOGOA [29] into the multi-group and co-evolution framework, as shown in Figure 3. Using the same structure as MOGOA, this paper also constructs an external archive to keep optimal solutions from all subpopulations in each iteration and chooses the first n best performing optimal solutions as the final Pareto solution set in lines 5–6. Using the idea of co-evolutionary operation in algorithm 1, the evolutionary direction of all subpopulations is updated in line 7.

    Figure 3.  The pseudo code of MOGOA-MC.

    The final results of the multi-objective optimization problem are a group of solutions, which can not be directly compared with each other to get one optimal solution. So, we construct an external archive to store the optimal solution set, then compare and update the archives of all subpopulations together to obtain a global optimal solution set by adopting the elitist scheme in line 6. When the total amount of optimization solutions searched is greater than the size defined by the external archive, the average distance between each solution and all other solutions is figured using Eq (19), and the distances are sorted. The top n solutions with the largest average distance are selected and kept in the archive. Each new solution obtained from the next iteration needs to be compared with the average distance of other solutions in the archive, so as to constantly update the archive.

    A_dist(xi)=1KKj=1jidist(xixj) (19)

    where A_dist(xi) refers to the average distance between optimal solution xi and all other solutions, K is the size of the external archive.

    This method of selecting the optimal solutions has a disadvantage: in the process of archive updating, it can not guarantee that the final selected optimization solutions are those with the largest average distance among all searched optimal solutions. In other words, among all the optimal solutions obtained, the solutions with larger average distance may be deleted in the update process. This disadvantage can be improved by increasing the size of external archive, but the increase of capacity will increase the cost of optimal solutions ranking. The size of external archive is related to the specific optimization objectives, and a balance needs to be made between the cost of computation and the diversity of optimal solutions.

    This section first benchmark the performance of the proposed GOA-MC algorithm to verify the effectiveness of the multi-group and co-evolution framework in Section 5.1. Then in Section 5.2, several standard multi-objects test functions with different characteristics are applied to verify the performance of MOGOA-MC algorithm. The details of test functions employed in this work are presented in Tables A1–A6 of the Appendix. For the purpose of verifying the outcomes, GOA [1] and MOGOA [14] which have very good performance in the literatures of optimization problems are employed to compare with GOA-MC and MOGOA-MC respectively. All the experiments were carried out in this PC Configuration: System, Windows 10; CPU, 3.00 GHz; RAM, 16.00 GB; Language, Matlab 2016.

    For the purpose of verifying the improvement of convergence accuracy, convergence speed and search ability by the multi-group and co-evolution framework, a series of test functions [1,67] with different characteristics are applied, where test functions F1–F7 are unimodal benchmark functions, F9–F13 are multimodal benchmark functions, and F14–F19 are composite benchmark functions. Then the sensitivity analysis of the main parameters is carried out to verify the effectiveness of this framework.

    We compared and analyzed the performance of GOA with three different forms of parameter c and the GOA-MC with two different strategies. GOA-1, GOA-2, and GOA-3 are the original grasshopper optimization algorithm without using the multi-group and co-evolution framework. GOA-1, GOA-2, and GOA-3 respectively use the linear self-adaptive, cosine self-adaptive, and arc self-adaptive parameter c in Eqs (14)–(16). GOA-F and GOA-R are based on the multi-group and co-evolution framework respectively using fixed and randomly assigned parameter c in Eqs (14)–(16). For the sake of fairness, the main control parameters used in these algorithms refer to the parameter settings in [1] and [14], which are displayed in Table 2. Except for the number of subpopulations, other parameter settings are the same.

    Table 2.  Initial values for the control parameters of algorithms.
    Algorithm Parameter Value
    GOA-1
    GOA-2
    GOA-3
    size of population
    number of subpopulations
    number of iterations
    cmax, cmin
    120
    1
    300
    1, 0.00001
    GOA-F
    GOA-R
    size of population
    number of subpopulations
    number of iterations
    cmax, cmin
    120
    3
    300
    1, 0.00001

     | Show Table
    DownLoad: CSV

    Each test function runs 20 times independently for generating the statistical results in Table 3 and Figure 4. The average convergence curve of F1–F19 are presented in Figure 4. The steeper the curve descends, the faster the convergence rate will be. The closer the curve is to the x-axis, the better the convergence accuracy is. Compared with GOA-1, GOA-2 and GOA-3, GOA-F and GOA-R have significant performance improvement for most benchmark test functions. The way regulating the balance mechanism of exploration and exploitation is different by different forms of parameter c, such as LCS or NCS. In the GOA-F and GOA-R, subpopulations with different forms of parameter c search the optimal value at the same time. Therefore, in the process of searching the optimal value, a variety of balance mechanisms are used, then the search information exchange among subpopulations is completed through the co-evolution mechanism, which improves the ability of GOA to find the optimal value quickly and accurately. These results in Table 3 also prove that the optimization capability of GOA can be effectively improved by the multi-group and co-evolution framework.

    Table 3.  Results of benchmark test functions.
    GOA-1 GOA-2 GOA-3 GOA-F GOA-R
    AV. STD. AV. STD. AV. STD. AV. STD. AV. STD.
    F1 2.1065 × 10−8 6.8522 × 10−16 2.2761 × 10−13 9.6419 × 10−26 4.2179 × 10−15 4.8483 × 10−29 7.3512 × 10−15 2.1538 × 10−28 4.5877 × 10−16 5.3536 × 10−31
    F2 0.3737 0.2482 0.5282 0.8587 0.7809 1.1635 0.0116 6.4328 × 10−4 0.03344 0.0032
    F3 2.1711 × 10−7 8.7347 × 10−14 9.3245 × 10−12 2.7784 × 10−22 2.3598 × 10−14 6.8114 × 10−28 3.0693 × 10−11 7.9237 × 10−22 1.8549 × 10−14 2.1420 × 10−28
    F4 8.5561 × 10−5 1.2551 × 10−9 1.2373 × 10−7 1.2896 × 10−15 3.5723× 10−8 6.3449 × 10−16 1.2469 × 10−6 9.0957 × 10−13 2.583 × 10−8 2.0412 × 10−16
    F5 7.6481 135.3689 30.3799 3.5077 × 103 2.923 8.5927 0.47497 0.1125 1.9511 4.1055
    F6 2.9434 × 10−8 2.1365 × 10−16 2.3324 × 10−14 2.3266 × 10−28 2.9358 × 10−15 7.4365 × 10−30 5.721 × 10−12 4.7247 × 10−23 1.2923 × 10−15 3.7892 × 10−31
    F7 0.0120 2.4379 × 10−4 0.0083 4.6566 × 10−5 0.0216 0.0014 0.0003 4.0160 × 10−8 0.0011 8.4289 × 10−7
    F8 −1905.8277 1.7892 × 104 −1663.675 6.1265 × 10−4 −1651.5299 8.2914 × 103 −1822.201 4.6223 × 103 −1928.4939 2.5448 × 104
    F9 7.2976 19.5902 4.544 10.9820 5.966 31.0130 6.4971 67.4180 5.4234 27.8018
    F10 1.1123 0.7640 0.80252 1.0805 0.6585 0.7227 8.5829 × 10−7 6.6770 × 10−13 0.32926 0.4818
    F11 0.18848 0.0069 0.16064 0.0111 0.21093 0.0183 0.10766 0.0035 0.16875 0.0120
    F12 6.4353 × 10−5 2.0701 × 10−8 1.2925 × 10−8 7.9405 × 10−16 6.0977 × 10−10 1.7061 × 10−18 6.8682 × 10−9 2.3364× 10−16 2.2574 × 10−12 1.3613 × 10−23
    F13 0.0044 3.6223 × 10−5 6.9124 × 10−10 2.0558 × 10−18 1.7435 × 10−10 1.5146 × 10−19 3.7631 × 10−9 6.8730 × 10−17 1.3716 × 10−12 8.7490 × 10−24
    F14 0.998 7.0566 × 10−31 0.998 4.9304 × 10−32 0.998 2.4652 × 10−32 0.998 2.4652 × 10−32 0.998 2.4652 × 10−32
    F15 0.016592 7.1283 × 10−5 0.0052 7.1981 × 10−5 0.0049 7.4909 × 10−5 0.0007 5.1379 × 10−8 0.0012 4.2462 × 10−7
    F16 −1.0316 1.8413 × 10−24 −1.0316 0 −1.0316 2.4652 × 10−32 −1.0316 1.2326 × 10−32 −1.0316 3.6978 × 10−32
    F17 0.3979 1.0239 × 10−15 0.3979 1.9615 × 10−24 0.3979 2.9031 × 10−29 0.3979 6.2699 × 10−28 0.3979 0
    F18 3 8.1748 × 10−21 3 4.4275 × 10−29 3 4.4866× 10−30 3 7.6470 × 10−29 3 1.0354 × 10−29
    F19 −3.8628 7.7236 × 10−6 −3.8628 5.4281 × 10−12 −3.8628 3.7091 × 10−12 −3.8628 2.0292 × 10−15 −3.8628 5.3004 × 10−15
    *Notes: AV.: Average fitness value; STD.: Standard Deviation of fitness values obtained from 20 times independently running

     | Show Table
    DownLoad: CSV
    Figure 4.  Average convergence curve of F1–F19.

    For unimodal benchmark test functions F1–F7 which have only one global optimum, the outcomes in Table 3 and Figure 4 prove that the multi-group and co-evolution framework significantly improves convergence accuracy and speed of GOA. On the multimodal benchmark functions F8–F13 with several local optimal solutions, except for F9, GOA-F and GOA-R outperform others. The results of F8–F13 in Figure 4 also show the superiority of the multi-group and co-evolution framework in convergence speed. Due to the existence of local optimal solutions, the optimization algorithm needs to further balance exploration and exploitation and jumps out of local optimum through exploration. Therefore, the comprehensive performance in the multimodal benchmark functions shows that the grouping mechanism of the framework can further improve the exploration capacity of the algorithm, and the co-evolution mechanism simultaneously ensures the exploitation ability.

    Composite benchmark functions are more complex than other general multimodal benchmark functions. According to the composite benchmark functions F14–F19 in Table 3, it can be observed that all the algorithms have generated the same best results. In the process of calculation, we found an interesting phenomenon: all five algorithms will converge to the same optimal value after a certain amount of iterations, even if we adjust the parameters in Table 2 by a large margin. But in the convergence speed, they are still quite different. As can be observed from Figure 4, GOA-F and GOA-R have faster convergence speed than GOA-1, GOA-2 and GOA-3 on F14-F19. Because each subpopulation has different form of parameter c, GOA-F and GOA-R can simultaneously strengthen the ability of exploration and exploitation in different stages of the optimization process to accelerate convergence.

    In this section, the sensitivity analysis of main parameters is discussed in detail. The size of subpopulation and the number of groups are two main parameters affecting the performance of the multi-group and co-evolution framework. The impact on convergence accuracy and speed of proposed approach is examined by a series of experiments on test functions F1–F19. Table 4 presents the settings of the main parameters for sensitivity analysis. The size of the subpopulation has 5 different scales. The number of groups has 10 levels. Therefore, on each test function, 50 different groups of experiments of GOA-R are run 10 times independently to conduct a comprehensive sensitivity analysis.

    Table 4.  The settings of the relevant parameters for sensitivity analysis.
    Parameters Parameters range
    Size of subpopulation Five population sizes are set: 10, 40, 70, 90 and 130
    Number of groups Ten different number of groups are set: 1–10

     | Show Table
    DownLoad: CSV

    The results can be observed in Figure 5 in detail. The x-axis means the population size, and the numbers 1–5 indicate the population size of 10, 40, 70,100 and 130 respectively. The y-axis means the number of subpopulations. The z-axis represents the average of the optimal values obtained by each group after 10 independently runs. It should be noted that the results are normalized between 0 and 1 for facilitating the sensitivity analysis. The optimization results will not be significantly improved with the increasing number and size of subpopulations, but sometimes gets worse, such as the results of F2, F5, F8, F11, F15 and F19 in Figure 5. Similar results are also found on other test functions. The search agent of GOA appears to attract each other frequently on the unimodal test functions and have high repulsion rate between each other when settling multi-modal and composite test functions [1]. Therefore, when the size of subpopulation increases, the high repulsion rate will have negative influence on the convergence. Moreover, when the size of population continues to increase, this influence will gradually offset the performance improvement brought by the multi-group and co-evolution framework.

    Figure 5.  Sensitivity analysis on the population size and the number of subpopulations.

    As shown in Figure 6, with the increase in the population size and the number of subpopulations, the running time increases as well on test function F1. The similar results are also found on other test functions. The increase of the number of subpopulations increases the amount of information interaction between subpopulations, and the increase of population size enhances the amount of information interaction within subpopulations.

    Figure 6.  The running time change with the population size and the number of subpopulations.

    It can be seen from the above analysis that excessive grouping and large population size can not improve the convergence accuracy, but will significantly increase the running time. A practicable suggestion is that the size of subpopulation should be controlled between 40 and 100, and the number of groups should be controlled between 3 and 5.

    In this section, several standard multi-objective tests functions with different features are applied to verify the performance of the presented approach in multi-objective optimization. These standard test functions are used in many literatures on multi-objective optimization: ZDT [68], DTLZ [69], and CEC2009 [70].

    For the purpose of intuitively assess the performance of MOGOA-MC in convergence, accuracy, and diversity of optimal solutions, two performance indicators are employed: generation distance (GD) and inverse generation distance (IGD) [14]. IGD metric calculates the Euclidean distance between the obtained Pareto optimal solution and the true Pareto optimal solution in the reference set, which can measure the convergence and diversity of the algorithm.

    The smaller the value of IGD is, the better the overall performance of the algorithm is.

    IGD(P,P)=xPminyP dist(x,y)|P| (20)

    where P is the Pareto optimal solution acquired by the algorithm, P* is a group of uniformly distributed reference points sampled from the true Pareto optimal solutions, and dist (x, y) refer to the Euclidean distance.

    As a convergence evaluation indicator, GD metric measures the closeness between the obtained Pareto optimal solutions and the true Pareto optimal solutions. The closer GD is to 0, the better the convergence is.

    GD(P,P)=yPminxPdist(x,y)2|P| (21)

    where the definitions of P, P*, and dist (x, y) are the same as those in IGD.

    Similar to GOA-MC, for the purpose of verifying the advantage of the multi-group and co-evolution framework in settling multi-objective optimization problems, we designed a comparative experiment between the MOGOA combined with the multi-group and co-evolution framework (MOGOA-MC) and the original MOGOA. The original MOGOA separately adopts three different forms of parameter c, which are MOGOA-1 using the parameter c in Eq (14), MOGOA-2 using the parameter c in Eq (15) and MOGOA-3 using the parameter c in Eq (16). MOGOA-MC separately adopts two different assignment strategies of parameter c, which are MOGOA-F and MOGOA-R. MOGOA-F adopts fixed assignment strategy to assign a fixed parameter c to each subpopulation from Eq (17). MOGOA-R using random assignment strategy to randomly assign a parameter c for each subpopulation from Eq (18). Other parameters are set uniformly as follows: the amount of population is set to 120, the maximum numbers of iterations is set to 100, the number of subpopulations is 3, cmax and cmin are set to 1 and 0.00001 respectively.

    The quantitative results on ZDT and DTLZ are presented in Tables 57. All algorithms are run independently 20 times. The average, standard deviation, best, and worst values of IGD are presented in Table 5, and the values of GD are presented in Table 6. The obtained Pareto fronts are qualitatively illustrated in Figures 7 and 8.

    Table 5.  Statistical results for IGD on ZDT1, ZDT2, ZDT3, ZDT4.
    IGD AV. STD. Worst Best AV. STD. Worst Best
    ZDT1 ZDT2
    MOGOA-1 0.008110 4.7486 × 10−5 0.01781 0.001834 0.02181 9.1174 × 10−5 0.03179 0.004372
    MOGOA-2 0.007148 2.4368 × 10−6 0.02056 0.002063 0.005184 4.7499 × 10−5 0.006806 0.003169
    MOGOA-3 0.008415 3.4927 × 10−5 0.01019 0.002604 0.01787 5.7181 × 10−5 0.02821 0.003130
    MOGOA-F 0.001804 3.0666 × 10−7 0.003185 0.001253 0.005893 8.6611 × 10−7 0.02075 0.001781
    MOGOA-R 0.001596 3.5401 × 10−8 0.002075 0.001159 0.005196 7.8079 × 10−7 0.01625 0.001880
    ZDT3 ZDT4
    MOGOA-1 0.007790 2.3325 × 10−5 0.01105 0.003316 0.08460 0.0026 0.1368 0.03810
    MOGOA-2 0.01659 1.6409 × 10−5 0.03471 0.004913 0.1329 0.0024 0.2114 0.02060
    MOGOA-3 0.02568 1.2805 × 10−6 0.04807 0.009530 0.1198 0.0048 0.1887 0.03327
    MOGOA-F 0.004040 2.6464 × 10−7 0.008719 0.002587 0.06507 0.0020 0.1041 0.03486
    MOGOA-R 0.003725 5.6783 × 10−6 0.004526 0.002669 0.06345 0.0006 0.1062 0.02554
    *Notes: AV.: Average IGD value; STD.: Standard Deviation of IGD obtained from 20 times independently running; Best: the best values of IGD; Worst: the worst values of IGD

     | Show Table
    DownLoad: CSV
    Table 6.  Statistical results for GD on ZDT1, ZDT2, ZDT3, ZDT4.
    GD Average STD. Worst Best Average STD. Worst Best
    ZDT1 ZDT2
    MOGOA-1 0.01670 2.6674 × 10−5 0.02259 0.01021 0.03076 6.7313 × 10−5 0.04413 0.02325
    MOGOA-2 0.03256 9.7455 × 10−5 0.05450 0.01714 0.03358 7.2653 × 10−5 0.04381 0.01948
    MOGOA-3 0.02952 2.2262 × 10−5 0.04720 0.01451 0.02260 3.3174 × 10−4 0.03280 0.01217
    MOGOA-F 0.01138 1.9063 × 10−5 0.01340 0.007802 0.01892 5.8046 × 10−5 0.02858 0.01394
    MOGOA-R 0.01207 6.6201 × 10−6 0.01498 0.009695 0.01933 1.2821 × 10−5 0.02801 0.01195
    ZDT3 ZDT4
    MOGOA-1 0.01762 8.0160 × 10−5 0.03189 0.008943 0.5279 0.0408 0.8772 0.3554
    MOGOA-2 0.01432 1.4615 × 10−4 0.01124 0.01872 0.6291 0.0367 0.8792 0.4846
    MOGOA-3 0.02113 1.3262 × 10−5 0.03587 0.009765 0.5735 0.0095 0.9137 0.3339
    MOGOA-F 0.009671 2.6231 × 10−6 0.01316 0.007496 0.2822 0.0101 0.3790 0.1958
    MOGOA-R 0.008808 4.8387 × 10−6 0.01167 0.005164 0.2960 0.0058 0.3931 0.1758
    *Notes: AV.: Average GD value, STD.: Standard Deviation of GD obtained from 20 times independently running; Best: the best values of GD; Worst: the worst values of GD

     | Show Table
    DownLoad: CSV
    Table 7.  Statistical results for IGD and GD on DTLZ1.
    DTLZ1 IGD GD
    AV. STD. Worst Best AV. STD. Worst Best
    MOGOA-1 1.5255 0.0059 1.6343 1.4330 8.7203 0.1935 9.1334 8.0797
    MOGOA-2 1.4481 0.0096 1.5687 1.3351 10.7424 4.4340 14.1027 8.7297
    MOGOA-3 1.4337 0.0171 1.5761 1.2453 10.2828 1.0573 10.2828 8.1977
    MOGOA-F 1.2910 0.0022 1.3548 1.2312 7.1866 0.0513 7.5682 7.0019
    MOGOA-R 1.3062 4.7446 × 10−4 1.3297 1.2826 7.1640 0.0064 7.2661 7.0701
    *Notes: AV.: Average IGD or GD value; STD.: Standard Deviation of IGD or GD values obtained from 20 times independently running; Best: the best values of IGD or GD; Worst: the worst values of IGD or GD

     | Show Table
    DownLoad: CSV
    Figure 7.  Pareto optimal front on ZDT.
    Figure 8.  Pareto optimal front on DTLZ1.

    Each benchmark test function has its own characteristics, and different settings of key parameters will have an impact on the performance of algorithms. Such as, on ZDT1, ZDT2 and ZDT4, MOGOA-2 utilizing the parameter c with cosine adaption form of NCS have better performance than MOGOA-1 and MOGOA-3; On ZDT3 which has a nonconvex and discontinuous Pareto front, MOGOA-1 utilizing the parameter c with line adaption form of LCS have better performance than MOGOA-2 and MOGOA-3. However, in practical application, it is very difficult to choose the appropriate parameter settings when the optimization problem is not well understood. Therefore, in this paper, in order to solve the problem of parameter selection and setting, a variety of different settings of one key parameter are comprehensively applied in the optimization search process. The effectiveness of the proposed method is also proved by the results in Table 5. Such as, IGD values of MOGOA-F and MOGOA-R are significantly better than MOGOA-1, MOGOA-2 and MOGOA-3 on ZDT1, ZDT3 and ZDT4. On ZDT2, compared with the best algorithm MOGOA-2, MOGOA-F and MOGOA-R also show strong competitiveness.

    Through the co-evolution mechanism, the optimization information can be exchanged between subpopulations, so that all search individuals can converge to the true Pareto front as soon as possible. Therefore, the global optimization solutions can be found faster and more accurately. The conclusions can be drawn from the values of GD in Table 6, MOGOA-F and MOGOA-R have significantly better performance than others.

    Practical optimization problems usually involve more than two objectives. The more objectives there are, the more complex the optimization problem is, so it is difficult to get its Pareto front and to visualize it. In order to verify the effectiveness of the method in the multi-objective optimization problem with more than two objectives, we benchmark it on DTLZ1. DTLZ1 has three objectives that need to be optimized at the same time, and its Pareto front is a hyperplane. The quantitative results in Table 7 also show the good IGD and GD performance of MOGOA-F and MOGOA-R on DTLZ1. Although the performance improvement is not as good as ZDT series test functions, the effect is still significant.

    The quantitative results in Tables 57 verify the superiority of the algorithms with the multi-group and co-evolution framework and prove that the framework can significantly improve the diversity and convergence of Pareto optimal solutions. The qualitative distribution presentation of best Pareto optimal fronts in Figures 7 and 8 also supports this conclusion. In our method, different subpopulations have different settings of key parameters, and the optimization mechanism of search individuals is also different. Therefore, more diverse optimization solutions can be found, and the distribution of optimization schemes is more uniform. The Pareto optimal fronts obtained from MOGOA-F and MOGOA-R are more equally distributed in the true Pareto optimal fronts than that of MOGOA-1, which also prove that the multi-group and co-evolution framework improves the diversity of Pareto optimal solutions.

    In this section, the CEC2009 test functions are applied to further confirm the performance of the proposed methods. UF1–UF7 test functions are bi-objective, UF8–UF10 test functions are tri-objective. The results on CEC2009 test functions are presented in Tables 8 and 9.

    Table 8.  Statistical results for IGD on UF1 to UF10.
    IGD AV. STD. Worst Best Average STD. Worst Best
    UF1 UF2
    MOGOA-1 0.1151 6.554 × 10−4 01597 0.0880 0.0594 6.612 × 10−5 0.0729 0.0468
    MOGOA-2 0.0985 1.697 × 10−4 0.1337 0.0866 0.0645 9.783 × 10−5 0.0815 0.0463
    MOGOA-3 0.1133 9.301 × 10−5 0.1266 0.0997 0.0816 1.842 × 10−4 0.1105 0.0635
    MOGOA-F 0.0844 1.136 × 10−5 0.0901 0.0800 0.0439 1.496 × 10−5 0.0506 0.0377
    MOGOA-R 0.0865 1.866 × 10−5 0.0942 0.0795 0.0433 2.161 × 10−5 0.0537 0.0371
    UF3 UF4
    MOGOA-1 0.4206 0.0062 0.5074 0.3117 0.1464 1.478 × 10−4 0.1617 0.1212
    MOGOA-2 0.4534 0.0123 0.6390 0.3373 0.1695 1.623 × 10−3 0.2355 0.1511
    MOGOA-3 0.4240 0.0141 0.5824 0.2429 0.1196 1.118 × 10−4 0.1306 0.0974
    MOGOA-F 0.3937 0.0189 0.5836 0.1989 0.0979 4.529 × 10−5 0.1074 0.0867
    MOGOA-R 0.4129 0.0111 0.5460 0.2836 0.0952 1.134 × 10−5 0.0998 0.0903
    UF5 UF6
    MOGOA-1 0.6389 0.0181 0.8746 0.4432 0.7147 0.0289 0.9301 0.4911
    MOGOA-2 0.7470 0.3465 2.3977 0.3523 0.6317 0.0342 0.9505 0.4775
    MOGOA-3 0.7433 0.0369 1.0504 0.4615 0.6590 0.0350 0.8884 0.4050
    MOGOA-F 0.6062 0.0152 0.7732 0.4153 0.5581 0.0245 0.8470 0.3660
    MOGOA-R 0.7570 0.0457 1.1485 0.4526 0.5351 0.0333 0.8000 0.2680
    UF7 UF8
    MOGOA-1 0.0718 4.958 × 10−5 0.0856 0.0619 0.2415 4.804 × 10−4 0.2800 0.2142
    MOGOA-2 0.0935 7.649 × 10−4 0.1363 0.0595 0.2301 7.129 × 10−4 0.2697 0.1905
    MOGOA-3 0.0834 1.507 × 10−4 0.1066 0.0647 0.2622 8.372 × 10−4 0.3121 0.2193
    MOGOA-F 0.0652 3.215 × 10−5 0.0740 0.0584 0.1650 7.117 × 10−5 0.1785 0.1499
    MOGOA-R 0.0629 2.813 × 10−5 0.0745 0.0552 0.1770 2.142 × 10−4 0.2077 0.1601
    UF9 UF10
    MOGOA-1 0.2485 2.385 × 10−3 0.3604 0.1980 0.3965 9.223 × 10−3 0.5928 0.2685
    MOGOA-2 0.2381 1.305 × 10−3 0.3325 0.2174 0.3935 8.569 × 10−3 0.5442 0.2206
    MOGOA-3 0.2874 1.112 × 10−3 0.3568 0.2414 0.6113 8.244 × 10−3 1.0740 0.3165
    MOGOA-F 0.1836 2.932 × 10−4 0.2122 0.1618 0.2583 6.211 × 10−4 0.2997 0.2171
    MOGOA-R 0.1824 3.376 × 10−4 0.2100 0.1515 0.2377 9.941 × 10−5 0.2552 0.2254
    *Notes: AV.: Average IGD value; STD.: Standard Deviation of IGD obtained from 20 times independently running; Best: the best values of IGD; Worst: the worst values of IGD

     | Show Table
    DownLoad: CSV
    Table 9.  Statistical results for GD on UF1 to UF10.
    GD Average STD. Worst Best Average STD. Worst Best
    UF1 UF2
    MOGOA-1 0.0179 1.359 × 10−4 0.0439 0.0074 0.0194 5.765 × 10−4 0.0701 0.0035
    MOGOA-2 0.0126 4.841 × 10−5 0.0254 0.0064 0.0104 1.638 × 10−4 0.0465 0.0040
    MOGOA-3 0.0176 1.815 × 10−4 0.0516 0.0042 0.0174 2.290 × 10−4 0.0472 0.0049
    MOGOA-F 0.0081 1.557 × 10−5 0.0163 0.0031 0.0081 2.046 × 10−5 0.0149 0.0037
    MOGOA-R 0.0118 6.625 × 10−5 0.0317 0.0033 0.0080 6.252 × 10−5 0.0299 0.0038
    UF3 UF4
    MOGOA-1 0.1126 0.0057 0.2702 0.0359 0.0184 2.052 × 10−5 0.0278 0.0124
    MOGOA-2 0.1066 0.0027 0.1804 0.0369 0.0236 9.309 × 10−5 0.0442 0.0131
    MOGOA-3 0.1126 0.0130 0.4048 0.0417 0.0131 2.589 × 10−6 0.0157 0.0103
    MOGOA-F 0.0603 0.0019 0.1375 0.0138 0.0144 8.153 × 10−6 0.0181 0.0105
    MOGOA-R 0.0657 0.0012 0.1187 0.0298 0.0130 1.071 × 10−6 0.0153 0.0119
    UF5 UF6
    MOGOA-1 0.1698 0.0018 0.1701 2.046 × 10−4 0.1313 0.0112 0.3372 0.0087
    MOGOA-2 0.1545 0.0221 0.5399 0.0307 0.1056 0.0036 0.2272 0.0149
    MOGOA-3 0.1208 0.0088 0.2931 9.984 × 10−13 0.1786 0.0302 0.5893 0.0412
    MOGOA-F 0.1299 0.0058 0.3013 0.0198 0.1181 0.0226 0.2127 0.0226
    MOGOA-R 0.1270 0.0031 0.2334 0.0773 0.1225 0.0476 0.2446 0.0476
    UF7 UF8
    MOGOA-1 0.0074 2.416 × 10−5 0.0169 0.0013 0.0434 3.596 × 10−4 0.0693 0.0153
    MOGOA-2 0.0083 3.660 × 10−5 0.0222 0.0019 0.0299 2.001 × 10−4 0.0571 0.0132
    MOGOA-3 0.0084 1.075 × 10−5 0.0144 0.0034 0.0389 3.710 × 10−4 0.0691 0.0121
    MOGOA-F 0.0051 6.742 × 10−6 0.0098 0.0023 0.0227 6.346 × 10−5 0.0354 0.0134
    MOGOA-R 0.0044 3.043 × 10−6 0.0082 0.0024 0.0289 5.846 × 10−4 0.0941 0.0115
    UF9 UF10
    MOGOA-1 0.0459 4.682 × 10−4 0.0994 0.0245 0.2103 6.276 × 10−3 0.3394 0.1107
    MOGOA-2 0.0455 1.671 × 10−4 0.0643 0.0259 0.2502 1.831 × 10−3 0.5226 0.0447
    MOGOA-3 0.0612 1.127 × 10−3 0.1237 0.0162 0.2603 2.171 × 10−3 0.5838 0.0926
    MOGOA-F 0.0374 1.021 × 10−4 0.0523 0.0232 0.2697 3.392 × 10−3 0.3572 0.1663
    MOGOA-R 0.0424 2.216 × 10−4 0.0659 0.0166 0.2306 3.232 × 10−3 0.3497 0.1622
    *Notes: AV.: Average GD value; STD.: Standard Deviation of GD obtained from 20 times independently running; Best: the best values of GD; Worst: the worst values of GD

     | Show Table
    DownLoad: CSV

    The conclusions on CEC2009 test functions are consistent with those of ZDT series test functions and DTLZ test functions. The multi-group mechanism in the multi-group and co-evolution framework gives more differences to the search individuals, which can help to search more diverse optimization solutions; the co-evolution mechanism makes the subpopulations with better performance can transmit information to the whole population, so as to speed up the convergence speed and improve the search efficiency and convergence accuracy.

    The IGD values in Table 8 show that the performance of MOGOA-F and MOGOA-R are also better than that of MOGOA-1, MOGOA-2 and MOGOA-3 on the CEC2009 test suite. The GD values in Table 9 also verify that the proposed algorithms have better convergence than original MOGOA. However, on UF5, UF6 and UF10, although the proposed method is very competitive, it does not show the best performance. Since the Pareto fronts of UF5 and UF6 are discontinuous, and the metric of GD measures the average Euclidean distance from each solution obtained from algorithm to the solution on the nearest true Pareto front, the performance of the proposed methods on this kind of optimization problems are not significant.

    In terms of overall performance, these outcomes in Tables 8 and 9 can affirm that the multi-group and co-evolution framework is able to ameliorate the diversity and the convergence of the Pareto optimal solutions.

    The average, standard deviation, maximum, and minimum in the previous tables are obtained from the results of 20 independent runs. They reflect the average performance of the algorithm. To show that the outcomes were not acquired accidentally, the Wilcoxon rank-sum test is employed. The p-values are the results of the Wilcoxon rank-sum test. When p-values are less than 0.05, it shows that rank sum rejects the null hypothesis of equal medians at the default 5% significance level. The higher the p value, the more competitive the algorithm is, even if it is not the best one. The p-values of the algorithm with the best average performance in each test function is "N/A". Then the best performing algorithm is chosen to test with other algorithms in pairs. The p-values in Table 10 show that MOGOA-F and MOGOA-R based on the multi-group and co-evolution framework performs well on most of the test functions. The p-values between MOGOA-F and MOGOA-R are all greater than 0.05 on GD and IGD, which indicates that the two algorithms are highly competitive with each other. While MOGOA-1, MOGOA-2 or MOGOA-3 are not the algorithm with the best average performance, their p-values are less than 0.05 in most test functions, so they are not competitive.

    Table 10.  P-values acquired from the rank sum test on UF1 to UF10.
    Test functions GD IGD
    M-1 M-2 M-3 M-F M-R M-1 M-2 M-3 M-F M-R
    UF1 0.0073 0.1212 0.0211 N/A 0.3075 0.0001 0.0001 0.0001 N/A 0.3075
    UF2 0.2730 0.4274 0.0257 0.9698 N/A 0.0001 0.0001 0.0001 0.6232 N/A
    UF3 0.0757 0.0376 0.1405 N/A 0.6232 0.6776 0.3447 0.6232 N/A 0.7337
    UF4 0.0017 0.0001 0.5205 0.2730 N/A 0.0001 0.0001 0.0001 0.1212 N/A
    UF5 0.1620 0.9698 N/A 0.6776 0.7337 0.6232 0.7913 0.1212 N/A 0.1041
    UF6 0.9097 N/A 0.4727 0.9097 0.6232 0.0452 0.3847 0.1859 0.9097 N/A
    UF7 0.0922 0.0708 0.0036 0.8501 N/A 0.0073 0.0058 0.0001 0.5205 N/A
    UF8 0.0140 0.0166 0.0241 N/A 0.8501 0.0001 0.0001 0.0001 N/A 0.0376
    UF9 0.2725 0.1315 0.0257 N/A 0.3881 0.0008 0.0001 0.0001 0.9097 N/A
    UF10 N/A 0.4272 0.6232 0.1041 0.3075 0.0001 0.0028 0.0001 0.0211 N/A
    *Notes: M-1: MOGOA-1; M-2: MOGOA-2; M-3: MOGOA-3; M-F: MOGOA-F; M-R: MOGOA-R

     | Show Table
    DownLoad: CSV

    In general, the above experimental outcomes indicate that the improvement of the multi-group and co-evolution framework to the diversity and convergence of MOGOA is very significant. Although MOGOA has proved its superiority in reference [14], which are compared with other competitive multi-objective optimization algorithms, such as MOPSO, MOEA/D. The proposed methods are greatly effective for settling optimization problems with multiple objectives. This is mainly attributed to these two features of the framework: firstly, the grouping of subpopulations with different evolutionary mechanisms increases the diversity of search agents and strengthens the exploration capability of the optimization algorithm; secondly, the co-evolution mechanism between subpopulations ensures the convergence of the algorithm, and accelerates the convergence speed.

    In this paper, a multi-group and co-evolution framework for meta-heuristic methods is proposed, which is employed to improve the performance of GOA. With the fast convergence and high accuracy, the promising performance of the framework is verified on the benchmark test functions. The sensitivity of the main parameters is analyzed to check the feasibility and validation of the proposed framework. For the multi-objective optimization problems, the proposed framework was used to extend the MOGOA algorithm. To evaluate the performance improvement brought by the proposed framework, the numerical evaluations were also conducted by comparing with the original MOGOA on a series of test suits. GD and IGD were calculated to prove the advancement of diversity and accuracy of solutions by the proposed framework. Moreover, the Wilcoxon rank-sum tests were also conducted to show the statistically significant. It could be said that these improvements were due to the modifications in the difference and interaction of search agents, which led to more convenient exploration and more active exploitation. The proposed methods in this paper can find a fine balance between exploration and exploitation, which is particularly important in solving multimodal search space and large-scale optimization problems without deep understanding, such as feature selection and neural network training in machine learning, job-shop scheduling and control of power systems in engineering applications. Experimental results show the effectiveness of the proposed method, but there are still some limitations. Because of no free lunch theorem in the optimization domain, the proposed method needs further adjustment and modification to adapt to the actual specific problems, such as binary, dynamic, discrete, and others. Due to the existence of co-evolution among subpopulations, the proposed method spends more time than original algorithm in the optimization process. The multi-group and co-evolution framework is only applicable to the same kind of swarm intelligence algorithm at present. Therefore, for future research, it is suggested to further improve the multi-group and co-evolution framework for integrating various swarm intelligence algorithms. The settings of key parameters should also be quantitatively explained according to specific problems in this framework. It will be focused on the applications on more complex engineering problems.

    This work is partially supported by the Natural Science Foundation of China under Grant (31771679, 31671589), the Anhui Foundation for Science and Technology Major Project, China, under Grant (16030701092, 18030901034, 201904e01020006), the Key Laboratory of Agricultural Electronic, Commerce, Ministry of Agriculture of China under Grant (AEC2018003, AEC2018006), the 2019 Anhui University collaborative innovation project (GXXT-2019-013), the Hefei Major Research Project of Key Technology (J2018G14), Natural Science Research Project of Anhui Provincial Department of Education (KJ2020A0107).

    The authors declare no conflict of interest.

    Test functions utilized in this paper.

    Table A1.  Unimodal benchmark functions.
    Function Dim Range fmin
    f1(x)=ni=1x2i 30 [–100,100] 0
    f2(x)=ni=1|xi|+ni=1|xi| 30 [–10, 10] 0
    f3(x)=ii=1(ij1xj)2 30 [–100,100] 0
    f4=max{|xi|,1in} 30 [–100,100] 0
    f5(x)=n1i=1[100(xi+1x2i)2+(xi1)2] 30 [–30, 30] 0
    f6(x)=ni=1([xi+0.5])2 30 [–100,100] 0
    f7(x)=ni=1ix4irandom[0,1) 30 [–128,128] 0

     | Show Table
    DownLoad: CSV
    Table A2.  Multimodal benchmark functions.
    Function Dim Range fmin
    F8(x)=ni=1xi sin (|xi|) 30 [–500,500] –418.9829 × Dim
    F9(x)=ni=1[x2i10 cos (2πxi)+10] 30 [–5.12, 5.12] 0
    F10(x)=20exp(0.21nni=1x2i)exp(1nni=1cos(2πxi))+20+e 30 [–32, 32] 0
    F11(x)=14000ni=1x2ini=1cos(xii)+1 30 [–600,600] 0
    F12(x)=πn{10(sin πy1)+n1i=1(yi1)2[1+10sin2(πyi+1)]+(yn1)2}                        +ni=1u(xi,10,100,4)+ni=1u(xi,10,100,4)
    yi=1+xi+14
    u(xi,a,k,m)={k(xia)mxi>a0a<xi<ak(xia)mxi<a
    30 [–50, 50] 0
    F13(x)=0.1{sin2(3πx1)                              +ni=1(x11)2[1+sin2(3πxi+1)]                              +(xn1)2[1+sin2(2πxn)]}+ni=1u(xi,10,100,4) 30 [–50, 50] 0

     | Show Table
    DownLoad: CSV
    Table A3.  Composite benchmark functions.
    Function Dim Range fmin
    F14(CF1):f1,f2,f3,...,f10=Sphere Function[σ1,σ2,σ3,...,σ10]=[1,1,1,...,1][λ1,λ2,λ3,...,λ10]=[5/100,5/100,5/100,...,5/100] 30 [–5, 5] 0
    F15(CF2):f1,f2,f3,...,f10=Griewanks Function[σ1,σ2,σ3,...,σ10]=[1,1,1,...,1][λ1,λ2,λ3,...,λ10]=[5/100,5/100,5/100,...,5/100] 30 [–5, 5] 0
    F16(CF3):f1,f2,f3,...,f10=Griewanks Function[σ1,σ2,σ3,...,σ10]=[1,1,1,...,1][λ1,λ2,λ3,...,λ10]=[1,1,1,...,1] 30 [–5, 5] 0
    f17(CF4):f1,f2=Ackleys Functionf3,f4=Rastrigins Functionf5,f6=Weierstrasss Functionf7,f8=Griewanks Functionf9,f10=Sphere Function[σ1,σ2,σ3,...,σ10]=[1,1,1,...,1][λ1,λ2,λ3,...,λ10]=[5/32,5/32,1,1,5/0.5,5/0.5,5/100,5/100,5/100,5/100] 30 [–5, 5] 0
    f18(CF5):f1,f2=Rastrigins Functionf3,f4=Weierstrasss Functionf5,f6=Griewanks Functionf7,f8=Ackleys Functionf9,f10=Sphere Function[σ1,σ2,σ3,...,σ10]=[1,1,1,...,1][λ1,λ2,λ3,...,λ10]=[1/5,1/5,5/0.5,5/0.5,5/100,5/100,5/32,5/32,5/100,5/100] 30 [–5, 5] 0
    f19(CF6):f1,f2=Rastrigins Functionf3,f4=Weierstrasss Functionf5,f6=Griewanks Functionf7,f8=Ackleys Functionf9,f10=Sphere Function[σ1,σ2,σ3,...,σ10]=[0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8,0.9,1][λ1,λ2,λ3,...,λ10]=[0.11/5,0.21/5,0.35/0.5,0.45/0.5,0.55/100]0.65/100,0.75/32,0.85/32,0.95/100,15/100 30 [–5, 5] 0

     | Show Table
    DownLoad: CSV
    Table A4.  Multi-objective benchmark functions (ZDT & DTLZ1).
    Name Function
    ZDT1 minf1(x1)=x1
    minf2(x)=g(1(f1/g))
    g(x)=1+9mi=2x1/(m1)
    s.t.0xi1,i=1,2,...,30
    ZDT2 minf1(x1)=x1
    minf2(x)=g(1(f1/g)2)
    g(x)=1+9mi=2xi/(m1)
    s.t.0xi1,i=1,2,...,30
    ZDT3 minf1(x1)=x1
    minf2(x)=g(1f1/g(f1/g)sin(10πf1))
    g(x)=1+9mi=2xi/(m1)
    s.t.0xi1,i=1,2,...,30
    ZDT4 minf1(x1)=x1
    minf2(x)=g(1(f1/g))
    g(x)=1+10(n1)+mi=2(xi10cos(4πxi))
    s.t.0x11,5xi5,i=2,...,9
    DTLZ1 minf1(x1)=0.5x1x2(1+g)
    minf2(x)=0.5x1(1x2)(1+g)
    minf3(x)=0.5(1x1)(1+g)
    g(x)=100(5+mi=3(xi0.5))2cos(20πxi0.5)
    0xi1,i=1,2,...,30

     | Show Table
    DownLoad: CSV
    Table A5.  Bi-objective test problems (CEC2009).
    NameFunction
    UF1 f1=x1+2| J1 |j J1 [xjsin(6πx1+jπn)]2,f2=1x+2| J2 |j J2 [xjsin(6πx1+jπn)]2
     J1 ={j|j is odd and 2jn}, J2 ={j|j is even and 2jn}
    UF2 f1=x1+2| J1 |j J1 y2j,f2=1x+2| J2 |j J2 y2j
     J1 ={j|j is odd and 2jn}, J2 ={j|j is even and 2jn}
    yj={xj[0.3x21 cos (24πx1+4jπn)+0.6x1] cos (6πx1+jπn)if j J1 xj[0.3x21 cos (24πx1+4jπn)+0.6x1] sin (6πx1+jπn)     if j J2 
    UF3 f1=x1+2| J1 |(4j J1 y2j2j J1  cos (20yjπj)+2)
    f2=x1+2| J2 |(4j J1 y2j2j J2  cos (20yjπj)+2)
     J1 and J2 are the same as those of UF1,yj=xjx0.5(1.0+3(j2)n2)1, j=2,3,,n
    UF4 f1=x1+2| J1 |j J1 h(yj),f2=1x2+2| J2 |j J2 h(yj)
     J1 and J2 are the same as those of UF1,yj=xj sin (6πx1+jπn),j=2,3,,n, h(t)=|t|1+e2|t|
    UF5 f1=x1+(12N+ϵ)| sin (2Nπx1)|+2| J1 |jϵ J1 h(yj),f2=x1+(12N+ϵ)| sin (2Nπx1)|+2| J2 |jϵ J2 h(yj)
     J1 and J2 are identical to those of UF1,yj=xj sin (6πx1+jπn),
    j=2,3,,n, h(t)=|t|1+e2|t|
    h(t)=2t2 cos (4πt)+1
    UF6 f1=x1+max{0,2(12N+ϵ)sin(2Nπx1)}+2| J1 |(4j J1 y2j2j J1  cos (20yjπj)+1)
    f2=1x1+max{0,2(12N+ϵ)sin(2Nπx1)}+2| J2 |(4j J2 y2j2j J2  cos (20yjπj)+1)
     J1 and J2 are identical to those of UF1, ϵ>0,yj=xj sin (6πx1+jπn),  j=2,3,,n
    UF7 f1=5x1+2| J1 |j J1 y2j,f2=15x1+2| J2 |j J2 y2j
     J1 and J2 are identical to those of UF1, ϵ>0,yj=xj sin (6πx1+jπn),  j=2,3,,n

     | Show Table
    DownLoad: CSV
    Table A6.  Tri-objective test problems (CEC2009).
    NameFunction
    UF8 f1= cos (0.5x1π) cos (0.5x2π)+2| J1 |j J1 (xj2x2 sin (2πx1+jπn)2)
    f2= cos (0.5x1π) sin (0.5x2π)+2| J2 |j J2 (xj2x2 sin (2πx1+jπn)2)
    f3= sin (0.5x1π)+2|J3|jJ3(xj2x2 sin (2πx1+jπn)2)
     J1 ={j|3jn, and j1 is a multiplication of 3}
     J2 ={j|3jn, and j2 is a multiplication of 3}
    J3={j|3jn, and j is a multiplication of 3}
     J1 ={j|j is odd and 2jn}, J2 ={j|j is even and 2jn}
    UF9 f1=0.5[max{0,(1+ϵ)(14(2x11)2)}+2x1]x2+2| J1 |j J1 (xj2x2 sin (2πx1+jπn)2)
    f2=0.5[max{0,(1+ϵ)(14(2x11)2)}+2x1]x2+2| J2 |j J2 (xj2x2 sin (2πx1+jπn)2)
    f3=1x2+2|J3|jJ3(xj2x2 sin (2πx1+jπn)2)
     J1 ={j|3jn, and j1 is a multiplication of 3},
     J2 ={j|3jn, and j2 is a multiplication of 3},
    J3={j|3jn, and j is a multiplication of 3},  ϵ=0.1
    UF10 f1= cos (0.5x1π)cos(0.5x2π)+2| J1 |j J1 [4y2i cos (8πyi)+1]
    f2= cos (0.5x1π)sin (0.5x2π)+2| J2 |j J2 [4y2i cos (8πyi)+1]
    f3= sin (0.5x1π)+2|J3|jJ3[4y2j cos (8πyj)+1]
     J1 ={j|3jn, and j1 is a multiplication of 3},
     J2 ={j|3jn, and j2 is a multiplication of 3},
    J3={j|3jn, and j is a multiplication of 3}

     | Show Table
    DownLoad: CSV


    [1] Joy TJ, Foster GR, Renard KG (2002) Soil erosion: Processes, prediction, measurement and control. John Wiley Sons Inc
    [2] Devátý J, Dostál T, Hösl R, et al. (2019) Effects of historical land use and land pattern changes on soil erosion-Case studies from Lower Austria and Central Bohemia. Land Use Policy 82: 674–685. https://doi.org/10.1016/j.landusepol.2018.11.058 doi: 10.1016/j.landusepol.2018.11.058
    [3] Smith P, House JI, Bustamante M. et al. (2016) Global change pressures on soils from land use and management. Global Change Biol 22: 1008–1028. https://doi.org/10.1111/gcb.13068 doi: 10.1111/gcb.13068
    [4] Shojaei S, Kalantari Z, Rodrigo-Comino J (2020) Prediction of factors affecting activation of soil erosion by mathematical modeling at pedon scale under laboratory conditions. Sci Rep 10. https://doi.org/10.1038/s41598-020-76926-1 doi: 10.1038/s41598-020-76926-1
    [5] Borrelli P, Robinson DA, Panagos P, et al. (2020) Land use and climate change impacts on global soil erosion by water (2015–2070). PNAS https://doi.org/10.1073/pnas.2001403117 doi: 10.1073/pnas.2001403117
    [6] Nearing MA, Pruski FF, O'neal MR (2004) Expected climate change impacts on soil erosion rates: A review. J Soil Water Conserv 59: 43–50.
    [7] Congo-Rwanda DR, Karamage F, Shao H, et al. (2016) Deforestation Effects on Soil Erosion in the Lake Kivu Basin, Forests.
    [8] Wenger AS, Atkinson S, Santini T, et al. (2018) Predicting the impact of wlogging activitieson soil erosion and water quality in steep, forested tropical islands. Environ Res Lett 13. https://doi.org/10.1088/1748-9326/aab9eb doi: 10.1088/1748-9326/aab9eb
    [9] Zhao L, Hou R (2019) Human causes of soil loss in rural karst environments: a case study of Guizhou, China. Sci Rep 9: 3225. https://doi.org/10.1038/s41598-018-35808-3 doi: 10.1038/s41598-018-35808-3
    [10] Lal R (2001) Soil degradation by erosion. Land Degrad Dev 12: 519–539. https://doi.org/10.1002/ldr.472 doi: 10.1002/ldr.472
    [11] Sthiannopkao S, Takizawa S, Wirojanagud W (2006) Effects of soil erosion on water quality and water uses in the upper Phong watershed. Water Sci Technol https://doi.org/10.2166/wst.2006.037 doi: 10.2166/wst.2006.037
    [12] Acharya AK, Kafle N (2009) Land degradation issues in Nepal and its management through agroforestry. J Agric Environ 10: 133–143. https://doi.org/10.3126/aej.v10i0.2138 doi: 10.3126/aej.v10i0.2138
    [13] Issaka S, Ashraf MA (2017) Impact of soil erosion and degradation on water quality: a review. Geol Ecol Landsc https://doi.org/10.1080/24749508.2017.1301053 doi: 10.1080/24749508.2017.1301053
    [14] Chalise D, Kumar L, Kristiansen P (2019) Land degradation by soil erosion in Nepal: A Review. Soil Systems 3: 12. https://doi.org/10.3390/soilsystems3010012 doi: 10.3390/soilsystems3010012
    [15] Camara M, Jamil NR, Abdullah AFB (2019) Impact of land uses on water quality in Malaysia: a review. Ecol Process https://doi.org/10.1186/s13717-019-0164-x doi: 10.1186/s13717-019-0164-x
    [16] Wischmeier WH, Smith DD (1965) Prediction Rainfall Erosion Losses from Cropland East of the Rocky Mountains: A Guide for Selection of Practices for Soil and Water Conservation. Agr Handb 282.
    [17] Kenneth GR, George RF, Glenn AW, Jeffrey PP (1991) RUSLE: Revised universal soil loss equation. J Soil Water Conserv 46: 30–33.
    [18] Williams JR (1975) Sediment-yield prediction with Universal Equation using runoff energy factor. In: Present and Prospective Technology for Predicting Sediment Yield and Sources. US Dept Agrie 244–252.
    [19] Gavrilovic Z (1988) The use of empirical method (erosion potential method) for calculating sediment production and transportation in unstudied or torrential streams (Editor White W.R. In: International Conference on River Regime) John Wiley Sons 411–422.
    [20] Emmanouloudis DA, Christou OP, Filippidis E (2003) Quantitative estimation of degradation in the Alikamon river basin using GIS. Erosion Prediction in Ungauged Basins: Integrating Methods and Techniques (Proceedings of symposium HS01 held during IUGG2003 at Sapporo). IAHS Publ. no. 279.
    [21] Haghizadeh A, Teang L, Godarzi E (2009) Forecasting Sediment with Erosion Potential Method with Emphasis on Land Use Changes at Basin. Electronic J Geotech Engn 14.
    [22] Tazioli A (2009) Evaluation of erosion in equipped basins, preliminary results of a comparison between the Gavrilovic model and direct measurements of sediment transport. Environ Geol 56: 825–831. https://doi.org/10.1007/s00254-007-1183-y doi: 10.1007/s00254-007-1183-y
    [23] Milanesi L, Pilotti M, Clerici A (2014) The Application of the Erosion Potential Method to Alpine Areas: Methodological Improvements and Test Case. Engin Geolr Soc Terr https://doi.org/10.1007/978-3-319-09054-2_73 doi: 10.1007/978-3-319-09054-2_73
    [24] Milanesi L, Pilotti M, Clerici A, et al. (2015) Application of an improved version of the erosion potential method in alpine areas. Ital J Engn Geol Enviro 1.
    [25] Lense G, Parreiras T, Moreira R, et al (2019) Estimates of soil losses by the erosion potential method in tropical latosols. Agri Sci https://doi.org/10.1590/1413-7054201943012719 doi: 10.1590/1413-7054201943012719
    [26] Marko O, Gjipalaj J, Shkodrani N (2022) Application of the Erosion Potential Method in Vithkuqi Watersheds (Southeastern Albania). J Ecol Eng 23: 17–24. https://doi.org/10.12911/22998993/146131 doi: 10.12911/22998993/146131
    [27] Blinkov I, Kostadinov S (2010) Applicability of various erosion risk assessment methods for engineering purposes, BALWOIS conference, Ohrid, Macedonia.
    [28] Blinkov I, Kostadinov S, Marinov I (2013) Comparison of erosion and erosion control works in Macedonia, Serbia and Bulgaria. Int Soil Water Conserv Res https://doi.org/10.1016/S2095-6339(15)30027-7 doi: 10.1016/S2095-6339(15)30027-7
    [29] Vujacic D, Barovic G, Tanaskovikj V, et al. (2015) Calculation of runoff and sediment yield in the Pisevska Rijeka Watershed, Polimlje, Montenegro. Agric For 61: 225–234. https://doi.org/10.17707/AgricultForest.61.2.20 doi: 10.17707/AgricultForest.61.2.20
    [30] Spalevic V, Barovic G, Mitrovic M, et al. (2015) Assessment of sediment yield using the Erosion Potential Method (EPM) in the Karlicica watershed of Montenegro. Conference Paper.
    [31] Vujacic D, Spalevic V (2016) Assessment of Runoff and Soil Erosion in the Radulicka Rijeka Watershed, Polimlje, Montenegro. Agric For 62: 283–292. https://doi.org/10.17707/AgricultForest.62.2.25 doi: 10.17707/AgricultForest.62.2.25
    [32] Maliqi E, Sing SK (2019) Quantitative Estimation of Soil Erosion Using Open-Access Earth Observation Data Sets and Erosion Potential Model. Water Conserv Sci Eng 4: 187–200. https://doi.org/10.1007/s41101-019-00078-1 doi: 10.1007/s41101-019-00078-1
    [33] Gocic M, Dragicevic S, Radivojevic A, et al. (2020) Changes in Soil Erosion Intensity Caused by Land Use and Demographic Changes in the Jablanica River Basin, Serbia. Agriculture 10: 345. https://doi.org/10.3390/agriculture10080345 doi: 10.3390/agriculture10080345
    [34] Marko O, Lako A, Çobani E (2011) Evaluation of soil erosion in the area of Kallmet Lezha District. Geotech SP 1474–1482. https://doi.org/10.1061/41165(397)151 doi: 10.1061/41165(397)151
    [35] Schiemer F, Drescher A, Hauer C, et al. (2018) The Vjosa River corridor: a riverine ecosystem of Europe significance. Acta ZooBot Austria 155: 1–40.
    [36] Zemljic M (1971) Calcul du debit solide - Evaluation de la vegetation comme un des facteursantierosifs. In: International Symposium Interpraevent, Villach, Austria.
    [37] Dragičević N, Karleuša B, Ožanić N (2017) Erosion Potential Method (Gavrilović Method) Sensitivity Analysis. Soil Water Res https://doi.org/10.17221/27/2016-SWR doi: 10.17221/27/2016-SWR
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1995) PDF downloads(178) Cited by(2)

Figures and Tables

Figures(8)  /  Tables(4)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog