
In this paper, we mainly study the equivalence and computing between Nash equilibria and the solutions to the system of equations. First, we establish a new equivalence theorem between Nash equilibria of n-person noncooperative games and solutions of algebraic equations with parameters, that is, finding a Nash equilibrium point of the game is equivalent to solving a solution of the system of equations, which broadens the methods of finding Nash equilibria and builds a connection between these two types of problems. Second, an adaptive differential evolution algorithm based on cultural algorithm (ADECA) is proposed to compute the system of equations. The ADECA algorithm applies differential evolution (DE) algorithm to the population space of cultural algorithm (CA), and increases the efficiency by adaptively improving the mutation factor and crossover operator of the DE algorithm and applying new mutation operation. Then, the convergence of the ADECA algorithm is proved by using the finite state Markov chain. Finally, the new equivalence of solving Nash equilibria and the practicability and effectiveness of the algorithm proposed in this paper are verified by computing three classic games.
Citation: Huimin Li, Shuwen Xiang, Shunyou Xia, Shiguo Huang. Finding the Nash equilibria of n-person noncooperative games via solving the system of equations[J]. AIMS Mathematics, 2023, 8(6): 13984-14007. doi: 10.3934/math.2023715
[1] | Chengqing Pan, Haishu Lu . On the existence of solutions for systems of generalized vector quasi-variational equilibrium problems in abstract convex spaces with applications. AIMS Mathematics, 2024, 9(11): 29942-29973. doi: 10.3934/math.20241447 |
[2] | Huimin Li, Shuwen Xiang, Yanlong Yang, Chenwei Liu . Differential evolution particle swarm optimization algorithm based on good point set for computing Nash equilibrium of finite noncooperative game. AIMS Mathematics, 2021, 6(2): 1309-1323. doi: 10.3934/math.2021081 |
[3] | Radu Precup, Andrei Stan . Stationary Kirchhoff equations and systems with reaction terms. AIMS Mathematics, 2022, 7(8): 15258-15281. doi: 10.3934/math.2022836 |
[4] | Chenwei Liu, Shuwen Xiang, Yanlong Yang . Existence and essential stability of Nash equilibria for biform games with Shapley allocation functions. AIMS Mathematics, 2022, 7(5): 7706-7719. doi: 10.3934/math.2022432 |
[5] | Jaicer López-Rivero, Hugo Cruz-Suárez, Carlos Camilo-Garay . Nash equilibria in risk-sensitive Markov stopping games under communication conditions. AIMS Mathematics, 2024, 9(9): 23997-24017. doi: 10.3934/math.20241167 |
[6] | Jun Wang, Dan Wang, Yuan Yuan . Research on low-carbon closed-loop supply chain strategy based on differential games-dynamic optimization analysis of new and remanufactured products. AIMS Mathematics, 2024, 9(11): 32076-32101. doi: 10.3934/math.20241540 |
[7] | Haishu Lu, Kai Zhang, Rong Li . Collectively fixed point theorems in noncompact abstract convex spaces with applications. AIMS Mathematics, 2021, 6(11): 12422-12459. doi: 10.3934/math.2021718 |
[8] | Zifu Fan, Kexin Fan, Wei Zhang . Research on the data integration strategy of local governments and enterprises under central government subsidies. AIMS Mathematics, 2022, 7(6): 10143-10164. doi: 10.3934/math.2022564 |
[9] | Jewaidu Rilwan, Poom Kumam, Idris Ahmed . Overtaking optimality in a discrete-time advertising game. AIMS Mathematics, 2022, 7(1): 552-568. doi: 10.3934/math.2022035 |
[10] | Zifu Fan, Youpeng Tao, Wei Zhang, Kexin Fan, Jiaojiao Cheng . Research on open and shared data from government-enterprise cooperation based on a stochastic differential game. AIMS Mathematics, 2023, 8(2): 4726-4752. doi: 10.3934/math.2023234 |
In this paper, we mainly study the equivalence and computing between Nash equilibria and the solutions to the system of equations. First, we establish a new equivalence theorem between Nash equilibria of n-person noncooperative games and solutions of algebraic equations with parameters, that is, finding a Nash equilibrium point of the game is equivalent to solving a solution of the system of equations, which broadens the methods of finding Nash equilibria and builds a connection between these two types of problems. Second, an adaptive differential evolution algorithm based on cultural algorithm (ADECA) is proposed to compute the system of equations. The ADECA algorithm applies differential evolution (DE) algorithm to the population space of cultural algorithm (CA), and increases the efficiency by adaptively improving the mutation factor and crossover operator of the DE algorithm and applying new mutation operation. Then, the convergence of the ADECA algorithm is proved by using the finite state Markov chain. Finally, the new equivalence of solving Nash equilibria and the practicability and effectiveness of the algorithm proposed in this paper are verified by computing three classic games.
Game theory is a mathematical theory that studies interactions between decision makers. The n-person noncooperative game is an important part of game theory [1,2,3]. As we all know, Von Neumann [4] and Nash [5,6] established the models of matrix games and n-person finite noncooperative games respectively, and Nash proposed the core equilibrium concept in noncooperative games, called Nash equilibrium, and proved that this kind game has at least one mixed strategies equilibrium point. At present, the n-person noncooperative game and its related researches have been widely used in mathematics, economics, operations research, biology, textile industry and other fields [7,8,9,10,11]. In these researches and applications, the key point is how to find Nash equilibrium points effectively, which largely depends on the good mathematical description of the problem. Therefore, it is very important to find a suitable mathematical model when solving games.
At present, there are many mathematical methods to solve the Nash equilibria, mainly by equating finding Nash equilibria to a fixed point problem, an optimization problem or a tensor complementarity problem. The finding Nash equilibria is equivalent to the fixed point problem, which is mainly handled by the Brouwer fixed point theorem, the Kakutani fixed point theorem and the KyFan point theorem. Many scholars have made a lot of researches in this field [12,13,14,15]. Furthermore, the research that finding Nash equilibria is equivalent to the optimization problem is springing up. Because the equilibrium points are natural actions of self interested behavior in competitive situations. In other words, at an equilibrium point, each competitor is maximizing his payoff against all other competitive actions in the situation. Therefore, finding a Nash equilibrium is generally as a nonconvex optimization problem, which has been sufficiently studied in [16,17,18,19,20]. Further, some scholars also have equated the finding Nash equilibria to the tensor complementarity problem [21,22,23,24], and the player's payoff function is given by a homogeneous polynomial defined by this player's payoff tensor. In addition to the above usual methods for finding Nash equilibria, in recent years, with the extensive research of multi-agent systems in social science and engineering systems, using distributed algorithms seeking Nash equilibria has become an emerging research topic. This method provides solutions for complex engineering games[25,26,27,28,29,30,31], such as generalized games with coupled constraints, aggregative games with non-smooth objective functions and uncertain games, etc.
Although the above research methods and calculation methods have solved a large number of games. With the complexity of real problems and the diversification of games, the above methods have some disadvantages in the computing process, such as high computational complexity, cumbersome derivation process and complex algorithm design. Therefore, we propose a new equivalent method of finding Nash equilibria in this paper, which directly applies the definition of Nash equilibrium to establish an equivalent relationship with the system of equations, and transform the problem of finding the Nash equilibria into solving the system of equations. So that it can be calculated using the developed method of solving the system of equations. Therefore, compared with finding the Nash equilibria equivalent to the fixed point problem and nonconvex optimization problem, the proposed method does not need to define complex set valued mapping, no have tedious theoretical derivation and proof. Compared with find the Nash equilibria equivalent to the tensor complementarity problem, the proposed method does not need to consider the nonlinear complementarity and nonsmooth of the payoff function. This method provides a new idea for find the Nash equilibria, which can be easily understood and applied by readers with weak mathematical foundation. On the one hand, this method broadens the expression form of Nash equilibria and provides a new method for finding Nash equilibria. It reduces the difficulty of seeking Nash equilibria, and only needs to consider the calculation methods of the equivalent equations. On the other hand, this equivalent method also has certain significance for the realization of Nash equilibrium: since the equations are derived from the definition of Nash equilibrium, each equation essentially represents the rational behavior of the player. In each equation, the probability of each component of the Nash equilibrium is greater than 0 is meaning that the corresponding pure strategy is in the direction of increasing the profit, that is, the direction of Nash equilibrium realization. Therefore, this method has certain research significance.
This paper is organized as follows: In Section 2, we give the basic definitions of the game and an equivalence theorem between the Nash equilibria of n-person noncooperative game and the solutions of equations. In Section 3, we propose the ADECA algorithm and prove its convergence theoretically through the finite state Markov chain. This algorithm applies the DE algorithm to the population space of the CA algorithm, and uses the belief space of the CA to extract hidden evolutionary knowledge and guides the evolution of the population space. Additionally, the DE algorithm is adaptively improved to enhance the performance of the ADECA algorithm. The equivalence theorem of the Nash equilibria and the superiority of the ADECA algorithm proposed in this paper are verified by examples in Section 4.
Definition 2.1. [32] We consider a n-person noncooperative game Γ={(N,Si,Pi,Xi,Ui),i=1,…,n}
where
(1) N={1,…,n} is the set of players and n is the number of players;
(2) Si={si1,…,simi},∀i∈N is the pure strategy set of player i, mi represents the number of strategies available to player i, S=∏ni=1Si;
(3) Pi:S→R,∀i∈N represents the payoff function of player i;
(4) Xi={xi=(xi1,…,xiK,…,ximi):xiK≥0,∑miK=1xiK=1}, ∀i∈N is the set of mixed strategies. X=∏ni=1Xi, and each mixed strategy profile meets x=(x1,x2,…,xn)∈X;
(5) Ui:X→R,∀i∈N represents the expected payoff function of player i.
ui(x)=m1∑K1=1…mn∑Kn=1pi(s1K1,…,snKn)n∏i=1xiKi, |
where ui(x) represents the expected payoff value of player i when he chooses a mixed strategy xi=(xi1,…,ximi)∈Xi. pi(siKi,…,snKn) represents the payoff value of player i when each player chooses pure strategy siKi∈Si,i=1,…,n.
Denote by
ui(x‖siKi)≜ui(x1,…,xi−1,siKi,xi+1,…,xn)≜m1∑K1=1…mi−1∑Ki−1=1mi+1∑Ki+1=1…mn∑Kn=1pi(s1K1,…,snKn)x1K1…xi−1Ki−1xi+1Ki+1…xnKn, |
where siKi(1≤Ki≤mi) is pure strategy of player i. (x‖siKi) represents siKi instead of xi of player i, and the other players do not change their own mixed strategy.
Definition 2.2. If x∗=(x∗1,…,x∗n), such that ui(x∗i,x∗i∧)=maxyi∈Xiui(yi,x∗i∧),∀i∈N, then x∗ is a Nash equilibrium point of n-person noncooperative game, where i∧=N∖{i}, ∀i∈N.
Conclusion 2.1. [20] Mixed strategy x∗=(x∗1,…,x∗n)∈X is a Nash equilibrium point of a game Γ if and only if ui(x∗)≥ui(x∗‖siKi), where, Ki=m1,…,mn,i=1,…,n.
Theorem 2.1. For any n-person finite noncooperative game, the mixed strategy profile,
x=(x1,x2,…,xn)∈X,xi=(xi1,…,xiEi,…,ximi),∀i∈N |
is a Nash equilibrium if and only if there is a set of real numbers
λiEiKi≥0,i=1,2,…,n,Ei≠Ki,Ei,Ki∈{1,2,…,mi} |
such that λiEiKi and x=(x1,x2,…,xn) satisfy the following equations:
λ1E1K1+x1E1(m2∑E2=1…mn∑En=1p1(s1K1,s2E2,…,snEn)⋅x2E2⋅…⋅xnEn−m2∑E2=1…mn∑En=1p1(s1E1,s2E2,…,snEn)⋅x2E2⋅…⋅xnEn)=0λ1E1K1≥0,E1≠K1,E1,K1∈{1,2,…,m1}⋮λiEiKi+xiEi(m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siKi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn−m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn)=0λiEiKi≥0,Ei≠Ki,Ei,Ki∈{1,2,…,mi}⋮λnEnKn+xnEn(m1∑E1=1…mn−1∑En−1=1pn(s1E1,…,sn−1En−1,snKn)⋅x1E1⋅…⋅xn−1En−1−m1∑E1=1…mn−1∑En−1=1pn(s1E1,…,sn−1En−1,snEn)⋅x1E1⋅…⋅xn−1En−1)=0λnEnKn≥0,En≠Kn,En,Kn∈{1,2,…,mn}. | (2.1) |
Proof. Necessity: Assume that x=(x1,x2,…,xn)∈X,xi=(xi1,…,xiEi,…,ximi),∀i∈N is a Nash equilibrium. Then the expected payoff of player i under this equilibrium is
ui(x)=m1∑E1=1…mn∑En=1pi(s1E1,s2E2,…,snEn)⋅x1E1⋅x2E2⋅…⋅xnEn=mi∑Ei=1xiEi[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]=xi1[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,si1,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]+xi2[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,si2,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]+mi∑Ei=3xiEi[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn], | (2.2) |
if xiEh1,…,xiEhl>0,xiEhl+1,…,xiEhmi=0, then for each Ei∈{Eh1,…,Ehl}, Ki∈{1,…,mi}, and Ki≠Ei, we have
m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn≥m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siKi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn. | (2.3) |
(Contradiction): Assuming that the above formula (2.3) does not hold, then there is a certain Ei∈{Eh1,…,Ehl}, Ki≠Ei, where Ki∈{1,…,mi}, such that
m1∑E1=1…mi−1∑Ei−1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn<m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siKi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn. |
For the convenience of writing, let Ei=1, Ki=2, that is,
m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,si1,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn<m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,si2,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn. | (2.4) |
Taking y=(x1,…,xi−1,yi,xi+1,…,xn),yi=(0,xi1+xi2,xi3,…,xiEi,…,ximi), and combining inequality (2.4) with the formula (2.2), we obtain
ui(y)=m1∑E1=1…mn∑En=1pi(s1E1,s2E2,…,snEn)⋅y1E1⋅y2E2⋅…⋅ynEn=mi∑Ei=1yiEi[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]=(xi1+xi2)[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,si2,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]+mi∑Ei=3xiEi[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]=xi1[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,si2,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]+xi2[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,si2,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn] |
+mi∑Ei=3xiEi[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]>xi1[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,si1,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]+xi2[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,si2,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]+mi∑Ei=3xiEi[m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn]=ui(x), |
it contradicts that x is a Nash equilibrium. So we get that for each Nash equilibrium xiEi>0, and ∀Ki∈{1,…,mi},
m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn≥m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siKi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn. |
In particular, if Ei,Ki∈{Eh1,…,Ehl}, xiEi,xiKi>0, then
m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn=m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siKi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn. |
The following proof shows that the Nash equilibrium x=(x1,x2,…,xn)∈X satisfies the system of equations in Theorem 2.1.
(ⅰ) For every i∈N, if xiEi=0, then for each Ki∈{1,…,mi},Ki≠Ei, and take λiEiKi=0, we have
λiEiKi+xiEi(m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siKi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn−m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn)=0. | (2.5) |
(ⅱ) If xiEi>0, then for each Ki∈{1,…,mi},Ki≠Ei, that is formula (2.3),
m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn≥m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siKi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn. | (2.6) |
(ⅱ-1) If the equality sign of formula (2.6) holds, take λiEiKi=0, then formula (2.5) holds.
(ⅱ-2) If the equality sign of formula (2.6) does not hold, that is
m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn>m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siKi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn, |
then take
λiEiKi=xiEi(m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siEi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn−m1∑E1=1…mi−1∑Ei−1=1mi+1∑Ei+1=1…mn∑En=1pi(s1E1,…,siKi,…,snEn)⋅x1E1⋅…⋅xi−1Ei−1⋅xi+1Ei+1⋅…⋅xnEn), |
therefore
\begin{equation*} \begin{aligned} \lambda_{\mathscr{\mathfrak{E}}_{i} \mathscr{K}_{i}}^{i}+& x_{\mathscr{\mathfrak{E}}_{i}}^{i}\left(\sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{K}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\right.\\ &\left.-\sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{\mathfrak{E}}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) = 0. \end{aligned} \end{equation*} |
The necessity has been proved. Next, we show the sufficiency of Theorem 2.1.
Sufficiency: If there is a set of real numbers \lambda_{\mathscr{\mathfrak{E}}_{i} \mathscr{K}_{i}}^{i} \geq 0, i = 1, 2, \ldots, n, \mathscr{\mathfrak{E}}_{i} \neq \mathscr{K}_{i}, \mathscr{\mathfrak{E}}_{i}, \mathscr{K}_{i} \in\left\{1, 2, \ldots, m_{i}\right\} and mixed strategy profile x = \left(x^{1}, x^{2}, \ldots, x^{n}\right)\in X, x^{i} = \left(x_{1}^{i}, \ldots, x_{\mathscr{\mathfrak{E}}_{i}}^{i}, \ldots, x_{m_{i}}^{i}\right), i \in N , satisfying the system of equations in Theorem 2.1, The following proof shows that x is a Nash equilibrium. Let x^{i} = \left(x_{1}^{i}, \ldots, x_{\mathscr{\mathfrak{E}}_{i_k}}^{i}, x_{\mathscr{\mathfrak{E}}_{i_{k+1}}}^{i}, \ldots, x_{m_{i}}^{i}\right) , where x_{1}^{i}, \ldots, x_{\mathscr{\mathfrak{E}}_{i_k}}^{i} > 0, x_{\mathscr{\mathfrak{E}}_{i_{k+1}}}^{i}, \ldots, x_{m_{i}}^{i} = 0 . Since x satisfies the system of equations in Theorem 2.1, then for any \mathscr{\mathfrak{E}}_i\in\{1, \ldots, \mathscr{\mathfrak{E}}_{i_k}\} , we get
\begin{equation*} \begin{aligned} \lambda_{\mathscr{\mathfrak{E}}_{i} \mathscr{K}_{i}}^{i} &+x_{\mathscr{\mathfrak{E}}_{i}}^{i}\left(\sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{K}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\right.\\ &\left.-\sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{\mathfrak{E}}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) = 0. \end{aligned} \end{equation*} |
Because x_{1}^{i}, \ldots, x_{\mathscr{\mathfrak{E}}_{i_{k}}}^{i} > 0 , for \forall \mathscr{K}_i\in\{1, 2, \ldots, m_i\} , we have
\begin{equation*} \begin{aligned} \sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{\mathfrak{E}}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\\ \geq \sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{K}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}, \end{aligned} \end{equation*} |
and because for any player i , the expected payoff for pure strategy is
\begin{equation*} u_{i}\left(x \| s_{\mathscr{K}_{i}}^{i}\right) = \sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{K}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}, \end{equation*} |
therefore
\begin{equation*} \begin{aligned} u_{i}(x) & = \sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{\mathfrak{E}}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot x_{\mathscr{\mathfrak{E}}_{2}}^{2} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n} \\ & = \sum\limits_{\mathscr{\mathfrak{E}}_{i} = 1}^{m_{i}} x_{\mathscr{\mathfrak{E}}_{i}}^{i}\left[\sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{\mathfrak{E}}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\right] \\ & = \sum\limits_{\mathscr{\mathfrak{E}}_{i} = 1}^{\mathscr{\mathfrak{E}}_{i_{k}}} x_{\mathscr{\mathfrak{E}}_{i}}^{i}\left[\sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{\mathfrak{E}}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\right]\\ &\geq \sum\limits_{\mathscr{\mathfrak{E}}_{i} = 1}^{\mathscr{\mathfrak{E}}_{i_k}} x_{\mathscr{\mathfrak{E}}_{i}}^{i}\left[\sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{K}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\right]\\ & = \left[\sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{K}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\right] \cdot \sum\limits_{\mathscr{\mathfrak{E}}_{i} = 1}^{\mathscr{\mathfrak{E}}_{i_{k}}} x_{\mathscr{\mathfrak{E}}_i}^{i}\\ & = \sum\limits_{\mathscr{\mathfrak{E}}_{1} = 1}^{m_{1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{i-1} = 1}^{m_{i-1}} \sum\limits_{\mathscr{\mathfrak{E}}_{i+1} = 1}^{m_{i+1}} \ldots \sum\limits_{\mathscr{\mathfrak{E}}_{n} = 1}^{m_{n}} p_{i}\left(s_{\mathscr{\mathfrak{E}}_{1}}^{1}, \ldots, s_{\mathscr{K}_{i}}^{i}, \ldots, s_{\mathscr{\mathfrak{E}}_{n}}^{n}\right) \cdot x_{\mathscr{\mathfrak{E}}_{1}}^{1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{i-1}}^{i-1} \cdot x_{\mathscr{\mathfrak{E}}_{i+1}}^{i+1} \cdot \ldots \cdot x_{\mathscr{\mathfrak{E}}_{n}}^{n}\\ & = u_{i}\left(x \| s_{\mathscr{K}_{i}}^{i}\right). \end{aligned} \end{equation*} |
Judging by Conclusion 2.1 that x = (x^1, x^2, \ldots, x^n) is a Nash equilibrium.
According to the preparatory knowledge of n -person finite noncooperative game and Theorem 2.1, we can get the equivalence theorem of the bimatrix game.
Corollary 2.1. Suppose the mixed strategy of the two-player finite noncooperative game is x = (x^1, x^2) , where x^1 = (x^1_1, x^1_2, \ldots, x^1_{m_1}), x^1 = (x^2_1, x^2_2, \ldots, x^2_{m_2}) , and \sum_{i = 1}^{m_1}x^1_{i} = 1 , \sum_{i = 1}^{m_2}x^2_{i} = 1 . The payoff matrices are A and B respectively:
A = \begin{bmatrix} a_{11} \; & \; \ldots \; & \; a_{1m_2} \\ \vdots \; & \; \vdots \; & \; \vdots \\ a_{m_11} \; & \; \ldots \; & \; a_{m_1m_2} \end{bmatrix}, \; \; \; B = \begin{bmatrix} b_{11} \; & \; \ldots \; & \; b_{1m_2} \\ \vdots \; & \; \vdots \; & \; \vdots \\ b_{m_11} \; & \; \ldots \; & \; b_{m_1m_2} \end{bmatrix}, |
then, x is a Nash equilibrium if and only if there is a set of real numbers \lambda_{\mathscr{\mathfrak{E}}_i \mathscr{K}_i}^{i} \geq 0, i = 1, 2, \mathscr{\mathfrak{E}}_{i} \neq \mathscr{K}_{i}, \mathscr{\mathfrak{E}}_{i}, \mathscr{K}_{i} \in\left\{1, 2, \ldots, m_{i}\right\} , such that \lambda_{\mathscr{\mathfrak{E}}_i \mathscr{K}_i}^{i} and x = (x^{1}, x^{2}) satisfy the following equations:
\begin{equation} \left\{ \begin{aligned} &\lambda_{\mathscr{K}_1\mathscr{K}_2}^1+x_{\mathscr{K}_1}^1(A_{\mathscr{K}_2}-A_{\mathscr{K}_1})[x^2]^{\prime} = 0, \\ &\lambda_{\mathscr{\mathfrak{E}}_1\mathscr{\mathfrak{E}}_2}^2+x_{\mathscr{\mathfrak{E}}_1}^2x^1(B_{\mathscr{\mathfrak{E}}_2}-B_{\mathscr{\mathfrak{E}}_1}) = 0, \end{aligned} \right. \end{equation} | (2.7) |
where A_{\mathscr{K}_t}(t = 1, 2) is the \mathscr{K}_t th row of matrix A , B_{\mathscr{\mathfrak{E}}_t}(t = 1, 2) is the \mathscr{\mathfrak{E}}_t th column of matrix B , x^1 = (x^1_1, \ldots, x^1_{m_1}), x^2 = (x^2_1, \ldots, x^2_{m_2}) , and \mathscr{K}_1\neq \mathscr{K}_2 \in\{1, \ldots, m_1\}, \mathscr{\mathfrak{E}}_1\neq \mathfrak{E}_2 \in\{1, \ldots, m_2\} .
For the convenience of writing, formula (2.1) can be simplified as:
\begin{equation} \left\{ \begin{aligned} &f_1(\lambda_1, x_1, \ldots, x_\mathscr{K}) = 0, \\ &f_2(\lambda_2, x_1, \ldots, x_\mathscr{K}) = 0, \\ &\vdots\\ &f_m(\lambda_m, x_1, \ldots, x_\mathscr{K}) = 0, \end{aligned} \right. \end{equation} | (2.8) |
where m = \sum_{i = 1}^nm_i(m_i-1) , \mathscr{K} = \sum_{i = 1}^nm_i , \lambda_i\geq 0, i = 1, 2, \ldots, m , and 0\leq x_i\leq1, i = 1, 2, \ldots, \mathscr{K} , \sum_{i = 1}^{m_1}x_i = 1, \sum_{i = m_1+1}^{m_1+m_2}x_i = 1, \ldots, \sum_{i = m_{n-1}+1}^{m_{n-1}+m_n}x_i = 1 .
According to the equivalent methods of solving the equations, the solution of the Eq (2.8) is usually equivalent to the following optimization problem, and the optimal value is 0:
\begin{equation} min \; F(\lambda_1, \ldots, \lambda_m, x_1, \ldots, x_\mathscr{K}) = \sum\limits_{i = 1}^m|f_i(\lambda_i, x_1, \ldots, x_\mathscr{K})|. \end{equation} | (2.9) |
An algorithm for finding the Nash equilibria based on the system of equations is given below.
In 1994, Reynolds [33] simulated the evolution of human society and proposed a bionic intelligent computing method with a bilevel evolutionary mechanism, cultural algorithm (CA). This algorithm is composed of independently evolved population space and belief space. The population space simulates the evolution process of individuals according to certain behavioral criteria from the microscopic perspective, while the belief space simulates the evolution process of culture formation, transmission and comparison from the macroscopic perspective. Population space and belief space are two relatively independent evolutionary processes, and the two spaces are connected according to special protocols (acceptance function and influence function). Those interactions are depicted in Figure 1.
The population space is used to implement any population evolutionary algorithm. On the one hand, it evaluates the fitness value of individuals, and implements evolutionary operations such as selection, crossover and mutation for population; On the other hand, excellent individuals are provided to the belief space as samples. The belief space selects the sample individuals from the evaluated populations in the population space through the acceptance function. Under the function of knowledge updating, the hidden information carried by the sample individuals is processed and summarized, described and stored in the form of knowledge. Finally, all kinds of knowledge react on the population space through the influence function to guide the evolution of the population space, so as to accelerate the evolutionary convergence and improve the adaptability of the algorithm. The details on the interaction and influence between belief space and population space will be given below. (The following discussions are based on the assumption of solving the minimum optimization problem).
The CA algorithm provides a computational model of a multi evolutionary processes. Therefore, from the perspective of computational model, an evolutionary algorithm that meets the requirements of CA algorithm can be regarded as an evolutionary process of population space. In this paper, the proposed algorithm uses DE algorithm with simple operation in the population space of CA algorithm.
The DE algorithm is a new simple and robust evolutionary algorithm. It is first introduced by Storm and Price [34]. There are three main operations: mutation, crossover and selection operation.
(1) Mutation
The mutation operation is mainly executed to distinguish DE from other evolutionary algorithms. The mutation individual V = (v_{i1}, \ldots, v_{ij}, \ldots, v_{iD}), i = 1, \ldots, N, j = 1, \ldots, D is generated by the following equation:
\begin{equation} v_{i}^{t+1} = x_{r1}^{t}+F\cdot(x_{r2}^{t}-x_{r3}^{t}), \end{equation} | (3.1) |
where N and D denote population size and space dimension respectively, r1 , r2 and r3 are randomly generated integers within [1, N] , and r1\neq r2\neq r3\neq i , F is a constriction factor to control the size of difference of two individuals and t is the current generation.
(2) Crossover
We use the crossover between the parent and offspring with the given probability for generating new individual U = (u_{i1}, \ldots, u_{ij}, \ldots, u_{iD}) :
\begin{equation} u_{ij}^{t+1} = \left\{ \begin{aligned} v_{ij}^{t+1} & , & if\; (rand(j)\leq CR)\; or\; (j = rnbr(i)), \\ x_{ij} & , & otherwise, \end{aligned} \right. \end{equation} | (3.2) |
where rand(j)\in [0, 1] is random values, CR\in [0, 1] is crossover operator, rnbr(i) is a randomly selected integer on [1, D] , which ensure that a new individual gets at least one component value from the mutation vector.
(3) Selection
The offspring X_{ij}^{t+1} is generated by selecting the individual and parent according to the following formula:
\begin{equation} X_{ij}^{t+1} = \left\{ \begin{aligned} U_{ij}^{t+1} & , & if\; (f(U_{ij}^{t+1}) < f(U_{ij}^{t})), \\ X_{ij}^{t} & , & otherwise, \end{aligned} \right. \end{equation} | (3.3) |
where f(\cdot) is the fitness function.
Although the DE algorithm is widely used in optimization problems, with the increase of the complexity of solving problems, the DE algorithm also has some disadvantages, such as slow convergence, low accuracy and weak stability. Therefore, in order to better apply to CA algorithm, the DE algorithm is improved.
Many scholars have verified that the performance of the DE algorithm mainly depends on the selection of mutation operation and related parameters. Furthermore, the mutation factor F and crossover operator CR directly affect the searching capability and solving efficiency of the DE algorithm [35]. According to formula (3.1), the mutation factor F is used to control the influence of the difference vector on the mutation individual v_i . When F is large, the difference vector (x_{r2}-x_{r3}) has a great influence on v_i and causes a large disturbance, which is beneficial for maintaining population diversity and global search of the algorithm, but at the same time it will reduce the convergence speed; On the contrary, when F is small, the disturbance is small, that is, the diversity of the population is small, which is conducive to local search of the algorithm and makes the convergence speed faster. However, the algorithm is easy to fall into the local optimal and the "premature". According to formula (3.2), when CR is larger, the more contribution v_i makes to u_i , which is beneficial to expanding new spaces and accelerating the convergence of the algorithm. However, in the later period, the mutation individuals tend to be the same, which is detrimental for maintaining population diversity, making the algorithm prone to "premature" phenomenon; Conversely, CR is smaller, the more contribution x_i makes, which is beneficial for maintaining population diversity. However, in this way, the ability of the algorithm to explore new spaces is weakened, and the convergence speed is relatively slow.
In order to make the algorithm have better global search capability and convergence speed, adaptive mutation and crossover operator are adopted:
\begin{equation} \left\{ \begin{aligned} &\lambda = e^{1-\frac{T}{T+1-t}}, \\ &F = F_0\cdot 2^\lambda, \\ &CR = CR_0\cdot 2^\lambda, \end{aligned} \right. \end{equation} | (3.4) |
where t is the current iterate time, T is the maximum number of iteration, F_0 and CR_0 respectively are the initial mutation factor and crossover operator. The adaptive mutation factor and crossover operator are proposed, which can adaptively determine the mutation rate and crossover rate according to the search progress of the algorithm. According to formula (3.4), since t\in [1, T] , then \lambda\in [e^{1-T}, 1]\in (0, 1] , so, F = F_0\cdot 2^\lambda\in (F_0, 2F_0] , that is, 2F_0 is taken when t = 1 , which has a larger value to avoid the algorithm "premature"; With the progress of the algorithm, the value of F gradually decreases, until later close to F_0 , which retains good information, avoids the destruction of the optimal solution, and increases the possibility of searching for the global optimal solution. The CR = CR_0\cdot 2^\lambda\in (CR_0, 2CR_0] is similar: CR has a large value in the early stage, which increases the exploration ability of the algorithm and avoids the algorithm falling into local optimal. In the later stage, the crossover rate is gradually reduced to retain good individuals. Therefore, the probability of finding the global optimal solution is increased.
There are two main ways of traditional mutation operations in the DE algorithm [36]:
\begin{align*} &(i)\; \; DE/rand/1/bin: v_{i}^{t+1} = x_{r1}^{t}+F\cdot(x_{r2}^{t}-x_{r3}^{t}), \\ &(ii)\; \; DE/best/1/bin:v_{i}^{t+1} = x_{best}^{t}+F\cdot(x_{r2}^{t}-x_{r3}^{t}), \end{align*} |
where x_{best}^{t} represents the best individual in the current generation, that is, the optimal position searched by this individual so far. The first mutation method has a strong global search capability but slow convergence speed. The second method has a fast convergence speed, but it is easy to fall into local optimum values. In order to overcome these shortcomings, a new mutation operation is proposed,
\begin{equation} v_{i}^{t+1} = \lambda x_{r1}^{t}+(1-\lambda)x_{best}^{t}+F\cdot(x_{r2}^{t}-x_{r3}^{t}). \end{equation} | (3.5) |
In the population space of this paper, in order to jump out of the local optimum and accelerate the convergence of the algorithm. Formulas (3.4) and (3.5) will be used to improve the DE algorithm.
Belief space adopts the \langle S, N\rangle structure proposed in the literature [33]. Where S = \{s^t_1, s^t_2, \ldots, s^t_i, \ldots, s^t_m\} is the set of optimal individuals, s^t_i represents the i th optimal individual in the t th generation, and m represents the size of the set. N = \langle X_1, X_2, \ldots, X_j, \ldots, X_n\rangle is the normative knowledge, representing the feasible search area of the problems, where X_j = \langle I_j, L_j, U_j\rangle and n is the number of variables. I_j = [l_j, u_j] = \{x|\; l_j\leq x_j\leq u_j\} , l_j and u_j respectively are the upper and lower bound of variable x_j ; L_j represents the fitness value corresponding to the variable l_j ; U_j represents the fitness value of the variable u_j .
In order to simplify the operation process of the algorithm, only the current optimal individual s^t is used to update the S in belief space, and update it according to the following formula:
\begin{equation} \begin{aligned} s^{t} = \begin{cases} x_{best}^{t}, &\quad f(x_{best}^t) < f(s^t), \\ s^{t}, &\quad otherwise, \\ \end{cases} \end{aligned} \end{equation} | (3.6) |
where x_{best}^t is the optimal individual of the t th generation.
With the process of evolution, the search space gradually gathers in the dominant region. Therefore, when the individuals exceed the current search space, the normative knowledge N is updated according to the following formula:
\begin{equation} \begin{aligned} l^{t+1}_j = \begin{cases} x_{ij}^{t}, &\quad x_{ij}^t < l_{j}^{t} \; or\; f(x_{ij}^{t}) < L_j^t, \\ l_j^{t}, &\quad otherwise, \end{cases} \end{aligned} \end{equation} | (3.7) |
\begin{equation} \begin{aligned} L^{t+1}_j = \begin{cases} f(x_{j}^{t}), &\quad x_{ij}^t < l_{j}^{t} \; or\; f(x_{ij}^{t}) < L_j^t, \\ L_j^{t}, &\quad otherwise, \end{cases} \end{aligned} \end{equation} | (3.8) |
\begin{equation} \begin{aligned} u^{t+1}_j = \begin{cases} x_{ij}^{t}, &\quad x_{ij}^t > u_{j}^{t} \; or\; f(x_{ij}^{t}) < U_j^t, \\ u_j^{t}, &\quad otherwise, \end{cases} \end{aligned} \end{equation} | (3.9) |
\begin{equation} \begin{aligned} U^{t+1}_j = \begin{cases} f(x_{j}^{t}), &\quad x_{ij}^t > u_{j}^{t} \; or\; f(x_{ij}^{t}) < U_j^t, \\ U_j^{t}, &\quad otherwise. \end{cases} \end{aligned} \end{equation} | (3.10) |
Formulas (3.7) and (3.9) are the updating process of upper and lower bound of variable x_j , formulas (3.8) and (3.10) are the updating process of fitness value of corresponding variable x_j .
Acceptance function and influence function are the links between belief space and population space. The acceptance function selects better individuals from the population space and submits them to the belief space for knowledge updating, and the core of its research is to select better individuals. In the early stage of evolution, the better individuals have more valuable information, so more individuals can be selected for the belief space. In the later period of evolution, the algorithm gradually converges, and the dominant individuals and their implied information are similar. In order to maintain the diversity of knowledge, the number of accepting individuals should be reduced. Therefore, the dynamic acceptance function proposed by Saleem [37] is used, the number of accepted individuals decreases while the number of generations increases. We get the number of accepted individuals with the following expression:
\begin{equation} accept() = P\%\cdot N+(P\%\cdot N)/t, \end{equation} | (3.11) |
where N is the population size, t is the current iterate moment, in this paper, we take P\% = 20\% , and the number of an interval to execute the accept function is usually preset (according to the maximum number of iterations, we take the number of the interval as 5).
The main purpose of the influence function is to guide the population evolution by using knowledge of the belief space. According to the ADE model (Formula (3.5), (3.2), (3.3)), if the parent individuals are within the interval given by the normative knowledge, the particles will be updated according to the ADE algorithm. If it is outside the given interval, the offspring generation rule (formula (3.3)) becomes the following form:
\begin{equation} \begin{aligned} X_{ij}^{t+1} = \begin{cases} rand(0, 1)\cdot (u_j-l_j)+l_j, & if\; x^t_{ij} < l_j \; or\; x^t_{ij} > u_j, \\ X_{ij}^{t}, &\quad otherwise.\\ \end{cases} \end{aligned} \end{equation} | (3.12) |
The pseudo code of the ADECA algorithm is shown in Algorithm 1, and the specific implementation processes are described in detail as follows:
Algorithm 1 ADECA | |
Input: Parameters N , D , CR_0 , F_0 , l , u , P , T , \varepsilon Output: The best vector (Solution) \ldots \Delta 1: t \gets 0 (Initializing population space and belief space) 2: x_{ij}^t = rand(0, 1)\cdot (u_{j}-l_{j})+l_{j} 3: L = min(accept \; f(\cdot)) 4: U = max(accept \; f(\cdot)) 5: l = min(accept \; x_{ij}) 6: u = max(accept \; x_{ij}) 7: while |f(\Delta)|\geq \varepsilon or t\leq T do 8: for i = 1 to N do 9: (Mutation and Crossover) 10: for j = 1 to D do 11: v_{ij}^t = Mutation\; (x_{ij}^t) (formula (3.5)) 12: u_{ij}^t = Crossover\; (v_{ij}^t, x_{ij}^t) 13: end for 14: (Selection) |
15: if f(u_{ij}^t) < f(x_{ij}^t) then
16: x_{ij}^t\leftarrow u_{ij}^t 17: else { f(x_{ij}^t) < f(\Delta) } 18: \Delta \leftarrow x_{ij}^t 19: end if 20: end for 21: (Updating the normative knowledge) 22: if x_{ij}^t < l_j^t or f(x_{ij}^t) < L_j^t) then 23: l_{j}^t\leftarrow x_{ij}^t 24: L_{j}^t\leftarrow f(x_{ij}^t) 25: else 26: l_{j}^t\leftarrow l_{j}^t 27: L_{j}^t\leftarrow L_{j}^t 28: end if 29: t = t+1 30: end while |
Step 1. Set the parameters of the ADECA algorithm, such as N , D , CR_0 , F_0 , l , u , P , T and accuracy \varepsilon .
Step 2. Initializing population space and belief space. Randomly generate N initial individuals P(0) in population space, and each individual satisfies \sum_{\mathscr{K}_{i} = 1}^{m_{i}}x_{\mathscr{K}_{i}}^i = 1, x_{\mathscr{K}_{i}}^i\geq0, x_{\mathscr{K}_{i}}^i\in X^{i}, i = 1, \ldots, N; \mathscr{K}_{i} = 1, \ldots, m_{i} . The search range defined by the corresponding normative knowledge in the belief space and as the whole search space.
Step 3. Using the acceptance function (formula (3.11)) to accept better individuals and update the belief space.
Step 4. Calculating the fitness function value f(\cdot) of each individual in population P(t) and determining the x_{pbest}^{t} .
Step 5. The next generation population P(t+1) is generated by mutation operation (formula (3.5)), crossover operation (formula (3.2)), selection operation (formula (3.3)).
Step 6. Updating the normative knowledge of belief space by formulas (3.7), (3.8), (3.9), (3.10).
Step 7. Determining whether to end according to the accuracy and the maximum number of iteration, and output the optimal value, otherwise, turn to step 3.
As a random optimization algorithm based on population, the evolution process of the adaptive differential evolution algorithm in the population space can be described by a random process. For the convenience of discussion, considering mutation and crossover operations as a difference operator (DO), the selection operation is regarded as a selection operator (SO). The following uses the finite state Markov chain to analyze the search process of the ADECA algorithm, and use the axiomatic model [38] to study the probability distribution of the adaptive difference population in the decision space. Then, it is proved that the ADECA algorithm weakly converges to the global optimal solution in probability.
Definition 3.1. The DO operation randomly selects three different individuals from the population to execute formula (3.5) to generate the mutation vector V , and then performs crossover operation with the original target individual X according to formula (3.2) to generate the middle vector U . This process can be regarded as a random mapping \psi_D:\Omega \times S\rightarrow S^2 , that is
\begin{equation} \psi_D(\omega, X_i) = \psi_D(\langle X, V\rangle) = U, \end{equation} | (3.13) |
where \Omega is a nonempty abstract set, its element \omega is a basic event and the S is the solution space.
Definition 3.2. The SO operation is a process of selecting the individual with better fitness from the middle vector U_i and the target vector X_i according to the greedy selection method, which is recorded as a mapping \psi_S:S^2\rightarrow S ,
\begin{equation} \psi_S(\langle X_i, U_i\rangle) = \min\{f(X_i), f(U_i)\}. \end{equation} | (3.14) |
According to the above two definitions, one iteration of ADECA's population space can be seen as a mapping \psi = (\psi_S\circ \psi_D):\Omega \times S\rightarrow S .
Definition 3.3. [39] Let X = (X_n, n = 0, 1, 2, \cdots) is a random process, and its state space S is a finite set. If X has Markov property defined by the following formula, that is, for any n\geq 0 and any state i_0, i_1, \cdots, i_{n+1}\in S , and P\{X_0 = i_0, X_1 = i_1, \cdots, X_n = i_n\} > 0 , if
\begin{equation} P\{X_{n+1} = i_{n+1}|X_0 = i_0, X_1 = i_1, \cdots, X_n = i_n\} = P\{X_{n+1} = i_{n+1}|X_n = i_n\} \end{equation} | (3.15) |
is established, X is called finite Markov chain.
Here, S represents the range of random variables X_n, n = 0, 1, 2, \cdots , that is, the set of all possible values. According to the above definition, the finite Markov chain is expressed as: in a finite state, the state of the next moment is only related to the state of the current moment.
In the ADECA algorithm, X is the population space, S is the solution space, X_i and X_j are the i th and j th generation of evolutionary individuals, respectively. Knowledge-guided differential operation and selection operation can be regarded as a random transformation from state X_{i}\in S to state X_{j}\in S . In the process of state transition, state X(t + 1) is only related to state X(t) , therefore, according to the definition of finite Markov chain, the evolution process of ADECA algorithm is a finite Markov chain. Assume that the state transition matrix of the difference operation guided by knowledge is P_D , and the state transition matrix of the selection operation is P_S . Since the knowledge that guides the difference operation and selection operation in each generation contains different information, the state transition matrix P(t) = P_D(t)P_S(t) is different, that is, P is related to the evolutionary generation t . Therefore, the evolutionary process of ADECA algorithm is a non-homogeneous Markov chain (process). The following uses the axiomatic model to analyze the random search process of the ADECA algorithm.
Definition 3.4. [40] Random variable sequence \{X(t)\} weakly converges to the global optimal solution set S^{\ast} in probability, if it satisfies
\begin{equation} \lim\limits_{t \to \infty}P\{X(t)\cap S^{\ast}\neq \emptyset\} = 1, \end{equation} | (3.16) |
where S^{\ast} = \{X|\forall Y\in S, f(X)\leq f(Y)\} is the global optimal solution set of the fitness function f(\cdot) .
Definition 3.5. [40] If for any X\in B, Y\notin B , there is f(X) < f(Y) , then B\subset S is called the satisfactory solution set.
According to the definition, the fitness value of each individual in the satisfactory solution set is smaller than the fitness value of the individual outside the satisfactory solution set. Obviously, the global optimal solution set S^{\ast} is the satisfactory solution set, and is the minimum satisfactory solution set.
Theorem 3.1. [40] If for any satisfactory solution set B\subset S , there is
\begin{gather} P\{X(t+1)\cap B = \emptyset|X(t) = X\} \leq a(t), X \cap B\neq \emptyset, \end{gather} | (3.17) |
\begin{gather} P\{X(t+1)\cap B = \emptyset|X(t) = X\} \leq b(t), X \cap B = \emptyset, \end{gather} | (3.18) |
and satisfy: (1) \sum_{t = 1}^{\infty}(1-b(t)) = \infty , (2) \frac{a(t)}{1-b(t)}\rightarrow0 , then \{X(t)\} weakly converges to the global optimal solution set S^{\ast} in probability.
Corollary 3.1. [40] If the transition probability of \{X(t)\} satisfies the condition:
\begin{equation} \begin{aligned} P\{X(t+1)\cap B\neq \emptyset|X(t) = X\} = \begin{cases} 1, &X \cap B\neq \emptyset, \\ \geq \delta , &X \cap B = \emptyset.\\ \end{cases} \end{aligned} \end{equation} | (3.19) |
Then \{X(t)\} weakly converges to the global optimal solution set S^{\ast} in probability.
Theorem 3.2. The ADECA algorithm weakly converges to the global optimal solution set S^{\ast} in probability.
Proof. In the iteration of the ADECA algorithm, assuming the current state is X(t) = X_i , after the differential operation and selection operation, the next generation state is X(t+1) = X_j .
In the ADECA algorithm, the normative knowledge K_{nor} represents the change of the feasible search space. With the evolution, the search range should be concentrated in the dominant area. In each iteration, the current optimal individual recorded by the normative knowledge is retained in the next generation population, and the optimal individual will no longer participate in the next difference operation and selection operation.
(1) When X_i\cap B\neq \emptyset , it means that the current optimal individual must be included in the satisfactory set B . From the optimal retention strategy, it can be seen that no matter which generation t is, there is
\begin{equation} P\{X(t+1)\cap B\neq \emptyset|X(t) = X_i\} = 1, \; \; \; \; X_i\cap B\neq \emptyset. \end{equation} | (3.20) |
(2) When X_i\cap B = \emptyset , it indicates that the current population X(t) = X_i does not contain individuals of satisfactory set B . Then we consider each population X_j after X_i , if X_j\cap B = \emptyset , it is obviously meaningless. Therefore, the next generation population X_j\cap B\neq \emptyset is further discussed below.
When the state X_j satisfies X_j\cap B\neq \emptyset , that is, X_j contains the better individuals of the satisfactory set B . The transition probability generated by the difference operation under the guidance of normative knowledge is
\begin{alignat} {30} p_{ij}^D& = P_D\{X(t+1) = X_j|X(t) = X_i\} \\ & = \prod\limits_{k = 1}^{n}P_D\{(X_k(t+1) = x_{jk}|X_k(t) = x_{ik})|(Nor_k)\}P(Nor_k)\\ & = \prod\limits_{k = 1}^{n}P_D\{(X_k(t+1) = x_{jk}|X_k(t) = x_{ik})|(Nor_k)\}(accept(Nor)), \end{alignat} | (3.21) |
where Nor_k represents that the k th individual is affected by the normative knowledge, and P(Nor_k) is the probability of the next generation of differential operation, accept(Nor) represents the population acceptance proportion under the guidance of normative knowledge.
According to the meaning of the normative knowledge, each dimension of the individual after the difference operation does not exceed the feasible region. Therefore
\begin{equation} P_D\{(X_k(t+1) = x_{jk}|X_k(t) = x_{ik})|(Nor_k)\} > 0, \end{equation} | (3.22) |
and because of accpet(Nor) > 0 , so p_{ij}^D > 0 .
Considering that the selection mechanism of survival of the fittest is used in the ADECA algorithm to obtain the next generation population, the transition probability p_{ij}^S > 0 , so we can get
\begin{equation} P\{X(t+1)\cap B\neq \emptyset|X(t) = X_i\} = p_{ij}^Dp_{ij}^S > 0, \; \; \; \; X_i\cap B = \emptyset. \end{equation} | (3.23) |
It can be seen from Eqs (3.20) and (3.23) that the transition probability of the population evolution process of ADECA algorithm meets the condition of Corollary 3.1. It can be obtained that the state vector \{X(t)\} of the ADECA algorithm weakly converges to the global optimal solution set S^{\ast} in probability.
In this section, some matrix games are calculated to demonstrate the effectiveness of the proposed method. Taking the classic prisoner's dilemma game, the husband and wife game and the rock-paper-scissors game as examples. The algorithm ADECA, DECA and DE are used to find the Nash equilibria based on the system of equations, and the simulation comparison figures of them are given. The parameters of the algorithm are set as: N = 100 , T = 300 , F_0 = 0.4 , CR_0 = 0.9 , P = 0.2 , \varepsilon = 10^{-8} , etc. The following is the process of computing these games:
Example 4.1. Prisoner's dilemma game \Gamma_1(S^{1}, S^{2}, X^{1}, X^{2}, A, B) , where S^1 = \{s^1_1, s^1_2\} = \{confess, not\; confess\} , S^2 = \{s^2_1, s^2_2\} = \{confess, not\; confess\} , X^1 = \{x^1_1, x^1_2\} , X^2 = \{x^2_1, x^2_2\} , and x_1^1, x_1^2 respectively represent the probability of player 1, 2 choosing to confess, x_2^1, x_2^2 respectively represent the probability of player 1, 2 choosing not to confess, A and B are payoff matrices of the player in this game:
A = \left[ \begin{array}{ccc} -5 \; \; & 0 \\ -8 \; \; & -1 \end{array} \right ], \ \ B = \left[ {\begin{array}{ccc} -5 \; \; & -8 \\ 0\; \; & -1 \end{array}} \right]. |
According to formula (2.7), the above game can be transformed into the following equations:
\begin{equation} \left\{ \begin{aligned} &\lambda_{12}^1+x_1^1(-3x_1^2-x_2^2) = 0, \\ &\lambda_{21}^1+x_2^1(3x_1^2+x_2^2) = 0, \\ &\lambda_{12}^2+x_1^2(-3x_1^1-x_2^1) = 0, \\ &\lambda_{21}^2+x_2^2(3x_1^1+x_2^1) = 0. \end{aligned} \right. \end{equation} | (4.1) |
Denote each equation of the above equations by f_i , i = 1, 2, 3, 4 . According to the formula (2.9), the Eq (4.1) is equivalent to the following optimization problem:
\begin{gather} min \; F = |f_1(\lambda_{12}^1, x_1^1, x_1^2, x_2^2)|+|f_2(\lambda_{21}^1, x_2^1, x_1^2, x_2^2)|+ |f_3(\lambda_{12}^2, x_1^2, x_1^1, x_2^1)|+|f_4(\lambda_{21}^2, x_2^2, x_1^1, x_2^1)|, \\ x^1_{1}+x^1_{2} = 1, \\ x^2_{1}+x^2_{2} = 1. \end{gather} | (4.2) |
Making use of the ADECA algorithm, the DECA algorithm and the DE algorithm, we show that the solution of this problem is x^1 = \{1, 0\}, x^2 = \{1, 0\}, \lambda_{12}^1 = \lambda_{12}^2 = 3, \lambda_{21}^1 = \lambda_{21}^2 = 0 , that is, the prisoner's dilemma game has only one pure strategy Nash equilibrium (confess, confess) .
Example 4.2. Husband and wife game \Gamma_2(S^{1}, S^{2}, X^{1}, X^{2}, A, B) , where S^1 = \{s^1_1, s^1_2\} = \{football \; match, fashion \; show\} , S^2 = \{s^2_1, s^2_2\} = \{football \; match, fashion \; show\} , X^1 = \{x^1_1, x^1_2\} , X^2 = \{x^2_1, x^2_2\} , and A and B are the payoff matrices of player 1, 2 respectively, as shown below:
A = \left[ \begin{array}{ccc} 1 \; \; & 0 \\ 0 \; \; & 2 \end{array} \right ], \ \ B = \left[ {\begin{array}{ccc} 2 \; \; & 0 \\ 0\; \; & 1 \end{array}} \right]. |
According to the above payoff matrices and the calculation steps of Example 1, it is found that this problem has two pure strategy Nash equilibria and one mixed strategy Nash equilibrium: x^1 = (1, 0), x^2 = (1, 0); x^1 = (0, 1), x^2 = (0, 1); x^1 = (1/3, 2/3), x^2 = (2/3, 1/3) .
Example 4.3. Rock-paper-scissors game \Gamma_3(S^{1}, S^{2}, X^{1}, X^{2}, A, B) , where S^1 = \{s^1_1, s^1_2, s^1_3\} = \{rock, paper, scissors\} , S^2 = \{s^2_1, s^2_2, s^2_3\} = \{rock, paper, scissors\} , X^1 = \{x^1_1, x^1_2, x^1_3\} , X^2 = \{x^2_1, x^2_2, x^2_3\} and A, B are the payoff matrices of player 1, 2 respectively, as shown below:
A = \left[ \begin{array}{ccc} 0 \; \; & 1 \; \; &-1\\ -1 \; \; & 0 \; \; &1\\ 1 \; \; & -1 \; \; &0 \end{array} \right ], \ \ B = \left[ {\begin{array}{ccc} 0 \; \; & -1 \; \; &1\\ 1 \; \; & 0 \; \; &-1\\ -1 \; \; & 1 \; \; &0 \end{array}} \right]. |
Calculating this example by using the ADECA, DECA and DE algorithms, we find there is only one mixed strategy Nash equilibrium: x^1 = (1/3, 1/3, 1/3), x^2 = (1/3, 1/3, 1/3) .
The following Figures 2–4 are comparison figures of solving the above games with the ADECA, DECA and DE algorithms.
It can be clearly seen from Figures 2–4 that the ADECA algorithm has the best performance in the calculation process, followed by DECA algorithm, and DE algorithm is the worst, especially Figure 4 is the most obvious, where the ADECA algorithm finds the global optimal solution in the first 25 iterations, while the DECA and DE algorithms reach the optimal solution after 200 iterations. On the one hand, we found that the DECA algorithm and the DE algorithm have similar performance. This seems rational because the DECA algorithm only adds one more belief space in the implementation process, which increases the operation of extracting individual hidden information, so it is only a little faster than DE algorithm in convergence speed. On the other hand, with the increase of the number of iterations, the diversity of individuals decreases and the algorithm easily falls into a local optimum. Therefore, neither the DE algorithm nor the DECA algorithm avoids this result. However, the adaptive parameters and improved mutation operations of the ADECA algorithm have played a role, which not only maintaining the diversity of the population in the early stage of iteration, but also avoiding the destruction of the optimal solution in the later stage of iteration, and speeding up the search for the global optimal solution.
By analyzing the results obtained above, it can be seen that the ADECA algorithm proposed in this paper can well find the equilibrium solutions of matrix games based on the system of equations. Further, comparing this algorithm with the DECA and DE algorithms, we find that this algorithm has a faster convergence speed and avoids falling into the local optimum. It provides a reference for future studies on related issues.
In this paper, we study the Nash equilibria based on the system of equations. First, we give the equivalence theorem between the Nash equilibria and the solutions of the system of equations, that is, the Nash equilibria of the n -person finite noncooperative game are equivalent to the solutions of the system of algebraic equations with a series of relaxation parameters. This method is a new method for solving Nash equilibria except that the finding Nash equilibria is equivalent to solving optimization problems and fixed point problems. It not only builds a bridge between these two types of problems so that people can use the theories and methods of studying one problem to study another problem, broadens the methods for finding the Nash equilibria and lays a mathematical foundation for finding the Nash equilibria based on equations in the future. However, the equivalence form is also derived from the definition of Nash equilibrium, which implies the rational behavior of players. Second, we provide an equivalent form of solving matrix games and propose the ADECA algorithm. This algorithm uses a simple DE algorithm as the population space of CA algorithm, and in order to improve the convergence ability of the algorithm and global search capability, the mutation factor and crossover operator of the DE algorithm are adaptively improved, and a more globally convergent mutation operation is also proposed. Then, its convergence is proved theoretically by using finite Markov chain. Finally, the superiority of the ADECA algorithm is verified through three classic matrix game examples, and we also show that this algorithm can be used to find the Nash equilibria based on the equations, which lays the foundation for future research.
Although the method proposed in this paper can solve n-person noncooperative games theoretically, and provides specific steps for solving matrix games. It has certain research significance. However, when calculating specific multiplayer games or more complex games, the fitness function becomes increasingly complex due to the increase of the number of participants and the number of strategies, which may lead to the result falling into the local optimal solution and ending the computing in advance. Therefore, in order to overcome the limitations of the algorithm in solving multiplayer games or more complex games, the algorithm needs to be further improved in the future.
This study was supported by the National Natural Science Foundation of China (Grant Nos. [71961003] and [12061020]), the Scientific Research Foundation of Guizhou University (Grant No. [2019]49).
The authors declare no potential conflict of interests.
[1] |
M. J. Beckmann, Spatial oligopoly as a noncooperative game, Int. J. Game Theory, 2 (1973), 263–268. https://doi.org/10.1007/BF01737575 doi: 10.1007/BF01737575
![]() |
[2] |
A. Muthoo, M. J. Osborne, A. Rubinstein, A course in game theory, Economica, 63 (1996), 164–165. https://doi.org/10.2307/2554642 doi: 10.2307/2554642
![]() |
[3] |
E. Braggion, N. Gatti, R. Lucchetti, T. Sandholm, B. von Stengel, Strong Nash equilibria and mixed strategies, Int. J. Game Theory, 49 (2020), 699–710. https://doi.org/10.1007/s00182-020-00723-3 doi: 10.1007/s00182-020-00723-3
![]() |
[4] | J. V. Neumann, O. Morgenstern, The theory of games and economic behaviour, Princeton: Princeton University Press, 1944. |
[5] |
J. F. Nash, Equilibrium points in n-person games, Proc. Natl. Acad. Sci. USA, 36 (1950), 48–49. https://doi.org/10.1073/pnas.36.1.48 doi: 10.1073/pnas.36.1.48
![]() |
[6] |
J. F. Nash, Non-cooperative games, Ann. Math., 54 (1951), 286–295. https://doi.org/10.2307/1969529 doi: 10.2307/1969529
![]() |
[7] |
I. M. Bomze, Non-cooperative two-person games in biology: a classification, Int. J. Game Theory, 15 (1986), 31–57. https://doi.org/10.1007/BF01769275 doi: 10.1007/BF01769275
![]() |
[8] |
S. Govindan, R. Wilson, A global newton method to compute Nash equilibria, J. Econ. Theory, 110 (2003), 65–86. https://doi.org/10.1016/S0022-0531(03)00005-X doi: 10.1016/S0022-0531(03)00005-X
![]() |
[9] |
C. E. Lemke, J. T. Howson Jr., Equilibrium points of bimatrix games, Journal of the Society for Industrial and Applied Mathematics, 12 (1964), 413–423. https://doi.org/10.1137/0112033 doi: 10.1137/0112033
![]() |
[10] |
T. Kunieda, K. Nishimura, Finance and economic growth in a dynamic game, Dyn. Games Appl., 8 (2018), 588–600. https://doi.org/10.1007/s13235-018-0249-7 doi: 10.1007/s13235-018-0249-7
![]() |
[11] |
X. Wang, K. L. Teo, Generalized Nash equilibrium problem over a fuzzy strategy set, Fuzzy Set. Syst., 434 (2022), 172–184. https://doi.org/10.1016/j.fss.2021.06.006 doi: 10.1016/j.fss.2021.06.006
![]() |
[12] |
S. W. Xiang, S. Y. Xia, J. H. He, Y. L. Yang, C. W. Liu, Stability of fixed points of set-valued mappings and strategic stability of Nash equilibria, J. Nonlinear Sci. Appl., 10 (2017), 3599–3611. https://doi.org/10.22436/jnsa.010.07.20 doi: 10.22436/jnsa.010.07.20
![]() |
[13] |
S. W. Xiang, D. J. Zhang, R. Li, Y. L. Yang, Strongly essential set of KyFan's points and the stability of Nash equilibrium, J. Math. Anal. Appl., 459 (2018), 839–851. https://doi.org/10.1016/j.jmaa.2017.11.009 doi: 10.1016/j.jmaa.2017.11.009
![]() |
[14] |
J. P. Torres-Martínez, Fixed points as Nash equilibria, Fixed Point Theory Appl., 2006 (2006), 36135. https://doi.org/10.1155/FPTA/2006/36135 doi: 10.1155/FPTA/2006/36135
![]() |
[15] |
M. Schaefer, D. Štefankovič, Fixed points, Nash equilibria, and the existential theory of the reals, Theory Comput. Syst., 60 (2015), 172–193. https://doi.org/10.1007/s00224-015-9662-0 doi: 10.1007/s00224-015-9662-0
![]() |
[16] |
H. Mills, Equilibrium points in finite games, Journal of the Society for Industrial and Applied Mathematics, 8 (1960), 397–402. https://doi.org/10.1137/0108026 doi: 10.1137/0108026
![]() |
[17] |
N. G. Pavlidis, K. E. Parsopoulos, M. N. Vrahatis, Computing Nash equilibria through computational intelligence methods, J. Comput. Appl. Math., 175 (2005), 113–136. https://doi.org/10.1016/j.cam.2004.06.005 doi: 10.1016/j.cam.2004.06.005
![]() |
[18] |
A. S. Strekalovskii, R. Enkhbat, Polymatrix games and optimization problems, Autom. Remote Control, 75 (2014), 632–645. https://doi.org/10.1134/S0005117914040043 doi: 10.1134/S0005117914040043
![]() |
[19] | E. Rentsen, N. Tungalag, A. Gornov, A. Anikin, The curvilinear search algorithm for solving three-person game, In: Supplementary Proceedings of the 9th International Conference on Discrete Optimization and Operations Research and Scientific School (DOOR 2016), Vladivostok, Russia, September 19–23, 2016,574–583. |
[20] |
H. M. Li, S. W. Xiang, Y. L. Yang, C. W. Liu, Differential evolution particle swarm optimization algorithm based on good point set for computing nash equilibrium of finite noncooperative game, AIMS Mathematics, 6 (2021), 1309–1323. https://doi.org/10.3934/math.2021081 doi: 10.3934/math.2021081
![]() |
[21] |
J. T. Howson Jr., Equilibria of polymatrix games, Manage. Sci., 18 (1972), 312–318. https://doi.org/10.1287/mnsc.18.5.312 doi: 10.1287/mnsc.18.5.312
![]() |
[22] | J. Y. Han, N. H. Xiu, H. D. Qi, Nonlinear complementarity theory and algorithm, (Chinese), Shanghai: Shanghai Scienceand Technology Press, 2006. |
[23] |
Z. H. Huang, L. Qi, Formulating an n-person noncooperative game as a tensor complementarity problem, Comput. Optim. Appl., 66 (2017), 557–576. https://doi.org/10.1007/s10589-016-9872-7 doi: 10.1007/s10589-016-9872-7
![]() |
[24] |
C. Chen, L. Zhang, Finding nash equilibrium for a class of multi-person noncooperative games via solving tensor complementarity problem, Appl. Numer. Math., 145 (2019), 458–468. https://doi.org/10.1016/j.apnum.2019.05.013 doi: 10.1016/j.apnum.2019.05.013
![]() |
[25] |
F. Salehisadaghiani, L. Pavel, Distributed nash equilibrium seeking: a gossip-based algorithm, Automatica, 72 (2016), 209–216. https://doi.org/10.1016/j.automatica.2016.06.004 doi: 10.1016/j.automatica.2016.06.004
![]() |
[26] |
G. Chen, Y. Ming, Y. Hong, P. Yi, Distributed algorithm for \varepsilon-generalized Nash equilibria with uncertain coupled constraints, Automatica, 123 (2021), 109313. https://doi.org/10.1016/j.automatica.2020.109313 doi: 10.1016/j.automatica.2020.109313
![]() |
[27] |
Y. Zou, B. Huang, Z. Meng, W. Ren, Continuous-time distributed nash equilibrium seeking algorithms for non-cooperative constrained games, Automatica, 127 (2021), 109535. https://doi.org/10.1016/j.automatica.2021.109535 doi: 10.1016/j.automatica.2021.109535
![]() |
[28] |
C. Cenedese, G. Belgioioso, S. Grammatico, M. Cao, An asynchronous distributed and scalable generalized nash equilibrium seeking algorithm for strongly monotone games, Eur. J. Control, 58 (2021), 143–151. https://doi.org/10.1016/j.ejcon.2020.08.006 doi: 10.1016/j.ejcon.2020.08.006
![]() |
[29] | S. Liang, P. Yi, Y. Hong, Distributed Nash equilibrium seeking of a class of aggregative games, In: 2017 13th IEEE International Conference on Control & Automation (ICCA), Ohrid, Macedonia, 2017, 58–63. https://doi.org/10.1109/ICCA.2017.8003035 |
[30] |
M. Ye, G. Hu, Distributed Nash equilibrium seeking by a consensus based approach, IEEE Trans. Automat. Contr., 62 (2017), 4811–4818. https://doi.org/10.1109/TAC.2017.2688452 doi: 10.1109/TAC.2017.2688452
![]() |
[31] |
X. Wang, A computational approach to dynamic generalized Nash equilibrium problem with time delay, Commun. Nonlinear Sci. Numer. Simulat., 117 (2023), 106954. https://doi.org/10.1016/j.cnsns.2022.106954 doi: 10.1016/j.cnsns.2022.106954
![]() |
[32] | J. Yu, Selection of game theory, (Chinese), Beijing: Science Press, 2014. |
[33] | R. G. Reynolds, W. Sverdlik, Problem solving using cultural algorithms, In: Proceedings of the First IEEE Conference on Evolutionary Computation, IEEE World Congress on Computational Intelligence, Orlando, FL, USA, 1994,645–650. https://doi.org/10.1109/ICEC.1994.349983 |
[34] | R. Storn, K. Price, Minimizing the real functions of the icec'96 contest by differential evolution, In: Proceedings of IEEE international conference on evolutionary computation, Nagoya, Japan, 1996,842–844. https://doi.org/10.1109/ICEC.1996.542711 |
[35] | K. Price, P. M. Storn, J. A. Lampinen, Differential evolution: a practical approach to global optimization, Berlin, Heidelberg: Springer, 2005. https://doi.org/10.1007/3-540-31306-0 |
[36] | N. Noman, H. Iba, Enhancing differential evolution performance with local search for high dimensional function optimization, In: Proceedings of the 7th annual conference on Genetic and evolutionary computation, 2005,967–974. https://doi.org/10.1145/1068009.1068174 |
[37] | S. M. Saleem, Knowledge-based solution to dynamic optimization problems using cultural algorithms, PhD thesis, Wayne State University, Detroit, Michigan, 2001. |
[38] |
Z. B. Xu, Z. P. Chen, X. S. Zhang, Theoretical development on genetic algorithms: a review, (Chinese), Advances in Mathematics, 29 (2000), 97–114. https://doi.org/10.3969/j.issn.1000-0917.2000.02.001 doi: 10.3969/j.issn.1000-0917.2000.02.001
![]() |
[39] | B. Zhang, J. X. Zhang, Applied stochastic processes, (Chinese), Beijing: Tsinghua University Press, 2006. |
[40] | W. X. Zhang, Y. Liang, Mathematical foundation of geneticalgorithms, (Chinese), Xi'an: Xi'an Jiaotong University Press, 2001. |
1. | Jin Sha, Sisi Zheng, Fabio Tramontana, Research on Nash Equilibrium of Dual-Channel Supply Chain Based on Wholesale Price Discrimination and Sales Efforts, 2024, 2024, 1607-887X, 1, 10.1155/2024/2045842 | |
2. | Kai Wang, Wensheng Jia, A new intelligent algorithm for solving generalized Nash equilibrium problem, 2025, 123, 11100168, 17, 10.1016/j.aej.2025.03.044 |