
Reservoir computing has emerged as a powerful and efficient machine learning tool especially in the reconstruction of many complex systems even for chaotic systems only based on the observational data. Though fruitful advances have been extensively studied, how to capture the art of hyper-parameter settings to construct efficient RC is still a long-standing and urgent problem. In contrast to the local manner of many works which aim to optimize one hyper-parameter while keeping others constant, in this work, we propose a global optimization framework using simulated annealing technique to find the optimal architecture of the randomly generated networks for a successful RC. Based on the optimized results, we further study several important properties of some hyper-parameters. Particularly, we find that the globally optimized reservoir network has a largest singular value significantly larger than one, which is contrary to the sufficient condition reported in the literature to guarantee the echo state property. We further reveal the mechanism of this phenomenon with a simplified model and the theory of nonlinear dynamical systems.
Citation: Bin Ren, Huanfei Ma. Global optimization of hyper-parameters in reservoir computing[J]. Electronic Research Archive, 2022, 30(7): 2719-2729. doi: 10.3934/era.2022139
[1] | Bolin Zhao . Seeking optimal parameters for achieving a lightweight reservoir computing: A computational endeavor. Electronic Research Archive, 2022, 30(8): 3004-3018. doi: 10.3934/era.2022152 |
[2] | Shan Jiang, Li Liang, Meiling Sun, Fang Su . Uniform high-order convergence of multiscale finite element computation on a graded recursion for singular perturbation. Electronic Research Archive, 2020, 28(2): 935-949. doi: 10.3934/era.2020049 |
[3] | Bin Wang . Random periodic sequence of globally mean-square exponentially stable discrete-time stochastic genetic regulatory networks with discrete spatial diffusions. Electronic Research Archive, 2023, 31(6): 3097-3122. doi: 10.3934/era.2023157 |
[4] | Shaochen Lin, Xuyang Liu, Xiujuan Ma, Hongliang Mao, Zijian Zhang, Salabat Khan, Liehuang Zhu . The impact of network delay on Nakamoto consensus mechanism. Electronic Research Archive, 2022, 30(10): 3735-3754. doi: 10.3934/era.2022191 |
[5] | Xinzheng Xu, Yanyan Ding, Zhenhu Lv, Zhongnian Li, Renke Sun . Optimized pointwise convolution operation by Ghost blocks. Electronic Research Archive, 2023, 31(6): 3187-3199. doi: 10.3934/era.2023161 |
[6] | Tej Bahadur Shahi, Cheng-Yuan Xu, Arjun Neupane, Dayle B. Fleischfresser, Daniel J. O'Connor, Graeme C. Wright, William Guo . Peanut yield prediction with UAV multispectral imagery using a cooperative machine learning approach. Electronic Research Archive, 2023, 31(6): 3343-3361. doi: 10.3934/era.2023169 |
[7] | Ramziya Rifhat, Kai Wang, Lei Wang, Ting Zeng, Zhidong Teng . Global stability of multi-group SEIQR epidemic models with stochastic perturbation in computer network. Electronic Research Archive, 2023, 31(7): 4155-4184. doi: 10.3934/era.2023212 |
[8] | Ruiping Wen, Liang Zhang, Yalei Pei . A hybrid singular value thresholding algorithm with diagonal-modify for low-rank matrix recovery. Electronic Research Archive, 2024, 32(11): 5926-5942. doi: 10.3934/era.2024274 |
[9] | Ruyang Yin, Jiping Xing, Pengli Mo, Nan Zheng, Zhiyuan Liu . BO-B&B: A hybrid algorithm based on Bayesian optimization and branch-and-bound for discrete network design problems. Electronic Research Archive, 2022, 30(11): 3993-4014. doi: 10.3934/era.2022203 |
[10] | Xiaodong Wen, Xiangdong Liu, Cunhui Yu, Haoning Gao, Jing Wang, Yongji Liang, Jiangli Yu, Yan Bai . IOOA: A multi-strategy fusion improved Osprey Optimization Algorithm for global optimization. Electronic Research Archive, 2024, 32(3): 2033-2074. doi: 10.3934/era.2024093 |
Reservoir computing has emerged as a powerful and efficient machine learning tool especially in the reconstruction of many complex systems even for chaotic systems only based on the observational data. Though fruitful advances have been extensively studied, how to capture the art of hyper-parameter settings to construct efficient RC is still a long-standing and urgent problem. In contrast to the local manner of many works which aim to optimize one hyper-parameter while keeping others constant, in this work, we propose a global optimization framework using simulated annealing technique to find the optimal architecture of the randomly generated networks for a successful RC. Based on the optimized results, we further study several important properties of some hyper-parameters. Particularly, we find that the globally optimized reservoir network has a largest singular value significantly larger than one, which is contrary to the sufficient condition reported in the literature to guarantee the echo state property. We further reveal the mechanism of this phenomenon with a simplified model and the theory of nonlinear dynamical systems.
Recently, reservoir computing (RC) [1], also known as a generalization of echo-state network (ESN) [2] or liquid state machine (LSM) [3], has emerged as a powerful and efficient machine learning tool in reconstruction or/and prediction of many complex physical systems even for chaotic systems only based on the observational data of time series data [4,5,6]. In sharp contrast to its great efficacy, as a special variant of a recurrent neural network (RNN), RC has a surprisingly contracted architecture, where only three weight matrices are involved: the input weight matrix and a reservoir network matrix are randomly generated and fixed, leaving only one output weight matrix for training, as shown in Figure 1. As such, simple and efficient least squares optimization methods rather than the resource-consuming back propagation algorithm are good enough for the training process [7]. Thus, a question arises naturally: how to capture the art of hyper-parameter settings for RC's networks? As a matter of fact, this is a long-standing and urgent problem and great attentions have been attracted to carry out various discussions, i.e., from the topology and distribution of the random connections [8,9] to the spectral radius and singular value of the random network [10,11]. Generally, the existing studies are mainly based on a variable control experimentation manner, i.e., optimizing one hyper-parameter while leaving all the other hyper-parameters constant. In such a way, the obtained results are mainly local with specific settings but sometimes cannot be generalized to global hyper-parameters space.
In this work, we propose a global optimization framework using simulated annealing technique to find the optimal architecture of the randomly generated networks for a successful RC. With the optimized results, we further study several important properties of some hyper-parameters, i.e., the sparsity and distribution of the networks, the spectral radius and the largest singular value of the reservoir networks. Interestingly, we find that the globally optimized reservoir network has a spectral radius near one and a largest singular value significantly larger than one, which is contrary to the sufficient condition reported in the literature to guarantee the echo state property. We further study the mechanism of this phenomenon with a simplified model and the theory of nonlinear dynamical systems.
For the job of nonlinear dynamics reconstruction based on time series data, a general framework of RC could be sketched in Figure 1. Here, the input data xk∈Rn represents the observation vector of a dynamical system sampled at time step k with a specific smooth observe function such that xk=h(zk) where zk is the state vector. The underlying dynamical system is assumed to be evolving on a compact manifold M with the evolution operator φ∈Diff2(M):zk+1=φ(zk). The reservoir network is composed of m reservoir neurons whose connections are represented by the reservoir network matrix Wres and the vector rk∈Rm represents the state of reservoir neurons at time step k. The input layer weight matrix Win and the reservoir network matrix Wres are, respectively, m×n and m×m random matrices generated according to certain distribution laws.
The dynamical evolution of the reservoir neurons is governed by:
rk=(1−α)rk−1+αtanh(Wresrk−1+Winxk), | (2.1) |
where α is the leakage factor, and tanh∈C2(R,(−1,1)) is a sigmoid function. The reservoir vector and the output vector yk∈Rl is connected by the output weight matrix Wout∈Rl×m such that yk=Woutrk. Given the time series, denoted by xk,k=1,⋯,N, as training data, the target is to train the output weight matrix Wout so as to approximate the one-step dynamics prediction, i.e., yk≈xk+1. To this, a loss function is designed as
L=N∑k=1‖xk+1−Woutrk‖2+β‖Wout‖2, | (2.2) |
where β>0, the L2-regularization coefficient, is introduced to make optimization robust. The output weight matrix Wout is thus generally obtained by minimizing the loss function (2.2) over the training data set. After training, one can fix the output weight matrix Wout and redirect the output yk=Woutrk as an approximation of xk+1 into the input layer of the network and thus the autonomous dynamics for xk with k>N could be generated.
The RC framework is distinguished from usual RNNs due to the fact that the input weight matrix Win and the reservoir network matrix Wres are randomly generated rather than being trained. Therefore several properties e.g., the sparsity, the distribution, the spectral radius, and the largest singular value of the randomly generated weight matrices will undoubtedly affect the performance and careful choice a priori is required, i.e., they are hyper-parameters in a RC framework. Among all the hyper-parameters, the spectral radius ρ, defined as the largest absolute eigenvalue, of Wres, is generally believed as a key to the success of reservoir computing. The seminal works [1] and [10] conclude that spectral radius is related to the echo state property(ESP), a necessary condition for a RC to work properly, and the memory capacity, a capacity evaluating time series reconstruction ability, and therefore spectral radius is suggested to be less than 1. However, several other works [8,11,12] also imply that the optimal spectral radius varies case from case and spectral radius larger than 1 sometimes shows best performance. To further theoretically study the echo state property, the largest singular value σ is introduced in [2] and σ(Wres)<1 is adopted as a sufficient condition to ensure ESN. Besides ρ and σ, the sparsity, topology, and distribution of the randomly generated matrices are also usually considered as key hyper-parameters of a RC. It is suggested in the seminal work [7] that in order to generate rich variety of dynamics, the reservoir network should be sparse and recent works also confirms that low connectivity is beneficial for forecasting chaotic systems [9]. As for the topology and distribution of the randomly generated matrices, besides the commonly used Erdös-Rényi random network, both small-world and scale free networks have been discussed in [13], and analogously, other than the commonly adopted uniform distribution, both Gaussian and even Bernoulli distribution have been studied in [14].
Though many efforts have been made to understand and reveal the optimal choice of these hyper-parameters, the existing studies are mainly reported in a variable control experimentation manner, i.e., optimizing one hyper-parameter while leaving all the other hyper-parameters constant. In such a way, the obtained optimal results may be local with specific settings. As a matter of fact, in [8,13], it is reported that different topologies of reservoir networks may yield significantly different conclusions for optimal spectral radiuses. Therefore, a global optimization study is urgent and necessary.
Simulated annealing (SA) is a probabilistic technique for approximating global optimization in a large search space for a given function L. The basic idea of SA is that, in each iteration step, the SA heuristically considers some neighboring state s∗ of the current state s and probabilistically decides whether moving the system to the new state s∗ or staying in the current state s. Such probability is decided by both the improvement L(s∗)−L(s) and a decaying temperature T, and typically the iterated step is repeated until the system reaches a state good enough for the application.
In this work, we take the chaotic Lorenz system as a benchmark:
˙x=a(y−x),˙y=−xz+bx−y,˙z=xy−cz, | (2.3) |
where a=10,b=28, and c=8/3. We consider the job of dynamics reconstruction using RC, i.e., after training period, the autonomous dynamics xt=[x(t),y(t),z(t)]T is generated by the trained RC and the reconstruction is evaluated by the forecasting horizon (the length of precise predictions with error less than a threshold). To find the global optimization for the key hyper-parameters of RC, we take all the weights in the two matrices Win and Wres as state variables s=[Win,Wres] in the SA framework and design the target function as L=h−α‖Wres‖1 where h is the forecasting horizon expressed in Lyapunov times and α>0 is a regularization coefficient to make Wres sparse as generally required for RC [7,9]. Thus the SA algorithm adopted in this work could be summarized in Algorithm 1 as follow.
Algorithm 1. Generate an initial state s0, train the RC, and evaluate the target function L=L(s0).
Repeat the iteration until k⩾kMAX or forecasting horizon h keeps unchanged for RMAX iterations:
1) Generate a random neighboring state s∗ based on the current s, train the RC and evaluate the target function L∗=L(s∗).
2) IF L∗>L, Pt=1.
ELSE Pt=exp(−(L−L∗)T(k))
3) IF Pt⩾Random(0,1), accept the new state, i.e., let s=s∗ and L=L∗.
Here k stands for the kth iteration, kMAX is the largest number of iterations.
In this work, we set the size of Win and Wres as 100×3 and 100×100 respectively and thus there are in total 100,300 parameters in the SA state s. To accelerate the optimization, inspired by the drop out strategy in deep learning, we only update a set of randomly picked 1000 parameters in each iteration and update the decaying temperature T after each 100 iterations, i.e., T(k)=1∗0.95k|100.
In order to avoid unbounded states, inspired by the work [15], we also restrict the update of states in the following way
s∗i={β∗ub+(1−β)si,si>ub,si,si∈[lb,ub],β∗lb+(1−β)si,si<lb, | (2.4) |
where the upper bound and lower bound are ub=0.8,lb=−0.8 and β is a uniformly random number between 0 and 1.
A typical process of SA and the optimized results are illustrated in Figure 2, and in order to guarantee convergence, we set kMAX=20,000 and RMAX=5000 empirically.
To make the results robust, we carry out the SA global optimization algorithm 50 independent times to find the optimal hyper-parameters for the RC and obtain 50 sets of optimized Win and Wres respectively. Based on such a result we discuss the choice of several key hyper-parameters, i.e., the topology of the reservoir matrix, the distribution of the weight, the spectral radius ρ, and the largest singular value σ.
Before we discuss the topology of the optimized random network, we note that though the L1 regulation has been introduced into the SA algorithm, there are still several very small non-zeros weights which will contaminate the topology analysis. Therefore, we run the sparsity test as illustrated in Figure 3(a) and set 10−3.5 as the threshold for non-zero weight in the random network. The degree distributions of such pre-trimmed Wres are shown in Figure 3(b), where the majority shows a symmetric style, and this character implies that the optimized networks Wres in this job are generally Erdös-Rényi random networks since the degree distribution of an ER random network obeys Poisson distribution while Poisson distribution with high mean values appears in a symmetric style [16]. Additionally, we also carry out a test to study whether there is any small-world property in the optimized Wres. In Figure 4(a), we show the clustering coefficients for the 50 independent optimized Wres as well as the ones for the corresponding random networks. Here the corresponding random network refers to the random network that has the same numbers of nodes and same number of edges per node as this network. It is clear that the clustering coefficients of Wres are not significantly larger than the ones for the corresponding random networks, implying that the optimized Wres does not have small-world property.
When generating random matrices, the weight distribution is another essential hyper-parameter. To test whether the weight distribution obeys normal distribution for the optimized networks, we use the Quantile-Quantile Plots (QQ Plots) as a test. To be concrete, we first linearly scale the weight distribution into zero mean and one deviation, and then plot the quantile against the standard normal distribution. If more than 80% points falls in the confident interval, i.e., ± 0.3 of the fitted line, we regard the weight distribution as normal, and two typical QQ Plots are illustrated in Figure 3(c), (d). In order to avoid the affection of initial conditions and the random neighborhood used in the SA optimization algorithm, we have tried both normal distribution and uniform distribution to generate initial states as well as the random neighborhood, and we come to the conclusion that the weights of all the optimized Wres obey normal distribution while the weights of Win highly depends on the choice of initial state distribution, as shown in Table 1. Actually, this observation coincides with the existing results where many works prefer normal distribution for Wres while both normal distribution and uniform distribution are adopted for Win [5,9,12,17,18,19].
![]() |
Uniform | Normal |
Uniform | Win: 0/50, Wres: 50/50 | Win: 0/50, Wres: 50/50 |
Normal | Win: 39/50, Wres: 50/50 | Win: 36/50, Wres: 50/50 |
Based on the 50 independent runs for optimized Wres, the respective values for spectral radius ρ and the largest singular value σ are illustrated in Figure 4(b). It is clear that the optimized values of ρ are all less than one and clustered around the mean value 0.87, which coincides with the conclusion suggested by the seminal works [1,10] that a spectral radius less than one but close to one is a good choice for both echo state property and the memory capacity.
While for the largest singular values, we find that the optimized values of σ are all larger than one with a mean value as large as 1.71. Since the theoretical analysis in [2] shows σ<1 is a rigorous sufficient condition to ensure echo state property, we further check that whether the echo state property holds for the RC using our optimized σ which is larger than one. The echo state property means that the effect of initial conditions should vanish as time passes, i.e., starting from different initial values, the dynamics of internal neurons should synchronize with each other rapidly [10]. In Figure 4(c), a typical result is shown that no matter what the neurons' initial values are, their dynamics quickly synchronize with each other under the same input, i.e., the echo state property still holds.
In order to study the mechanism of this phenomenon, we first carry out a simple analysis analogous to the work in [2]. Starting with arbitrary two different initial conditions r10 and r20, consider two trajectories r1k and r2k of RC neurons generated by Eq (2.1) under the same input xk. The evolution of difference between two trajectories Δrk=r1k−r2k could be estimated as
‖Δrk‖=‖(1−α)Δrk−1+α(tanh(Wresr1k−1+Winxk)−tanh(Wresr2k−1+Winxk))‖≤(1−α)‖Δrk−1‖+α‖WresΔrk−1‖≤(1−α+ασ)‖Δrk−1‖, | (3.1) |
thus as long as σ<1 we have ‖Δrk‖≤β‖Δrk−1‖ where β=1−(1−σ)α<1, and consequently Δrk tends to zero exponentially, i.e., r1k and r2k synchronizes with each other quickly. However when σ>1, the above estimation no longer holds. To further reveal the mechanism why echo state property still generally holds when σ>1, we consider a simple model as an illustration. We consider the 1-D situation, i.e., there is only one neuron in the reservoir, and let the leaky coefficient α=1 which is also adopted in many works [5]. Then the dynamics of the single neuron could be described by
rk+1=tanh(σrk+Winxk+1). |
The map rk+1=tanh(σrk) has three fixed points when σ>1, as illustrated in Figure 5, where c0=0 is repelling while c+>0 and c−<0 are attractive. It is clear from Figure 5 that (0,+∞) is the attractive basin for c+ while (−∞,0) is the attractive basin for c−. Thus, as long as the initial values r10 and r20 fall in the same attractive basin, two trajectories r1k and r2k under map rk+1=tanh(σrk) will soon converge to the same fixed point c+ or c−. Note the fact ‖Winxk‖≪1 and it could be noted as a perturbation γk. Thus the two trajectories r1k and r2k under map rk+1=tanh(σrk+γk) will generically oscillate around the same fixed point c+ or c−. Further note that the derivative of the map around c+ or c− is less than one, i.e., the map is contractive around c+ or c− as highlighted in the pink region of Figure 5, thus the two trajectories will synchronize with each other soon, and therefore the echo state property still holds.
In this paper, we carry out an optimization schema using simulated annealing (SA) technique to find the globally optimized hyper-parameters for reservoir computing in the job of chaotic dynamics reconstruction. Specifically, we discuss the choice of random network topology for the reservoir network Wres, the type of weight distribution for both the input layer network Win and Wres, and we further study the spectral radius and largest singular value of the reservoir network Wres which are closely related to the echo state property and memory capacity of the RC. Most of the globally optimized hyper-parameters coincide with the main stream of the existing results, i.e., the topology of the random network Wres satisfy ER random network, the weight distributions for Wres and Win are mainly normal distribution, and the spectral radius ρ is less than one but close to one. On one side, such results confirm the effectiveness of our proposed SA schema as a global optimization method, and on the other side, these results also provide a way to choose/initialize Win and Wres when considering these hyper-parameters.
However, we also find the optimized largest singular value σ is significantly larger than one, which does not satisfy the theoretical sufficient condition to ensure the echo state property. To study this phenomenon, we further confirm that the echo state property still holds even with σ>1 and RC works well. With a simple illustrative model, we reveal the mechanism of RC with σ>1 using the theory of nonlinear dynamical systems. Indeed, such analysis is simple and heuristic, it is a challenging work to consider more complicated situations, e.g., if the initial values r10 and r20 fall in different attractive basins, or even high-dimensional cases with saddles. Some tools in nonlinear dynamical systems such as noise induced synchronization may be introduced to understand the mechanism and some recently developed methods [20,21] may be helpful for further applications. However this is beyond the topic in this paper and the quantitative analysis for these problems is still open and forms the direction for the future works.
This work is supported by NSFC with grant No. 12171350.
The authors declare no conflict of interests.
[1] |
H. Jaeger, H. Haas, Harnessing nonlinearity: Predicting chaotic systems and saving energy in wireless communication, Science, 304 (2004), 78–80. https://doi.org/10.1126/science.1091277 doi: 10.1126/science.1091277
![]() |
[2] | H. Jaeger, The "echo state" approach to analysing and training recurrent neural networks-with an erratum note, German National Research Center for Information Technology GMD Technical Report, 34 (2001), 148. |
[3] |
W. Maass, T. Natschläger, H. Markram, Real-time computing without stable states: A new framework for neural computation based on perturbations, Neural Comput., 14 (2002), 2531–2560. https://doi.org/10.1162/089976602760407955 doi: 10.1162/089976602760407955
![]() |
[4] |
J. Pathak, B. Hunt, M. Girvan, Z. Lu, E. Ott, Model-free prediction of large spatiotemporally chaotic systems from data: A reservoir computing approach, Phys. Rev. Lett., 120 (2018), 024102. https://doi.org/10.1103/PhysRevLett.120.024102 doi: 10.1103/PhysRevLett.120.024102
![]() |
[5] |
G. Tanaka, T. Yamane, J. B. Héroux, R. Nakane, N. Kanazawa, S. Takeda, et al., Recent advances in physical reservoir computing: A review, Neural Networks, 115 (2019), 100–123. https://doi.org/10.1016/j.neunet.2019.03.005 doi: 10.1016/j.neunet.2019.03.005
![]() |
[6] |
Q. Zhu, H. F. Ma, W. Lin, Detecting unstable periodic orbits based only on time series: When adaptive delayed feedback control meets reservoir computing, Chaos, 29 (2019), 093125. https://doi.org/10.1063/1.5120867 doi: 10.1063/1.5120867
![]() |
[7] | H. Jaeger, Tutorial on training recurrent neural networks, covering BPPT, RTRL, EKF and the "echo state network" approach, GMD-Forschungszentrum Informationstechnik Bonn, 5 (2002). |
[8] |
A. Haluszczynski, C. Räth, Good and bad predictions: Assessing and improving the replication of chaotic attractors by means of reservoir computing, Chaos, 29 (2019), 103143. https://doi.org/10.1063/1.5118725 doi: 10.1063/1.5118725
![]() |
[9] |
A. Griffith, A. Pomerance, D. J. Gauthier, Forecasting chaotic systems with very low connectivity reservoir computers, Chaos, 29 (2019), 123108. https://doi.org/10.1063/1.5120710 doi: 10.1063/1.5120710
![]() |
[10] |
M. Lukoševičius, H. Jaeger, Reservoir computing approaches to recurrent neural network training, Comput. Sci. Rev., 3 (2009), 127–149. https://doi.org/10.1016/j.cosrev.2009.03.005 doi: 10.1016/j.cosrev.2009.03.005
![]() |
[11] |
J. Jiang, Y. C. Lai, Model-free prediction of spatiotemporal dynamical systems with recurrent neural networks: Role of network spectral radius, Phys. Rev. Res., 1 (2019), 033056. https://doi.org/10.1103/PhysRevResearch.1.033056 doi: 10.1103/PhysRevResearch.1.033056
![]() |
[12] |
D. Verstraeten, B. Schrauwen, M. d'Haene, D. Stroobandt, An experimental unification of reservoir computing methods, Neural Networks, 20 (2007), 391–403. https://doi.org/10.1016/j.neunet.2007.04.003 doi: 10.1016/j.neunet.2007.04.003
![]() |
[13] |
H. Cui, X. Liu, L. Li, The architecture of dynamic reservoir in the echo state network, Chaos, 22 (2012), 033127. https://doi.org/10.1063/1.4746765 doi: 10.1063/1.4746765
![]() |
[14] |
B. Zhang, D. J. Miller, Y. Wang, Nonlinear system modeling with random matrices: echo state networks revisited, IEEE Trans. Neural Networks Learn. Syst., 23 (2011), 175–182. https://doi.org/10.1109/TNNLS.2011.2178562 doi: 10.1109/TNNLS.2011.2178562
![]() |
[15] |
M. Ji, Z. Jin, H. Tang, An improved simulated annealing for solving the linear constrained optimization problems, Appl. Math. Comput., 183 (2006), 251–259. https://doi.org/10.1016/j.amc.2006.05.070 doi: 10.1016/j.amc.2006.05.070
![]() |
[16] |
G. F. de Arruda, F. A. Rodrigues, Y. Moreno, Fundamentals of spreading processes in single and multilayer complex networks, Phys. Rep., 756 (2018), 1–59. https://doi.org/10.1016/j.physrep.2018.06.007 doi: 10.1016/j.physrep.2018.06.007
![]() |
[17] |
Z. Lu, J. Pathak, B. Hunt, M. Girvan, R. Brockett, E. Ott, Reservoir observers: Model-free inference of unmeasured variables in chaotic systems, Chaos, 27 (2017), 041102. https://doi.org/10.1063/1.4979665 doi: 10.1063/1.4979665
![]() |
[18] |
X. Dutoit, B. Schrauwen, J. Van Campenhout, D. Stroobandt, H. Van Brussel, M. Nuttin, Pruning and regularization in reservoir computing, Neurocomputing, 72 (2009), 1534–1546. https://doi.org/10.1016/j.neucom.2008.12.020 doi: 10.1016/j.neucom.2008.12.020
![]() |
[19] | D. Verstraeten, J. Dambre, X. Dutoit, B. Schrauwen, Memory versus non-linearity in reservoirs, in The 2010 International Joint Conference on Neural Networks (IJCNN), IEEE, (2010), 1–8. https://doi.org/10.1109/IJCNN.2010.5596492 |
[20] |
X. Ying, S. Y. Leng, H. F. Ma, Q. Nie, Y. C. Lai, W. Lin, Continuity scaling: A rigorous framework for detecting and quantifying causality accurately, Research, 2022 (2022), 9870149. https://doi.org/10.34133/2022/9870149 doi: 10.34133/2022/9870149
![]() |
[21] |
J. W. Hou, H. F. Ma, D. He, J. Sun, Q. Nie, W. Lin, Harvesting random embedding for high-frequency change-point detection in temporal complex systems, Natl. Sci. Rev., 9 (2022), nwab228. https://doi.org/10.1093/nsr/nwab228 doi: 10.1093/nsr/nwab228
![]() |
1. | Ruizhi Cao, Chun Guan, Zhongxue Gan, Siyang Leng, Reviving the Dynamics of Attacked Reservoir Computers, 2023, 25, 1099-4300, 515, 10.3390/e25030515 | |
2. | Ji Xia, Junyu Chu, Siyang Leng, Huanfei Ma, Reservoir computing decoupling memory–nonlinearity trade-off, 2023, 33, 1054-1500, 10.1063/5.0156224 | |
3. | Zhihao Zuo, Ruizhi Cao, Zhongxue Gan, Jiawen Hou, Chun Guan, Siyang Leng, Feedback coupling induced synchronization of neural networks, 2023, 549, 09252312, 126457, 10.1016/j.neucom.2023.126457 | |
4. | Zequn Lin, Zhaofan Lu, Zengru Di, Ying Tang, Learning noise-induced transitions by multi-scaling reservoir computing, 2024, 15, 2041-1723, 10.1038/s41467-024-50905-w | |
5. | Suguru Shimomura, 2024, Chapter 11, 978-981-99-5071-3, 227, 10.1007/978-981-99-5072-0_11 | |
6. | In-mo Lee, Yoojeung Kim, Taehoon Kim, Hayoung Choi, Seung Yeop Yang, Yunho Kim, Recursive reservoir concatenation for salt-and-pepper denoising, 2025, 160, 00313203, 111196, 10.1016/j.patcog.2024.111196 | |
7. | Kaushal Kumar, Forecasting Crude Oil Prices Using Reservoir Computing Models, 2024, 0927-7099, 10.1007/s10614-024-10797-w | |
8. | Ji Xia, Luonan Chen, Huan-Fei Ma, Enhanced chaotic communication with machine learning, 2024, 14, 2158-3226, 10.1063/5.0237990 |
![]() |
Uniform | Normal |
Uniform | Win: 0/50, Wres: 50/50 | Win: 0/50, Wres: 50/50 |
Normal | Win: 39/50, Wres: 50/50 | Win: 36/50, Wres: 50/50 |