Abstract

The echo state property (ESP) is a key concept for understanding the working principle of the most widely used reservoir computing model, the echo state network (ESN). The ESP is achieved most of the operation time under general conditions, yet the property is lost when a combination of driving input signals and intrinsic reservoir dynamics causes unfavorable conditions for forgetting the initial transient state. A widely used treatment, setting the spectral radius of the weight matrix below the unity, is not sufficient as it may not properly account for the nature of driving inputs. Here, we characterize how noisy driving inputs affect the dynamical properties of an ESN and the empirical evaluation of the ESP. The standard ESN with a hyperbolic tangent activation function is tested using the MNIST handwritten digit datasets at different additive white Gaussian noise levels. The correlations among the neurons, input mapping, and memory capacity of the reservoir nonlinearly decrease with the noise level. These trends agree with the deterioration of the MNIST classification accuracy against noise. In addition, the ESP index for noisy driving input is developed as a tool to help easily assess ESPs in practical applications. Bifurcation analysis explicates how the noise destroys an asymptotical convergence in an ESN and confirms that the proposed index successfully captures the ESP against noise. These results pave the way for developing noise-robust reservoir computing systems, which may promote the validity and utility of reservoir computing for real-world machine learning applications.

1. Introduction

Reservoir computing (RC) provides a supervised learning framework for recurrent neural networks (RNNs) that help overcome practical limitations of RNNs such as the convergence issue on training and vanishing/exploding gradient problems [17]. Among various types of RC models, the echo state network (ESN) originally developed by Herbert Jaeger in 2001 [5, 8] has been widely used. In an ESN system, a random sparsely connected RNN (acting as a reservoir) maps the time-varying driving inputs into the high-dimensional spatiotemporal information in the reservoir, and then a readout layer is used to link the reservoir information and desired output [4, 5, 810]. In most cases, the connectivity matrix of the reservoir network is fixed, yet the weights of connections between the reservoir and the readout layer are trainable in a supervised manner. This efficient learning strategy enables high performance at a low computational cost, compared to conventional RNN settings. This basic working principle is shared with another widely used RC model, the liquid state machine (LSM) which was developed by Maass, Natschlager, and Markram (independently from the ESN) [4, 9, 10]. These RC models have been successfully applied to a wide range of real-world problems including stock market prediction [11, 12], biomedical application [1317], speech recognition [18, 19], and handwriting recognition [2023], as well as fundamental problems in physics such as critical transition dynamics [2427], network link inferring [28, 29], and stochastic/chaotic time series prediction [7, 3041].

The echo state property (ESP) was originally defined by Jaeger [5] to grasp the computational capabilities and mechanisms of ESNs, which may be used as a design principle of the reservoirs. Intuitively, a reservoir having this property can be successfully entrained by the driving input to generate high-dimensional, nonlinear, and rich memory properties for computations. In other words, the reservoir dynamics asymptotically washes out its transient initial state induced by the driving input; this is also referred to as “input forgetting” or “state forgetting” [5]. The original ESP [5] claims that an ESN meets the ESP if all state vectors driven by any input sequence from a compact set U asymptotically converge to the same state. While it is a sufficient condition for achieving ESP that the largest singular value of the weight matrix is smaller than unity, it is a necessary condition that the spectral radius of the matrix is smaller than unity. Some following works elaborated on the ESP definition and its sufficient conditions by taking into account the nature of driving input signals. The work by Yildiz et al. [42] considered the distribution of the driving input signals to refine the original ESP concept. This work had important ramifications for the conditions for ESP, a spectral radius greater than unity does not necessarily imply a loss of the ESP, and thus, the commonly used procedure of scaling the spectral radius to under unity to ensure ESPs can be flawed. Meanwhile, the condition on the largest singular value of the weight matrix was found to be too restrictive, leading to poor performance, and alternative conditions were formulated [42, 43]. Subsequently, less restrictive sufficient conditions for the ESP to prevent the fast washing-out problem have been developed [42, 43]. More recently, Manjunath and Jaeger [44] provided an alternative formulation where the ESP is defined concerning a specific input signal rather than a range of possible inputs. For a given input signal, the formulation prescribes the spectral properties of the network weight matrix W that satisfies the ESP. However, work still needs to be done to elucidate when ESP is satisfied given a broad distribution of input signals, e.g., inputs in the presence of noise. Wainrib and Galtier [45] developed a cheap algorithm to establish a local and operational version of the ESP through the computation of the largest Lyapunov exponent. Basterrech [46] presented an empirical analysis of the accuracy and the input mapping of reservoirs. Kubota et al. [47] experimentally demonstrated that cultured neuronal networks can have ESPs to serve as physical reservoir computers.

As stated above, the ESP concept has undergone several refinements since its introduction, yet the validity of widely used literature conditions for the ESP is limited because they do not always properly account for the nature of driving input signals. Of course, the simple technical treatment (i.e., setting the (effective) spectral radius of the weight matrix below the unity) often fails to guarantee the ESP when a combination of driving input and intrinsic reservoir dynamics causes unfavorable conditions for forgetting the transient initial state. For this, the empirical assessment of ESP in practice needs more elaboration in providing sufficient conditions for the ESP, considering that the ESP is a cooperative phenomenon of the intrinsic reservoir dynamics and a set of admissible driving inputs. This study specifically describes numerical simulations and an analytical characterization of the ESP in the presence of noisy driving inputs. The standard ESN model with a tangent hyperbolic activation function is used to examine the dynamical properties and the resulting ESP during the MNIST handwritten digit classification tasks at different additive white Gaussian noise levels. The effects of the noise on the reservoir dynamics are characterized by various measures including the correlations among the neuronal activities within a reservoir, the mapping of the noisy input to the reservoir, and the memory capacity. These dynamical properties are related to the MNIST classification accuracy and bifurcation dynamics of the reservoir. In addition, the ESP index for noisy driving input is developed based on the work by Gallicchio [48] to help easily assess the property in practical applications. Bifurcation analysis was employed to capture the underlying dynamical properties of the reservoir and to prove the validity of the proposed ESP index.

By way of outline, Section 2 reviews the theoretical frameworks for understanding ESNs and the ESP. Section 3 describes the computational and theoretical methods including the definition of noisy input-driven ESP, bifurcation analysis, and the MNIST classification task. Section 4 begins by describing the changes in the dynamical properties of the reservoir against noise, followed by bifurcation dynamics of the reservoir, and these are related to the deterioration of classification accuracy and the ESP index. Section 5 combines the discussion and conclusions.

2. Theoretical Framework

2.1. Standard Echo State Network

The standard echo state network (ESN) model proposed by Herbert Jaeger in 2001 [5] with reservoir units, inputs, and outputs is defined as follows:where , , and are the internal, input, and output vectors at time , respectively, is the internal weight matrix of the reservoir, is the input matrix, is the feedback matrix, and is the output matrix. The state activation function is a sigmoid function (usually is a hyperbolic tangent function) applied component-wise with and the output activation function is , where each is usually the identity or a sigmoid function.

The compactness condition means is defined on , where and are compact sets and satisfies with , . Let and denote the set of left-infinite input and state vector sequences, respectively. We say is compatible with when .

The standard ESN with a hyperbolic tangent activation function without feedback (i.e., and ) is given by

2.2. Echo State Property for Standard ESNs

The echo state property (ESP) is first defined by Jaeger in his original paper that proposed the ESN [5]: a network with the compactness condition has the ESP with respect to if for any left infinite input sequence and any two state vector sequences compatible with , it holds that . It is important to note that the ESP is not a property of a reservoir (or network) per se, but a property of a pair (the reservoir and the set of admissible inputs). The original definition of the ESP takes into account only the range of admissible inputs, not the probability distribution of the input process. In practice, however, it is this distribution that determines the admissible range of spectral radius for almost all input sequences—which are the practically relevant ones, not the “pathological” ones, which destroy the ESP but occur with zero probability [42]. Therefore, a more useful definition of the ESP is proposed considering the probability of occurrence for admissible inputs [42]: a network satisfies the echo state property with respect to a stochastic process (where the random variables take values in a set ) if with probability one, and compatible with , it holds that . The definition of ESP for a specific input signal that respects the nature of the expected input signals in more detail is suggested by Manjunath and Jaeger [44]: a network is said to have the echo state property with respect to an input sequence if , , and , then for all .

The ESP is connected to the spectral properties of the weight matrix , and some work has been devoted to stating and refining sufficient/necessary conditions for the ESP of the standard ESN [5, 4244]. A rather restrictive sufficient condition for the ESP was given by Jaeger [5] as , where denotes the maximum singular value of . Since this condition is too restrictive and the input is washed out very fast, it is not commonly used in practice. A less restrictive sufficient condition known to date is that is diagonally Schur stable [42, 43]. More recently, Manjunath and Jaeger [44] provided an improved formulation of the sufficient condition for the ESP linked to an input: , where , and is the indicator function that is 1 when its argument is true, and 0 otherwise. In the presence of zero input, a necessary condition for the ESP is , where denotes the spectral radius of [5]. However, for nonzero input signals, the condition is neither sufficient nor necessary for the ESP [42].

Conditions for the ESP used in the literature typically fail to properly account for the effects of driving input signals, often limiting the potentialities of the RC approach. Gallicchio [48] introduced an empirical ESP index that enables analysis of the stability regimes of reservoirs:where is the number of randomly generated initial states, the first and the last time step used for the calculation of the ESP index are denoted by and , respectively, and represents the Euclidean distance between two vectors and . A process on a state space X is defined as follows:where .

2.3. Bifurcation Analysis in ESNs

The dynamical properties of ESNs have been investigated using bifurcation theory; it has been considered the standard method to examine the qualitative changes of the dynamical systems such as phase transitions and instabilities [49]. Yildiz et al. [42] used bifurcation analysis to prove that the spectral radius condition is not a necessary condition for ESP under the zero input environment [42]. With zero input , the origin becomes a trivial fixed point with a sigmoid activation function. Since stationary origin state is compatible with zero input, the problem is reduced to the stability and existence problem of additional fixed points, which can be analyzed with bifurcation theory in an autonomous dynamical system. In two-dimensional systems consisting of two nodes, the weight matrix should be in the stable triangle region [50]. For convenience, they assumed the one component in the weight matrix zero . By fixing , where is positive, new fixed points emerge if and disappear if . It is called degenerate bifurcation which generates two more fixed points away from the origin in a two-dimensional system. Moreover, increasing determinant with fixed trace induces additional nontrivial fixed points away from the origin. This bifurcation analysis ensures the existence of nontrivial fixed points—at least four—away from the origin which can be either asymptotically stable or unstable. Stabilities of these fixed points in reservoir dynamics should be separated from stabilities in bifurcation analysis.

Using computational simulations, one can easily check the stabilities of these points by observing the basin of points. Yildiz et al. [42] showed the case that there exist three asymptotically stable fixed points including the origin and two fixed points from the degenerate bifurcation and two saddle points from the pitchfork bifurcation. These results can be extended to higher dimensions using appropriate block matrices. However, the bifurcations of the higher dimensional nonautonomous dynamical systems are difficult to solve analytically: in higher dimensions, ESNs can exhibit the Neimar–Sacker bifurcation, indicating fixed points other than the origin can exist at , and the input noise significantly affects the dynamical properties of reservoirs and in turn the training and inference performance of ESNs. The mathematical theory called stochastic bifurcation theory has dealt with this type of dynamical systems perturbed by additive stochastic noise [51]. However, artificial neural networks including ESNs allow the input signal to be mapped into networks through a nonlinear sigmoid-like activation function. This nonlinearity applied to stochastic terms makes analytic approaches much more challenging due to the difficulty of linearly separating the deterministic and stochastic terms. To circumvent these problems, we resort to numerical simulations to examine the behavior of the stochastic and nonlinear dynamical systems.

3. Methods

We refine the literature definitions of ESP [5, 4244] and ESP index [48] to adequately reflect the effects of noisy driving input , where is the input sequence from a compact set U and the additive white Gaussian noise (AWGN). We also provide a sufficient condition of ESP considering the noise.

3.1. Echo State Property upon Noisy Driving Input

Definition 1. (ESP with respect to noisy driving input). A network (with the compactness condition) has the echo state property of tolerance and confidence with respect to and noise process : if for any left infinite noisy input sequence , , and any two state vector sequences compatible with different realizations of , it holds that .

Here, denotes the norm defined on the state space . If , , and , the definition becomes the original one by Jaeger [5].

A sufficient condition for the ESP with respect to noisy driving input is given by following proposition—sufficient condition for ESP with respect to noisy driving input. For a standard ESN model in equation (2), if the following conditions are satisfied:(i) and are different realizations of the noise process;(ii);(iii),

the ESN has ESP of tolerance ϵ and confidence γ with respect to noisy driving input. The proof of the proposition is presented in Appendix A. When the probability distribution of the noise is given, the regime of spectral properties of and where ESP holds can be calculated using the given proposition.

Finally, we define the ESP index noisy driving input to empirically assess the property, which leads to:where is a process defined using the instead of (see equation (4)).

3.2. Bifurcation Analysis

For an ESN that satisfies the ESP, different reservoir states induced by the same driving input from a compact set U should asymptotically converge. If the input is given zero (i.e., ), the reservoir starting from any initial state should converge to the origin which is a unique global stable fixed point. The stability and uniqueness of the fixed point are analytically provable, yet the analytical approaches seem to be challenging in the case of the nonlinear mapping of input signal containing noise because the problem of separating the stochastic part from the deterministic term is not easy to solve. Instead, we use numerical simulations to examine the nonlinear nonautonomous dynamical properties of the ESN upon noisy driving input.

The ESNs upon a driving input with white Gaussian noise evolve following the relation:where is an internal weight matrix, is an input matrix, is reservoir state, and is the K-dimensional vector whose component is a independent white Gaussian noise. The input matrix is fixed to when noise is given for convenience, and without input. As discussed above, if there is no input (i.e. ), the ESNs with ESP will converge to the origin. Reservoir states were updated at each step following equation (6). After one update, all states are bounded within since the activation function is bounded. Hence, initial states within are enough to examine the full dynamics of the reservoirs. In one-dimensional ESNs, 50 initial states within with equal intervals randomly chosen from the uniform distribution and the reservoirs were updated for 300 steps. In two-dimensional ESNs, 100 initial states are chosen randomly in —each component is extracted from a uniform distribution —and updated for 300 steps. Increasing the spectral radius of (by multiplying constants), we observe the stability in ESNs to track the ESP. In the simulations, the input noise was given in the same way as the MNIST classification task (Section 3.3).

3.3. Handwritten Digit Classification Tasks

The modified National Institute of Standards and Technology (MNIST) image database for handwritten digits (“0,” “1,” …, and “9”) is used to test the performances of the standard ESN. The training and test accuracies are obtained using different numbers of examples: the sizes of the training sets vary from 120 to 240, …, 30,000, and the test sets are one-sixth of the corresponding training set size, i.e., from 20 to 40, …, 5,000 examples. Each training or testing set contains an equal number of samples from each digit. The pixel dimension of a digit number image was 28 × 28 = 784 pixels. The AWGN was added to each image with different signal-to-noise ratios (SNRs). The SNR was defined as , where is the pixel intensity of the pixel () in the image without noise, and denotes a white Gaussian noise with a mean of zero. The variance of is determined by the given SNR value. Pixel intensities of the noisy image are given as follows:

Each image is transformed into 28 temporal driving input signals by presenting the normalized intensities of 28 pixels in a column at each time step, i.e., element () of the input signal at the time step is given by . The readout layer consists of 10 simple neurons with a linear activation function, where the output signal of readout neuron represents the likelihood that the input image at time t belongs to class D. We train the output to copy a target signal , where if the current image at time t belongs to class D, otherwise. For the test, class D with the highest probability is chosen.

4. Results

Figure 1 exhibits the effect of noise on the handwritten digit classification accuracy using the standard ESN model (see equation (2)). For each SNR level, the training and test accuracies converged as the training/test dataset size increased (left side of Figure 1(a); training and test accuracies colored cyan and blue, respectively), where the accuracies declined as the SNR level decreased (i.e., increasing AWGN levels). The neuronal activities in the reservoir fluctuated more frequently as the SNR decreased (Figure 1(a)) and these caused the decrease of the mean absolute value of cross-correlations with a lag of zero () between all pairs of neuronal activities in the reservoir (Figure 1(b)). The dependence of the training and test accuracies on the SNR was nonlinear, as displayed in Figure 1(c): both the training and test accuracies drastically dropped at SNR < ∼1.0.

The short-term memory capacity (MC) of the reservoir, which is the primary measure of the network’s ability to store the past input information [52], nonlinearly decreased upon reducing the SNR level (i.e., promoting AWGN) (Figure 2(a)). The MC is the sum of -delay short-term memory capacity () against SNR levels (i.e., ), where the means the coefficient of determination () of the linear regression with the reservoir state vector to predict driving input signal with a delay of ; was set to 20 in Figure 2(a) and each of values against SNR levels are displayed in Appendix B. The mapping score () of the driving input signal also decreased by promoting the AWGN levels (Figure 2(b)). The score quantifies how much information of the driving input signal is mapped to the reservoir against noise by comparing the reservoir state induced by the driving inputs with and without the noise. It is defined as , where D denotes the Euclidean distance and and mean the state vector by the input sequence without noise () and that by the noisy input sequence (), respectively. The trends of these two information processing indices of the reservoir are in good agreement with the deterioration pattern of the classification accuracy due to the noise (Figure 1(c)). This indicates that the addition of noise strongly influences the information processing of the reservoir via changing the neural activities, which in turn impacts the computational performance.

As a way to assess the computational capability of the reservoir, the echo state property (ESP) index for noisy driving input which reflects the dynamical changes of the reservoir is devised based on the work by Gallicchio [48]. The ESP index measures the average deviation of noise-driven trajectories induced by random initial states from a noiseless trajectory starting from the zero state (see Section 3.1 for a formal definition). It thus captures the existence of the ESP: intuitively, an ESP index value close to zero strongly suggests that the ESN possesses the ESP, while a larger value means the reservoir is far from having the ESP. Figure 3 displays the ESP index and test accuracy for noisy driving input as functions of spectral radius and input scaling at different noise conditions—no addition of noise and adding the AWGN with SNR from 4.0 to 0.1. The reservoirs were generated by varying spectral radii and input scaling values as follows: the elements of the input matrix were randomly sampled from a uniform distribution on the interval , where represents the input scaling. The internal weight matrix was randomly generated so that 1% of the elements are nonzero, and these nonzero values are uniformly sampled from the range . Then, was rescaled to achieve the desired spectral radius. For the case without noise (Figure 3(a)), both the ESP index for noisy driving input and the original index by Gallicchio [48] generally agree with the distribution of the test accuracies on the spectral radius-input scaling plane: the index value becomes higher (i.e., less tendency to have ESP) as the spectral radius increases and the input scaling decreases (i.e., towards the top-left corner on the plane). However, when increasing the noise level (Figures 3(b)3(d)), the ESP indices for noisy input well capture the collapse of computational properties of the reservoirs (i.e., deterioration of test accuracies), while the original index did not capture the effect of noise on the computation.

The detailed dynamical properties of ESNs upon noisy driving input are investigated using the one-dimensional (N = 1) and two-dimensional (N = 2) ESN models in agreement with Yildiz’s method for zero input cases [42]. In N = 1 case, is a constant scalar and it is the spectral radius itself. is a trivial fixed point in this system for arbitrary . The stability of a trivial fixed point is determined by the spectral radius; is stable if and unstable if ; additional nonzero stable fixed points emerge for . The evolution of the reservoir states obtained by numerical simulations confirmed these results. Figure 4 exhibits the dynamics of one-dimensional ESNs with zero input and with only the white Gaussian noise. For the zero input case, one-dimensional ESNs converge to trivial fixed point for (first column in Figure 4(a)) or two nonzero fixed points for (second column in Figure 4(a)). The reservoir states manifested a pitchfork bifurcation (last column in Figure 4(a)): the states converged to a stable fixed point for . As spectral radius increases, an unstable fixed point emerges around and two nonzero stable fixed points emerge for (Figure 4(a)). Upon white Gaussian noise, while the overall tendencies were similar to the zero input case, the reservoir states fluctuated by the noise: different initial states, in turn, converged to an asymptotically same trajectory for all noise levels (SNR = 4.0, 1.0, and 0.1), and the degree of fluctuation increased with the noise level (Figures 4(b)4(d)). In the cases of (first column in Figures 4(b)4(d)), reservoir states fluctuate around and the degree of fluctuation increases as the noise level increases. In the cases of (second column in Figures 4(b)4(d)), states flow into the vicinity of one of the two original fixed points. As the noise level increases (or SNR decreases), noise induces more blurring of fixed points, while we can still see the footprints of fixed points observed in the dynamics of ESNs without input. The last 100 steps are used to test the convergence of the reservoir states upon 10 different white Gaussian noise levels (third column in Figures 4(b)4(d)). For given spectral radius , states are colored with dark green if states are bounded within tolerance , i.e., for . Otherwise, states are colored light green if they are not bounded. In a noisy environment, states can deviate from the origin (i.e., ) even at .

We extend our analysis into two-dimensional ESNs where two neighboring nodes (N = 2) are connected. In N = 2 case, is a 2 × 2 matrix and the reservoir state in each step k can be represented by a two-dimensional vector . Reflecting that the stabilities of N = 2 cases are not solely determined by , the models exhibit various types of bifurcations compared to the N = 1 systems. In addition to constant rescaling, there are more bifurcation parameters such as [42]. Here, we analyze the system which exhibits a Hopf bifurcation under a noise-free environment. In a Hopf bifurcation, a stable fixed point becomes unstable, and a limit cycle arises around the fixed point as the bifurcation parameter crosses the critical value. Internal weight matrix was used and spectral radius is rescaled by multiplying the constant . The reservoir states in two-dimensional ESNs without noise converge to the trivial fixed point for as in the one-dimensional case (first column in Figure 5(a)). However, the reservoir states in two-dimensional ESNs with oscillate (second column in Figure 5(a)). The last 100 steps from different initial states converge (third column in Figure 5(a)). They exhibit a Hopf bifurcation; a stable fixed point becomes unstable, and a limit cycle arises around the fixed point as crosses 1. For spectral radii , the states are distributed around the origin, implying the existence of a limit cycle. This result indicates that ESP does not hold if . In the case of white Gaussian noise with different levels (i.e., SNR = 4.0, 1.0, 0.1), the reservoir states in two-dimensional ESNs start to fluctuate (Figures 5(b)5(d)). For (first column in Figures 5(b)5(d)), reservoir states fluctuate around the fixed point and the degree of fluctuation increases as the noise level increases. In the cases of (second column in Figures 5(b)5(d)), the states quickly oscillate within the bounded range. The last 100 steps upon 10 different white Gaussian noise levels are examined to confirm the convergence of the reservoir states (third column in Figures 5(b)5(d)). For given spectral radius , states are colored with dark green if states are bounded within tolerance ; for (third column in Figures 5(b)5(d)). Otherwise, states are colored light green if they are not bounded. As in the case of the one-dimensional model, ESNs for some initial states can deviate from the origin (i.e., ) even for as the noise level increases.

Results from the bifurcation analysis are related to the ESP index for noisy driving input (equation (5)) and the original index by Gallicchio [48]. The reservoir states in bifurcation analysis (i.e., last 250 time steps ()) in Figures 4 and 5 were used to compute the ESP indices. As expected, both ESP indices well agreed with the case of zero input (i.e., no addition of white Gaussian noise) (Figure 6(a)). The values become drastically increased as the spectral radius is promoted beyond the unity (i.e., ), whereby ESP is easily destroyed. In the presence of noise, only the ESP index for noisy driving input could capture the deterioration of the ESP with the noise level for ; the original ESP index remained almost zero for this regime (Figures 6(b)–6(d)). For , both indices increase with the spectral radius.

5. Discussion and Conclusions

Despite the importance of the empirical assessment of the ESP for the logical design and optimal operation of reservoir computers, the commonly used conditions for the ESP do not explicitly account for the interference from the input noise which can significantly affect their performance. To provide useful information about the empirical and analytical assessment of the computational capabilities of ESNs, a series of extensive numerical simulations and analytic characterization of the ESP were performed.

The analysis began with the comparison of the primary dynamical properties of the standard ESN model with different input noise levels (i.e., no addition of noise and adding the AWGN with SNR from 4.0 to 0.1). The significant and distinct relationship between the dynamical measures and the MNIST classification accuracy indicated that the noise-induced dynamical changes and the computational capability of the reservoir are fundamentally intertwined. We then provided the ESP index for noisy driving input, reflecting these dynamical changes based on the work by Gallicchio [48], to help easily assess the computational capability of ESNs in practical applications. We have extended the bifurcation analysis of Yildiz et al. [42] using the one-dimensional and two-dimensional ESN models, by taking into account the effects of AWGN on reservoir dynamics to explicate the underlying physics of the noise effects and to confirm the validity of the proposed ESP index. For both cases in one-dimensional ESN (Figure 4) and two-dimensional ESN (Figure 5), the convergence of ESN bifurcates with an increasing spectral radius of the internal weight matrix : pitchfork bifurcations and Hopf bifurcations were observed and the origin was a unique fixed point for small without noise. However, the fixed point was distracted by the AWGN (Figures 4 and 5); this means that all state vectors driven by any input sequence from a compact set U would not asymptotically converge to the same state due to the interference of the noise. For both one- and two-dimensional ESNs, when they have a fixed point at , the reservoir states were bounded within the tolerance of without AWGN (Figures 4(a) and 5(a)), yet some reservoir states deviated from the fixed point under SNR = 4, 1, 0.1 (Figures 4(b)4(d) and 5(b)5(d)). This loss of ESP by the AWGN significantly changed the dynamical properties and information processing of the reservoir as captured by the neural correlations (Figure 1(b)), the memory capacity (Figure 2(a)), and the mapping of the noisy input to the reservoir (Figure 2(b)), and these, in turn, led to the collapse of computational capability as indicated by the deterioration of MNIST classification accuracy against noise (Figure 1(c)). The proposed ESP index was defined as the average deviation of noise-driven trajectories from a noiseless trajectory (see Section 3.1 for a formal definition), and the index well characterized the collapse of the computational properties of the reservoirs (Figures 3 and 6). While the original ESP index [48] was designed for noise-free conditions, the proposed ESP index considered the changes in the dynamical properties due to the input noise (Figures 1, 2, 4, and 5) to better characterize the computational capabilities of ESNs, especially in dealing with real-world problems where the interference from the input noise significantly affect the performance of ESNs. As exhibited in Figures 4 and 5, the reservoir states are strongly entrained by the noise and the strongly entrained fluctuating patterns of the reservoir states would wash out their transient initial states (Figures 4(b)4(d) and 5(b)5(d)). While the original index does not differentiate these effects of noise from the convergence of relevant information of the driving input, the newly proposed ESP index can circumvent these problems by incorporating both trajectories of the reservoir states induced by driving inputs with and without noise in defining the index (equation (5)).

Our work provides a framework to understand ESP in the context of a noisy driving input—the proposed definitions of the ESP and ESP index may enable the empirical assessments of the computational capabilities of the reservoirs for noisy input conditions. This may promote the validity, reliability, and utility of reservoir computers for real-world machine-learning applications.

Appendix

A. The Proof of Proposition for Defining the Echo State Property upon Noisy Driving Input

Proof. For any two state vector sequences and compatible with and , respectively.Applying the inequality above, we obtainTherefore, if , then .

B. Short-Term Memory Capacity for Each Delay

Short-term memory capacity for each delay is provided in Figure 7.

Data Availability

The codes used to support the findings of this study have been deposited in the GitHub repository (https://github.com/LCNP-KIST/Echo-state-property-upon-noisy-driving-input).

Disclosure

The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Conflicts of Interest

The authors declare that they have no conflicts of interest regarding the publication of this paper.

Authors’ Contributions

K.H. conceptualized the study. J.W., H.K., and K.H. contributed to modeling and simulations. J.W., H.K., S.H.K., and K.H. performed analysis. K.H., H.K., J.W., and S.H.K. wrote and prepared the original draft. K.H. wrote, reviewed, and edited the article. K.H. performed supervision. All authors have read and agreed to the published version of the manuscript. Junhyuk Woo and Hyeongmo Kim contributed equally to this work.

Acknowledgments

This research was funded by the Korea Institute of Science and Technology (KIST) Institutional Program (Project nos. 2E32921 and 2E32163), National R&D Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (2021M3F3A2A01037808), and Institute of Information and Communications Technology Planning and Evaluation (IITP) grant funded by the Korea government (MSIT) (no. 2022-0-00159, Development of AI edge devices for the prediction of senior cardio-cerebrovascular disease and dementia).