1
|
Ehlers PJ, Nurdin HI, Soh D. Improving the performance of echo state networks through state feedback. Neural Netw 2025; 184:107101. [PMID: 39778290 DOI: 10.1016/j.neunet.2024.107101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2023] [Revised: 07/16/2024] [Accepted: 12/23/2024] [Indexed: 01/11/2025]
Abstract
Reservoir computing, using nonlinear dynamical systems, offers a cost-effective alternative to neural networks for complex tasks involving processing of sequential data, time series modeling, and system identification. Echo state networks (ESNs), a type of reservoir computer, mirror neural networks but simplify training. They apply fixed, random linear transformations to the internal state, followed by nonlinear changes. This process, guided by input signals and linear regression, adapts the system to match target characteristics, reducing computational demands. A potential drawback of ESNs is that the fixed reservoir may not offer the complexity needed for specific problems. While directly altering (training) the internal ESN would reintroduce the computational burden, an indirect modification can be achieved by redirecting some output as input. This feedback can influence the internal reservoir state, yielding ESNs with enhanced complexity suitable for broader challenges. In this paper, we demonstrate that by feeding some component of the reservoir state back into the network through the input, we can drastically improve upon the performance of a given ESN. We rigorously prove that, for any given ESN, feedback will almost always improve the accuracy of the output. For a set of three tasks, each representing different problem classes, we find that with feedback the average error measures are reduced by 30%-60%. Remarkably, feedback provides at least an equivalent performance boost to doubling the initial number of computational nodes, a computationally expensive and technologically challenging alternative. These results demonstrate the broad applicability and substantial usefulness of this feedback scheme.
Collapse
Affiliation(s)
- Peter J Ehlers
- Wyant College of Optical Sciences, University of Arizona, Tuscon, AZ, USA.
| | - Hendra I Nurdin
- School of Electrical Engineering and Telecommunications, University of New South Wales, Sydney, Australia.
| | - Daniel Soh
- Wyant College of Optical Sciences, University of Arizona, Tuscon, AZ, USA.
| |
Collapse
|
2
|
Ehlers PJ, Nurdin HI, Soh D. Stochastic reservoir computers. Nat Commun 2025; 16:3070. [PMID: 40157931 PMCID: PMC11955002 DOI: 10.1038/s41467-025-58349-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2024] [Accepted: 03/20/2025] [Indexed: 04/01/2025] Open
Abstract
Reservoir computing is a form of machine learning that utilizes nonlinear dynamical systems to perform complex tasks in a cost-effective manner when compared to typical neural networks. Recent advancements in reservoir computing, in particular quantum reservoir computing, use reservoirs that are inherently stochastic. In this paper, we investigate the universality of stochastic reservoir computers which use the probabilities of each stochastic reservoir state as the readout instead of the states themselves. This allows the number of readouts to scale exponentially with the size of the reservoir hardware, offering the advantage of compact device size. We prove that classes of stochastic echo state networks form universal approximating classes. We also investigate the performance of two practical examples in classification and chaotic time series prediction. While shot noise is a limiting factor, we show significantly improved performance compared to a deterministic reservoir computer with similar hardware when noise effects are small.
Collapse
Affiliation(s)
- Peter J Ehlers
- Wyant College of Optical Sciences, University of Arizona, Tuscon, AZ, US.
| | - Hendra I Nurdin
- School of Electrical Engineering and Telecommunications, University of New South Wales, Sydney, Australia
| | - Daniel Soh
- Wyant College of Optical Sciences, University of Arizona, Tuscon, AZ, US.
| |
Collapse
|
3
|
Picco E, Jaurigue L, Lüdge K, Massar S. Efficient optimisation of physical reservoir computers using only a delayed input. COMMUNICATIONS ENGINEERING 2025; 4:3. [PMID: 39827312 PMCID: PMC11742992 DOI: 10.1038/s44172-025-00340-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/18/2024] [Accepted: 01/07/2025] [Indexed: 01/22/2025]
Abstract
Reservoir computing is a machine learning algorithm for processing time dependent data which is well suited for experimental implementation. Tuning the hyperparameters of the reservoir is a time-consuming task that limits is applicability. Here we present an experimental validation of a recently proposed optimisation technique in which the reservoir receives both the input signal and a delayed version of the input signal. This augments the memory of the reservoir and improves its performance. It also simplifies the time-consuming task of hyperparameter tuning. The experimental system is an optoelectronic setup based on a fiber delay loop and a single nonlinear node. It is tested on several benchmark tasks and reservoir operating conditions. Our results demonstrate the effectiveness of the delayed input method for experimental implementation of reservoir computing systems.
Collapse
Affiliation(s)
- Enrico Picco
- Laboratoire d'Information Quantique CP224, Université libre de Bruxelles (ULB), Av. F. D. Roosevelt 50, 1050, Bruxelles, Belgium
| | - Lina Jaurigue
- Institute of Physics, Technische Universität Ilmenau, Weimarer Straße 25, 98693, Ilmenau, Germany
| | - Kathy Lüdge
- Institute of Physics, Technische Universität Ilmenau, Weimarer Straße 25, 98693, Ilmenau, Germany
| | - Serge Massar
- Laboratoire d'Information Quantique CP224, Université libre de Bruxelles (ULB), Av. F. D. Roosevelt 50, 1050, Bruxelles, Belgium.
| |
Collapse
|
4
|
Zhong D, Wu Q, Zhang J, Wang T, Chen Y, Zeng H, Ren Z, Wang Y, Qiu C. Exploration of a brain-inspired photon reservoir computing network based on quantum-dot spin-VCSELs. OPTICS EXPRESS 2024; 32:28441-28461. [PMID: 39538661 DOI: 10.1364/oe.527428] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 07/15/2024] [Indexed: 11/16/2024]
Abstract
Based on small-world network theory, we have developed a brain-inspired photonic reservoir computing (RC) network system utilizing quantum dot spin-vertical-cavity surface-emitting lasers (QD spin-VCSELs) and formulated a comprehensive theoretical model for it. This innovative network system comprises input layers, a reservoir network layer, and output layers. The reservoir network layer features four distinct reservoir modules that are asymmetrically coupled. Each module is represented by a QD spin-VCSEL, characterized by optical feedback and optical injection. Within these modules, four chaotic polarization components, emitted from both the ground and excited states of the QD Spin-VCSEL, form four distinct reservoirs through a process of asymmetric coupling. Moreover, these components, when emitted by the ground and excited states of a driving QD spin-VCSEL within a specific parameter space, act as targets for prediction. Delving further, we investigated the correlation between various system parameters, such as the sampling period, the interval between virtual nodes, the strengths of optical injection and feedback, frequency detuning, and the predictive accuracy of each module's four photonic RCs concerning the four designated predictive targets. We also examined how these parameters influence the memory storage capabilities of the four photonics RCs within each module. Our findings indicate that when a module receives coupling injections from more than two other modules, and an RC within this module is also subject to coupling injections from over two other RCs, the system displays reduced predictive errors and enhanced memory storage capacities when the system parameters are fixed. Namely, the superior performance of the reservoir module in predictive accuracy and memory capacities follows from its complex interaction with multiple light injections and coupling injections, with its three various PCs benefiting from three, two, and one coupling injections respectively. Conversely, variations in optical injection and feedback strength, as well as frequency detuning, introduce only marginal fluctuations in the predictive errors across the four photonics RCs within each module and exert minimal impact on the memory storage capacity of individual photonics RCs within the modules. Our investigated results contribute to the development of photonic reservoir computing towards fast response biological neural networks.
Collapse
|
5
|
Hart JD. Attractor reconstruction with reservoir computers: The effect of the reservoir's conditional Lyapunov exponents on faithful attractor reconstruction. CHAOS (WOODBURY, N.Y.) 2024; 34:043123. [PMID: 38579149 DOI: 10.1063/5.0196257] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/05/2024] [Accepted: 03/22/2024] [Indexed: 04/07/2024]
Abstract
Reservoir computing is a machine learning framework that has been shown to be able to replicate the chaotic attractor, including the fractal dimension and the entire Lyapunov spectrum, of the dynamical system on which it is trained. We quantitatively relate the generalized synchronization dynamics of a driven reservoir during the training stage to the performance of the trained reservoir computer at the attractor reconstruction task. We show that, in order to obtain successful attractor reconstruction and Lyapunov spectrum estimation, the maximal conditional Lyapunov exponent of the driven reservoir must be significantly more negative than the most negative Lyapunov exponent of the target system. We also find that the maximal conditional Lyapunov exponent of the reservoir depends strongly on the spectral radius of the reservoir adjacency matrix; therefore, for attractor reconstruction and Lyapunov spectrum estimation, small spectral radius reservoir computers perform better in general. Our arguments are supported by numerical examples on well-known chaotic systems.
Collapse
Affiliation(s)
- Joseph D Hart
- U.S. Naval Research Laboratory, Code 5675, Washington, DC 20375, USA
| |
Collapse
|
6
|
Nazerian A, Nathe C, Hart JD, Sorrentino F. Synchronizing chaos using reservoir computing. CHAOS (WOODBURY, N.Y.) 2023; 33:103121. [PMID: 37832520 PMCID: PMC10576628 DOI: 10.1063/5.0161076] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/08/2023] [Accepted: 09/14/2023] [Indexed: 10/15/2023]
Abstract
We attempt to achieve complete synchronization between a drive system unidirectionally coupled with a response system, under the assumption that limited knowledge on the states of the drive is available at the response. Machine-learning techniques have been previously implemented to estimate the states of a dynamical system from limited measurements. We consider situations in which knowledge of the non-measurable states of the drive system is needed in order for the response system to synchronize with the drive. We use a reservoir computer to estimate the non-measurable states of the drive system from its measured states and then employ these measured states to achieve complete synchronization of the response system with the drive.
Collapse
Affiliation(s)
- Amirhossein Nazerian
- Mechanical Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131, USA
| | - Chad Nathe
- Mechanical Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131, USA
| | | | - Francesco Sorrentino
- Mechanical Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131, USA
| |
Collapse
|
7
|
Miao W, Narayanan V, Li JS. Interpretable Design of Reservoir Computing Networks Using Realization Theory. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:6379-6389. [PMID: 34982700 PMCID: PMC10567084 DOI: 10.1109/tnnls.2021.3136495] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
The reservoir computing networks (RCNs) have been successfully employed as a tool in learning and complex decision-making tasks. Despite their efficiency and low training cost, practical applications of RCNs rely heavily on empirical design. In this article, we develop an algorithm to design RCNs using the realization theory of linear dynamical systems. In particular, we introduce the notion of α -stable realization and provide an efficient approach to prune the size of a linear RCN without deteriorating the training accuracy. Furthermore, we derive a necessary and sufficient condition on the irreducibility of the number of hidden nodes in linear RCNs based on the concepts of controllability and observability from systems theory. Leveraging the linear RCN design, we provide a tractable procedure to realize RCNs with nonlinear activation functions. We present numerical experiments on forecasting time-delay systems and chaotic systems to validate the proposed RCN design methods and demonstrate their efficacy.
Collapse
|
8
|
Chauhan S, Mandal S, Yadav V, Jaiswal PK, Priya M, Shrimali MD. Machine learning based prediction of phase ordering dynamics. CHAOS (WOODBURY, N.Y.) 2023; 33:061103. [PMID: 37327496 DOI: 10.1063/5.0156611] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/01/2023] [Accepted: 05/31/2023] [Indexed: 06/18/2023]
Abstract
Machine learning has proven exceptionally competent in numerous applications of studying dynamical systems. In this article, we demonstrate the effectiveness of reservoir computing, a famous machine learning architecture, in learning a high-dimensional spatiotemporal pattern. We employ an echo-state network to predict the phase ordering dynamics of 2D binary systems-Ising magnet and binary alloys. Importantly, we emphasize that a single reservoir can be competent enough to process the information from a large number of state variables involved in the specific task at minimal computational training cost. Two significant equations of phase ordering kinetics, the time-dependent Ginzburg-Landau and Cahn-Hilliard-Cook equations, are used to depict the result of numerical simulations. Consideration of systems with both conserved and non-conserved order parameters portrays the scalability of our employed scheme.
Collapse
Affiliation(s)
- Swati Chauhan
- Department of Physics, Central University of Rajasthan, Ajmer, Rajasthan 305 817, India
| | - Swarnendu Mandal
- Department of Physics, Central University of Rajasthan, Ajmer, Rajasthan 305 817, India
| | - Vijay Yadav
- Department of Physics, Indian Institute of Technology Jodhpur, Jodhpur, Rajasthan 342 030, India
| | - Prabhat K Jaiswal
- Department of Physics, Indian Institute of Technology Jodhpur, Jodhpur, Rajasthan 342 030, India
| | - Madhu Priya
- Department of Physics, Birla Institute of Technology Mesra, Ranchi, Jharkhand 835 215, India
| | - Manish Dev Shrimali
- Department of Physics, Central University of Rajasthan, Ajmer, Rajasthan 305 817, India
| |
Collapse
|
9
|
Shi L, Yan Y, Wang H, Wang S, Qu SX. Predicting chaotic dynamics from incomplete input via reservoir computing with (D+1)-dimension input and output. Phys Rev E 2023; 107:054209. [PMID: 37329034 DOI: 10.1103/physreve.107.054209] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2022] [Accepted: 04/20/2023] [Indexed: 06/18/2023]
Abstract
Predicting future evolution based on incomplete information of the past is still a challenge even though data-driven machine learning approaches have been successfully applied to forecast complex nonlinear dynamics. The widely adopted reservoir computing (RC) can hardly deal with this since it usually requires complete observations of the past. In this paper, a scheme of RC with (D+1)-dimension input and output (I/O) vectors is proposed to solve this problem, i.e., the incomplete input time series or dynamical trajectories of a system, in which certain portion of states are randomly removed. In this scheme, the I/O vectors coupled to the reservoir are changed to (D+1)-dimension, where the first D dimensions store the state vector as in the conventional RC, and the additional dimension is the corresponding time interval. We have successfully applied this approach to predict the future evolution of the logistic map and Lorenz, Rössler, and Kuramoto-Sivashinsky systems, where the inputs are the dynamical trajectories with missing data. The dropoff rate dependence of the valid prediction time (VPT) is analyzed. The results show that it can make forecasting with much longer VPT when the dropoff rate θ is lower. The reason for the failure at high θ is analyzed. The predictability of our RC is determined by the complexity of the dynamical systems involved. The more complex they are, the more difficult they are to predict. Perfect reconstructions of chaotic attractors are observed. This scheme is a pretty good generalization to RC and can treat input time series with regular and irregular time intervals. It is easy to use since it does not change the basic architecture of conventional RC. Furthermore, it can make multistep-ahead prediction just by changing the time interval in the output vector into a desired value, which is superior to conventional RC that can only do one-step-ahead forecasting based on complete regular input data.
Collapse
Affiliation(s)
- Lufa Shi
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710119, China
| | - Youfang Yan
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710119, China
| | - Hengtong Wang
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710119, China
| | - Shengjun Wang
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710119, China
| | - Shi-Xian Qu
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710119, China
| |
Collapse
|
10
|
Patel D, Ott E. Using machine learning to anticipate tipping points and extrapolate to post-tipping dynamics of non-stationary dynamical systems. CHAOS (WOODBURY, N.Y.) 2023; 33:023143. [PMID: 36859201 DOI: 10.1063/5.0131787] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/24/2022] [Accepted: 01/31/2023] [Indexed: 06/18/2023]
Abstract
The ability of machine learning (ML) models to "extrapolate" to situations outside of the range spanned by their training data is crucial for predicting the long-term behavior of non-stationary dynamical systems (e.g., prediction of terrestrial climate change), since the future trajectories of such systems may (perhaps after crossing a tipping point) explore regions of state space which were not explored in past time-series measurements used as training data. We investigate the extent to which ML methods can yield useful results by extrapolation of such training data in the task of forecasting non-stationary dynamics, as well as conditions under which such methods fail. In general, we find that ML can be surprisingly effective even in situations that might appear to be extremely challenging, but do (as one would expect) fail when "too much" extrapolation is required. For the latter case, we show that good results can potentially be obtained by combining the ML approach with an available inaccurate conventional model based on scientific knowledge.
Collapse
Affiliation(s)
- Dhruvit Patel
- The Institute for Research in Electronics and Applied Physics, University of Maryland, College Park, Maryland 26742, USA
| | - Edward Ott
- The Institute for Research in Electronics and Applied Physics, University of Maryland, College Park, Maryland 26742, USA
| |
Collapse
|
11
|
Goldmann M, Mirasso CR, Fischer I, Soriano MC. Learn one size to infer all: Exploiting translational symmetries in delay-dynamical and spatiotemporal systems using scalable neural networks. Phys Rev E 2022; 106:044211. [PMID: 36397530 DOI: 10.1103/physreve.106.044211] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2021] [Accepted: 09/27/2022] [Indexed: 06/16/2023]
Abstract
We design scalable neural networks adapted to translational symmetries in dynamical systems, capable of inferring untrained high-dimensional dynamics for different system sizes. We train these networks to predict the dynamics of delay-dynamical and spatiotemporal systems for a single size. Then, we drive the networks by their own predictions. We demonstrate that by scaling the size of the trained network, we can predict the complex dynamics for larger or smaller system sizes. Thus, the network learns from a single example and by exploiting symmetry properties infers entire bifurcation diagrams.
Collapse
Affiliation(s)
- Mirko Goldmann
- Instituto de Física Interdisciplinar y Sistemas Complejos (IFISC, UIB-CSIC), Campus Universitat de les Illes Balears E-07122, Palma de Mallorca, Spain
| | - Claudio R Mirasso
- Instituto de Física Interdisciplinar y Sistemas Complejos (IFISC, UIB-CSIC), Campus Universitat de les Illes Balears E-07122, Palma de Mallorca, Spain
| | - Ingo Fischer
- Instituto de Física Interdisciplinar y Sistemas Complejos (IFISC, UIB-CSIC), Campus Universitat de les Illes Balears E-07122, Palma de Mallorca, Spain
| | - Miguel C Soriano
- Instituto de Física Interdisciplinar y Sistemas Complejos (IFISC, UIB-CSIC), Campus Universitat de les Illes Balears E-07122, Palma de Mallorca, Spain
| |
Collapse
|
12
|
Kanno K, Haya AA, Uchida A. Reservoir computing based on an external-cavity semiconductor laser with optical feedback modulation. OPTICS EXPRESS 2022; 30:34218-34238. [PMID: 36242440 DOI: 10.1364/oe.460016] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/24/2022] [Accepted: 08/22/2022] [Indexed: 06/16/2023]
Abstract
We numerically and experimentally investigate reservoir computing based on a single semiconductor laser with optical feedback modulation. In this scheme, an input signal is injected into a semiconductor laser via intensity or phase modulation of the optical feedback signal. We perform a chaotic time-series prediction task using the reservoir and compare the performances of intensity and phase modulation schemes. Our results indicate that the feedback signal of the phase modulation scheme outperforms that of the intensity modulation scheme. Further, we investigate the performance dependence of reservoir computing on parameter values and observe that the prediction error improves for large injection currents, unlike the results in a semiconductor laser with an optical injection input. The physical origin of the superior performance of the phase modulation scheme is analyzed using external cavity modes obtained from steady-state analysis in the phase space. The analysis indicates that high-dimensional mapping can be achieved from the input signal to the trajectory of the response laser output by using phase modulation of the feedback signal.
Collapse
|
13
|
Carroll TL, Hart JD. Time shifts to reduce the size of reservoir computers. CHAOS (WOODBURY, N.Y.) 2022; 32:083122. [PMID: 36049918 DOI: 10.1063/5.0097850] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/02/2022] [Accepted: 07/22/2022] [Indexed: 06/15/2023]
Abstract
A reservoir computer is a type of dynamical system arranged to do computation. Typically, a reservoir computer is constructed by connecting a large number of nonlinear nodes in a network that includes recurrent connections. In order to achieve accurate results, the reservoir usually contains hundreds to thousands of nodes. This high dimensionality makes it difficult to analyze the reservoir computer using tools from the dynamical systems theory. Additionally, the need to create and connect large numbers of nonlinear nodes makes it difficult to design and build analog reservoir computers that can be faster and consume less power than digital reservoir computers. We demonstrate here that a reservoir computer may be divided into two parts: a small set of nonlinear nodes (the reservoir) and a separate set of time-shifted reservoir output signals. The time-shifted output signals serve to increase the rank and memory of the reservoir computer, and the set of nonlinear nodes may create an embedding of the input dynamical system. We use this time-shifting technique to obtain excellent performance from an opto-electronic delay-based reservoir computer with only a small number of virtual nodes. Because only a few nonlinear nodes are required, construction of a reservoir computer becomes much easier, and delay-based reservoir computers can operate at much higher speeds.
Collapse
Affiliation(s)
| | - Joseph D Hart
- U.S. Naval Research Laboratory, Washington D.C. 20375, USA
| |
Collapse
|
14
|
Whiteaker B, Gerstoft P. Reducing echo state network size with controllability matrices. CHAOS (WOODBURY, N.Y.) 2022; 32:073116. [PMID: 35907714 DOI: 10.1063/5.0071926] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/17/2021] [Accepted: 06/21/2022] [Indexed: 06/15/2023]
Abstract
Echo state networks are a fast training variant of recurrent neural networks excelling at approximating nonlinear dynamical systems and time series prediction. These machine learning models act as nonlinear fading memory filters. While these models benefit from quick training and low complexity, computation demands from a large reservoir matrix are a bottleneck. Using control theory, a reduced size replacement reservoir matrix is found. Starting from a large, task-effective reservoir matrix, we form a controllability matrix whose rank indicates the active sub-manifold and candidate replacement reservoir size. Resulting time speed-ups and reduced memory usage come with minimal error increase to chaotic climate reconstruction or short term prediction. Experiments are performed on simple time series signals and the Lorenz-1963 and Mackey-Glass complex chaotic signals. Observing low error models shows variation of active rank and memory along a sequence of predictions.
Collapse
Affiliation(s)
- Brian Whiteaker
- Scripps Institution of Oceanography, University of California at San Diego, La Jolla, California 92093-0238, USA
| | - Peter Gerstoft
- Scripps Institution of Oceanography, University of California at San Diego, La Jolla, California 92093-0238, USA
| |
Collapse
|
15
|
Gavrilov A, Loskutov E, Feigin A. Data-driven stochastic model for cross-interacting processes with different time scales. CHAOS (WOODBURY, N.Y.) 2022; 32:023111. [PMID: 35232042 DOI: 10.1063/5.0077302] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/01/2021] [Accepted: 01/20/2022] [Indexed: 06/14/2023]
Abstract
In this work, we propose a new data-driven method for modeling cross-interacting processes with different time scales represented by time series with different sampling steps. It is a generalization of a nonlinear stochastic model of an evolution operator based on neural networks and designed for the case of time series with a constant sampling step. The proposed model has a more complex structure. First, it describes each process by its own stochastic evolution operator with its own time step. Second, it takes into account possible nonlinear connections within each pair of processes in both directions. These connections are parameterized asymmetrically, depending on which process is faster and which process is slower. They make this model essentially different from the set of independent stochastic models constructed individually for each time scale. All evolution operators and connections are trained and optimized using the Bayesian framework, forming a multi-scale stochastic model. We demonstrate the performance of the model on two examples. The first example is a pair of coupled oscillators, with the couplings in both directions which can be turned on and off. Here, we show that inclusion of the connections into the model allows us to correctly reproduce observable effects related to coupling. The second example is a spatially distributed data generated by a global climate model running in the middle 19th century external conditions. In this case, the multi-scale model allows us to reproduce the coupling between the processes which exists in the observed data but is not captured by the model constructed individually for each process.
Collapse
Affiliation(s)
- A Gavrilov
- Institute of Applied Physics of the Russian Academy of Sciences, Nizhny Novgorod 603950, Russia
| | - E Loskutov
- Institute of Applied Physics of the Russian Academy of Sciences, Nizhny Novgorod 603950, Russia
| | - A Feigin
- Institute of Applied Physics of the Russian Academy of Sciences, Nizhny Novgorod 603950, Russia
| |
Collapse
|
16
|
Carroll TL. Optimizing memory in reservoir computers. CHAOS (WOODBURY, N.Y.) 2022; 32:023123. [PMID: 35232031 DOI: 10.1063/5.0078151] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/10/2021] [Accepted: 01/31/2022] [Indexed: 06/14/2023]
Abstract
A reservoir computer is a way of using a high dimensional dynamical system for computation. One way to construct a reservoir computer is by connecting a set of nonlinear nodes into a network. Because the network creates feedback between nodes, the reservoir computer has memory. If the reservoir computer is to respond to an input signal in a consistent way (a necessary condition for computation), the memory must be fading; that is, the influence of the initial conditions fades over time. How long this memory lasts is important for determining how well the reservoir computer can solve a particular problem. In this paper, I describe ways to vary the length of the fading memory in reservoir computers. Tuning the memory can be important to achieve optimal results in some problems; too much or too little memory degrades the accuracy of the computation.
Collapse
Affiliation(s)
- T L Carroll
- US Naval Research Lab, Washington DC 20375, USA
| |
Collapse
|
17
|
Del Frate E, Shirin A, Sorrentino F. Reservoir computing with random and optimized time-shifts. CHAOS (WOODBURY, N.Y.) 2021; 31:121103. [PMID: 34972324 PMCID: PMC8684442 DOI: 10.1063/5.0068941] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/28/2021] [Accepted: 11/24/2021] [Indexed: 06/14/2023]
Abstract
We investigate the effects of application of random time-shifts to the readouts of a reservoir computer in terms of both accuracy (training error) and performance (testing error). For different choices of the reservoir parameters and different "tasks," we observe a substantial improvement in both accuracy and performance. We then develop a simple but effective technique to optimize the choice of the time-shifts, which we successfully test in numerical experiments.
Collapse
Affiliation(s)
- Enrico Del Frate
- Mechanical Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131, USA
| | - Afroza Shirin
- Mechanical Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131, USA
| | - Francesco Sorrentino
- Mechanical Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131, USA
| |
Collapse
|
18
|
Kobayashi MU, Nakai K, Saiki Y, Tsutsumi N. Dynamical system analysis of a data-driven model constructed by reservoir computing. Phys Rev E 2021; 104:044215. [PMID: 34781491 DOI: 10.1103/physreve.104.044215] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2021] [Accepted: 09/16/2021] [Indexed: 11/07/2022]
Abstract
This study evaluates data-driven models from a dynamical system perspective, such as unstable fixed points, periodic orbits, chaotic saddle, Lyapunov exponents, manifold structures, and statistical values. We find that these dynamical characteristics can be reconstructed much more precisely by a data-driven model than by computing directly from training data. With this idea, we predict the laminar lasting time distribution of a particular macroscopic variable of chaotic fluid flow, which cannot be calculated from a direct numerical simulation of the Navier-Stokes equation because of its high computational cost.
Collapse
Affiliation(s)
| | - Kengo Nakai
- Faculty of Marine Technology, Tokyo University of Marine Science and Technology, Tokyo 135-8533, Japan
| | - Yoshitaka Saiki
- Graduate School of Business Administration, Hitotsubashi University, Tokyo 186-8601, Japan
| | - Natsuki Tsutsumi
- Faculty of Commerce and Management, Hitotsubashi University, Tokyo 186-8601, Japan
| |
Collapse
|
19
|
Gao H, Wang A, Wang L, Jia Z, Guo Y, Gao Z, Yan L, Qin Y, Wang Y. 0.75 Gbit/s high-speed classical key distribution with mode-shift keying chaos synchronization of Fabry-Perot lasers. LIGHT, SCIENCE & APPLICATIONS 2021; 10:172. [PMID: 34456335 PMCID: PMC8403675 DOI: 10.1038/s41377-021-00610-w] [Citation(s) in RCA: 26] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/09/2021] [Revised: 08/04/2021] [Accepted: 08/08/2021] [Indexed: 05/31/2023]
Abstract
High-speed physical key distribution is diligently pursued for secure communication. In this paper, we propose and experimentally demonstrate a scheme of high-speed key distribution using mode-shift keying chaos synchronization between two multi-longitudinal-mode Fabry-Perot lasers commonly driven by a super-luminescent diode. Legitimate users dynamically select one of the longitudinal modes according to private control codes to achieve mode-shift keying chaos synchronization. The two remote chaotic light waveforms are quantized to generate two raw random bit streams, and then those bits corresponding to chaos synchronization are sifted as shared keys by comparing the control codes. In this method, the transition time, i.e., the chaos synchronization recovery time is determined by the rising time of the control codes rather than the laser transition response time, so the key distribution rate is improved greatly. Our experiment achieved a 0.75-Gbit/s key distribution rate with a bit error rate of 3.8 × 10-3 over 160-km fiber transmission with dispersion compensation. The entropy rate of the laser chaos is evaluated as 16 Gbit/s, which determines the ultimate final key rate together with the key generation ratio. It is therefore believed that the method pays a way for Gbit/s physical key distribution.
Collapse
Affiliation(s)
- Hua Gao
- Key Laboratory of Advanced Transducers and Intelligent Control System, Ministry of Education and Shanxi Province, Taiyuan, China
- College of Physics and Optoelectronics, Taiyuan University of Technology, Taiyuan, China
| | - Anbang Wang
- Key Laboratory of Advanced Transducers and Intelligent Control System, Ministry of Education and Shanxi Province, Taiyuan, China.
- College of Physics and Optoelectronics, Taiyuan University of Technology, Taiyuan, China.
| | - Longsheng Wang
- Key Laboratory of Advanced Transducers and Intelligent Control System, Ministry of Education and Shanxi Province, Taiyuan, China
- College of Physics and Optoelectronics, Taiyuan University of Technology, Taiyuan, China
| | - Zhiwei Jia
- Key Laboratory of Advanced Transducers and Intelligent Control System, Ministry of Education and Shanxi Province, Taiyuan, China
- College of Physics and Optoelectronics, Taiyuan University of Technology, Taiyuan, China
| | - Yuanyuan Guo
- Key Laboratory of Advanced Transducers and Intelligent Control System, Ministry of Education and Shanxi Province, Taiyuan, China
- College of Physics and Optoelectronics, Taiyuan University of Technology, Taiyuan, China
| | - Zhensen Gao
- School of Information Engineering, Guangdong University of Technology, Guangzhou, China
- Guangdong Provincial Key Laboratory of Photonics Information Technology, Guangzhou, China
| | - Lianshan Yan
- Center for Information Photonics and Communications, School of Information Science and Technology, Southwest Jiaotong University, Chengdu, China
| | - Yuwen Qin
- School of Information Engineering, Guangdong University of Technology, Guangzhou, China
- Guangdong Provincial Key Laboratory of Photonics Information Technology, Guangzhou, China
| | - Yuncai Wang
- School of Information Engineering, Guangdong University of Technology, Guangzhou, China
- Guangdong Provincial Key Laboratory of Photonics Information Technology, Guangzhou, China
| |
Collapse
|
20
|
Carroll TL. Optimizing Reservoir Computers for Signal Classification. Front Physiol 2021; 12:685121. [PMID: 34220549 PMCID: PMC8249854 DOI: 10.3389/fphys.2021.685121] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2021] [Accepted: 05/24/2021] [Indexed: 11/13/2022] Open
Abstract
Reservoir computers are a type of recurrent neural network for which the network connections are not changed. To train the reservoir computer, a set of output signals from the network are fit to a training signal by a linear fit. As a result, training of a reservoir computer is fast, and reservoir computers may be built from analog hardware, resulting in high speed and low power consumption. To get the best performance from a reservoir computer, the hyperparameters of the reservoir computer must be optimized. In signal classification problems, parameter optimization may be computationally difficult; it is necessary to compare many realizations of the test signals to get good statistics on the classification probability. In this work, it is shown in both a spiking reservoir computer and a reservoir computer using continuous variables that the optimum classification performance occurs for the hyperparameters that maximize the entropy of the reservoir computer. Optimizing for entropy only requires a single realization of each signal to be classified, making the process much faster to compute.
Collapse
|
21
|
Carroll TL. Low dimensional manifolds in reservoir computers. CHAOS (WOODBURY, N.Y.) 2021; 31:043113. [PMID: 34251231 DOI: 10.1063/5.0047006] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/09/2021] [Accepted: 03/26/2021] [Indexed: 06/13/2023]
Abstract
A reservoir computer is a complex dynamical system, often created by coupling nonlinear nodes in a network. The nodes are all driven by a common driving signal. Reservoir computers can contain hundreds to thousands of nodes, resulting in a high dimensional dynamical system, but the reservoir computer variables evolve on a lower dimensional manifold in this high dimensional space. This paper describes how this manifold dimension depends on the parameters of the reservoir computer, and how the manifold dimension is related to the performance of the reservoir computer at a signal estimation task. It is demonstrated that increasing the coupling between nodes while controlling the largest Lyapunov exponent of the reservoir computer can optimize the reservoir computer performance. It is also noted that the sparsity of the reservoir computer network does not have any influence on performance.
Collapse
Affiliation(s)
- T L Carroll
- U.S. Naval Research Lab, Washington, DC 20375, USA
| |
Collapse
|
22
|
Patel D, Canaday D, Girvan M, Pomerance A, Ott E. Using machine learning to predict statistical properties of non-stationary dynamical processes: System climate,regime transitions, and the effect of stochasticity. CHAOS (WOODBURY, N.Y.) 2021; 31:033149. [PMID: 33810745 DOI: 10.1063/5.0042598] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/31/2020] [Accepted: 03/05/2021] [Indexed: 06/12/2023]
Abstract
We develop and test machine learning techniques for successfully using past state time series data and knowledge of a time-dependent system parameter to predict the evolution of the "climate" associated with the long-term behavior of a non-stationary dynamical system, where the non-stationary dynamical system is itself unknown. By the term climate, we mean the statistical properties of orbits rather than their precise trajectories in time. By the term non-stationary, we refer to systems that are, themselves, varying with time. We show that our methods perform well on test systems predicting both continuous gradual climate evolution as well as relatively sudden climate changes (which we refer to as "regime transitions"). We consider not only noiseless (i.e., deterministic) non-stationary dynamical systems, but also climate prediction for non-stationary dynamical systems subject to stochastic forcing (i.e., dynamical noise), and we develop a method for handling this latter case. The main conclusion of this paper is that machine learning has great promise as a new and highly effective approach to accomplishing data driven prediction of non-stationary systems.
Collapse
Affiliation(s)
- Dhruvit Patel
- The Institute for Research in Electronics and Applied Physics, University of Maryland, College Park, Maryland 26742, USA
| | - Daniel Canaday
- Potomac Research LLC, Alexandria, Virginia 22311-1311, USA
| | - Michelle Girvan
- The Institute for Research in Electronics and Applied Physics, University of Maryland, College Park, Maryland 26742, USA
| | | | - Edward Ott
- The Institute for Research in Electronics and Applied Physics, University of Maryland, College Park, Maryland 26742, USA
| |
Collapse
|
23
|
Zhong D, Yang H, Xi J, Zeng N, Xu Z, Deng F. Predictive learning of multi-channel isochronal chaotic synchronization by utilizing parallel optical reservoir computers based on three laterally coupled semiconductor lasers with delay-time feedback. OPTICS EXPRESS 2021; 29:5279-5294. [PMID: 33726067 DOI: 10.1364/oe.418202] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/22/2020] [Accepted: 01/28/2021] [Indexed: 06/12/2023]
Abstract
In this work, we utilize three parallel optical reservoir computers to model three optical dynamic systems, respectively. Here, the three laser-elements in the response laser array with both delay-time feedback and optical injection are utilized as nonlinear nodes to realize three optical chaotic reservoir computers (RCs). The nonlinear dynamics of three laser-elements in the driving laser array are predictively learned by these three parallel RCs. We show that these three parallel reservoir computers can reproduce the nonlinear dynamics of the three laser-elements in the driving laser array with self-feedback. Very small training errors for their predictions can be realized by the optimization of two key parameters such as the delay-time and the interval of the virtual nodes. Moreover, these three parallel RCs to be trained will well synchronize with three chaotic laser-elements in the driving laser array, respectively, even when there are some parameter mismatches between the response laser array and the driving laser array. Our findings show that optical reservoir computing approach possibly provide a successful path for the realization of the high-quality chaotic synchronization between the driving laser and the response laser when their rate-equations imperfectly match each other.
Collapse
|
24
|
Gulina M, Mauroy A. Two methods to approximate the Koopman operator with a reservoir computer. CHAOS (WOODBURY, N.Y.) 2021; 31:023116. [PMID: 33653036 DOI: 10.1063/5.0026380] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/25/2020] [Accepted: 01/20/2021] [Indexed: 06/12/2023]
Abstract
The Koopman operator provides a powerful framework for data-driven analysis of dynamical systems. In the last few years, a wealth of numerical methods providing finite-dimensional approximations of the operator have been proposed [e.g., extended dynamic mode decomposition (EDMD) and its variants]. While convergence results for EDMD require an infinite number of dictionary elements, recent studies have shown that only a few dictionary elements can yield an efficient approximation of the Koopman operator, provided that they are well-chosen through a proper training process. However, this training process typically relies on nonlinear optimization techniques. In this paper, we propose two novel methods based on a reservoir computer to train the dictionary. These methods rely solely on linear convex optimization. We illustrate the efficiency of the method with several numerical examples in the context of data reconstruction, prediction, and computation of the Koopman operator spectrum. These results pave the way for the use of the reservoir computer in the Koopman operator framework.
Collapse
Affiliation(s)
- Marvyn Gulina
- Department of Mathematics and Namur Institute for Complex Systems (naXys), University of Namur, 5000 Namur, Belgium
| | - Alexandre Mauroy
- Department of Mathematics and Namur Institute for Complex Systems (naXys), University of Namur, 5000 Namur, Belgium
| |
Collapse
|
25
|
Bollt E. On explaining the surprising success of reservoir computing forecaster of chaos? The universal machine learning dynamical system with contrast to VAR and DMD. CHAOS (WOODBURY, N.Y.) 2021; 31:013108. [PMID: 33754755 DOI: 10.1063/5.0024890] [Citation(s) in RCA: 38] [Impact Index Per Article: 9.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/14/2020] [Accepted: 11/30/2020] [Indexed: 06/12/2023]
Abstract
Machine learning has become a widely popular and successful paradigm, especially in data-driven science and engineering. A major application problem is data-driven forecasting of future states from a complex dynamical system. Artificial neural networks have evolved as a clear leader among many machine learning approaches, and recurrent neural networks are considered to be particularly well suited for forecasting dynamical systems. In this setting, the echo-state networks or reservoir computers (RCs) have emerged for their simplicity and computational complexity advantages. Instead of a fully trained network, an RC trains only readout weights by a simple, efficient least squares method. What is perhaps quite surprising is that nonetheless, an RC succeeds in making high quality forecasts, competitively with more intensively trained methods, even if not the leader. There remains an unanswered question as to why and how an RC works at all despite randomly selected weights. To this end, this work analyzes a further simplified RC, where the internal activation function is an identity function. Our simplification is not presented for the sake of tuning or improving an RC, but rather for the sake of analysis of what we take to be the surprise being not that it does not work better, but that such random methods work at all. We explicitly connect the RC with linear activation and linear readout to well developed time-series literature on vector autoregressive (VAR) averages that includes theorems on representability through the Wold theorem, which already performs reasonably for short-term forecasts. In the case of a linear activation and now popular quadratic readout RC, we explicitly connect to a nonlinear VAR, which performs quite well. Furthermore, we associate this paradigm to the now widely popular dynamic mode decomposition; thus, these three are in a sense different faces of the same thing. We illustrate our observations in terms of popular benchmark examples including Mackey-Glass differential delay equations and the Lorenz63 system.
Collapse
Affiliation(s)
- Erik Bollt
- Department of Electrical and Computer Engineering, Clarkson University, Potsdam, New York 13699, USA and Clarkson Center for Complex Systems Science (C3S2), Potsdam, New York 13699, USA
| |
Collapse
|
26
|
Carroll TL. Do reservoir computers work best at the edge of chaos? CHAOS (WOODBURY, N.Y.) 2020; 30:121109. [PMID: 33380041 DOI: 10.1063/5.0038163] [Citation(s) in RCA: 22] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/19/2020] [Accepted: 12/02/2020] [Indexed: 06/12/2023]
Abstract
It has been demonstrated that cellular automata had the highest computational capacity at the edge of chaos [N. H. Packard, in Dynamic Patterns in Complex Systems, edited by J. A. S. Kelso, A. J. Mandell, and M. F. Shlesinger (World Scientific, Singapore, 1988), pp. 293-301; C. G. Langton, Physica D 42(1), 12-37 (1990); J. P. Crutchfield and K. Young, in Complexity, Entropy, and the Physics of Information, edited by W. H. Zurek (Addison-Wesley, Redwood City, CA, 1990), pp. 223-269], the parameter at which their behavior transitioned from ordered to chaotic. This same concept has been applied to reservoir computers; a number of researchers have stated that the highest computational capacity for a reservoir computer is at the edge of chaos, although others have suggested that this rule is not universally true. Because many reservoir computers do not show chaotic behavior but merely become unstable, it is felt that a more accurate term for this instability transition is the "edge of stability." Here, I find two examples where the computational capacity of a reservoir computer decreases as the edge of stability is approached: in one case because generalized synchronization breaks down and in the other case because the reservoir computer is a poor match to the problem being solved. The edge of stability as an optimal operating point for a reservoir computer is not in general true, although it may be true in some cases.
Collapse
Affiliation(s)
- T L Carroll
- U.S. Naval Research Lab, Washington, DC 20375, USA
| |
Collapse
|
27
|
Mabrouk YA, Räth C. Calibrated reservoir computers. CHAOS (WOODBURY, N.Y.) 2020; 30:113134. [PMID: 33261366 DOI: 10.1063/5.0030651] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/23/2020] [Accepted: 10/28/2020] [Indexed: 06/12/2023]
Abstract
We observe the presence of infinitely fine-scaled alternations within the performance landscape of reservoir computers aimed for chaotic data forecasting. We investigate the emergence of the observed structures by means of variations of the transversal stability of the synchronization manifold relating the observational and internal dynamical states. Finally, we deduce a simple calibration method in order to attenuate the thus evidenced performance uncertainty.
Collapse
Affiliation(s)
- Y A Mabrouk
- Fakultät für Physik, Ludwig-Maximilians-Universität München, Schellingstraße 4, 80779 München, Germany
| | - C Räth
- Fakultät für Physik, Ludwig-Maximilians-Universität München, Schellingstraße 4, 80779 München, Germany
| |
Collapse
|
28
|
Chen X, Weng T, Yang H, Gu C, Zhang J, Small M. Mapping topological characteristics of dynamical systems into neural networks: A reservoir computing approach. Phys Rev E 2020; 102:033314. [PMID: 33075895 DOI: 10.1103/physreve.102.033314] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/18/2020] [Accepted: 09/09/2020] [Indexed: 11/07/2022]
Abstract
Significant advances have recently been made in modeling chaotic systems with the reservoir computing approach, especially for prediction. We find that although state prediction of the trained reservoir computer will gradually deviate from the actual trajectory of the original system, the associated geometric features remain invariant. Specifically, we show that the typical geometric metrics including the correlation dimension, the multiscale entropy, and the memory effect are nearly identical between the trained reservoir computer and its learned chaotic systems. We further demonstrate this fact on a broad range of chaotic systems ranging from discrete and continuous chaotic systems to hyperchaotic systems. Our findings suggest that the successfully reservoir computer may be topologically conjugate to an observed dynamical system.
Collapse
Affiliation(s)
- Xiaolu Chen
- Business School, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Tongfeng Weng
- Alibaba Research Center for Complexity Sciences, Hangzhou Normal University, Hangzhou 311121, People's Republic of China
| | - Huijie Yang
- Business School, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Changgui Gu
- Business School, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Jie Zhang
- Institute of Science and Technology for Brain-Inspired Intellegence, Fudan University, Shanghai 200433, People's Republic of China
| | - Michael Small
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia.,Mineral Resources, CSIRO, Kensington, Western Australia 6151, Australia
| |
Collapse
|
29
|
Carroll TL. Path length statistics in reservoir computers. CHAOS (WOODBURY, N.Y.) 2020; 30:083130. [PMID: 32872832 DOI: 10.1063/5.0014643] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/20/2020] [Accepted: 08/03/2020] [Indexed: 06/11/2023]
Abstract
Because reservoir computers are high dimensional dynamical systems, designing a good reservoir computer is difficult. In many cases, the designer must search a large nonlinear parameter space, and each step of the search requires simulating the full reservoir computer. In this work, I show that a simple statistic based on the mean path length between nodes in the reservoir computer is correlated with better reservoir computer performance. The statistic predicts the diversity of signals produced by the reservoir computer, as measured by the covariance matrix of the reservoir computer. This statistic by itself is not sufficient to predict reservoir computer performance because not only must the reservoir computer produce a diverse set of signals, it must be well matched to the training signals. Nevertheless, this path length statistic allows the designer to eliminate some network configurations from consideration without having to actually simulate the reservoir computer, reducing the complexity of the design process.
Collapse
Affiliation(s)
- T L Carroll
- U.S. Naval Research Lab, Washington, DC 20375, USA
| |
Collapse
|
30
|
Haluszczynski A, Aumeier J, Herteux J, Räth C. Reducing network size and improving prediction stability of reservoir computing. CHAOS (WOODBURY, N.Y.) 2020; 30:063136. [PMID: 32611106 DOI: 10.1063/5.0006869] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/06/2020] [Accepted: 06/05/2020] [Indexed: 06/11/2023]
Abstract
Reservoir computing is a very promising approach for the prediction of complex nonlinear dynamical systems. Besides capturing the exact short-term trajectories of nonlinear systems, it has also proved to reproduce its characteristic long-term properties very accurately. However, predictions do not always work equivalently well. It has been shown that both short- and long-term predictions vary significantly among different random realizations of the reservoir. In order to gain an understanding on when reservoir computing works best, we investigate some differential properties of the respective realization of the reservoir in a systematic way. We find that removing nodes that correspond to the largest weights in the output regression matrix reduces outliers and improves overall prediction quality. Moreover, this allows to effectively reduce the network size and, therefore, increase computational efficiency. In addition, we use a nonlinear scaling factor in the hyperbolic tangent of the activation function. This adjusts the response of the activation function to the range of values of the input variables of the nodes. As a consequence, this reduces the number of outliers significantly and increases both the short- and long-term prediction quality for the nonlinear systems investigated in this study. Our results demonstrate that a large optimization potential lies in the systematical refinement of the differential reservoir properties for a given dataset.
Collapse
Affiliation(s)
- Alexander Haluszczynski
- Department of Physics, Ludwig-Maximilians-Universität, Schellingstraße 4, 80799 Munich, Germany
| | - Jonas Aumeier
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft- und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| | - Joschka Herteux
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft- und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| | - Christoph Räth
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft- und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| |
Collapse
|
31
|
Carroll TL. Dimension of reservoir computers. CHAOS (WOODBURY, N.Y.) 2020; 30:013102. [PMID: 32013466 DOI: 10.1063/1.5128898] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/23/2019] [Accepted: 12/11/2019] [Indexed: 06/10/2023]
Abstract
A reservoir computer is a complex dynamical system, often created by coupling nonlinear nodes in a network. The nodes are all driven by a common driving signal. In this work, three dimension estimation methods, false nearest neighbor, covariance dimension, and Kaplan-Yorke dimension, are used to estimate the dimension of the reservoir dynamical system. It is shown that the signals in the reservoir system exist on a relatively low dimensional surface. Changing the spectral radius of the reservoir network can increase the fractal dimension of the reservoir signals, leading to an increase in a testing error.
Collapse
Affiliation(s)
- T L Carroll
- US Naval Research Lab, Washington, DC 20375, USA
| |
Collapse
|
32
|
Banerjee A, Pathak J, Roy R, Restrepo JG, Ott E. Using machine learning to assess short term causal dependence and infer network links. CHAOS (WOODBURY, N.Y.) 2019; 29:121104. [PMID: 31893648 DOI: 10.1063/1.5134845] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/04/2019] [Accepted: 12/05/2019] [Indexed: 06/10/2023]
Abstract
We introduce and test a general machine-learning-based technique for the inference of short term causal dependence between state variables of an unknown dynamical system from time-series measurements of its state variables. Our technique leverages the results of a machine learning process for short time prediction to achieve our goal. The basic idea is to use the machine learning to estimate the elements of the Jacobian matrix of the dynamical flow along an orbit. The type of machine learning that we employ is reservoir computing. We present numerical tests on link inference of a network of interacting dynamical nodes. It is seen that dynamical noise can greatly enhance the effectiveness of our technique, while observational noise degrades the effectiveness. We believe that the competition between these two opposing types of noise will be the key factor determining the success of causal inference in many of the most important application situations.
Collapse
Affiliation(s)
- Amitava Banerjee
- Department of Physics and Institute for Research in Electronics and Applied Physics, University of Maryland, College Park, Maryland 20742, USA
| | - Jaideep Pathak
- Department of Physics and Institute for Research in Electronics and Applied Physics, University of Maryland, College Park, Maryland 20742, USA
| | - Rajarshi Roy
- Department of Physics and Institute for Research in Electronics and Applied Physics, University of Maryland, College Park, Maryland 20742, USA
| | - Juan G Restrepo
- Department of Applied Mathematics, University of Colorado, Boulder, Colorado 80309, USA
| | - Edward Ott
- Department of Physics and Institute for Research in Electronics and Applied Physics, University of Maryland, College Park, Maryland 20742, USA
| |
Collapse
|
33
|
Maslennikov OV, Nekorkin VI. Collective dynamics of rate neurons for supervised learning in a reservoir computing system. CHAOS (WOODBURY, N.Y.) 2019; 29:103126. [PMID: 31675797 DOI: 10.1063/1.5119895] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/14/2019] [Accepted: 09/26/2019] [Indexed: 06/10/2023]
Abstract
In this paper, we study collective dynamics of the network of rate neurons which constitute a central element of a reservoir computing system. The main objective of the paper is to identify the dynamic behaviors inside the reservoir underlying the performance of basic machine learning tasks, such as generating patterns with specified characteristics. We build a reservoir computing system which includes a reservoir-a network of interacting rate neurons-and an output element that generates a target signal. We study individual activities of interacting rate neurons, while implementing the task and analyze the impact of the dynamic parameter-a time constant-on the quality of implementation.
Collapse
Affiliation(s)
- Oleg V Maslennikov
- Institute of Applied Physics of the Russian Academy of Sciences, 46 Ulyanov Street, 603950 Nizhny Novgorod, Russia
| | - Vladimir I Nekorkin
- Institute of Applied Physics of the Russian Academy of Sciences, 46 Ulyanov Street, 603950 Nizhny Novgorod, Russia
| |
Collapse
|
34
|
Shirin A, Klickstein IS, Sorrentino F. Stability analysis of reservoir computers dynamics via Lyapunov functions. CHAOS (WOODBURY, N.Y.) 2019; 29:103147. [PMID: 31675840 DOI: 10.1063/1.5123733] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/08/2019] [Accepted: 10/08/2019] [Indexed: 06/10/2023]
Abstract
A Lyapunov design method is used to analyze the nonlinear stability of a generic reservoir computer for both the cases of continuous-time and discrete-time dynamics. Using this method, for a given nonlinear reservoir computer, a radial region of stability around a fixed point is analytically determined. We see that the training error of the reservoir computer is lower in the region where the analysis predicts global stability but is also affected by the particular choice of the individual dynamics for the reservoir systems. For the case that the dynamics is polynomial, it appears to be important for the polynomial to have nonzero coefficients corresponding to at least one odd power (e.g., linear term) and one even power (e.g., quadratic term).
Collapse
Affiliation(s)
- Afroza Shirin
- Mechanical Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131, USA
| | - Isaac S Klickstein
- Mechanical Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131, USA
| | - Francesco Sorrentino
- Mechanical Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131, USA
| |
Collapse
|
35
|
Haluszczynski A, Räth C. Good and bad predictions: Assessing and improving the replication of chaotic attractors by means of reservoir computing. CHAOS (WOODBURY, N.Y.) 2019; 29:103143. [PMID: 31675800 DOI: 10.1063/1.5118725] [Citation(s) in RCA: 26] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/04/2019] [Accepted: 10/09/2019] [Indexed: 06/10/2023]
Abstract
The prediction of complex nonlinear dynamical systems with the help of machine learning techniques has become increasingly popular. In particular, reservoir computing turned out to be a very promising approach especially for the reproduction of the long-term properties of a nonlinear system. Yet, a thorough statistical analysis of the forecast results is missing. Using the Lorenz and Rössler system, we statistically analyze the quality of prediction for different parametrizations-both the exact short-term prediction as well as the reproduction of the long-term properties (the "climate") of the system as estimated by the correlation dimension and largest Lyapunov exponent. We find that both short- and long-term predictions vary significantly among the realizations. Thus, special care must be taken in selecting the good predictions as realizations, which deliver better short-term prediction also tend to better resemble the long-term climate of the system. Instead of only using purely random Erdös-Renyi networks, we also investigate the benefit of alternative network topologies such as small world or scale-free networks and show which effect they have on the prediction quality. Our results suggest that the overall performance with respect to the reproduction of the climate of both the Lorenz and Rössler system is worst for scale-free networks. For the Lorenz system, there seems to be a slight benefit of using small world networks, while for the Rössler system, small world and Erdös-Renyi networks performed equivalently well. In general, the observation is that reservoir computing works for all network topologies investigated here.
Collapse
Affiliation(s)
- Alexander Haluszczynski
- Department of Physics, Ludwig-Maximilians-Universität, Schellingstraße 4, 80799 Munich, Germany
| | - Christoph Räth
- Deutsches Zentrum für Luft- und Raumfahrt, Institut für Materialphysik im Weltraum, Münchner Str. 20, 82234 Wessling, Germany
| |
Collapse
|
36
|
Carroll TL, Pecora LM. Network structure effects in reservoir computers. CHAOS (WOODBURY, N.Y.) 2019; 29:083130. [PMID: 31472504 DOI: 10.1063/1.5097686] [Citation(s) in RCA: 31] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/27/2019] [Accepted: 08/06/2019] [Indexed: 06/10/2023]
Abstract
A reservoir computer is a complex nonlinear dynamical system that has been shown to be useful for solving certain problems, such as prediction of chaotic signals, speech recognition, or control of robotic systems. Typically, a reservoir computer is constructed by connecting a large number of nonlinear nodes in a network, driving the nodes with an input signal and using the node outputs to fit a training signal. In this work, we set up reservoirs where the edges (or connections) between all the network nodes are either +1 or 0 and proceed to alter the network structure by flipping some of these edges from +1 to -1. We use this simple network because it turns out to be easy to characterize; we may use the fraction of edges flipped as a measure of how much we have altered the network. In some cases, the network can be rearranged in a finite number of ways without changing its structure; these rearrangements are symmetries of the network, and the number of symmetries is also useful for characterizing the network. We find that changing the number of edges flipped in the network changes the rank of the covariance of a matrix consisting of the time series from the different nodes in the network and speculate that this rank is important for understanding the reservoir computer performance.
Collapse
Affiliation(s)
- T L Carroll
- US Naval Research Laboratory, Washington, DC 20375, USA
| | - L M Pecora
- US Naval Research Laboratory, Washington, DC 20375, USA
| |
Collapse
|
37
|
Nakai K, Saiki Y. Machine-learning inference of fluid variables from data using reservoir computing. Phys Rev E 2018; 98:023111. [PMID: 30253537 DOI: 10.1103/physreve.98.023111] [Citation(s) in RCA: 24] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2018] [Indexed: 06/08/2023]
Abstract
We infer both microscopic and macroscopic behaviors of a three-dimensional chaotic fluid flow using reservoir computing. In our procedure of the inference, we assume no prior knowledge of a physical process of a fluid flow except that its behavior is complex but deterministic. We present two ways of inference of the complex behavior: the first, called partial inference, requires continued knowledge of partial time-series data during the inference as well as past time-series data, while the second, called full inference, requires only past time-series data as training data. For the first case, we are able to infer long-time motion of microscopic fluid variables. For the second case, we show that the reservoir dynamics constructed from only past data of energy functions can infer the future behavior of energy functions and reproduce the energy spectrum. It is also shown that we can infer time-series data from only one measurement by using the delay coordinates. This implies that the obtained reservoir systems constructed without the knowledge of microscopic data are equivalent to the dynamical systems describing the macroscopic behavior of energy functions.
Collapse
Affiliation(s)
- Kengo Nakai
- Graduate School of Mathematical Sciences, The University of Tokyo, Tokyo 153-8914, Japan
| | - Yoshitaka Saiki
- Graduate School of Business Administration, Hitotsubashi University, Tokyo 186-8601, Japan
- JST, PRESTO, Saitama 332-0012, Japan
- Institute for Physical Science and Technology, University of Maryland, College Park, Maryland 20742, USA
| |
Collapse
|