1
|
Steinegger J, Räth C. Predicting three-dimensional chaotic systems with four qubit quantum systems. Sci Rep 2025; 15:6201. [PMID: 39979362 PMCID: PMC11842815 DOI: 10.1038/s41598-025-87768-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2024] [Accepted: 01/21/2025] [Indexed: 02/22/2025] Open
Abstract
Reservoir computing (RC) is among the most promising approaches for AI-based prediction models of complex systems. It combines superior prediction performance with very low CPU-needs for training. Recent results demonstrated that quantum systems are also well-suited as reservoirs in RC. Due to the exponential growth of the Hilbert space dimension obtained by increasing the number of quantum elements small quantum systems are already sufficient for time series prediction. Here, we demonstrate that three-dimensional systems can already well be predicted by quantum reservoir computing with a quantum reservoir consisting of the minimal number of qubits necessary for this task, namely four. This is achieved by optimizing the encoding of the data, using spatial and temporal multiplexing and recently developed read-out-schemes that also involve higher exponents of the reservoir response. We outline, test and validate our approach using eight prototypical three-dimensional chaotic systems. Both, the short-term prediction and the reproduction of the long-term system behavior (the system's "climate") are feasible with the same setup of optimized hyperparameters. Our results may be a further step towards the realization of a dedicated small quantum computer for prediction tasks in the NISQ-era.
Collapse
Affiliation(s)
- Joel Steinegger
- Institut für KI Sicherheit, Deutsches Zentrum für Luft- und Raumfahrt (DLR), Wilhelm-Runge-Straße 10, 89081, Ulm, Germany
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft- und Raumfahrt (DLR), Linder Höhe, 51170, Köln, Germany
| | - Christoph Räth
- Institut für KI Sicherheit, Deutsches Zentrum für Luft- und Raumfahrt (DLR), Wilhelm-Runge-Straße 10, 89081, Ulm, Germany.
| |
Collapse
|
2
|
Li X, Small M, Lei Y. Reservoir computing with higher-order interactive coupled pendulums. Phys Rev E 2023; 108:064304. [PMID: 38243442 DOI: 10.1103/physreve.108.064304] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2023] [Accepted: 11/28/2023] [Indexed: 01/21/2024]
Abstract
The reservoir computing approach utilizes a time series of measurements as input to a high-dimensional dynamical system known as a reservoir. However, the approach relies on sampling a random matrix to define its underlying reservoir layer, which leads to numerous hyperparameters that need to be optimized. Here, we propose a nonlocally coupled pendulum model with higher-order interactions as a novel reservoir, which requires no random underlying matrices and fewer hyperparameters. We use Bayesian optimization to explore the hyperparameter space within a minimal number of iterations and train the coupled pendulums model to reproduce the chaotic attractors, which simplifies complicated hyperparameter optimization. We illustrate the effectiveness of our technique with the Lorenz system and the Hindmarsh-Rose neuronal model, and we calculate the Pearson correlation coefficients between time series and the Hausdorff metrics in the phase space. We demonstrate the contribution of higher-order interactions by analyzing the interaction between different reservoir configurations and prediction performance, as well as computations of the largest Lyapunov exponents. The chimera state is found as the most effective dynamical regime for prediction. The findings, where we present a new reservoir structure, offer potential applications in the design of high-performance modeling of dynamics in physical systems.
Collapse
Affiliation(s)
- Xueqi Li
- School of Mathematics and Statistics, Northwestern Polytechnical University, Xi'an 710072, China
| | - Michael Small
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley WA 6009, Australia
- Mineral Resources, CSIRO, Kensington WA 6151, Australia
| | - Youming Lei
- School of Mathematics and Statistics, Northwestern Polytechnical University, Xi'an 710072, China
- Ministry of Industry and Information Technology Key Laboratory of Dynamics and Control of Complex Systems, Northwestern Polytechnical University, Xi'an 710072, China
| |
Collapse
|
3
|
Mahata A, Padhi R, Apte A. Variability of echo state network prediction horizon for partially observed dynamical systems. Phys Rev E 2023; 108:064209. [PMID: 38243433 DOI: 10.1103/physreve.108.064209] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Accepted: 11/10/2023] [Indexed: 01/21/2024]
Abstract
Study of dynamical systems using partial state observation is an important problem due to its applicability to many real-world systems. We address the problem by studying an echo state network (ESN) framework with partial state input with partial or full state output. Application to the Lorenz system and Chua's oscillator (both numerically simulated and experimental systems) demonstrate the effectiveness of our method. We show that the ESN, as an autonomous dynamical system, is capable of making short-term predictions up to a few Lyapunov times. However, the prediction horizon has high variability depending on the initial condition-an aspect that we explore in detail using the distribution of the prediction horizon. Further, using a variety of statistical metrics to compare the long-term dynamics of the ESN predictions with numerically simulated or experimental dynamics and observed similar results, we show that the ESN can effectively learn the system's dynamics even when trained with noisy numerical or experimental data sets. Thus, we demonstrate the potential of ESNs to serve as cheap surrogate models for simulating the dynamics of systems where complete observations are unavailable.
Collapse
Affiliation(s)
- Ajit Mahata
- Department of Data Science, Indian Institute of Science Education and Research, IISER Pune 411008, India
| | - Reetish Padhi
- Department of Data Science, Indian Institute of Science Education and Research, IISER Pune 411008, India
| | - Amit Apte
- Department of Data Science, Indian Institute of Science Education and Research, IISER Pune 411008, India
- International Centre for Theoretical Sciences (ICTS-TIFR), Bengaluru 560089, India
| |
Collapse
|
4
|
Ma H, Prosperino D, Räth C. A novel approach to minimal reservoir computing. Sci Rep 2023; 13:12970. [PMID: 37563235 PMCID: PMC10415382 DOI: 10.1038/s41598-023-39886-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2023] [Accepted: 08/01/2023] [Indexed: 08/12/2023] Open
Abstract
Reservoir computers are powerful machine learning algorithms for predicting nonlinear systems. Unlike traditional feedforward neural networks, they work on small training data sets, operate with linear optimization, and therefore require minimal computational resources. However, the traditional reservoir computer uses random matrices to define the underlying recurrent neural network and has a large number of hyperparameters that need to be optimized. Recent approaches show that randomness can be taken out by running regressions on a large library of linear and nonlinear combinations constructed from the input data and their time lags and polynomials thereof. However, for high-dimensional and nonlinear data, the number of these combinations explodes. Here, we show that a few simple changes to the traditional reservoir computer architecture further minimizing computational resources lead to significant and robust improvements in short- and long-term predictive performances compared to similar models while requiring minimal sizes of training data sets.
Collapse
Affiliation(s)
- Haochun Ma
- Department of Physics, Ludwig-Maximilians-Universität, Schellingstraße 4, 80799, Munich, Germany
| | - Davide Prosperino
- Department of Physics, Ludwig-Maximilians-Universität, Schellingstraße 4, 80799, Munich, Germany
| | - Christoph Räth
- Deutsches Zentrum für Luft- und Raumfahrt (DLR), Institut für KI Sicherheit, Wilhelm-Runge-Straße 10, 89081, Ulm, Germany.
| |
Collapse
|
5
|
Bosl WJ, Bosquet Enlow M, Lock EF, Nelson CA. A biomarker discovery framework for childhood anxiety. Front Psychiatry 2023; 14:1158569. [PMID: 37533889 PMCID: PMC10393248 DOI: 10.3389/fpsyt.2023.1158569] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/04/2023] [Accepted: 07/04/2023] [Indexed: 08/04/2023] Open
Abstract
Introduction Anxiety is the most common manifestation of psychopathology in youth, negatively affecting academic, social, and adaptive functioning and increasing risk for mental health problems into adulthood. Anxiety disorders are diagnosed only after clinical symptoms emerge, potentially missing opportunities to intervene during critical early prodromal periods. In this study, we used a new empirical approach to extracting nonlinear features of the electroencephalogram (EEG), with the goal of discovering differences in brain electrodynamics that distinguish children with anxiety disorders from healthy children. Additionally, we examined whether this approach could distinguish children with externalizing disorders from healthy children and children with anxiety. Methods We used a novel supervised tensor factorization method to extract latent factors from repeated multifrequency nonlinear EEG measures in a longitudinal sample of children assessed in infancy and at ages 3, 5, and 7 years of age. We first examined the validity of this method by showing that calendar age is highly correlated with latent EEG complexity factors (r = 0.77). We then computed latent factors separately for distinguishing children with anxiety disorders from healthy controls using a 5-fold cross validation scheme and similarly for distinguishing children with externalizing disorders from healthy controls. Results We found that latent factors derived from EEG recordings at age 7 years were required to distinguish children with an anxiety disorder from healthy controls; recordings from infancy, 3 years, or 5 years alone were insufficient. However, recordings from two (5, 7 years) or three (3, 5, 7 years) recordings gave much better results than 7 year recordings alone. Externalizing disorders could be detected using 3- and 5 years EEG data, also giving better results with two or three recordings than any single snapshot. Further, sex assigned at birth was an important covariate that improved accuracy for both disorder groups, and birthweight as a covariate modestly improved accuracy for externalizing disorders. Recordings from infant EEG did not contribute to the classification accuracy for either anxiety or externalizing disorders. Conclusion This study suggests that latent factors extracted from EEG recordings in childhood are promising candidate biomarkers for anxiety and for externalizing disorders if chosen at appropriate ages.
Collapse
Affiliation(s)
- William J. Bosl
- Center for AI & Medicine, University of San Francisco, San Francisco, CA, United States
- Computational Health Informatics Program, Boston Children’s Hospital, Boston, MA, United States
- Department of Pediatrics, Harvard Medical School, Boston, MA, United States
| | - Michelle Bosquet Enlow
- Department of Psychiatry and Behavioral Sciences, Boston Children’s Hospital, Boston, MA, United States
- Department of Psychiatry, Harvard Medical School, Boston, MA, United States
| | - Eric F. Lock
- Division of Biostatistics, School of Public Health, University of Minnesota, Minneapolis, MN, United States
| | - Charles A. Nelson
- Department of Pediatrics, Harvard Medical School, Boston, MA, United States
- Laboratories of Cognitive Neuroscience, Division of Developmental Medicine, Boston Children’s Hospital, Boston, MA, United States
- Harvard Graduate School of Education, Cambridge, MA, United States
| |
Collapse
|
6
|
Ma H, Prosperino D, Haluszczynski A, Räth C. Efficient forecasting of chaotic systems with block-diagonal and binary reservoir computing. CHAOS (WOODBURY, N.Y.) 2023; 33:2895979. [PMID: 37307160 DOI: 10.1063/5.0151290] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/20/2023] [Accepted: 05/12/2023] [Indexed: 06/14/2023]
Abstract
The prediction of complex nonlinear dynamical systems with the help of machine learning has become increasingly popular in different areas of science. In particular, reservoir computers, also known as echo-state networks, turned out to be a very powerful approach, especially for the reproduction of nonlinear systems. The reservoir, the key component of this method, is usually constructed as a sparse, random network that serves as a memory for the system. In this work, we introduce block-diagonal reservoirs, which implies that a reservoir can be composed of multiple smaller reservoirs, each with its own dynamics. Furthermore, we take out the randomness of the reservoir by using matrices of ones for the individual blocks. This breaks with the widespread interpretation of the reservoir as a single network. In the example of the Lorenz and Halvorsen systems, we analyze the performance of block-diagonal reservoirs and their sensitivity to hyperparameters. We find that the performance is comparable to sparse random networks and discuss the implications with regard to scalability, explainability, and hardware realizations of reservoir computers.
Collapse
Affiliation(s)
- Haochun Ma
- Department of Physics, Ludwig-Maximilians-Universität, Schellingstraße 4, 80799 Munich, Germany
- Allianz Global Investors, risklab, Seidlstraße 24, 80335 Munich, Germany
| | - Davide Prosperino
- Department of Physics, Ludwig-Maximilians-Universität, Schellingstraße 4, 80799 Munich, Germany
- Allianz Global Investors, risklab, Seidlstraße 24, 80335 Munich, Germany
| | | | - Christoph Räth
- Deutsches Zentrum für Luft- und Raumfahrt (DLR), Institut für KI Sicherheit, Wilhelm-Runge-Straße 10, 89081 Ulm, Germany
| |
Collapse
|
7
|
Tsuchiyama K, Röhm A, Mihana T, Horisaki R, Naruse M. Effect of temporal resolution on the reproduction of chaotic dynamics via reservoir computing. CHAOS (WOODBURY, N.Y.) 2023; 33:063145. [PMID: 37347641 DOI: 10.1063/5.0143846] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/27/2023] [Accepted: 06/05/2023] [Indexed: 06/24/2023]
Abstract
Reservoir computing is a machine learning paradigm that uses a structure called a reservoir, which has nonlinearities and short-term memory. In recent years, reservoir computing has expanded to new functions such as the autonomous generation of chaotic time series, as well as time series prediction and classification. Furthermore, novel possibilities have been demonstrated, such as inferring the existence of previously unseen attractors. Sampling, in contrast, has a strong influence on such functions. Sampling is indispensable in a physical reservoir computer that uses an existing physical system as a reservoir because the use of an external digital system for the data input is usually inevitable. This study analyzes the effect of sampling on the ability of reservoir computing to autonomously regenerate chaotic time series. We found, as expected, that excessively coarse sampling degrades the system performance, but also that excessively dense sampling is unsuitable. Based on quantitative indicators that capture the local and global characteristics of attractors, we identify a suitable window of the sampling frequency and discuss its underlying mechanisms.
Collapse
Affiliation(s)
- Kohei Tsuchiyama
- Department of Information Physics and Computing, Graduate School of Information Science and Technology, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, Japan
| | - André Röhm
- Department of Information Physics and Computing, Graduate School of Information Science and Technology, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, Japan
| | - Takatomo Mihana
- Department of Information Physics and Computing, Graduate School of Information Science and Technology, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, Japan
| | - Ryoichi Horisaki
- Department of Information Physics and Computing, Graduate School of Information Science and Technology, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, Japan
| | - Makoto Naruse
- Department of Information Physics and Computing, Graduate School of Information Science and Technology, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, Japan
| |
Collapse
|
8
|
Tan E, Algar S, Corrêa D, Small M, Stemler T, Walker D. Selecting embedding delays: An overview of embedding techniques and a new method using persistent homology. CHAOS (WOODBURY, N.Y.) 2023; 33:032101. [PMID: 37003815 DOI: 10.1063/5.0137223] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/01/2022] [Accepted: 02/07/2023] [Indexed: 06/19/2023]
Abstract
Delay embedding methods are a staple tool in the field of time series analysis and prediction. However, the selection of embedding parameters can have a big impact on the resulting analysis. This has led to the creation of a large number of methods to optimize the selection of parameters such as embedding lag. This paper aims to provide a comprehensive overview of the fundamentals of embedding theory for readers who are new to the subject. We outline a collection of existing methods for selecting embedding lag in both uniform and non-uniform delay embedding cases. Highlighting the poor dynamical explainability of existing methods of selecting non-uniform lags, we provide an alternative method of selecting embedding lags that includes a mixture of both dynamical and topological arguments. The proposed method, Significant Times on Persistent Strands (SToPS), uses persistent homology to construct a characteristic time spectrum that quantifies the relative dynamical significance of each time lag. We test our method on periodic, chaotic, and fast-slow time series and find that our method performs similar to existing automated non-uniform embedding methods. Additionally, n-step predictors trained on embeddings constructed with SToPS were found to outperform other embedding methods when predicting fast-slow time series.
Collapse
Affiliation(s)
- Eugene Tan
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| | - Shannon Algar
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| | - Débora Corrêa
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| | - Michael Small
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| | - Thomas Stemler
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| | - David Walker
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| |
Collapse
|
9
|
Pershin A, Beaume C, Li K, Tobias SM. Training a neural network to predict dynamics it has never seen. Phys Rev E 2023; 107:014304. [PMID: 36797895 DOI: 10.1103/physreve.107.014304] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2022] [Accepted: 12/15/2022] [Indexed: 01/24/2023]
Abstract
Neural networks have proven to be remarkably successful for a wide range of complicated tasks, from image recognition and object detection to speech recognition and machine translation. One of their successes lies in their ability to predict future dynamics given a suitable training data set. Previous studies have shown how echo state networks (ESNs), a type of recurrent neural networks, can successfully predict both short-term and long-term dynamics of even chaotic systems. This study shows that, remarkably, ESNs can successfully predict dynamical behavior that is qualitatively different from any behavior contained in their training set. Evidence is provided for a fluid dynamics problem where the flow can transition between laminar (ordered) and turbulent (seemingly disordered) regimes. Despite being trained on the turbulent regime only, ESNs are found to predict the existence of laminar behavior. Moreover, the statistics of turbulent-to-laminar and laminar-to-turbulent transitions are also predicted successfully. The utility of ESNs in acting as early-warning generators for transition is discussed. These results are expected to be widely applicable to data-driven modeling of temporal behavior in a range of physical, climate, biological, ecological, and finance models characterized by the presence of tipping points and sudden transitions between several competing states.
Collapse
Affiliation(s)
- Anton Pershin
- Atmospheric, Oceanic and Planetary Physics, University of Oxford, Oxford, United Kingdom and School of Mathematics, University of Leeds, Leeds, OX1 3PU United Kingdom
| | - Cédric Beaume
- School of Mathematics, University of Leeds, Leeds, LS2 9JT United Kingdom
| | - Kuan Li
- School of Mathematics, University of Leeds, Leeds, LS2 9JT United Kingdom
| | - Steven M Tobias
- School of Mathematics, University of Leeds, Leeds, LS2 9JT United Kingdom
| |
Collapse
|
10
|
You Y, Zhang L, Tao P, Liu S, Chen L. Spatiotemporal Transformer Neural Network for Time-Series Forecasting. ENTROPY (BASEL, SWITZERLAND) 2022; 24:1651. [PMID: 36421506 PMCID: PMC9689721 DOI: 10.3390/e24111651] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/05/2022] [Revised: 11/05/2022] [Accepted: 11/08/2022] [Indexed: 06/16/2023]
Abstract
Predicting high-dimensional short-term time-series is a difficult task due to the lack of sufficient information and the curse of dimensionality. To overcome these problems, this study proposes a novel spatiotemporal transformer neural network (STNN) for efficient prediction of short-term time-series with three major features. Firstly, the STNN can accurately and robustly predict a high-dimensional short-term time-series in a multi-step-ahead manner by exploiting high-dimensional/spatial information based on the spatiotemporal information (STI) transformation equation. Secondly, the continuous attention mechanism makes the prediction results more accurate than those of previous studies. Thirdly, we developed continuous spatial self-attention, temporal self-attention, and transformation attention mechanisms to create a bridge between effective spatial information and future temporal evolution information. Fourthly, we show that the STNN model can reconstruct the phase space of the dynamical system, which is explored in the time-series prediction. The experimental results demonstrate that the STNN significantly outperforms the existing methods on various benchmarks and real-world systems in the multi-step-ahead prediction of a short-term time-series.
Collapse
Affiliation(s)
- Yujie You
- College of Computer Science, Sichuan University, Chengdu 610065, China
| | - Le Zhang
- College of Computer Science, Sichuan University, Chengdu 610065, China
- Key Laboratory of Systems Biology, Hangzhou Institute for Advanced Study, University of Chinese Academy of Sciences, Hangzhou 310024, China
- Key Laboratory of Systems Health Science of Zhejiang Province, Hangzhou Institute for Advanced Study, University of Chinese Academy of Sciences, Hangzhou 310024, China
| | - Peng Tao
- Key Laboratory of Systems Health Science of Zhejiang Province, Hangzhou Institute for Advanced Study, University of Chinese Academy of Sciences, Hangzhou 310024, China
| | - Suran Liu
- College of Computer Science, Sichuan University, Chengdu 610065, China
| | - Luonan Chen
- Key Laboratory of Systems Health Science of Zhejiang Province, Hangzhou Institute for Advanced Study, University of Chinese Academy of Sciences, Hangzhou 310024, China
- State Key Laboratory of Cell Biology, Institute of Biochemistry and Cell Biology, Center for Excellence in Molecular Cell Science, Chinese Academy of Sciences, Shanghai 200031, China
- Guangdong Institute of Intelligence Science and Technology, Hengqin, Zhuhai 519031, China
- West China Biomedical Big Data Center, Med-X Center for Informatics, West China Hospital, Sichuan University, Chengdu 610041, China
| |
Collapse
|
11
|
Goldmann M, Mirasso CR, Fischer I, Soriano MC. Learn one size to infer all: Exploiting translational symmetries in delay-dynamical and spatiotemporal systems using scalable neural networks. Phys Rev E 2022; 106:044211. [PMID: 36397530 DOI: 10.1103/physreve.106.044211] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2021] [Accepted: 09/27/2022] [Indexed: 06/16/2023]
Abstract
We design scalable neural networks adapted to translational symmetries in dynamical systems, capable of inferring untrained high-dimensional dynamics for different system sizes. We train these networks to predict the dynamics of delay-dynamical and spatiotemporal systems for a single size. Then, we drive the networks by their own predictions. We demonstrate that by scaling the size of the trained network, we can predict the complex dynamics for larger or smaller system sizes. Thus, the network learns from a single example and by exploiting symmetry properties infers entire bifurcation diagrams.
Collapse
Affiliation(s)
- Mirko Goldmann
- Instituto de Física Interdisciplinar y Sistemas Complejos (IFISC, UIB-CSIC), Campus Universitat de les Illes Balears E-07122, Palma de Mallorca, Spain
| | - Claudio R Mirasso
- Instituto de Física Interdisciplinar y Sistemas Complejos (IFISC, UIB-CSIC), Campus Universitat de les Illes Balears E-07122, Palma de Mallorca, Spain
| | - Ingo Fischer
- Instituto de Física Interdisciplinar y Sistemas Complejos (IFISC, UIB-CSIC), Campus Universitat de les Illes Balears E-07122, Palma de Mallorca, Spain
| | - Miguel C Soriano
- Instituto de Física Interdisciplinar y Sistemas Complejos (IFISC, UIB-CSIC), Campus Universitat de les Illes Balears E-07122, Palma de Mallorca, Spain
| |
Collapse
|
12
|
Whiteaker B, Gerstoft P. Reducing echo state network size with controllability matrices. CHAOS (WOODBURY, N.Y.) 2022; 32:073116. [PMID: 35907714 DOI: 10.1063/5.0071926] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/17/2021] [Accepted: 06/21/2022] [Indexed: 06/15/2023]
Abstract
Echo state networks are a fast training variant of recurrent neural networks excelling at approximating nonlinear dynamical systems and time series prediction. These machine learning models act as nonlinear fading memory filters. While these models benefit from quick training and low complexity, computation demands from a large reservoir matrix are a bottleneck. Using control theory, a reduced size replacement reservoir matrix is found. Starting from a large, task-effective reservoir matrix, we form a controllability matrix whose rank indicates the active sub-manifold and candidate replacement reservoir size. Resulting time speed-ups and reduced memory usage come with minimal error increase to chaotic climate reconstruction or short term prediction. Experiments are performed on simple time series signals and the Lorenz-1963 and Mackey-Glass complex chaotic signals. Observing low error models shows variation of active rank and memory along a sequence of predictions.
Collapse
Affiliation(s)
- Brian Whiteaker
- Scripps Institution of Oceanography, University of California at San Diego, La Jolla, California 92093-0238, USA
| | - Peter Gerstoft
- Scripps Institution of Oceanography, University of California at San Diego, La Jolla, California 92093-0238, USA
| |
Collapse
|
13
|
Roy M, Senapati A, Poria S, Mishra A, Hens C. Role of assortativity in predicting burst synchronization using echo state network. Phys Rev E 2022; 105:064205. [PMID: 35854538 DOI: 10.1103/physreve.105.064205] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2021] [Accepted: 05/11/2022] [Indexed: 06/15/2023]
Abstract
In this study, we use a reservoir computing based echo state network (ESN) to predict the collective burst synchronization of neurons. Specifically, we investigate the ability of ESN in predicting the burst synchronization of an ensemble of Rulkov neurons placed on a scale-free network. We have shown that a limited number of nodal dynamics used as input in the machine can capture the real trend of burst synchronization in this network. Further, we investigate the proper selection of nodal inputs of degree-degree (positive and negative) correlated networks. We show that for a disassortative network, selection of different input nodes based on degree has no significant role in the machine's prediction. However, in the case of assortative network, training the machine with the information (i.e., time series) of low degree nodes gives better results in predicting the burst synchronization. The results are found to be consistent with the investigation carried out with a continuous time Hindmarsh-Rose neuron model. Furthermore, the role of hyperparameters like spectral radius and leaking parameter of ESN on the prediction process has been examined. Finally, we explain the underlying mechanism responsible for observing these differences in the prediction in a degree correlated network.
Collapse
Affiliation(s)
- Mousumi Roy
- Department of Applied Mathematics, University of Calcutta, 92, A.P.C. Road, Kolkata 700009, India
| | - Abhishek Senapati
- Center for Advanced Systems Understanding (CASUS), 02826 Görlitz, Germany
| | - Swarup Poria
- Department of Applied Mathematics, University of Calcutta, 92, A.P.C. Road, Kolkata 700009, India
| | - Arindam Mishra
- Division of Dynamics, Lodz University of Technology, Stefanowskiego 1/15, 90924 Lodz, Poland
| | - Chittaranjan Hens
- Physics and Applied Mathematics Unit, Indian Statistical Institute, Kolkata 700108, India
| |
Collapse
|
14
|
Wang L, Fan H, Xiao J, Lan Y, Wang X. Criticality in reservoir computer of coupled phase oscillators. Phys Rev E 2022; 105:L052201. [PMID: 35706173 DOI: 10.1103/physreve.105.l052201] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2021] [Accepted: 04/12/2022] [Indexed: 06/15/2023]
Abstract
Accumulating evidence shows that the cerebral cortex is operating near a critical state featured by power-law size distribution of neural avalanche activities, yet evidence of this critical state in artificial neural networks mimicking the cerebral cortex is still lacking. Here we design an artificial neural network of coupled phase oscillators and, by the technique of reservoir computing in machine learning, train it for predicting chaos. It is found that when the machine is properly trained, oscillators in the reservoir are synchronized into clusters whose sizes follow a power-law distribution. This feature, however, is absent when the machine is poorly trained. Additionally, it is found that despite the synchronization degree of the original network, once properly trained, the reservoir network is always developed to the same critical state, exemplifying the "attractor" nature of this state in machine learning. The generality of the results is verified in different reservoir models and by different target systems, and it is found that the scaling exponent of the distribution is independent of the reservoir details and the bifurcation parameters of the target system, but is modified when the dynamics of the target system is changed to a different type. The findings shed light on the nature of machine learning, and are helpful to the design of high-performance machines in physical systems.
Collapse
Affiliation(s)
- Liang Wang
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710062, China
| | - Huawei Fan
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710062, China
| | - Jinghua Xiao
- School of Science, Beijing University of Posts and Telecommunications, Beijing 100876, China
| | - Yueheng Lan
- School of Science, Beijing University of Posts and Telecommunications, Beijing 100876, China
| | - Xingang Wang
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710062, China
| |
Collapse
|
15
|
Ren B, Ma H. Global optimization of hyper-parameters in reservoir computing. ELECTRONIC RESEARCH ARCHIVE 2022; 30:2719-2729. [DOI: 10.3934/era.2022139] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/01/2023]
Abstract
<abstract><p>Reservoir computing has emerged as a powerful and efficient machine learning tool especially in the reconstruction of many complex systems even for chaotic systems only based on the observational data. Though fruitful advances have been extensively studied, how to capture the art of hyper-parameter settings to construct efficient RC is still a long-standing and urgent problem. In contrast to the local manner of many works which aim to optimize one hyper-parameter while keeping others constant, in this work, we propose a global optimization framework using simulated annealing technique to find the optimal architecture of the randomly generated networks for a successful RC. Based on the optimized results, we further study several important properties of some hyper-parameters. Particularly, we find that the globally optimized reservoir network has a largest singular value significantly larger than one, which is contrary to the sufficient condition reported in the literature to guarantee the echo state property. We further reveal the mechanism of this phenomenon with a simplified model and the theory of nonlinear dynamical systems.</p></abstract>
Collapse
|
16
|
Tan E, Corrêa D, Stemler T, Small M. Grading your models: Assessing dynamics learning of models using persistent homology. CHAOS (WOODBURY, N.Y.) 2021; 31:123109. [PMID: 34972316 DOI: 10.1063/5.0073722] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/03/2021] [Accepted: 11/18/2021] [Indexed: 06/14/2023]
Abstract
Assessing model accuracy for complex and chaotic systems is a non-trivial task that often relies on the calculation of dynamical invariants, such as Lyapunov exponents and correlation dimensions. Well-performing models are able to replicate the long-term dynamics and ergodic properties of the desired system. We term this phenomenon "dynamics learning." However, existing estimates based on dynamical invariants, such as Lyapunov exponents and correlation dimensions, are not unique to each system, not necessarily robust to noise, and struggle with detecting pathological errors, such as errors in the manifold density distribution. This can make meaningful and accurate model assessment difficult. We explore the use of a topological data analysis technique, persistent homology, applied to uniformly sampled trajectories from constructed reservoir models of the Lorenz system to assess the learning quality of a model. A proposed persistent homology point summary, conformance, was able to identify models with successful dynamics learning and detect discrepancies in the manifold density distribution.
Collapse
Affiliation(s)
- Eugene Tan
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| | - Débora Corrêa
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| | - Thomas Stemler
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| | - Michael Small
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia
| |
Collapse
|
17
|
Röhm A, Gauthier DJ, Fischer I. Model-free inference of unseen attractors: Reconstructing phase space features from a single noisy trajectory using reservoir computing. CHAOS (WOODBURY, N.Y.) 2021; 31:103127. [PMID: 34717323 DOI: 10.1063/5.0065813] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/05/2021] [Accepted: 09/25/2021] [Indexed: 06/13/2023]
Abstract
Reservoir computers are powerful tools for chaotic time series prediction. They can be trained to approximate phase space flows and can thus both predict future values to a high accuracy and reconstruct the general properties of a chaotic attractor without requiring a model. In this work, we show that the ability to learn the dynamics of a complex system can be extended to systems with multiple co-existing attractors, here a four-dimensional extension of the well-known Lorenz chaotic system. We demonstrate that a reservoir computer can infer entirely unexplored parts of the phase space; a properly trained reservoir computer can predict the existence of attractors that were never approached during training and, therefore, are labeled as unseen. We provide examples where attractor inference is achieved after training solely on a single noisy trajectory.
Collapse
Affiliation(s)
- André Röhm
- Instituto de Física Interdisciplinar y Sistemas Complejos, IFISC (CSIC-UIB), Campus Universitat Illes Balears, E-07122 Palma de Mallorca, Spain
| | - Daniel J Gauthier
- Department of Physics, The Ohio State University, 191 West Woodruff Ave., Columbus, Ohio 43210, USA
| | - Ingo Fischer
- Instituto de Física Interdisciplinar y Sistemas Complejos, IFISC (CSIC-UIB), Campus Universitat Illes Balears, E-07122 Palma de Mallorca, Spain
| |
Collapse
|
18
|
Flynn A, Herteux J, Tsachouridis VA, Räth C, Amann A. Symmetry kills the square in a multifunctional reservoir computer. CHAOS (WOODBURY, N.Y.) 2021; 31:073122. [PMID: 34340331 DOI: 10.1063/5.0055699] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/01/2021] [Accepted: 06/22/2021] [Indexed: 06/13/2023]
Abstract
The learning capabilities of a reservoir computer (RC) can be stifled due to symmetry in its design. Including quadratic terms in the training of a RC produces a "square readout matrix" that breaks the symmetry to quell the influence of "mirror-attractors," which are inverted copies of the RC's solutions in state space. In this paper, we prove analytically that certain symmetries in the training data forbid the square readout matrix to exist. These analytical results are explored numerically from the perspective of "multifunctionality," by training the RC to specifically reconstruct a coexistence of the Lorenz attractor and its mirror-attractor. We demonstrate that the square readout matrix emerges when the position of one attractor is slightly altered, even if there are overlapping regions between the attractors or if there is a second pair of attractors. We also find that at large spectral radius values of the RC's internal connections, the square readout matrix reappears prior to the RC crossing the edge of chaos.
Collapse
Affiliation(s)
- Andrew Flynn
- School of Mathematical Sciences, University College Cork, Cork T12 XF62, Ireland
| | - Joschka Herteux
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft-und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| | | | - Christoph Räth
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft-und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| | - Andreas Amann
- School of Mathematical Sciences, University College Cork, Cork T12 XF62, Ireland
| |
Collapse
|
19
|
Controlling nonlinear dynamical systems into arbitrary states using machine learning. Sci Rep 2021; 11:12991. [PMID: 34155228 PMCID: PMC8217470 DOI: 10.1038/s41598-021-92244-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2021] [Accepted: 06/07/2021] [Indexed: 11/08/2022] Open
Abstract
Controlling nonlinear dynamical systems is a central task in many different areas of science and engineering. Chaotic systems can be stabilized (or chaotified) with small perturbations, yet existing approaches either require knowledge about the underlying system equations or large data sets as they rely on phase space methods. In this work we propose a novel and fully data driven scheme relying on machine learning (ML), which generalizes control techniques of chaotic systems without requiring a mathematical model for its dynamics. Exploiting recently developed ML-based prediction capabilities, we demonstrate that nonlinear systems can be forced to stay in arbitrary dynamical target states coming from any initial state. We outline and validate our approach using the examples of the Lorenz and the Rössler system and show how these systems can very accurately be brought not only to periodic, but even to intermittent and different chaotic behavior. Having this highly flexible control scheme with little demands on the amount of required data on hand, we briefly discuss possible applications ranging from engineering to medicine.
Collapse
|
20
|
Racca A, Magri L. Robust Optimization and Validation of Echo State Networks for learning chaotic dynamics. Neural Netw 2021; 142:252-268. [PMID: 34034072 DOI: 10.1016/j.neunet.2021.05.004] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2021] [Revised: 05/03/2021] [Accepted: 05/06/2021] [Indexed: 11/26/2022]
Abstract
An approach to the time-accurate prediction of chaotic solutions is by learning temporal patterns from data. Echo State Networks (ESNs), which are a class of Reservoir Computing, can accurately predict the chaotic dynamics well beyond the predictability time. Existing studies, however, also showed that small changes in the hyperparameters may markedly affect the network's performance. The overarching aim of this paper is to improve the robustness in the selection of hyperparameters in Echo State Networks for the time-accurate prediction of chaotic solutions. We define the robustness of a validation strategy as its ability to select hyperparameters that perform consistently between validation and test sets. The goal is three-fold. First, we investigate routinely used validation strategies. Second, we propose the Recycle Validation, and the chaotic versions of existing validation strategies, to specifically tackle the forecasting of chaotic systems. Third, we compare Bayesian optimization with the traditional grid search for optimal hyperparameter selection. Numerical tests are performed on prototypical nonlinear systems that have chaotic and quasiperiodic solutions, such as the Lorenz and Lorenz-96 systems, and the Kuznetsov oscillator. Both model-free and model-informed Echo State Networks are analysed. By comparing the networks' performance in learning chaotic (unpredictable) versus quasiperiodic (predictable) solutions, we highlight fundamental challenges in learning chaotic solutions. The proposed validation strategies, which are based on the dynamical systems properties of chaotic time series, are shown to outperform the state-of-the-art validation strategies. Because the strategies are principled - they are based on chaos theory such as the Lyapunov time - they can be applied to other Recurrent Neural Networks architectures with little modification. This work opens up new possibilities for the robust design and application of Echo State Networks, and Recurrent Neural Networks, to the time-accurate prediction of chaotic systems.
Collapse
Affiliation(s)
- Alberto Racca
- Department of Engineering, University of Cambridge, Trumpington Street, Cambridge CB2 1PZ, UK
| | - Luca Magri
- Department of Engineering, University of Cambridge, Trumpington Street, Cambridge CB2 1PZ, UK; Aeronautics Department, Imperial College London, Exhibition Rd, London, SW7 2AZ, UK; The Alan Turing Institute, 96 Euston Road, London, England, NW1 2DB, UK; Institute for Advanced Study, Technical University of Munich, Lichtenbergstrasse 2a, 85748 Garching, Germany(1).
| |
Collapse
|
21
|
Guo Y, Zhang H, Wang L, Fan H, Xiao J, Wang X. Transfer learning of chaotic systems. CHAOS (WOODBURY, N.Y.) 2021; 31:011104. [PMID: 33754764 DOI: 10.1063/5.0033870] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/20/2020] [Accepted: 12/24/2020] [Indexed: 06/12/2023]
Abstract
Can a neural network trained by the time series of system A be used to predict the evolution of system B? This problem, knowing as transfer learning in a broad sense, is of great importance in machine learning and data mining yet has not been addressed for chaotic systems. Here, we investigate transfer learning of chaotic systems from the perspective of synchronization-based state inference, in which a reservoir computer trained by chaotic system A is used to infer the unmeasured variables of chaotic system B, while A is different from B in either parameter or dynamics. It is found that if systems A and B are different in parameter, the reservoir computer can be well synchronized to system B. However, if systems A and B are different in dynamics, the reservoir computer fails to synchronize with system B in general. Knowledge transfer along a chain of coupled reservoir computers is also studied, and it is found that, although the reservoir computers are trained by different systems, the unmeasured variables of the driving system can be successfully inferred by the remote reservoir computer. Finally, by an experiment of chaotic pendulum, we demonstrate that the knowledge learned from the modeling system can be transferred and used to predict the evolution of the experimental system.
Collapse
Affiliation(s)
- Yali Guo
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710062, China
| | - Han Zhang
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710062, China
| | - Liang Wang
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710062, China
| | - Huawei Fan
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710062, China
| | - Jinghua Xiao
- School of Science, Beijing University of Posts and Telecommunications, Beijing 100876, China
| | - Xingang Wang
- School of Physics and Information Technology, Shaanxi Normal University, Xi'an 710062, China
| |
Collapse
|
22
|
Herteux J, Räth C. Breaking symmetries of the reservoir equations in echo state networks. CHAOS (WOODBURY, N.Y.) 2020; 30:123142. [PMID: 33380046 DOI: 10.1063/5.0028993] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/08/2020] [Accepted: 12/03/2020] [Indexed: 06/12/2023]
Abstract
Reservoir computing has repeatedly been shown to be extremely successful in the prediction of nonlinear time-series. However, there is no complete understanding of the proper design of a reservoir yet. We find that the simplest popular setup has a harmful symmetry, which leads to the prediction of what we call mirror-attractor. We prove this analytically. Similar problems can arise in a general context, and we use them to explain the success or failure of some designs. The symmetry is a direct consequence of the hyperbolic tangent activation function. Furthermore, four ways to break the symmetry are compared numerically: A bias in the output, a shift in the input, a quadratic term in the readout, and a mixture of even and odd activation functions. First, we test their susceptibility to the mirror-attractor. Second, we evaluate their performance on the task of predicting Lorenz data with the mean shifted to zero. The short-time prediction is measured with the forecast horizon while the largest Lyapunov exponent and the correlation dimension are used to represent the climate. Finally, the same analysis is repeated on a combined dataset of the Lorenz attractor and the Halvorsen attractor, which we designed to reveal potential problems with symmetry. We find that all methods except the output bias are able to fully break the symmetry with input shift and quadratic readout performing the best overall.
Collapse
Affiliation(s)
- Joschka Herteux
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft- und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| | - Christoph Räth
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft- und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| |
Collapse
|
23
|
Mabrouk YA, Räth C. Calibrated reservoir computers. CHAOS (WOODBURY, N.Y.) 2020; 30:113134. [PMID: 33261366 DOI: 10.1063/5.0030651] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/23/2020] [Accepted: 10/28/2020] [Indexed: 06/12/2023]
Abstract
We observe the presence of infinitely fine-scaled alternations within the performance landscape of reservoir computers aimed for chaotic data forecasting. We investigate the emergence of the observed structures by means of variations of the transversal stability of the synchronization manifold relating the observational and internal dynamical states. Finally, we deduce a simple calibration method in order to attenuate the thus evidenced performance uncertainty.
Collapse
Affiliation(s)
- Y A Mabrouk
- Fakultät für Physik, Ludwig-Maximilians-Universität München, Schellingstraße 4, 80779 München, Germany
| | - C Räth
- Fakultät für Physik, Ludwig-Maximilians-Universität München, Schellingstraße 4, 80779 München, Germany
| |
Collapse
|
24
|
Chen X, Weng T, Yang H, Gu C, Zhang J, Small M. Mapping topological characteristics of dynamical systems into neural networks: A reservoir computing approach. Phys Rev E 2020; 102:033314. [PMID: 33075895 DOI: 10.1103/physreve.102.033314] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/18/2020] [Accepted: 09/09/2020] [Indexed: 11/07/2022]
Abstract
Significant advances have recently been made in modeling chaotic systems with the reservoir computing approach, especially for prediction. We find that although state prediction of the trained reservoir computer will gradually deviate from the actual trajectory of the original system, the associated geometric features remain invariant. Specifically, we show that the typical geometric metrics including the correlation dimension, the multiscale entropy, and the memory effect are nearly identical between the trained reservoir computer and its learned chaotic systems. We further demonstrate this fact on a broad range of chaotic systems ranging from discrete and continuous chaotic systems to hyperchaotic systems. Our findings suggest that the successfully reservoir computer may be topologically conjugate to an observed dynamical system.
Collapse
Affiliation(s)
- Xiaolu Chen
- Business School, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Tongfeng Weng
- Alibaba Research Center for Complexity Sciences, Hangzhou Normal University, Hangzhou 311121, People's Republic of China
| | - Huijie Yang
- Business School, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Changgui Gu
- Business School, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Jie Zhang
- Institute of Science and Technology for Brain-Inspired Intellegence, Fudan University, Shanghai 200433, People's Republic of China
| | - Michael Small
- Complex Systems Group, Department of Mathematics and Statistics, The University of Western Australia, Crawley, Western Australia 6009, Australia.,Mineral Resources, CSIRO, Kensington, Western Australia 6151, Australia
| |
Collapse
|
25
|
Carroll TL. Path length statistics in reservoir computers. CHAOS (WOODBURY, N.Y.) 2020; 30:083130. [PMID: 32872832 DOI: 10.1063/5.0014643] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/20/2020] [Accepted: 08/03/2020] [Indexed: 06/11/2023]
Abstract
Because reservoir computers are high dimensional dynamical systems, designing a good reservoir computer is difficult. In many cases, the designer must search a large nonlinear parameter space, and each step of the search requires simulating the full reservoir computer. In this work, I show that a simple statistic based on the mean path length between nodes in the reservoir computer is correlated with better reservoir computer performance. The statistic predicts the diversity of signals produced by the reservoir computer, as measured by the covariance matrix of the reservoir computer. This statistic by itself is not sufficient to predict reservoir computer performance because not only must the reservoir computer produce a diverse set of signals, it must be well matched to the training signals. Nevertheless, this path length statistic allows the designer to eliminate some network configurations from consideration without having to actually simulate the reservoir computer, reducing the complexity of the design process.
Collapse
Affiliation(s)
- T L Carroll
- U.S. Naval Research Lab, Washington, DC 20375, USA
| |
Collapse
|
26
|
Haluszczynski A, Aumeier J, Herteux J, Räth C. Reducing network size and improving prediction stability of reservoir computing. CHAOS (WOODBURY, N.Y.) 2020; 30:063136. [PMID: 32611106 DOI: 10.1063/5.0006869] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/06/2020] [Accepted: 06/05/2020] [Indexed: 06/11/2023]
Abstract
Reservoir computing is a very promising approach for the prediction of complex nonlinear dynamical systems. Besides capturing the exact short-term trajectories of nonlinear systems, it has also proved to reproduce its characteristic long-term properties very accurately. However, predictions do not always work equivalently well. It has been shown that both short- and long-term predictions vary significantly among different random realizations of the reservoir. In order to gain an understanding on when reservoir computing works best, we investigate some differential properties of the respective realization of the reservoir in a systematic way. We find that removing nodes that correspond to the largest weights in the output regression matrix reduces outliers and improves overall prediction quality. Moreover, this allows to effectively reduce the network size and, therefore, increase computational efficiency. In addition, we use a nonlinear scaling factor in the hyperbolic tangent of the activation function. This adjusts the response of the activation function to the range of values of the input variables of the nodes. As a consequence, this reduces the number of outliers significantly and increases both the short- and long-term prediction quality for the nonlinear systems investigated in this study. Our results demonstrate that a large optimization potential lies in the systematical refinement of the differential reservoir properties for a given dataset.
Collapse
Affiliation(s)
- Alexander Haluszczynski
- Department of Physics, Ludwig-Maximilians-Universität, Schellingstraße 4, 80799 Munich, Germany
| | - Jonas Aumeier
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft- und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| | - Joschka Herteux
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft- und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| | - Christoph Räth
- Institut für Materialphysik im Weltraum, Deutsches Zentrum für Luft- und Raumfahrt, Münchner Str. 20, 82234 Wessling, Germany
| |
Collapse
|
27
|
Tang Y, Kurths J, Lin W, Ott E, Kocarev L. Introduction to Focus Issue: When machine learning meets complex systems: Networks, chaos, and nonlinear dynamics. CHAOS (WOODBURY, N.Y.) 2020; 30:063151. [PMID: 32611112 DOI: 10.1063/5.0016505] [Citation(s) in RCA: 37] [Impact Index Per Article: 7.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/04/2020] [Accepted: 06/05/2020] [Indexed: 06/11/2023]
Affiliation(s)
- Yang Tang
- Key Laboratory of Advanced Control and Optimization for Chemical Processes, Ministry of Education, East China University of Science and Technology, Shanghai, China
| | - Jürgen Kurths
- Potsdam Institute for Climate Impact Research, Potsdam 14473, Germany
| | - Wei Lin
- Center for Computational Systems Biology of ISTBI and Research Institute of Intelligent Complex Systems, Fudan University, Shanghai 200433, China
| | - Edward Ott
- Department of Physics, University of Maryland, College Park, Maryland 20742, USA
| | - Ljupco Kocarev
- Macedonian Academy of Sciences and Arts, 1000 Skopje, Macedonia
| |
Collapse
|