1
|
Guo X, Zheng Z, Cheong KH, Zou Q, Tiwari P, Ding Y. Sequence homology score-based deep fuzzy network for identifying therapeutic peptides. Neural Netw 2024; 178:106458. [PMID: 38901093 DOI: 10.1016/j.neunet.2024.106458] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2023] [Revised: 05/29/2024] [Accepted: 06/09/2024] [Indexed: 06/22/2024]
Abstract
The detection of therapeutic peptides is a topic of immense interest in the biomedical field. Conventional biochemical experiment-based detection techniques are tedious and time-consuming. Computational biology has become a useful tool for improving the detection efficiency of therapeutic peptides. Most computational methods do not consider the deviation caused by noise. To improve the generalization performance of therapeutic peptide prediction methods, this work presents a sequence homology score-based deep fuzzy echo-state network with maximizing mixture correntropy (SHS-DFESN-MMC) model. Our method is compared with the existing methods on eight types of therapeutic peptide datasets. The model parameters are determined by 10 fold cross-validation on their training sets and verified by independent test sets. Across the 8 datasets, the average area under the receiver operating characteristic curve (AUC) values of SHS-DFESN-MMC are the highest on both the training (0.926) and independent sets (0.923).
Collapse
Affiliation(s)
- Xiaoyi Guo
- Institute of Fundamental and Frontier Sciences, University of Electronic Science and Technology of China, Chengdu, 610054, PR China; Quzhou People's Hospital, Quzhou Affiliated Hospital of Wenzhou Medical University, Quzhou, 324000, PR China; Division of Mathematical Sciences, School of Physical and Mathematical Sciences, Nanyang Technological University, S637371, Singapore.
| | - Ziyu Zheng
- Department of Mathematical Sciences, University of Nottingham Ningbo, Ningbo, 315100, PR China.
| | - Kang Hao Cheong
- Division of Mathematical Sciences, School of Physical and Mathematical Sciences, Nanyang Technological University, S637371, Singapore; College of Computing and Data Science, Nanyang Technological University, S639798, Singapore.
| | - Quan Zou
- Institute of Fundamental and Frontier Sciences, University of Electronic Science and Technology of China, Chengdu, 610054, PR China; Yangtze Delta Region Institute (Quzhou), University of Electronic Science and Technology of China, Quzhou, 324000, PR China.
| | - Prayag Tiwari
- School of Information Technology, Halmstad University, Sweden.
| | - Yijie Ding
- Yangtze Delta Region Institute (Quzhou), University of Electronic Science and Technology of China, Quzhou, 324000, PR China.
| |
Collapse
|
2
|
Yan M, Huang C, Bienstman P, Tino P, Lin W, Sun J. Emerging opportunities and challenges for the future of reservoir computing. Nat Commun 2024; 15:2056. [PMID: 38448438 PMCID: PMC10917819 DOI: 10.1038/s41467-024-45187-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2023] [Accepted: 01/16/2024] [Indexed: 03/08/2024] Open
Abstract
Reservoir computing originates in the early 2000s, the core idea being to utilize dynamical systems as reservoirs (nonlinear generalizations of standard bases) to adaptively learn spatiotemporal features and hidden patterns in complex time series. Shown to have the potential of achieving higher-precision prediction in chaotic systems, those pioneering works led to a great amount of interest and follow-ups in the community of nonlinear dynamics and complex systems. To unlock the full capabilities of reservoir computing towards a fast, lightweight, and significantly more interpretable learning framework for temporal dynamical systems, substantially more research is needed. This Perspective intends to elucidate the parallel progress of mathematical theory, algorithm design and experimental realizations of reservoir computing, and identify emerging opportunities as well as existing challenges for large-scale industrial adoption of reservoir computing, together with a few ideas and viewpoints on how some of those challenges might be resolved with joint efforts by academic and industrial researchers across multiple disciplines.
Collapse
Affiliation(s)
- Min Yan
- Theory Lab, Central Research Institute, 2012 Labs, Huawei Technologies Co. Ltd., Hong Kong SAR, China
| | - Can Huang
- Theory Lab, Central Research Institute, 2012 Labs, Huawei Technologies Co. Ltd., Hong Kong SAR, China.
| | - Peter Bienstman
- Photonics Research Group, Department of Information Technology, Ghent University, Gent, Belgium
| | - Peter Tino
- School of Computer Science, The University of Birmingham, Birmingham, B15 2TT, United Kingdom
| | - Wei Lin
- Research Institute of Intelligent Complex Systems, Fudan University, Shanghai, 200433, China
- School of Mathematical Sciences, SCMS, SCAM, and CCSB, Fudan University, Shanghai, 200433, China
| | - Jie Sun
- Theory Lab, Central Research Institute, 2012 Labs, Huawei Technologies Co. Ltd., Hong Kong SAR, China.
| |
Collapse
|
3
|
Zhang L, Chen Z, Lu CT, Zhao L. Fast and adaptive dynamics-on-graphs to dynamics-of-graphs translation. Front Big Data 2023; 6:1274135. [PMID: 38045094 PMCID: PMC10691542 DOI: 10.3389/fdata.2023.1274135] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2023] [Accepted: 10/20/2023] [Indexed: 12/05/2023] Open
Abstract
Numerous networks in the real world change with time, producing dynamic graphs such as human mobility networks and brain networks. Typically, the "dynamics on graphs" (e.g., changing node attribute values) are visible, and they may be connected to and suggestive of the "dynamics of graphs" (e.g., evolution of the graph topology). Due to two fundamental obstacles, modeling and mapping between them have not been thoroughly explored: (1) the difficulty of developing a highly adaptable model without solid hypotheses and (2) the ineffectiveness and slowness of processing data with varying granularity. To solve these issues, we offer a novel scalable deep echo-state graph dynamics encoder for networks with significant temporal duration and dimensions. A novel neural architecture search (NAS) technique is then proposed and tailored for the deep echo-state encoder to ensure strong learnability. Extensive experiments on synthetic and actual application data illustrate the proposed method's exceptional effectiveness and efficiency.
Collapse
Affiliation(s)
- Lei Zhang
- Department of Computer Science, Virginia Tech, Falls Church, VA, United States
| | - Zhiqian Chen
- Department of Computer Science and Engineering, Mississippi State University, Mississippi, MS, United States
| | - Chang-Tien Lu
- Department of Computer Science, Virginia Tech, Falls Church, VA, United States
| | - Liang Zhao
- Department of Computer Science, Emory University, Atlanta, GA, United States
| |
Collapse
|
4
|
Na X, Ren W, Liu M, Han M. Hierarchical Echo State Network With Sparse Learning: A Method for Multidimensional Chaotic Time Series Prediction. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:9302-9313. [PMID: 35333719 DOI: 10.1109/tnnls.2022.3157830] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Echo state network (ESN), a type of special recurrent neural network with a large-scale randomly fixed hidden layer (called a reservoir) and an adaptable linear output layer, has been widely employed in the field of time series analysis and modeling. However, when tackling the problem of multidimensional chaotic time series prediction, due to the randomly generated rules for input and reservoir weights, not only the representation of valuable variables is enriched but also redundant and irrelevant information is accumulated inevitably. To remove the redundant components, reduce the approximate collinearity among echo-state information, and improve the generalization and stability, a new method called hierarchical ESN with sparse learning (HESN-SL) is proposed. The HESN-SL mines and captures the latent evolution patterns hidden from the dynamic system by means of layer-by-layer processing in stacked reservoirs, and leverage monotone accelerated proximal gradient algorithm to train a sparse output layer with variable selection capability. Meanwhile, we further prove that the HESN-SL satisfies the echo state property, which guarantees the stability and convergence of the proposed model when applied to time series prediction. Experimental results on two synthetic chaotic systems and a real-world meteorological dataset illustrate the proposed HESN-SL outperforms both original ESN and existing hierarchical ESN-based models for multidimensional chaotic time series prediction.
Collapse
|
5
|
Xia J, Chu J, Leng S, Ma H. Reservoir computing decoupling memory-nonlinearity trade-off. CHAOS (WOODBURY, N.Y.) 2023; 33:113120. [PMID: 37967262 DOI: 10.1063/5.0156224] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Accepted: 10/17/2023] [Indexed: 11/17/2023]
Abstract
Reservoir computing (RC), a variant recurrent neural network, has very compact architecture and ability to efficiently reconstruct nonlinear dynamics by combining both memory capacity and nonlinear transformations. However, in the standard RC framework, there is a trade-off between memory capacity and nonlinear mapping, which limits its ability to handle complex tasks with long-term dependencies. To overcome this limitation, this paper proposes a new RC framework called neural delayed reservoir computing (ND-RC) with a chain structure reservoir that can decouple the memory capacity and nonlinearity, allowing for independent tuning of them, respectively. The proposed ND-RC model offers a promising solution to the memory-nonlinearity trade-off problem in RC and provides a more flexible and effective approach for modeling complex nonlinear systems with long-term dependencies. The proposed ND-RC framework is validated with typical benchmark nonlinear systems and is particularly successful in reconstructing and predicting the Mackey-Glass system with high time delays. The memory-nonlinearity decoupling ability is further confirmed by several standard tests.
Collapse
Affiliation(s)
- Ji Xia
- School of Mathematical Sciences, Soochow University, Suzhou 215001, China
| | - Junyu Chu
- School of Mathematical Sciences, Soochow University, Suzhou 215001, China
| | - Siyang Leng
- Academy for Engineering and Technology and Research Institute of Intelligent Complex Systems, Fudan University, Shanghai 200433, China
| | - Huanfei Ma
- School of Mathematical Sciences, Soochow University, Suzhou 215001, China
| |
Collapse
|
6
|
Steiner P, Jalalvand A, Birkholz P. Cluster-Based Input Weight Initialization for Echo State Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:7648-7659. [PMID: 35120012 DOI: 10.1109/tnnls.2022.3145565] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Echo state networks (ESNs) are a special type of recurrent neural networks (RNNs), in which the input and recurrent connections are traditionally generated randomly, and only the output weights are trained. Despite the recent success of ESNs in various tasks of audio, image, and radar recognition, we postulate that a purely random initialization is not the ideal way of initializing ESNs. The aim of this work is to propose an unsupervised initialization of the input connections using the K -means algorithm on the training data. We show that for a large variety of datasets, this initialization performs equivalently or superior than a randomly initialized ESN while needing significantly less reservoir neurons. Furthermore, we discuss that this approach provides the opportunity to estimate a suitable size of the reservoir based on prior knowledge about the data.
Collapse
|
7
|
Gao R, Li R, Hu M, Suganthan PN, Yuen KF. Online dynamic ensemble deep random vector functional link neural network for forecasting. Neural Netw 2023; 166:51-69. [PMID: 37480769 DOI: 10.1016/j.neunet.2023.06.042] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2022] [Revised: 06/09/2023] [Accepted: 06/28/2023] [Indexed: 07/24/2023]
Abstract
This paper proposes a three-stage online deep learning model for time series based on the ensemble deep random vector functional link (edRVFL). The edRVFL stacks multiple randomized layers to enhance the single-layer RVFL's representation ability. Each hidden layer's representation is utilized for training an output layer, and the ensemble of all output layers forms the edRVFL's output. However, the original edRVFL is not designed for online learning, and the randomized nature of the features is harmful to extracting meaningful temporal features. In order to address the limitations and extend the edRVFL to an online learning mode, this paper proposes a dynamic edRVFL consisting of three online components, the online decomposition, the online training, and the online dynamic ensemble. First, an online decomposition is utilized as a feature engineering block for the edRVFL. Then, an online learning algorithm is designed to learn the edRVFL. Finally, an online dynamic ensemble method, which can measure the change in the distribution, is proposed for aggregating all layers' outputs. This paper evaluates and compares the proposed model with state-of-the-art methods on sixteen time series.
Collapse
Affiliation(s)
- Ruobin Gao
- School of Civil & Environmental Engineering, Nanyang Technological University, Singapore.
| | - Ruilin Li
- School of Electrical & Electronic Engineering, Nanyang Technological University, Singapore.
| | - Minghui Hu
- School of Electrical & Electronic Engineering, Nanyang Technological University, Singapore.
| | - P N Suganthan
- KINDI Center for Computing Research, College of Engineering, Qatar University, Doha, Qatar.
| | - Kum Fai Yuen
- School of Civil & Environmental Engineering, Nanyang Technological University, Singapore.
| |
Collapse
|
8
|
Mustaqeem, El Saddik A, Alotaibi FS, Pham NT. AAD-Net: Advanced end-to-end speech signal system for human emotion detection & recognition using attention-based deep echo state network. Knowl Based Syst 2023. [DOI: 10.1016/j.knosys.2023.110525] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/03/2023]
|
9
|
Deep echo state networks in data marketplaces. MACHINE LEARNING WITH APPLICATIONS 2023. [DOI: 10.1016/j.mlwa.2023.100456] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/25/2023] Open
|
10
|
Li Z, Liu Y, Tanaka G. Multi-Reservoir Echo State Networks with Hodrick–Prescott Filter for nonlinear time-series prediction. Appl Soft Comput 2023. [DOI: 10.1016/j.asoc.2023.110021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/19/2023]
|
11
|
Guo X, Tiwari P, Zou Q, Ding Y. Subspace projection-based weighted echo state networks for predicting therapeutic peptides. Knowl Based Syst 2023. [DOI: 10.1016/j.knosys.2023.110307] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/15/2023]
|
12
|
Wang X, Jin Y, Hao K. Computational Modeling of Structural Synaptic Plasticity in Echo State Networks. IEEE TRANSACTIONS ON CYBERNETICS 2022; 52:11254-11266. [PMID: 33760748 DOI: 10.1109/tcyb.2021.3060466] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Most existing studies on computational modeling of neural plasticity have focused on synaptic plasticity. However, regulation of the internal weights in the reservoir based on synaptic plasticity often results in unstable learning dynamics. In this article, a structural synaptic plasticity learning rule is proposed to train the weights and add or remove neurons within the reservoir, which is shown to be able to alleviate the instability of the synaptic plasticity, and to contribute to increase the memory capacity of the network as well. Our experimental results also reveal that a few stronger connections may last for a longer period of time in a constantly changing network structure, and are relatively resistant to decay or disruptions in the learning process. These results are consistent with the evidence observed in biological systems. Finally, we show that an echo state network (ESN) using the proposed structural plasticity rule outperforms an ESN using synaptic plasticity and three state-of-the-art ESNs on four benchmark tasks.
Collapse
|
13
|
Li Y, Li F. Growing deep echo state network with supervised learning for time series prediction. Appl Soft Comput 2022. [DOI: 10.1016/j.asoc.2022.109454] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
14
|
Optimal echo state network parameters based on behavioural spaces. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.06.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
15
|
Towards learning trustworthily, automatically, and with guarantees on graphs: An overview. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.04.072] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
16
|
Wang H, Liu Y, Lu P, Luo Y, Wang D, Xu X. Echo state network with logistic mapping and bias dropout for time series prediction. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.03.018] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
17
|
Jalalvand A, Abbate J, Conlin R, Verdoolaege G, Kolemen E. Real-Time and Adaptive Reservoir Computing With Application to Profile Prediction in Fusion Plasma. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:2630-2641. [PMID: 34115598 DOI: 10.1109/tnnls.2021.3085504] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Nuclear fusion is a promising alternative to address the problem of sustainable energy production. The tokamak is an approach to fusion based on magnetic plasma confinement, constituting a complex physical system with many control challenges. We study the characteristics and optimization of reservoir computing (RC) for real-time and adaptive prediction of plasma profiles in the DIII-D tokamak. Our experiments demonstrate that RC achieves comparable results to state-of-the-art (deep) convolutional neural networks (CNNs) and long short-term memory (LSTM) models, with a significantly easier and faster training procedure. This efficient approach allows for fast and frequent adaptation of the model to new situations, such as changing plasma conditions or different fusion devices.
Collapse
|
18
|
Paassen B, Schulz A, Stewart TC, Hammer B. Reservoir Memory Machines as Neural Computers. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:2575-2585. [PMID: 34255637 DOI: 10.1109/tnnls.2021.3094139] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Differentiable neural computers (DNCs) extend artificial neural networks with an explicit memory without interference, thus enabling the model to perform classic computation tasks, such as graph traversal. However, such models are difficult to train, requiring long training times and large datasets. In this work, we achieve some of the computational capabilities of DNCs with a model that can be trained very efficiently, namely, an echo state network with an explicit memory without interference. This extension enables echo state networks to recognize all regular languages, including those that contractive echo state networks provably cannot recognize. Furthermore, we demonstrate experimentally that our model performs comparably to its fully trained deep version on several typical benchmark tasks for DNCs.
Collapse
|
19
|
Pasa L, Navarin N, Sperduti A. Multiresolution Reservoir Graph Neural Network. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:2642-2653. [PMID: 34232893 DOI: 10.1109/tnnls.2021.3090503] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Graph neural networks are receiving increasing attention as state-of-the-art methods to process graph-structured data. However, similar to other neural networks, they tend to suffer from a high computational cost to perform training. Reservoir computing (RC) is an effective way to define neural networks that are very efficient to train, often obtaining comparable predictive performance with respect to the fully trained counterparts. Different proposals of reservoir graph neural networks have been proposed in the literature. However, their predictive performances are still slightly below the ones of fully trained graph neural networks on many benchmark datasets, arguably because of the oversmoothing problem that arises when iterating over the graph structure in the reservoir computation. In this work, we aim to reduce this gap defining a multiresolution reservoir graph neural network (MRGNN) inspired by graph spectral filtering. Instead of iterating on the nonlinearity in the reservoir and using a shallow readout function, we aim to generate an explicit k -hop unsupervised graph representation amenable for further, possibly nonlinear, processing. Experiments on several datasets from various application areas show that our approach is extremely fast and it achieves in most of the cases comparable or even higher results with respect to state-of-the-art approaches.
Collapse
|
20
|
Pedrelli L, Hinaut X. Hierarchical-Task Reservoir for Online Semantic Analysis From Continuous Speech. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:2654-2663. [PMID: 34570710 DOI: 10.1109/tnnls.2021.3095140] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
In this article, we propose a novel architecture called hierarchical-task reservoir (HTR) suitable for real-time applications for which different levels of abstraction are available. We apply it to semantic role labeling (SRL) based on continuous speech recognition. Taking inspiration from the brain, this demonstrates the hierarchies of representations from perceptive to integrative areas, and we consider a hierarchy of four subtasks with increasing levels of abstraction (phone, word, part-of-speech (POS), and semantic role tags). These tasks are progressively learned by the layers of the HTR architecture. Interestingly, quantitative and qualitative results show that the hierarchical-task approach provides an advantage to improve the prediction. In particular, the qualitative results show that a shallow or a hierarchical reservoir, considered as baselines, does not produce estimations as good as the HTR model would. Moreover, we show that it is possible to further improve the accuracy of the model by designing skip connections and by considering word embedding (WE) in the internal representations. Overall, the HTR outperformed the other state-of-the-art reservoir-based approaches and it resulted in extremely efficient with respect to typical recurrent neural networks (RNNs) in deep learning (DL) [e.g., long short term memory (LSTMs)]. The HTR architecture is proposed as a step toward the modeling of online and hierarchical processes at work in the brain during language comprehension.
Collapse
|
21
|
Kleyko D, Frady EP, Kheffache M, Osipov E. Integer Echo State Networks: Efficient Reservoir Computing for Digital Hardware. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:1688-1701. [PMID: 33351770 DOI: 10.1109/tnnls.2020.3043309] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
We propose an approximation of echo state networks (ESNs) that can be efficiently implemented on digital hardware based on the mathematics of hyperdimensional computing. The reservoir of the proposed integer ESN (intESN) is a vector containing only n -bits integers (where is normally sufficient for a satisfactory performance). The recurrent matrix multiplication is replaced with an efficient cyclic shift operation. The proposed intESN approach is verified with typical tasks in reservoir computing: memorizing of a sequence of inputs, classifying time series, and learning dynamic processes. Such architecture results in dramatic improvements in memory footprint and computational efficiency, with minimal performance loss. The experiments on a field-programmable gate array confirm that the proposed intESN approach is much more energy efficient than the conventional ESN.
Collapse
|
22
|
Clustered and deep echo state networks for signal noise reduction. Mach Learn 2022. [DOI: 10.1007/s10994-022-06135-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
23
|
Gallicchio C, Micheli A. Architectural richness in deep reservoir computing. Neural Comput Appl 2022. [DOI: 10.1007/s00521-021-06760-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
24
|
Bianchi F, Gallicchio C, Micheli A. Pyramidal Reservoir Graph Neural Network. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.04.131] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
25
|
Li Z, Tanaka G. Multi-reservoir echo state networks with sequence resampling for nonlinear time-series prediction. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.08.122] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
26
|
Paaßen B, Schulz A, Hammer B. Reservoir stack machines. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.05.106] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
27
|
Hu J, Zhao W, Tang J, Luo Q. Integrating a softened multi-interval loss function into neural networks for wind power prediction. Appl Soft Comput 2021. [DOI: 10.1016/j.asoc.2021.108009] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023]
|
28
|
Carta A, Sperduti A, Bacciu D. Encoding-based memory for recurrent neural networks. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.04.051] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
29
|
Barredo Arrieta A, Gil-Lopez S, Laña I, Bilbao MN, Del Ser J. On the post-hoc explainability of deep echo state networks for time series forecasting, image and video classification. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-06359-y] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
30
|
Functional deep echo state network improved by a bi-level optimization approach for multivariate time series classification. Appl Soft Comput 2021. [DOI: 10.1016/j.asoc.2021.107314] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
|
31
|
Suganthan PN, Katuwal R. On the origins of randomization-based feedforward neural networks. Appl Soft Comput 2021. [DOI: 10.1016/j.asoc.2021.107239] [Citation(s) in RCA: 32] [Impact Index Per Article: 10.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
32
|
Dettori S, Matino I, Colla V, Speets R. A Deep Learning-based approach for forecasting off-gas production and consumption in the blast furnace. Neural Comput Appl 2021; 34:911-923. [PMID: 33879977 PMCID: PMC8051551 DOI: 10.1007/s00521-021-05984-x] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2020] [Accepted: 03/25/2021] [Indexed: 10/29/2022]
Abstract
This article presents the application of a recent neural network topology known as the deep echo state network to the prediction and modeling of strongly nonlinear systems typical of the process industry. The article analyzes the results by introducing a comparison with one of the most common and efficient topologies, the long short-term memories, in order to highlight the strengths and weaknesses of a reservoir computing approach compared to one currently considered as a standard of recurrent neural network. As benchmark application, two specific processes common in the integrated steelworks are selected, with the purpose of forecasting the future energy exchanges and transformations. The procedures of training, validation and test are based on data analysis, outlier detection and reconciliation and variable selection starting from real field industrial data. The analysis of results shows the effectiveness of deep echo state networks and their strong forecasting capabilities with respect to standard recurrent methodologies both in terms of training procedures and accuracy. Supplementary Information The online version contains supplementary material available at 10.1007/s00521-021-05984-x.
Collapse
Affiliation(s)
- Stefano Dettori
- Scuola Superiore Sant'Anna - TeCIP Institute - ICT-COISP, Pisa, Italy
| | - Ismael Matino
- Scuola Superiore Sant'Anna - TeCIP Institute - ICT-COISP, Pisa, Italy
| | - Valentina Colla
- Scuola Superiore Sant'Anna - TeCIP Institute - ICT-COISP, Pisa, Italy
| | | |
Collapse
|
33
|
Impact of coupling topology upon noise robustness of small optical reservoirs. Sci Rep 2020; 10:14086. [PMID: 32839505 PMCID: PMC7445167 DOI: 10.1038/s41598-020-70775-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2018] [Accepted: 08/04/2020] [Indexed: 11/22/2022] Open
Abstract
In this work, we perform the numerical investigation of the performance of the small optical reservoir computing (RC) systems with four neurons using the commercial software for optical fiber communication system. The small optical RC system consists of the components of the optical fiber communication. The nonlinear function which is required in RC is provided by the erbium-doped optical fiber amplifiers (EDFA). We demonstrate that the EDFA should be operated in the saturated or non-linear regime to obtain a better performance of the small optical RC system. The performance of the small optical RC systems for different topological neuron structures is investigated. The results show that the interconnection between the neurons could offer a better performance than the systems without interconnection between the neurons. Moreover, the input signals with different noise levels are launched into the systems. The results show that the small optical RC system can classify the noisy input optical waveforms even when the signal-to-noise ratio is as low as − 2.55 dB.
Collapse
|
34
|
Jirak D, Tietz S, Ali H, Wermter S. Echo State Networks and Long Short-Term Memory for Continuous Gesture Recognition: a Comparative Study. Cognit Comput 2020. [DOI: 10.1007/s12559-020-09754-0] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
AbstractRecent developments of sensors that allow tracking of human movements and gestures enable rapid progress of applications in domains like medical rehabilitation or robotic control. Especially the inertial measurement unit (IMU) is an excellent device for real-time scenarios as it rapidly delivers data input. Therefore, a computational model must be able to learn gesture sequences in a fast yet robust way. We recently introduced an echo state network (ESN) framework for continuous gesture recognition (Tietz et al., 2019) including novel approaches for gesture spotting, i.e., the automatic detection of the start and end phase of a gesture. Although our results showed good classification performance, we identified significant factors which also negatively impact the performance like subgestures and gesture variability. To address these issues, we include experiments with Long Short-Term Memory (LSTM) networks, which is a state-of-the-art model for sequence processing, to compare the obtained results with our framework and to evaluate their robustness regarding pitfalls in the recognition process. In this study, we analyze the two conceptually different approaches processing continuous, variable-length gesture sequences, which shows interesting results comparing the distinct gesture accomplishments. In addition, our results demonstrate that our ESN framework achieves comparably good performance as the LSTM network but has significantly lower training times. We conclude from the present work that ESNs are viable models for continuous gesture recognition delivering reasonable performance for applications requiring real-time performance as in robotic or rehabilitation tasks. From our discussion of this comparative study, we suggest prospective improvements on both the experimental and network architecture level.
Collapse
|
35
|
A Soft Sensor Approach Based on an Echo State Network Optimized by Improved Genetic Algorithm. SENSORS 2020; 20:s20175000. [PMID: 32899330 PMCID: PMC7569782 DOI: 10.3390/s20175000] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/02/2020] [Revised: 08/17/2020] [Accepted: 08/31/2020] [Indexed: 02/02/2023]
Abstract
In the process of fault diagnosis and the health and safety operation evaluation of modern industrial processes, it is crucial to measure important state variables, which cannot be directly detected due to limitations of economy, technology, environment and space. Therefore, this paper proposes a data-driven soft sensor approach based on an echo state network (ESN) optimized by an improved genetic algorithm (IGA). Firstly, with an ESN, a data-driven model (DDM) between secondary variables and dominant variables is established. Secondly, in order to improve the prediction performance, the IGA is utilized to optimize the parameters of the ESN. Then, the immigration strategy is introduced and the crossover and mutation operators are changed adaptively to improve the convergence speed of the algorithm and address the problem that the algorithm falls into the local optimum. Finally, a soft sensor model of an ESN optimized by an IGA is established (IGA-ESN), and the advantages and performance of the proposed method are verified by estimating the alumina concentration in an aluminum reduction cell. The experimental results illustrated that the proposed method is efficient, and the error was significantly reduced compared with the traditional algorithm.
Collapse
|
36
|
Marcucci G, Pierangeli D, Conti C. Theory of Neuromorphic Computing by Waves: Machine Learning by Rogue Waves, Dispersive Shocks, and Solitons. PHYSICAL REVIEW LETTERS 2020; 125:093901. [PMID: 32915624 DOI: 10.1103/physrevlett.125.093901] [Citation(s) in RCA: 22] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/17/2019] [Revised: 04/14/2020] [Accepted: 07/09/2020] [Indexed: 06/11/2023]
Abstract
We study artificial neural networks with nonlinear waves as a computing reservoir. We discuss universality and the conditions to learn a dataset in terms of output channels and nonlinearity. A feed-forward three-layered model, with an encoding input layer, a wave layer, and a decoding readout, behaves as a conventional neural network in approximating mathematical functions, real-world datasets, and universal Boolean gates. The rank of the transmission matrix has a fundamental role in assessing the learning abilities of the wave. For a given set of training points, a threshold nonlinearity for universal interpolation exists. When considering the nonlinear Schrödinger equation, the use of highly nonlinear regimes implies that solitons, rogue, and shock waves do have a leading role in training and computing. Our results may enable the realization of novel machine learning devices by using diverse physical systems, as nonlinear optics, hydrodynamics, polaritonics, and Bose-Einstein condensates. The application of these concepts to photonics opens the way to a large class of accelerators and new computational paradigms. In complex wave systems, as multimodal fibers, integrated optical circuits, random, topological devices, and metasurfaces, nonlinear waves can be employed to perform computation and solve complex combinatorial optimization.
Collapse
Affiliation(s)
- Giulia Marcucci
- Institute for Complex Systems, National Research Council (ISC-CNR), Via dei Taurini 19, 00185 Rome, Italy and Department of Physics, Sapienza University, Piazzale Aldo Moro 2, 00185 Rome, Italy
| | - Davide Pierangeli
- Institute for Complex Systems, National Research Council (ISC-CNR), Via dei Taurini 19, 00185 Rome, Italy and Department of Physics, Sapienza University, Piazzale Aldo Moro 2, 00185 Rome, Italy
| | - Claudio Conti
- Institute for Complex Systems, National Research Council (ISC-CNR), Via dei Taurini 19, 00185 Rome, Italy and Department of Physics, Sapienza University, Piazzale Aldo Moro 2, 00185 Rome, Italy
| |
Collapse
|
37
|
Maglogiannis I, Iliadis L, Pimenidis E. Deep Echo State Networks in Industrial Applications. IFIP ADVANCES IN INFORMATION AND COMMUNICATION TECHNOLOGY 2020. [PMCID: PMC7256577 DOI: 10.1007/978-3-030-49186-4_5] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
This paper analyzes the impact of reservoir computing, and, in particular, of Deep Echo State Networks, to the modeling of highly non-linear dynamical systems that can be commonly found in the industry. Several applications are presented focusing on forecasting models related to energy content of steelwork byproduct gasses. Deep Echo State Network models are trained, validated and tested by exploiting datasets coming from a real industrial context, with good results in terms of accuracy of the predictions.
Collapse
Affiliation(s)
| | - Lazaros Iliadis
- Department of Civil Engineering, Lab of Mathematics and Informatics (ISCE), Democritus University of Thrace, Xanthi, Greece
| | - Elias Pimenidis
- Department of Computer Science and Creative Technologies, University of the West of England, Bristol, UK
| |
Collapse
|
38
|
Application of the deep learning for the prediction of rainfall in Southern Taiwan. Sci Rep 2019; 9:12774. [PMID: 31485008 PMCID: PMC6726605 DOI: 10.1038/s41598-019-49242-6] [Citation(s) in RCA: 41] [Impact Index Per Article: 8.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2018] [Accepted: 08/22/2019] [Indexed: 11/08/2022] Open
Abstract
Precipitation is useful information for assessing vital water resources, agriculture, ecosystems and hydrology. Data-driven model predictions using deep learning algorithms are promising for these purposes. Echo state network (ESN) and Deep Echo state network (DeepESN), referred to as Reservoir Computing (RC), are effective and speedy algorithms to process a large amount of data. In this study, we used the ESN and the DeepESN algorithms to analyze the meteorological hourly data from 2002 to 2014 at the Tainan Observatory in the southern Taiwan. The results show that the correlation coefficient by using the DeepESN was better than that by using the ESN and commercial neuronal network algorithms (Back-propagation network (BPN) and support vector regression (SVR), MATLAB, The MathWorks co.), and the accuracy of predicted rainfall by using the DeepESN can be significantly improved compared with those by using ESN, the BPN and the SVR. In sum, the DeepESN is a trustworthy and good method to predict rainfall; it could be applied to global climate forecasts which need high-volume data processing.
Collapse
|
39
|
Chouikhi N, Ammar B, Hussain A, Alimi AM. Bi-level multi-objective evolution of a Multi-Layered Echo-State Network Autoencoder for data representations. Neurocomputing 2019. [DOI: 10.1016/j.neucom.2019.03.012] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
|
40
|
|
41
|
Wikle CK. Comparison of Deep Neural Networks and Deep Hierarchical Models for Spatio-Temporal Data. JOURNAL OF AGRICULTURAL, BIOLOGICAL AND ENVIRONMENTAL STATISTICS 2019. [DOI: 10.1007/s13253-019-00361-7] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|