1
|
Chatzimiltis S, Agathocleous M, Promponas VJ, Christodoulou C. Post-processing enhances protein secondary structure prediction with second order deep learning and embeddings. Comput Struct Biotechnol J 2025; 27:243-251. [PMID: 39866664 PMCID: PMC11764030 DOI: 10.1016/j.csbj.2024.12.022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2024] [Revised: 12/20/2024] [Accepted: 12/21/2024] [Indexed: 01/28/2025] Open
Abstract
Protein Secondary Structure Prediction (PSSP) is regarded as a challenging task in bioinformatics, and numerous approaches to achieve a more accurate prediction have been proposed. Accurate PSSP can be instrumental in inferring protein tertiary structure and their functions. Machine Learning and in particular Deep Learning approaches show promising results for the PSSP problem. In this paper, we deploy a Convolutional Neural Network (CNN) trained with the Subsampled Hessian Newton (SHN) method (a Hessian Free Optimisation variant), with a two- dimensional input representation of embeddings extracted from a language model pretrained with protein sequences. Utilising a CNN trained with the SHN method and the input embeddings, we achieved on average a 79.96% per residue (Q3) accuracy on the CB513 dataset and 81.45% Q3 accuracy on the PISCES dataset (without any post-processing techniques applied). The application of ensembles and filtering techniques to the results of the CNN improved the overall prediction performance. The Q3 accuracy on the CB513 increased to 93.65% and for the PISCES dataset to 87.13%. Moreover, our method was evaluated using the CASP13 dataset where we showed that as the post-processing window size increased, the prediction performance increased as well. In fact, with the biggest post-processing window size (limited by the smallest CASP13 protein), we achieved a Q3 accuracy of 98.12% and a Segment Overlap (SOV) score of 96.98 on the CASP13 dataset when the CNNs were trained with the PISCES dataset. Finally, we showed that input representations from embeddings can perform equally well as representations extracted from multiple sequence alignments.
Collapse
Affiliation(s)
- Sotiris Chatzimiltis
- University of Cyprus, Department of Computer Science, Nicosia, Cyprus
- 5G/6GIC, Institute for Communication Systems (ICS), University of Surrey, Guildford, United Kingdom
| | - Michalis Agathocleous
- University of Cyprus, Department of Computer Science, Nicosia, Cyprus
- University of Nicosia, Department of Computer Science, Nicosia, Cyprus
| | | | | |
Collapse
|
2
|
Chang V, Xu QA, Chidozie A, Wang H. Predicting Economic Trends and Stock Market Prices with Deep Learning and Advanced Machine Learning Techniques. ELECTRONICS 2024; 13:3396. [DOI: 10.3390/electronics13173396] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/05/2025]
Abstract
The volatile and non-linear nature of stock market data, particularly in the post-pandemic era, poses significant challenges for accurate financial forecasting. To address these challenges, this research develops advanced deep learning and machine learning algorithms to predict financial trends, quantify risks, and forecast stock prices, focusing on the technology sector. Our study seeks to answer the following question: “Which deep learning and supervised machine learning algorithms are the most accurate and efficient in predicting economic trends and stock market prices, and under what conditions do they perform best?” We focus on two advanced recurrent neural network (RNN) models, long short-term memory (LSTM) and Gated Recurrent Unit (GRU), to evaluate their efficiency in predicting technology industry stock prices. Additionally, we integrate statistical methods such as autoregressive integrated moving average (ARIMA) and Facebook Prophet and machine learning algorithms like Extreme Gradient Boosting (XGBoost) to enhance the robustness of our predictions. Unlike classical statistical algorithms, LSTM and GRU models can identify and retain important data sequences, enabling more accurate predictions. Our experimental results show that the GRU model outperforms the LSTM model in terms of prediction accuracy and training time across multiple metrics such as RMSE and MAE. This study offers crucial insights into the predictive capabilities of deep learning models and advanced machine learning techniques for financial forecasting, highlighting the potential of GRU and XGBoost for more accurate and efficient stock price prediction in the technology sector.
Collapse
Affiliation(s)
- Victor Chang
- Department of Operations and Information Management, Aston Business School, Aston University, Birmingham B4 7ET, UK
| | - Qianwen Ariel Xu
- Department of Operations and Information Management, Aston Business School, Aston University, Birmingham B4 7ET, UK
| | - Anyamele Chidozie
- School of Computing, Engineering and Digital Technologies, Teesside University, Middlesbrough TS1 3BX, UK
| | - Hai Wang
- School of Computer Science and Digital Technologies, Aston University, Birmingham B4 7ET, UK
| |
Collapse
|
3
|
Lambrechts G, De Geeter F, Vecoven N, Ernst D, Drion G. Warming up recurrent neural networks to maximise reachable multistability greatly improves learning. Neural Netw 2023; 166:645-669. [PMID: 37604075 DOI: 10.1016/j.neunet.2023.07.023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2022] [Revised: 06/12/2023] [Accepted: 07/14/2023] [Indexed: 08/23/2023]
Abstract
Training recurrent neural networks is known to be difficult when time dependencies become long. In this work, we show that most standard cells only have one stable equilibrium at initialisation, and that learning on tasks with long time dependencies generally occurs once the number of network stable equilibria increases; a property known as multistability. Multistability is often not easily attained by initially monostable networks, making learning of long time dependencies between inputs and outputs difficult. This insight leads to the design of a novel way to initialise any recurrent cell connectivity through a procedure called "warmup" to improve its capability to learn arbitrarily long time dependencies. This initialisation procedure is designed to maximise network reachable multistability, i.e., the number of equilibria within the network that can be reached through relevant input trajectories, in few gradient steps. We show on several information restitution, sequence classification, and reinforcement learning benchmarks that warming up greatly improves learning speed and performance, for multiple recurrent cells, but sometimes impedes precision. We therefore introduce a double-layer architecture initialised with a partial warmup that is shown to greatly improve learning of long time dependencies while maintaining high levels of precision. This approach provides a general framework for improving learning abilities of any recurrent cell when long time dependencies are present. We also show empirically that other initialisation and pretraining procedures from the literature implicitly foster reachable multistability of recurrent cells.
Collapse
Affiliation(s)
- Gaspard Lambrechts
- Montefiore Institute, University of Liège, 10 allée de la découverte, Liège, 4000, Belgium.
| | - Florent De Geeter
- Montefiore Institute, University of Liège, 10 allée de la découverte, Liège, 4000, Belgium.
| | - Nicolas Vecoven
- Montefiore Institute, University of Liège, 10 allée de la découverte, Liège, 4000, Belgium
| | - Damien Ernst
- Montefiore Institute, University of Liège, 10 allée de la découverte, Liège, 4000, Belgium; LTCI, Telecom Paris, Institut Polytechnique de Paris, 19 place Marguerite Perey, Palaiseau, 91120, France.
| | - Guillaume Drion
- Montefiore Institute, University of Liège, 10 allée de la découverte, Liège, 4000, Belgium.
| |
Collapse
|
4
|
Shan D, Luo Y, Zhang X, Zhang C. DRRNets: Dynamic Recurrent Routing via Low-Rank Regularization in Recurrent Neural Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:2057-2067. [PMID: 34460403 DOI: 10.1109/tnnls.2021.3105818] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Recurrent neural networks (RNNs) continue to show outstanding performance in sequence learning tasks such as language modeling, but it remains difficult to train RNNs for long sequences. The main challenges lie in the complex dependencies, gradient vanishing or exploding, and low resource requirement in model deployment. In order to address these challenges, we propose dynamic recurrent routing neural networks (DRRNets), which can: 1) shorten the recurrent lengths by allocating recurrent routes dynamically for different dependencies and 2) reduce the number of parameters significantly by imposing low-rank constraints on the fully connected layers. A novel optimization algorithm via low-rank constraint and sparsity projection is developed to train the network. We verify the effectiveness of the proposed method by comparing it with multiple competitive approaches in several popular sequential learning tasks, such as language modeling and speaker recognition. The results in terms of different criteria demonstrate the superiority of our proposed method.
Collapse
|
5
|
Shan D, Zhang X, Zhang C. Spontaneous Temporal Grouping Neural Network for Long-Term Memory Modeling. IEEE Trans Cogn Dev Syst 2022. [DOI: 10.1109/tcds.2021.3050759] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Affiliation(s)
- Dongjing Shan
- Laboratory of Intelligent Information Processing, Army Engineering University, Nanjing, China
| | - Xiongwei Zhang
- Laboratory of Intelligent Information Processing, Army Engineering University, Nanjing, China
| | - Chao Zhang
- Key Laboratory of Machine Perception (MOE), Peking University, Beijing, China
| |
Collapse
|
6
|
Shan D, Zhang C, Nian Y. An optimization scheme for segmented-memory neural network. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.07.076] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
7
|
Zheng M, Li L, Peng H, Xiao J, Yang Y, Zhao H. Parameters estimation and synchronization of uncertain coupling recurrent dynamical neural networks with time-varying delays based on adaptive control. Neural Comput Appl 2016. [DOI: 10.1007/s00521-016-2822-6] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
8
|
Glüge S, Böck R, Palm G, Wendemuth A. Learning long-term dependencies in segmented-memory recurrent neural networks with backpropagation of error. Neurocomputing 2014. [DOI: 10.1016/j.neucom.2013.11.043] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|
9
|
Huynh TQ, Reggia JA. Symbolic representation of recurrent neural network dynamics. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2012; 23:1649-1658. [PMID: 24808009 DOI: 10.1109/tnnls.2012.2210242] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods.
Collapse
|
10
|
Tseng KH, Tsai JSH, Lu CY. Design of Delay-Dependent Exponential Estimator for T–S Fuzzy Neural Networks with Mixed Time-Varying Interval Delays Using Hybrid Taguchi-Genetic Algorithm. Neural Process Lett 2012. [DOI: 10.1007/s11063-012-9222-4] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
11
|
Gripon V, Berrou C. Sparse Neural Networks With Large Learning Diversity. ACTA ACUST UNITED AC 2011; 22:1087-96. [DOI: 10.1109/tnn.2011.2146789] [Citation(s) in RCA: 99] [Impact Index Per Article: 7.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|