1
|
Hodjat B, Shahrzad H, Miikkulainen R. Domain-Independent Lifelong Problem Solving Through Distributed ALife Actors. ARTIFICIAL LIFE 2024; 30:259-276. [PMID: 38048055 DOI: 10.1162/artl_a_00418] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/05/2023]
Abstract
A domain-independent problem-solving system based on principles of Artificial Life is introduced. In this system, DIAS, the input and output dimensions of the domain are laid out in a spatial medium. A population of actors, each seeing only part of this medium, solves problems collectively in it. The process is independent of the domain and can be implemented through different kinds of actors. Through a set of experiments on various problem domains, DIAS is shown able to solve problems with different dimensionality and complexity, to require no hyperparameter tuning for new problems, and to exhibit lifelong learning, that is, to adapt rapidly to run-time changes in the problem domain, and to do it better than a standard, noncollective approach. DIAS therefore demonstrates a role for ALife in building scalable, general, and adaptive problem-solving systems.
Collapse
Affiliation(s)
| | - Hormoz Shahrzad
- Cognizant AI Labs University of Texas at Austin Department of Computer Science
| | - Risto Miikkulainen
- Cognizant AI Labs University of Texas at Austin Department of Computer Science
| |
Collapse
|
2
|
Lussange J, Vrizzi S, Palminteri S, Gutkin B. Mesoscale effects of trader learning behaviors in financial markets: A multi-agent reinforcement learning study. PLoS One 2024; 19:e0301141. [PMID: 38557590 PMCID: PMC10984546 DOI: 10.1371/journal.pone.0301141] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2023] [Accepted: 03/08/2024] [Indexed: 04/04/2024] Open
Abstract
Recent advances in the field of machine learning have yielded novel research perspectives in behavioural economics and financial markets microstructure studies. In this paper we study the impact of individual trader leaning characteristics on markets using a stock market simulator designed with a multi-agent architecture. Each agent, representing an autonomous investor, trades stocks through reinforcement learning, using a centralized double-auction limit order book. This approach allows us to study the impact of individual trader traits on the whole stock market at the mesoscale in a bottom-up approach. We chose to test three trader trait aspects: agent learning rate increases, herding behaviour and random trading. As hypothesized, we find that larger learning rates significantly increase the number of crashes. We also find that herding behaviour undermines market stability, while random trading tends to preserve it.
Collapse
Affiliation(s)
- Johann Lussange
- Laboratoire des Neurosciences Cognitives, Département des Études Cognitives, INSERM U960, Paris, France
| | - Stefano Vrizzi
- Laboratoire des Neurosciences Cognitives, Département des Études Cognitives, INSERM U960, Paris, France
| | - Stefano Palminteri
- Laboratoire des Neurosciences Cognitives, Département des Études Cognitives, INSERM U960, Paris, France
- Center for Cognition and Decision Making, Department of Psychology, NU University Higher School of Economics, Moscow, Russia
| | - Boris Gutkin
- Laboratoire des Neurosciences Cognitives, Département des Études Cognitives, INSERM U960, Paris, France
- Center for Cognition and Decision Making, Department of Psychology, NU University Higher School of Economics, Moscow, Russia
| |
Collapse
|
3
|
Tsantekidis A, Passalis N, Tefas A. Modeling limit order trading with a continuous action policy for deep reinforcement learning. Neural Netw 2023; 165:506-515. [PMID: 37348431 DOI: 10.1016/j.neunet.2023.05.051] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2022] [Revised: 01/20/2023] [Accepted: 05/28/2023] [Indexed: 06/24/2023]
Abstract
Limit Orders allow buyers and sellers to set a "limit price" they are willing to accept in a trade. On the other hand, market orders allow for immediate execution at any price. Thus, market orders are susceptible to slippage, which is the additional cost incurred due to the unfavorable execution of a trade order. As a result, limit orders are often preferred, since they protect traders from excessive slippage costs due to larger than expected price fluctuations. Despite the price guarantees of limit orders, they are more complex compared to market orders. Orders with overly optimistic limit prices might never be executed, which increases the risk of employing limit orders in Machine Learning (ML)-based trading systems. Indeed, the current ML literature for trading almost exclusively relies on market orders. To overcome this limitation, a Deep Reinforcement Learning (DRL) approach is proposed to model trading agents that use limit orders. The proposed method (a) uses a framework that employs a continuous probability distribution to model limit prices, while (b) provides the ability to place market orders when the risk of no execution is more significant than the cost of slippage. Extensive experiments are conducted with multiple currency pairs, using hourly price intervals, validating the effectiveness of the proposed method and paving the way for introducing limit order modeling in DRL-based trading.
Collapse
Affiliation(s)
- Avraam Tsantekidis
- School of Informatics, Aristotle University of Thessaloniki, Thessaloniki, Greece.
| | - Nikolaos Passalis
- School of Informatics, Aristotle University of Thessaloniki, Thessaloniki, Greece.
| | - Anastasios Tefas
- School of Informatics, Aristotle University of Thessaloniki, Thessaloniki, Greece.
| |
Collapse
|
4
|
James N, Menzies M. Collective Dynamics, Diversification and Optimal Portfolio Construction for Cryptocurrencies. ENTROPY (BASEL, SWITZERLAND) 2023; 25:931. [PMID: 37372275 DOI: 10.3390/e25060931] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/18/2023] [Revised: 06/07/2023] [Accepted: 06/12/2023] [Indexed: 06/29/2023]
Abstract
Since its conception, the cryptocurrency market has been frequently described as an immature market, characterized by significant swings in volatility and occasionally described as lacking rhyme or reason. There has been great speculation as to what role it plays in a diversified portfolio. For instance, is cryptocurrency exposure an inflationary hedge or a speculative investment that follows broad market sentiment with amplified beta? We have recently explored similar questions with a clear focus on the equity market. There, our research revealed several noteworthy dynamics such as an increase in the market's collective strength and uniformity during crises, greater diversification benefits across equity sectors (rather than within them), and the existence of a "best value" portfolio of equities. In essence, we can now contrast any potential signatures of maturity we identify in the cryptocurrency market and contrast these with the substantially larger, older and better-established equity market. This paper aims to investigate whether the cryptocurrency market has recently exhibited similar mathematical properties as the equity market. Instead of relying on traditional portfolio theory, which is grounded in the financial dynamics of equity securities, we adjust our experimental focus to capture the presumed behavioral purchasing patterns of retail cryptocurrency investors. Our focus is on collective dynamics and portfolio diversification in the cryptocurrency market, and examining whether previously established results in the equity market hold in the cryptocurrency market and to what extent. The results reveal nuanced signatures of maturity related to the equity market, including the fact that correlations collectively spike around exchange collapses, and identify an ideal portfolio size and spread across different groups of cryptocurrencies.
Collapse
Affiliation(s)
- Nick James
- School of Mathematics and Statistics, University of Melbourne, Victoria 3010, Australia
| | - Max Menzies
- Beijing Institute of Mathematical Sciences and Applications, Tsinghua University, Beijing 101408, China
| |
Collapse
|
5
|
Online portfolio management via deep reinforcement learning with high-frequency data. Inf Process Manag 2023. [DOI: 10.1016/j.ipm.2022.103247] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/11/2023]
|
6
|
Sun S, Wang R, An B. Reinforcement Learning for Quantitative Trading. ACM T INTEL SYST TEC 2023. [DOI: 10.1145/3582560] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
Quantitative trading (QT), which refers to the usage of mathematical models and data-driven techniques in analyzing the financial market, has been a popular topic in both academia and financial industry since 1970s. In the last decade, reinforcement learning (RL) has garnered significant interest in many domains such as robotics and video games, owing to its outstanding ability on solving complex sequential decision making problems. RL’s impact is pervasive, recently demonstrating its ability to conquer many challenging QT tasks. It is a flourishing research direction to explore RL techniques’ potential on QT tasks. This paper aims at providing a comprehensive survey of research efforts on RL-based methods for QT tasks. More concretely, we devise a taxonomy of RL-based QT models, along with a comprehensive summary of the state of the art. Finally, we discuss current challenges and propose future research directions in this exciting field.
Collapse
Affiliation(s)
- Shuo Sun
- Nanyang Technological University, Singapore
| | | | - Bo An
- Nanyang Technological University, Singapore
| |
Collapse
|
7
|
Bonetti M, Bisi L, Restelli M. Risk-Averse Optimization of Reward-based Coherent Risk Measures. ARTIF INTELL 2023. [DOI: 10.1016/j.artint.2022.103845] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023]
|
8
|
He FF, Chen CT, Huang SH. A multi-agent virtual market model for generalization in reinforcement learning based trading strategies. Appl Soft Comput 2023. [DOI: 10.1016/j.asoc.2023.109985] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/15/2023]
|
9
|
Kwak D, Choi S, Chang W. Self-attention based deep direct recurrent reinforcement learning with hybrid loss for trading signal generation. Inf Sci (N Y) 2022. [DOI: 10.1016/j.ins.2022.12.042] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/27/2022]
|
10
|
Bisi L, Santambrogio D, Sandrelli F, Tirinzoni A, Ziebart BD, Restelli M. Risk-averse policy optimization via risk-neutral policy optimization. ARTIF INTELL 2022. [DOI: 10.1016/j.artint.2022.103765] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
11
|
Gunjan A, Bhattacharyya S. A brief review of portfolio optimization techniques. Artif Intell Rev 2022. [DOI: 10.1007/s10462-022-10273-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
12
|
Liu Z, Luo H, Chen P, Xia Q, Gan Z, Shan W. An efficient isomorphic CNN-based prediction and decision framework for financial time series. INTELL DATA ANAL 2022. [DOI: 10.3233/ida-216142] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Financial time series prediction and trading decision-making are priorities of computational intelligence for researchers in academia and the finance industry due to their broad application areas and substantial impact. However, these methods remain challenging because they retain various complex statistical properties, and the mechanism behind the processes is unknown to a large extent. A significant number of machine learning-based methods are proposed and demonstrate impressive results, especially deep learning-based models. Nevertheless, due to the high complexity of massive, nonlinear, and nonindependent data and the difficulties and time consumption of complicated training models of deep learning, the performance of online trading decisions is still inadequate for practical application. This paper proposes the Integrated Framework of Forecasting Based Online Trading Strategy (IFF-BOTS) to satisfy better prediction performance and dynamic decisions for real-world online trading systems. Our method adopts a novel isomorphic convolutional neural network (CNN)-based forecaster-classifier-executor architecture to exploit CNN-based price and trend integrated prediction and direct-reinforcement-learning-based trading decision-making. IFF-BOTS can also achieve better real-time performance for online trading. We empirically compare the proposed approach with state-of-the-art prediction and trading methods on real-world S&P and DJI datasets. The results show that the IFF-BOTS outperforms its competitors in predicting metrics, trading profits, and real-time performance.
Collapse
Affiliation(s)
- Zhongming Liu
- School of Computer and Software Engineering, Xihua University, Chengdu, Sichuan, China
| | - Hang Luo
- School of Economics, Xihua University, Chengdu, Sichuan, China
| | - Peng Chen
- School of Computer and Software Engineering, Xihua University, Chengdu, Sichuan, China
| | - Qibin Xia
- School of Computer and Software Engineering, Xihua University, Chengdu, Sichuan, China
| | - Zhihao Gan
- School of Computer and Software Engineering, Xihua University, Chengdu, Sichuan, China
| | - Wenyu Shan
- School of Computer and Software Engineering, Xihua University, Chengdu, Sichuan, China
| |
Collapse
|
13
|
Park K, Jung HG, Eom TS, Lee SW. Uncertainty-Aware Portfolio Management With Risk-Sensitive Multiagent Network. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; PP:362-375. [PMID: 35604996 DOI: 10.1109/tnnls.2022.3174642] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
As deep neural networks (DNNs) have gained considerable attention in recent years, there have been several cases applying DNNs to portfolio management (PM). Although some researchers have experimentally demonstrated its ability to make a profit, it is still insufficient to use in real situations because existing studies have failed to answer how risky investment decisions are. Furthermore, even though the objective of PM is to maximize returns within a risk tolerance, they overlook the predictive uncertainty of DNNs in the process of risk management. To overcome these limitations, we propose a novel framework called risk-sensitive multiagent network (RSMAN), which includes risk-sensitive agents (RSAs) and a risk adaptive portfolio generator (RAPG). Standard DNNs do not understand the risks of their decision, whereas RSA can take risk-sensitive decisions by estimating market uncertainty and parameter uncertainty. Acting as a trader, this agent is trained via reinforcement learning from dynamic trading simulations to estimate the distribution of reward and via unsupervised learning to assess parameter uncertainty without labeled data. We also present an RAPG that can generate a portfolio fitting the user's risk appetite without retraining by exploiting the estimated information from the RSAs. We tested our framework on the U.S. and Korean real financial markets to demonstrate the practicality of the RSMAN.
Collapse
|
14
|
A Novel Trading Strategy Framework Based on Reinforcement Deep Learning for Financial Market Predictions. MATHEMATICS 2021. [DOI: 10.3390/math9233094] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
The prediction of stocks is complicated by the dynamic, complex, and chaotic environment of the stock market. Investors put their money into the financial market, hoping to maximize profits by understanding market trends and designing trading strategies at the entry and exit points. Most studies propose machine learning models to predict stock prices. However, constructing trading strategies is helpful for traders to avoid making mistakes and losing money. We propose an automatic trading framework using LSTM combined with deep Q-learning to determine the trading signal and the size of the trading position. This is more sophisticated than traditional price prediction models. This study used price data from the Taiwan stock market, including daily opening price, closing price, highest price, lowest price, and trading volume. The profitability of the system was evaluated using a combination of different states of different stocks. The profitability of the proposed system was positive after a long period of testing, which means that the system performed well in predicting the rise and fall of stocks.
Collapse
|
15
|
Qiu Y, Qiu Y, Yuan Y, Chen Z, Lee R. QF-TraderNet: Intraday Trading via Deep Reinforcement With Quantum Price Levels Based Profit-And-Loss Control. Front Artif Intell 2021; 4:749878. [PMID: 34778753 PMCID: PMC8586520 DOI: 10.3389/frai.2021.749878] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2021] [Accepted: 09/21/2021] [Indexed: 11/13/2022] Open
Abstract
Reinforcement Learning (RL) based machine trading attracts a rich profusion of interest. However, in the existing research, RL in the day-trade task suffers from the noisy financial movement in the short time scale, difficulty in order settlement, and expensive action search in a continuous-value space. This paper introduced an end-to-end RL intraday trading agent, namely QF-TraderNet, based on the quantum finance theory (QFT) and deep reinforcement learning. We proposed a novel design for the intraday RL trader’s action space, inspired by the Quantum Price Levels (QPLs). Our action space design also brings the model a learnable profit-and-loss control strategy. QF-TraderNet composes two neural networks: 1) A long short term memory networks for the feature learning of financial time series; 2) a policy generator network (PGN) for generating the distribution of actions. The profitability and robustness of QF-TraderNet have been verified in multi-type financial datasets, including FOREX, metals, crude oil, and financial indices. The experimental results demonstrate that QF-TraderNet outperforms other baselines in terms of cumulative price returns and Sharpe Ratio, and the robustness in the acceidential market shift.
Collapse
Affiliation(s)
- Yifu Qiu
- Department of Computer Science and Technology, Division of Science and Technology, BNU-HKBU United International College, Zhuhai, China
| | - Yitao Qiu
- Department of Computer Science and Technology, Division of Science and Technology, BNU-HKBU United International College, Zhuhai, China
| | - Yicong Yuan
- Department of Computer Science and Technology, Division of Science and Technology, BNU-HKBU United International College, Zhuhai, China
| | - Zheng Chen
- Department of Computer Science and Technology, Division of Science and Technology, BNU-HKBU United International College, Zhuhai, China
| | - Raymond Lee
- Department of Computer Science and Technology, Division of Science and Technology, BNU-HKBU United International College, Zhuhai, China
| |
Collapse
|
16
|
|
17
|
Tsantekidis A, Passalis N, Toufa AS, Saitas-Zarkias K, Chairistanidis S, Tefas A. Price Trailing for Financial Trading Using Deep Reinforcement Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2021; 32:2837-2846. [PMID: 32516114 DOI: 10.1109/tnnls.2020.2997523] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Machine learning methods have recently seen a growing number of applications in financial trading. Being able to automatically extract patterns from past price data and consistently apply them in the future has been the focus of many quantitative trading applications. However, developing machine learning-based methods for financial trading is not straightforward, requiring carefully designed targets/rewards, hyperparameter fine-tuning, and so on. Furthermore, most of the existing methods are unable to effectively exploit the information available across various financial instruments. In this article, we propose a deep reinforcement learning-based approach, which ensures that consistent rewards are provided to the trading agent, mitigating the noisy nature of profit-and-loss rewards that are usually used. To this end, we employ a novel price trailing-based reward shaping approach, significantly improving the performance of the agent in terms of profit, Sharpe ratio, and maximum drawdown. Furthermore, we carefully designed a data preprocessing method that allows for training the agent on different FOREX currency pairs, providing a way for developing market-wide RL agents and allowing, at the same time, to exploit more powerful recurrent deep learning models without the risk of overfitting. The ability of the proposed methods to improve various performance metrics is demonstrated using a challenging large-scale data set, containing 28 instruments, provided by Speedlab AG.
Collapse
|
18
|
Wu ME, Syu JH, Lin JCW, Ho JM. Portfolio management system in equity market neutral using reinforcement learning. APPL INTELL 2021. [DOI: 10.1007/s10489-021-02262-0] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Abstract
AbstractPortfolio management involves position sizing and resource allocation. Traditional and generic portfolio strategies require forecasting of future stock prices as model inputs, which is not a trivial task since those values are difficult to obtain in the real-world applications. To overcome the above limitations and provide a better solution for portfolio management, we developed a Portfolio Management System (PMS) using reinforcement learning with two neural networks (CNN and RNN). A novel reward function involving Sharpe ratios is also proposed to evaluate the performance of the developed systems. Experimental results indicate that the PMS with the Sharpe ratio reward function exhibits outstanding performance, increasing return by 39.0% and decreasing drawdown by 13.7% on average compared to the reward function of trading return. In addition, the proposed model is more suitable for the construction of a reinforcement learning portfolio, but has 1.98 times more drawdown risk than the . Among the conducted datasets, the PMS outperforms the benchmark strategies in TW50 and traditional stocks, but is inferior to a benchmark strategy in the financial dataset. The PMS is profitable, effective, and offers lower investment risk among almost all datasets. The novel reward function involving the Sharpe ratio enhances performance, and well supports resource-allocation for empirical stock trading.
Collapse
|
19
|
Tsantekidis A, Passalis N, Tefas A. Diversity-driven knowledge distillation for financial trading using Deep Reinforcement Learning. Neural Netw 2021; 140:193-202. [PMID: 33774425 DOI: 10.1016/j.neunet.2021.02.026] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2020] [Revised: 12/08/2020] [Accepted: 02/22/2021] [Indexed: 11/18/2022]
Abstract
Deep Reinforcement Learning (RL) is increasingly used for developing financial trading agents for a wide range of tasks. However, optimizing deep RL agents is notoriously difficult and unstable, especially in noisy financial environments, significantly hindering the performance of trading agents. In this work, we present a novel method that improves the training reliability of DRL trading agents building upon the well-known approach of neural network distillation. In the proposed approach, teacher agents are trained in different subsets of RL environment, thus diversifying the policies they learn. Then student agents are trained using distillation from the trained teachers to guide the training process, allowing for better exploring the solution space, while "mimicking" an existing policy/trading strategy provided by the teacher model. The boost in effectiveness of the proposed method comes from the use of diversified ensembles of teachers trained to perform trading for different currencies. This enables us to transfer the common view regarding the most profitable policy to the student, further improving the training stability in noisy financial environments. In the conducted experiments we find that when applying distillation, constraining the teacher models to be diversified can significantly improve their performance of the final student agents. We demonstrate this by providing an extensive evaluation on various financial trading tasks. Furthermore, we also provide additional experiments in the separate domain of control in games using the Procgen environments in order to demonstrate the generality of the proposed method.
Collapse
Affiliation(s)
- Avraam Tsantekidis
- School of Informatics, Aristotle University of Thessaloniki, Thessaloniki 54124, Greece.
| | - Nikolaos Passalis
- School of Informatics, Aristotle University of Thessaloniki, Thessaloniki 54124, Greece.
| | - Anastasios Tefas
- School of Informatics, Aristotle University of Thessaloniki, Thessaloniki 54124, Greece.
| |
Collapse
|
20
|
|
21
|
Learning to trade in financial time series using high-frequency through wavelet transformation and deep reinforcement learning. APPL INTELL 2021. [DOI: 10.1007/s10489-021-02218-4] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
22
|
AbdelKawy R, Abdelmoez WM, Shoukry A. A synchronous deep reinforcement learning model for automated multi-stock trading. PROGRESS IN ARTIFICIAL INTELLIGENCE 2021. [DOI: 10.1007/s13748-020-00225-z] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
23
|
Abstract
This study investigated the performance of a trading agent based on a convolutional neural network model in portfolio management. The results showed that with real-world data the agent could produce relevant trading results, while the agent’s behavior corresponded to that of a high-risk taker. The data used were wide in comparison with earlier reported research and was based on the full set of the S&P 500 stock data for twenty-one years supplemented with selected financial ratios. The results presented are new in terms of the size of the data set used and with regards to the model used. The results provide direction and offer insight into how deep learning methods may be used in constructing automatic trading systems.
Collapse
|
24
|
Abstract
We present a model for active trading based on reinforcement machine learning and apply this to five major cryptocurrencies in circulation. In relation to a buy-and-hold approach, we demonstrate how this model yields enhanced risk-adjusted returns and serves to reduce downside risk. These findings hold when accounting for actual transaction costs. We conclude that real-world portfolio management application of the model is viable, yet, performance can vary based on how it is calibrated in test samples.
Collapse
|
25
|
Leem J, Kim HY. Action-specialized expert ensemble trading system with extended discrete action space using deep reinforcement learning. PLoS One 2020; 15:e0236178. [PMID: 32716945 PMCID: PMC7384672 DOI: 10.1371/journal.pone.0236178] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2020] [Accepted: 06/30/2020] [Indexed: 12/05/2022] Open
Abstract
Despite active research on trading systems based on reinforcement learning, the development and performance of research methods require improvements. This study proposes a new action-specialized expert ensemble method consisting of action-specialized expert models designed specifically for each reinforcement learning action: buy, hold, and sell. Models are constructed by examining and defining different reward values that correlate with each action under specific conditions, and investment behavior is reflected with each expert model. To verify the performance of this technique, profits of the proposed system are compared to those of single trading and common ensemble systems. To verify robustness and account for the extension of discrete action space, we compared and analyzed changes in profits of the three actions to our model's results. Furthermore, we checked for sensitivity with three different reward functions: profit, Sharpe ratio, and Sortino ratio. All experiments were conducted with S&P500, Hang Seng Index, and Eurostoxx50 data. The model was 39.1% and 21.6% more efficient than single and common ensemble models, respectively. Considering the extended discrete action space, the 3-action space was extended to 11- and 21-action spaces, and the cumulative returns increased by 427.2% and 856.7%, respectively. Results on reward functions indicated that our models are well trained; results of the Sharpe and Sortino ratios were better than the implementation of profit only, as in the single-model cases. The Sortino ratio was slightly better than the Sharpe ratio.
Collapse
Affiliation(s)
- JoonBum Leem
- Department of Financial Engineering, Ajou University, Yeongtong-gu, Suwon, Republic of Korea
| | - Ha Young Kim
- Graduate School of Information, Yonsei University, Seodaemun-gu, Seoul, Republic of Korea
| |
Collapse
|
26
|
Calabuig J, Falciani H, Sánchez-Pérez E. Dreaming machine learning: Lipschitz extensions for reinforcement learning on financial markets. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2020.02.052] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/21/2022]
|
27
|
Abstract
Recently there has been an exponential increase in the use of artificial intelligence for trading in financial markets such as stock and forex. Reinforcement learning has become of particular interest to financial traders ever since the program AlphaGo defeated the strongest human contemporary Go board game player Lee Sedol in 2016. We systematically reviewed all recent stock/forex prediction or trading articles that used reinforcement learning as their primary machine learning method. All reviewed articles had some unrealistic assumptions such as no transaction costs, no liquidity issues and no bid or ask spread issues. Transaction costs had significant impacts on the profitability of the reinforcement learning algorithms compared with the baseline algorithms tested. Despite showing statistically significant profitability when reinforcement learning was used in comparison with baseline models in many studies, some showed no meaningful level of profitability, in particular with large changes in the price pattern between the system training and testing data. Furthermore, few performance comparisons between reinforcement learning and other sophisticated machine/deep learning models were provided. The impact of transaction costs, including the bid/ask spread on profitability has also been assessed. In conclusion, reinforcement learning in stock/forex trading is still in its early development and further research is needed to make it a reliable method in this domain.
Collapse
|
28
|
|
29
|
Gokcesu K, Kozat SS. An Online Minimax Optimal Algorithm for Adversarial Multiarmed Bandit Problem. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:5565-5580. [PMID: 29994080 DOI: 10.1109/tnnls.2018.2806006] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
We investigate the adversarial multiarmed bandit problem and introduce an online algorithm that asymptotically achieves the performance of the best switching bandit arm selection strategy. Our algorithms are truly online such that we do not use the game length or the number of switches of the best arm selection strategy in their constructions. Our results are guaranteed to hold in an individual sequence manner, since we have no statistical assumptions on the bandit arm losses. Our regret bounds, i.e., our performance bounds with respect to the best bandit arm selection strategy, are minimax optimal up to logarithmic terms. We achieve the minimax optimal regret with computational complexity only log-linear in the game length. Thus, our algorithms can be efficiently used in applications involving big data. Through an extensive set of experiments involving synthetic and real data, we demonstrate significant performance gains achieved by the proposed algorithm with respect to the state-of-the-art switching bandit algorithms. We also introduce a general efficiently implementable bandit arm selection framework, which can be adapted to various applications.
Collapse
|
30
|
Emerging Technologies and Opportunities for Innovation in Financial Data Analytics: A Perspective. BIG DATA ANALYTICS 2018. [DOI: 10.1007/978-3-030-04780-1_9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022] Open
|
31
|
Wang H, Huang T, Liao X, Abu-Rub H, Chen G. Reinforcement Learning for Constrained Energy Trading Games With Incomplete Information. IEEE TRANSACTIONS ON CYBERNETICS 2017; 47:3404-3416. [PMID: 28885145 DOI: 10.1109/tcyb.2016.2539300] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
This paper considers the problem of designing adaptive learning algorithms to seek the Nash equilibrium (NE) of the constrained energy trading game among individually strategic players with incomplete information. In this game, each player uses the learning automaton scheme to generate the action probability distribution based on his/her private information for maximizing his own averaged utility. It is shown that if one of admissible mixed-strategies converges to the NE with probability one, then the averaged utility and trading quantity almost surely converge to their expected ones, respectively. For the given discontinuous pricing function, the utility function has already been proved to be upper semicontinuous and payoff secure which guarantee the existence of the mixed-strategy NE. By the strict diagonal concavity of the regularized Lagrange function, the uniqueness of NE is also guaranteed. Finally, an adaptive learning algorithm is provided to generate the strategy probability distribution for seeking the mixed-strategy NE.
Collapse
|
32
|
Deng Y, Bao F, Kong Y, Ren Z, Dai Q. Deep Direct Reinforcement Learning for Financial Signal Representation and Trading. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2017; 28:653-664. [PMID: 26890927 DOI: 10.1109/tnnls.2016.2522401] [Citation(s) in RCA: 107] [Impact Index Per Article: 15.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/20/2023]
Abstract
Can we train the computer to beat experienced traders for financial assert trading? In this paper, we try to address this challenge by introducing a recurrent deep neural network (NN) for real-time financial signal representation and trading. Our model is inspired by two biological-related learning concepts of deep learning (DL) and reinforcement learning (RL). In the framework, the DL part automatically senses the dynamic market condition for informative feature learning. Then, the RL module interacts with deep representations and makes trading decisions to accumulate the ultimate rewards in an unknown environment. The learning system is implemented in a complex NN that exhibits both the deep and recurrent structures. Hence, we propose a task-aware backpropagation through time method to cope with the gradient vanishing issue in deep training. The robustness of the neural system is verified on both the stock and the commodity future markets under broad testing conditions.
Collapse
|
33
|
Fallahpour S, Hakimian H, Taheri K, Ramezanifar E. Pairs trading strategy optimization using the reinforcement learning method: a cointegration approach. Soft comput 2016. [DOI: 10.1007/s00500-016-2298-4] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
34
|
Mousavi S, Esfahanipour A, Zarandi MHF. A novel approach to dynamic portfolio trading system using multitree genetic programming. Knowl Based Syst 2014. [DOI: 10.1016/j.knosys.2014.04.018] [Citation(s) in RCA: 38] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
35
|
Chen X, Gao Y, Wang R. Online selective kernel-based temporal difference learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2013; 24:1944-1956. [PMID: 24805214 DOI: 10.1109/tnnls.2013.2270561] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
In this paper, an online selective kernel-based temporal difference (OSKTD) learning algorithm is proposed to deal with large scale and/or continuous reinforcement learning problems. OSKTD includes two online procedures: online sparsification and parameter updating for the selective kernel-based value function. A new sparsification method (i.e., a kernel distance-based online sparsification method) is proposed based on selective ensemble learning, which is computationally less complex compared with other sparsification methods. With the proposed sparsification method, the sparsified dictionary of samples is constructed online by checking if a sample needs to be added to the sparsified dictionary. In addition, based on local validity, a selective kernel-based value function is proposed to select the best samples from the sample dictionary for the selective kernel-based value function approximator. The parameters of the selective kernel-based value function are iteratively updated by using the temporal difference (TD) learning algorithm combined with the gradient descent technique. The complexity of the online sparsification procedure in the OSKTD algorithm is O(n). In addition, two typical experiments (Maze and Mountain Car) are used to compare with both traditional and up-to-date O(n) algorithms (GTD, GTD2, and TDC using the kernel-based value function), and the results demonstrate the effectiveness of our proposed algorithm. In the Maze problem, OSKTD converges to an optimal policy and converges faster than both traditional and up-to-date algorithms. In the Mountain Car problem, OSKTD converges, requires less computation time compared with other sparsification methods, gets a better local optima than the traditional algorithms, and converges much faster than the up-to-date algorithms. In addition, OSKTD can reach a competitive ultimate optima compared with the up-to-date algorithms.
Collapse
|
36
|
Ou SL, Liu LYD, Ou YC. Using a genetic algorithm-based RAROC model for the performance and persistence of the funds. J Appl Stat 2013. [DOI: 10.1080/02664763.2013.856870] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
37
|
Tung WL, Quek C. Financial volatility trading using a self-organising neural-fuzzy semantic network and option straddle-based approach. EXPERT SYSTEMS WITH APPLICATIONS 2011; 38:4668-4688. [PMID: 32288336 PMCID: PMC7126939 DOI: 10.1016/j.eswa.2010.07.116] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Financial volatility refers to the intensity of the fluctuations in the expected return on an investment or the pricing of a financial asset due to market uncertainties. Hence, volatility modeling and forecasting is imperative to financial market investors, as such projections allow the investors to adjust their trading strategies in anticipation of the impending financial market movements. Following this, financial volatility trading is the capitalization of the uncertainties of the financial markets to realize investment profits in times of rising, falling and side-way market conditions. In this paper, an intelligent straddle trading system (framework) that consists of a volatility projection module (VPM) and a trade decision module (TDM) is proposed for financial volatility trading via the buying and selling of option straddles to help a human trader capitalizes on the underlying uncertainties of the Hong Kong stock market. Three different measures, namely: (1) the historical volatility (HV), (2) implied volatility (IV) and (3) model-based volatility (MV) of the Hang Seng Index (HSI) are employed to quantify the implicit volatility of the Hong Kong stock market. The TDM of the proposed straddle trading system combines the respective volatility measures with the well-established moving-averages convergence/divergence (MACD) principle to recommend trading actions to a human trader dealing in HSI straddles. However, the inherent limitation of the MACD trading rule is that it generates time-delayed trading signals due to the use of moving averages, which are essentially lagging trend indicators. This drawback is intuitively addressed in the proposed straddle trading system by applying the VPM to compute future projections of the volatility measures of the HSI prior to the activation of the TDM. The VPM is realized by a self-organising neural-fuzzy semantic network named the evolving fuzzy semantic memory (eFSM) model. As compared to existing statistical and computational intelligence based modeling techniques currently employed for financial volatility modeling and forecasting, eFSM possesses several desirable attributes such as: (1) an evolvable knowledge base to continuously address the non-stationary characteristics of the Hong Kong stock market; (2) highly formalized human-like information computations; and (3) a transparent structure that can be interpreted via a set of linguistic IF-THEN semantic fuzzy rules. These qualities provide added credence to the computed HSI volatility projections. The volatility modeling and forecasting performances of the eFSM, when benchmarked to several established modeling techniques, as well as the observed trading returns of the proposed straddle trading system, are encouraging.
Collapse
Affiliation(s)
- W L Tung
- Centre for Computational Intelligence, Block N4 #2A-32, School of Computer Engineering, Nanyang Technological University, Nanyang Avenue, Singapore 639798, Singapore
| | - C Quek
- Centre for Computational Intelligence, Block N4 #2A-32, School of Computer Engineering, Nanyang Technological University, Nanyang Avenue, Singapore 639798, Singapore
| |
Collapse
|
38
|
Abstract
Predictive models in regression and classification problems typically have a single model that covers most, if not all, cases in the data. At the opposite end of the spectrum is a collection of models, each of which covers a very small subset of the decision space. These are referred to as “small disjuncts.” The trade-offs between the two types of models have been well documented. Single models, especially linear ones, are easy to interpret and explain. In contrast, small disjuncts do not provides as clean or as simple an interpretation of the data, and have been shown by several researchers to be responsible for a disproportionately large number of errors when applied to out-of-sample data. This research provides a counterpoint, demonstrating that a portfolio of “simple” small disjuncts provides a credible model for financial market prediction, a problem with a high degree of noise. A related novel contribution of this article is a simple method for measuring the “yield” of a learning system, which is the percentage of in-sample performance that the learned model can be expected to realize on out-of-sample data. Curiously, such a measure is missing from the literature on regression learning algorithms. Pragmatically, the results suggest that for problems characterized by a high degree of noise and lack of a stable knowledge base it makes sense to reconstruct the portfolio of small rules periodically.
Collapse
|
39
|
Vassiliades V, Cleanthous A, Christodoulou C. Multiagent reinforcement learning: spiking and nonspiking agents in the iterated Prisoner's Dilemma. ACTA ACUST UNITED AC 2011; 22:639-53. [PMID: 21421435 DOI: 10.1109/tnn.2011.2111384] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
This paper investigates multiagent reinforcement learning (MARL) in a general-sum game where the payoffs' structure is such that the agents are required to exploit each other in a way that benefits all agents. The contradictory nature of these games makes their study in multiagent systems quite challenging. In particular, we investigate MARL with spiking and nonspiking agents in the Iterated Prisoner's Dilemma by exploring the conditions required to enhance its cooperative outcome. The spiking agents are neural networks with leaky integrate-and-fire neurons trained with two different learning algorithms: 1) reinforcement of stochastic synaptic transmission, or 2) reward-modulated spike-timing-dependent plasticity with eligibility trace. The nonspiking agents use a tabular representation and are trained with Q- and SARSA learning algorithms, with a novel reward transformation process also being applied to the Q-learning agents. According to the results, the cooperative outcome is enhanced by: 1) transformed internal reinforcement signals and a combination of a high learning rate and a low discount factor with an appropriate exploration schedule in the case of non-spiking agents, and 2) having longer eligibility trace time constant in the case of spiking agents. Moreover, it is shown that spiking and nonspiking agents have similar behavior and therefore they can equally well be used in a multiagent interaction setting. For training the spiking agents in the case where more than one output neuron competes for reinforcement, a novel and necessary modification that enhances competition is applied to the two learning algorithms utilized, in order to avoid a possible synaptic saturation. This is done by administering to the networks additional global reinforcement signals for every spike of the output neurons that were not "responsible" for the preceding decision.
Collapse
|
40
|
Joseph D, Gangadhar G, Srinivasa Chakravarthy V. ACE (Actor–Critic–Explorer) paradigm for reinforcement learning in basal ganglia: Highlighting the role of subthalamic and pallidal nuclei. Neurocomputing 2010. [DOI: 10.1016/j.neucom.2010.03.001] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
41
|
Weissensteiner A. A Q-learning approach to derive optimal consumption and investment strategies. IEEE TRANSACTIONS ON NEURAL NETWORKS 2009; 20:1234-43. [PMID: 19497814 DOI: 10.1109/tnn.2009.2020850] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
In this paper, we consider optimal consumption and strategic asset allocation decisions of an investor with a finite planning horizon. A Q-learning approach is used to maximize the expected utility of consumption. The first part of the paper presents conceptually the implementation of Q -learning in a discrete state-action space and illustrates the relation of the technique to the dynamic programming method for a simplified setting. In the second part of the paper, different generalization methods are explored and, compared to other implementations using neural networks, a combination with self-organizing maps (SOMs) is proposed. The resulting policy is compared to alternative strategies.
Collapse
Affiliation(s)
- Alex Weissensteiner
- Department of Banking and Finance, University of Innsbruck, 6020 Innsbruck, Austria.
| |
Collapse
|
42
|
Freitas FD, De Souza AF, de Almeida AR. Prediction-based portfolio optimization model using neural networks. Neurocomputing 2009. [DOI: 10.1016/j.neucom.2008.08.019] [Citation(s) in RCA: 76] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
43
|
|
44
|
Lee JW, Park J, O J, Lee J, Hong E. A Multiagent Approach to $Q$-Learning for Daily Stock Trading. ACTA ACUST UNITED AC 2007. [DOI: 10.1109/tsmca.2007.904825] [Citation(s) in RCA: 59] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
45
|
|
46
|
Xu X, Hu D, Lu X. Kernel-Based Least Squares Policy Iteration for Reinforcement Learning. ACTA ACUST UNITED AC 2007; 18:973-92. [PMID: 17668655 DOI: 10.1109/tnn.2007.899161] [Citation(s) in RCA: 167] [Impact Index Per Article: 9.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Abstract
In this paper, we present a kernel-based least squares policy iteration (KLSPI) algorithm for reinforcement learning (RL) in large or continuous state spaces, which can be used to realize adaptive feedback control of uncertain dynamic systems. By using KLSPI, near-optimal control policies can be obtained without much a priori knowledge on dynamic models of control plants. In KLSPI, Mercer kernels are used in the policy evaluation of a policy iteration process, where a new kernel-based least squares temporal-difference algorithm called KLSTD-Q is proposed for efficient policy evaluation. To keep the sparsity and improve the generalization ability of KLSTD-Q solutions, a kernel sparsification procedure based on approximate linear dependency (ALD) is performed. Compared to the previous works on approximate RL methods, KLSPI makes two progresses to eliminate the main difficulties of existing results. One is the better convergence and (near) optimality guarantee by using the KLSTD-Q algorithm for policy evaluation with high precision. The other is the automatic feature selection using the ALD-based kernel sparsification. Therefore, the KLSPI algorithm provides a general RL method with generalization performance and convergence guarantee for large-scale Markov decision problems (MDPs). Experimental results on a typical RL task for a stochastic chain problem demonstrate that KLSPI can consistently achieve better learning efficiency and policy quality than the previous least squares policy iteration (LSPI) algorithm. Furthermore, the KLSPI method was also evaluated on two nonlinear feedback control problems, including a ship heading control problem and the swing up control of a double-link underactuated pendulum called acrobot. Simulation results illustrate that the proposed method can optimize controller performance using little a priori information of uncertain dynamic systems. It is also demonstrated that KLSPI can be applied to online learning control by incorporating an initial controller to ensure online performance.
Collapse
Affiliation(s)
- Xin Xu
- Institute of Automation, College of Mechatronics and Automation, National University of Defense Technology, Changsha 410073, P. R. China.
| | | | | |
Collapse
|
47
|
Maximizing winning trades using a novel RSPOP fuzzy neural network intelligent stock trading system. APPL INTELL 2007. [DOI: 10.1007/s10489-007-0055-1] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
48
|
Solving Deep Memory POMDPs with Recurrent Policy Gradients. LECTURE NOTES IN COMPUTER SCIENCE 2007. [DOI: 10.1007/978-3-540-74690-4_71] [Citation(s) in RCA: 43] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/12/2023]
|
49
|
Ang KK, Quek C. Stock trading using RSPOP: a novel rough set-based neuro-fuzzy approach. ACTA ACUST UNITED AC 2006; 17:1301-15. [PMID: 17001989 DOI: 10.1109/tnn.2006.875996] [Citation(s) in RCA: 88] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
This paper investigates the method of forecasting stock price difference on artificially generated price series data using neuro-fuzzy systems and neural networks. As trading profits is more important to an investor than statistical performance, this paper proposes a novel rough set-based neuro-fuzzy stock trading decision model called stock trading using rough set-based pseudo outer-product (RSPOP) which synergizes the price difference forecast method with a forecast bottleneck free trading decision model. The proposed stock trading with forecast model uses the pseudo outer-product based fuzzy neural network using the compositional rule of inference [POPFNN-CRI(S)] with fuzzy rules identified using the RSPOP algorithm as the underlying predictor model and simple moving average trading rules in the stock trading decision model. Experimental results using the proposed stock trading with RSPOP forecast model on real world stock market data are presented. Trading profits in terms of portfolio end values obtained are benchmarked against stock trading with dynamic evolving neural-fuzzy inference system (DENFIS) forecast model, the stock trading without forecast model and the stock trading with ideal forecast model. Experimental results showed that the proposed model identified rules with greater interpretability and yielded significantly higher profits than the stock trading with DENFIS forecast model and the stock trading without forecast model.
Collapse
Affiliation(s)
- Kai Keng Ang
- Centre for Computational Intelligence, School of Computer Engineering, Nanyang Technological University, Singapore 639798, Singapore
| | | |
Collapse
|
50
|
O J, LEE J, LEE J, ZHANG B. Adaptive stock trading with dynamic asset allocation using reinforcement learning. Inf Sci (N Y) 2006. [DOI: 10.1016/j.ins.2005.10.009] [Citation(s) in RCA: 35] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|