2022
DOI: 10.3390/stats5020033
|View full text |Cite
|
Sign up to set email alerts
|

Quantitative Trading through Random Perturbation Q-Network with Nonlinear Transaction Costs

Abstract: In recent years, reinforcement learning (RL) has seen increasing applications in the financial industry, especially in quantitative trading and portfolio optimization when the focus is on the long-term reward rather than short-term profit. Sequential decision making and Markov decision processes are rather suited for this type of application. Through trial and error based on historical data, an agent can learn the characteristics of the market and evolve an algorithm to maximize the cumulative returns. In this… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 13 publications
0
5
0
Order By: Relevance
“…DRA (Briola et al 2021) uses LSTM and PPO. CDQNRP (Zhu and Zhu 2022) uses a random perturbation to increase the stability of training a DQN. However, these algorithms focus mainly on designing only one RL agent to conduct profitable trading in short-term scenarios, neglecting its failure to maintain performance over long periods.…”
Section: Rl For Quantitative Tradingmentioning
confidence: 99%
See 1 more Smart Citation
“…DRA (Briola et al 2021) uses LSTM and PPO. CDQNRP (Zhu and Zhu 2022) uses a random perturbation to increase the stability of training a DQN. However, these algorithms focus mainly on designing only one RL agent to conduct profitable trading in short-term scenarios, neglecting its failure to maintain performance over long periods.…”
Section: Rl For Quantitative Tradingmentioning
confidence: 99%
“…PPO and DQN. 4 DRA and CDQNRP• CDQNRP(Zhu and Zhu 2022) uses a random perturbed target frequency to enhance the stability during training. • MACD(Krug, Dobaj, and Macher 2022) is an upgraded method based on the traditional moving average method.…”
mentioning
confidence: 99%
“…In Equation (42), 𝜋𝜋(𝑠𝑠 𝑡𝑡 (𝑛𝑛)|𝛉𝛉 𝑎𝑎 ) is the output of DDPG's actor. DDPG uses a replay buffer ℬ that includes samples from older policies.…”
Section: Replay Buffermentioning
confidence: 99%
“…Other domains where RL has been used include hospital decision making [37], precision agriculture [38], and fluid mechanics [39]. The financial industry is another important sector where RL has been adopted for several scenarios [40][41][42]. It is of little surprise that RL has been extensively used to solve various problems in energy systems [43][44][45][46][47].…”
Section: Introductionmentioning
confidence: 99%
“…Reinforcement learning (RL) serves as an important branch of machine learning. As a powerful approach in decision and control theory, RL has attracted extensive focus, with wide applications in the fields of robotics 1 , quantitative finance 2 , computer vision 3 , healthcare 4 , career planning 5 , gaming 6 etc. The most common object of a games is to beat the opponents, whether they are computers or other human players.…”
Section: Introductionmentioning
confidence: 99%