4.7 Article

Diversity-driven knowledge distillation for financial trading using Deep Reinforcement Learning

期刊

NEURAL NETWORKS
卷 140, 期 -, 页码 193-202

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.neunet.2021.02.026

关键词

Deep Reinforcement Learning; Financial markets; Trading

资金

  1. European Union
  2. Greek national funds through the Operational Program Competitiveness, Entrepreneurship and Innovation [T2EDK-02094]
  3. State Scholarship Foundation (IKY), Greece
  4. Human Resources Development, Education and Lifelong Learning 2014-2020'' program

向作者/读者索取更多资源

The study introduced a novel method to improve the training reliability of DRL trading agents by using neural network distillation, with diversified teacher agents guiding the training of student agents to enhance performance in noisy financial environments.
Deep Reinforcement Learning (RL) is increasingly used for developing financial trading agents for a wide range of tasks. However, optimizing deep RL agents is notoriously difficult and unstable, especially in noisy financial environments, significantly hindering the performance of trading agents. In this work, we present a novel method that improves the training reliability of DRL trading agents building upon the well-known approach of neural network distillation. In the proposed approach, teacher agents are trained in different subsets of RL environment, thus diversifying the policies they learn. Then student agents are trained using distillation from the trained teachers to guide the training process, allowing for better exploring the solution space, while mimicking'' an existing policy/trading strategy provided by the teacher model. The boost in effectiveness of the proposed method comes from the use of diversified ensembles of teachers trained to perform trading for different currencies. This enables us to transfer the common view regarding the most profitable policy to the student, further improving the training stability in noisy financial environments. In the conducted experiments we find that when applying distillation, constraining the teacher models to be diversified can significantly improve their performance of the final student agents. We demonstrate this by providing an extensive evaluation on various financial trading tasks. Furthermore, we also provide additional experiments in the separate domain of control in games using the Procgen environments in order to demonstrate the generality of the proposed method. (C) 2021 Elsevier Ltd. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据