For financial reinforcement learning (FinRL), we've found a way to address the dreaded overfitting trap and increase your chances of success in the wild world of crypto. Our approach has been tested on 10 different currencies and during a market crash period, and has proven to be more profitable than the competition. So, don't just sit there, join us on our journey to the top of the crypto mountain!
Berend Gort is an Machine Learning Engineer. Berend recently worked as a research assistant at Columbia University and published the paper associated with this code at the conference AAAI '23, where he developed a solution that reduces overfitting in deep reinforcement learning models in finance by 46% compared to traditional methods. Berend is also passionate about meeting new people and is actively contributing the AI4Finance community.
The original paper authored by Berend Gort and Xiao-Yang Liu can be found here!
To reproduce the results in the paper, the codes are simplified as much as possible. You start with the settings inconfig_main.py
file, where you set all the settings for:
- The Walkforward, K-Cross Validation, and Combinatorial Purged Cross Validation (CPCV) methods.
- Set how many candles/data points you require for training and validation.
- Set which tickers you will download from Binance, the minimum buy limits.
- Set your technical indicators.
- Computes automatically the exact start and end dates for training and validation, respectively, based on your trade start date and end date.
A short description of each folder:
data
Contains all your training/validation data in the main folder, and a subfolder which containstrade_data
after download using both0_dl_trainval_data.py
and0_dl_trade_data.py
(more later)drl_agents
Contains the DRL framework ElegantRL which implements a series of model-free DRL algorithmsplots_and_metrics
Dump folder for all analysis images and performance metrics producedtrain
Holds all utility functions for DRL trainingtrain_results
After running either1_optimize_cpcv.py
/1_optimize_kcv.py
/1_optimize_wf.py
will have a folder with your trained DRL agents
Then, running and producing similar results to that in the paper are simple, following the numbered Python files as indicated by the number of the filename:
0_dl_trainval_data.py
Downloads the train and validation data according toconfig_main.py
0_dl_trade_data.py
Downloads the trade data according toconfig_main.py
1_optimize_cpcv.py
Optimizes hyperparameters with a Combinatorial Purged Cross-validation scheme1_optimize_kcv.py
Optimizes hyperparameters with a K-Fold Cross-validation scheme1_optimize_wf.py
Optimizes hyperparameters with a Walk-forward validation scheme2_validate.py
Shows insights about the training and validation process (select a results folder from train_results)4_backtestpy
Backtests trained DRL agents (enter multiple results folders from train_results in a list)5_pbo.py
Computes PBO for trained DRL agents (enter multiple results folders from train_results in a list)
Simply run the scripts in the above order. Please note the trained agents are auto-saved to the folder train_results
. That is where you can find your trained DRL agents!
@article{gort2022deep,
title={Deep reinforcement learning for cryptocurrency trading: Practical approach to address backtest overfitting},
author={Gort, Berend Jelmer Dirk and Liu, Xiao-Yang and Gao, Jiechao and Chen, Shuaiyu and Wang, Christina Dan},
journal={AAAI Bridge on AI for Financial Services},
year={2023}
}