A deep reinforcement learning framework for financial portfolio management with policy gradient optimization and backtesting tools.
PGPortfolio is a deep reinforcement learning framework specifically designed for financial portfolio management problems. It implements policy gradient optimization methods that treat portfolio management as an immediate reward optimization problem regularized by transaction costs, avoiding the need for Monte Carlo or bootstrapped gradient estimation. The framework includes tools for training configurable models, visualizing results with TensorBoard, and comparing against traditional financial algorithms.
Quantitative researchers, algorithmic trading developers, and academic researchers working on reinforcement learning applications in finance who need a specialized toolkit for portfolio optimization and backtesting.
Developers choose PGPortfolio because it offers a reinforcement learning approach specifically optimized for portfolio management with efficiency comparable to supervised learning, includes traditional financial algorithms for benchmarking, and provides a flexible, configurable research environment with visualization tools.
PGPortfolio: Policy Gradient Portfolio, the source code of "A Deep Reinforcement Learning Framework for the Financial Portfolio Management Problem"(https://arxiv.org/pdf/1706.10059.pdf).
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Implements a policy gradient method specifically designed for portfolio management, offering efficiency comparable to supervised learning by avoiding Monte Carlo gradient estimation, as detailed in the README.
Allows users to define network topology, training methods, and input data via JSON files, enabling customizable experimentation without code changes, as highlighted in the features.
Includes traditional financial algorithms from the OLPS toolbox for direct comparison, providing a comprehensive framework for evaluating RL against established methods.
Supports TensorBoard for monitoring training metrics and enables parallel runs for hyperparameter optimization, speeding up research iterations as mentioned in the README.
Relies on TensorFlow 1.x and tflearn, which are deprecated or less maintained, potentially causing compatibility issues and limiting access to modern ML features.
Primarily focused on backtesting with historical data; live trading requires additional broker API integration, as noted in the community contributions section and risk disclaimer.
README admits removal of internal discussions and versioning history, making it harder for new users to understand advanced features or contribute effectively.
Designed for academic experiments with configurable setups, but lacks optimizations for deployment, such as error handling or real-time data pipelines, as indicated by the focus on backtesting.