A framework for autonomous AI trading agents that self-improve their prompts through market feedback and Darwinian selection.
ATLAS is a framework for autonomous AI trading agents that self-improve through market feedback using an autoresearch loop inspired by Karpathy. It solves the problem of static AI models by continuously evolving agent prompts based on performance, training on different market regimes, and simulating reflexive futures to anticipate market changes.
Quantitative researchers, algorithmic traders, and fintech developers building adaptive AI-driven trading systems that require continuous self-improvement and regime-aware strategies.
Developers choose ATLAS for its unique integration of autoresearch, Darwinian agent weighting, and reflexive market simulation, enabling a self-evolving system that autonomously improves and adapts without manual intervention, all running cost-effectively on minimal infrastructure.
ATLAS by General Intelligence Capital โ Self-improving AI trading agents using Karpathy-style autoresearch
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Implements an autoresearch loop that modifies the worst-performing agent's prompt based on Sharpe ratio, with a 30% survival rate for modifications leading to continuous self-improvement without manual tuning.
Uses PRISM training to develop separate agent sets for distinct market conditions like bull markets or crises, enabling specialized survival strategies that adapt to different environments.
Runs on a $20/month Azure VM instead of GPUs, with full 18-month backtests costing $50-80, making iterative testing and deployment accessible for small teams.
The JANUS meta-layer algorithmically weights trained cohorts based on recent accuracy, with weight differentials emerging as a detector for novel or historical market regimes without explicit programming.
Trained agent prompts, evolved rules, and live data are proprietary and not included, forcing new users to start from scratch and lag behind by hundreds of iterations, as admitted in the README.
The 5-day testing period for prompt modifications is too slow for fast-moving markets, resulting in 0% modification survival in crisis and recovery cohorts, limiting effectiveness during rapid shifts.
Relies on multiple APIs (Anthropic, FMP, Finnhub) and the MiroFish engine, introducing setup complexity, ongoing costs, and potential points of failure in live deployment.