The dataset viewer is not available because its heuristics could not detect any supported data files. You can try uploading some data files, or configuring the data files location manually.
Reinforcement Learning Dataset for Volatility-Driven Futures Trading
This dataset collection contains four curated subsets of minute-level Binance Futures market data designed specifically for training, validating, testing, and backtesting reinforcement learning agents in high-volatility environments. The data is used in a production-grade project implementing a Dueling Double Deep Q-Network (D3QN) with Prioritized Experience Replay (PER), tailored for financial markets.
π‘ Project Context
The dataset supports a full RL pipeline aimed at developing an intelligent trading system capable of:
- Making profitable trading decisions in highly volatile conditions;
- Learning from localized market impulses rather than continuous streams;
- Operating in realistic conditions, including slippage and transaction fees.
π Full codebase: GitHub Repository
π Research article (English): RL Agent for Algorithmic Trading on Binance Futures β Architecture, Backtest, and Results
π Research article (Russian): RL-Π°Π³Π΅Π½Ρ Π΄Π»Ρ Π°Π»Π³ΠΎΡΠΈΡΠΌΠΈΡΠ΅ΡΠΊΠΎΠΉ ΡΠΎΡΠ³ΠΎΠ²Π»ΠΈ Π½Π° Binance Futures: Π°ΡΡ
ΠΈΡΠ΅ΠΊΡΡΡΠ°, Π±ΡΠΊΡΠ΅ΡΡ, ΡΠ΅Π·ΡΠ»ΡΡΠ°ΡΡ
π€ Real-time RL predictions: Telegram
π Dataset Structure
Each data sample is a 150-minute window centered around a strong volatility impulse.
- Shape:
(150, 7)β 150 minutes Γ 7 features - Channels:
open,high,low,close,volume,volume_weighted_average,num_trades - Format:
np.ndarraywrapped in.npz - Metadata: unique keys
(TICKER, datetime)per session
| Subset | Samples | Period | Purpose |
|---|---|---|---|
Train |
24,104 | 2020-01-14 β 2024-08-31 | Training |
Validation |
1,377 | 2024-09-01 β 2024-12-01 | Hyperparameter tuning |
Test |
3,400 | 2024-12-01 β 2025-03-01 | Final evaluation |
Backtest |
3,186 | 2025-03-01 β 2025-06-01 | Realistic simulation |
Each session contains:
- 90 minutes of pre-impulse history (for state construction)
- 60 minutes of post-impulse trading session (for reward computation)
π§ Dataset Motivation
This dataset departs from standard continuous sampling. Instead, it focuses only on high-volatility episodes that represent actual market decision points. Specifically:
- Price moves >5% within a 10-minute window
- Preceded by 90 minutes of relative stability
- Selected using a contrast ratio filter to remove noisy signals
These sessions serve as atomic training units for reinforcement learning agents operating in short-term trading strategies.
π§° Data Pipeline Tools
All preprocessing logic is encapsulated in reusable utilities (as part of the open project):
load_npz_dataset(path)β loads session list and metadataselect_and_arrange_channels(data, channels)β filters and arranges inputcalculate_normalization_stats(data)β computes per-channel statsapply_normalization(data, stats)β standardizes for agent consumption
π Utilities
All preprocessing tools are included in data_utils.py.
Data undergoes:
- Channel filtering
- Relative scaling
- Log transforms
- NaN/outlier protection
π Visualizations
Each sample comes with an optional visualization (example graphs available in the project):
- Line plots with volatility impulse marker at minute 90
- Session metadata in title (ticker + UTC timestamp)
- Used to audit signal quality and alignment
π License
License: MIT License
This dataset is released under the MIT license β you are free to use, modify, and distribute it for any purpose, including commercial use, provided that the original copyright and permission notice are included.
π¦ Download and Usage
You can load the dataset using HuggingFaceβs datasets library:
from datasets import load_dataset
# Example: load training split
train_dataset = load_dataset("ResearchRL/open-rl-trading-binance-dataset", split="train_data")
- Downloads last month
- 23