| | --- |
| | license: apache-2.0 |
| | datasets: |
| | - ustc-zyt/time-r1-data |
| | language: |
| | - en |
| | metrics: |
| | - mse |
| | - mae |
| | base_model: |
| | - Qwen/Qwen2.5-7B |
| | --- |
| | # π§ Time-R1 Reinforced Model Weights |
| |
|
| | These are the official **reinforcement learning (RL) fine-tuned model checkpoints** for the paper: |
| | **"Time Series Forecasting as Reasoning: A Slow-Thinking Approach with Reinforced LLMs"**. |
| |
|
| | --- |
| |
|
| | ## π¦ Model Details |
| |
|
| | * **Base Model**: Qwen2.5-7B |
| | * **Tuning Framework**: [Verl](https://github.com/volcengine/verl) + [LLaMA Factory](https://github.com/hiyouga/LLaMA-Factory) |
| | * **Final Stage**: Trained using GRIP (Group-based Relative Importance Policy optimization) |
| | * **Objective**: Multi-horizon time series forecasting with structured reasoning |
| |
|
| | --- |
| |
|
| | ## π¦ Files Included |
| |
|
| | This model follows the standard Hugging Face `transformers` format and uses the efficient `safetensors` backend. |
| |
|
| | ``` |
| | Time-R1/ |
| | βββ config.json |
| | βββ generation_config.json |
| | βββ model.safetensors.index.json |
| | βββ model-00001-of-00004.safetensors |
| | βββ model-00002-of-00004.safetensors |
| | βββ model-00003-of-00004.safetensors |
| | βββ model-00004-of-00004.safetensors |
| | βββ tokenizer_config.json |
| | βββ tokenizer.json |
| | βββ vocab.json |
| | ``` |
| |
|
| | β
Fully compatible with Hugging Face `transformers` and `AutoModelForCausalLM`. |