File size: 1,304 Bytes
1b55753 e3ef37a 1b55753 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 |
---
license: apache-2.0
datasets:
- ustc-zyt/time-r1-data
language:
- en
metrics:
- mse
- mae
base_model:
- Qwen/Qwen2.5-7B
---
# π§ Time-R1 Reinforced Model Weights
These are the official **reinforcement learning (RL) fine-tuned model checkpoints** for the paper:
**"Time Series Forecasting as Reasoning: A Slow-Thinking Approach with Reinforced LLMs"**.
---
## π¦ Model Details
* **Base Model**: Qwen2.5-7B
* **Tuning Framework**: [Verl](https://github.com/volcengine/verl) + [LLaMA Factory](https://github.com/hiyouga/LLaMA-Factory)
* **Final Stage**: Trained using GRIP (Group-based Relative Importance Policy optimization)
* **Objective**: Multi-horizon time series forecasting with structured reasoning
---
## π¦ Files Included
This model follows the standard Hugging Face `transformers` format and uses the efficient `safetensors` backend.
```
Time-R1/
βββ config.json
βββ generation_config.json
βββ model.safetensors.index.json
βββ model-00001-of-00004.safetensors
βββ model-00002-of-00004.safetensors
βββ model-00003-of-00004.safetensors
βββ model-00004-of-00004.safetensors
βββ tokenizer_config.json
βββ tokenizer.json
βββ vocab.json
```
β
Fully compatible with Hugging Face `transformers` and `AutoModelForCausalLM`. |