File size: 1,304 Bytes
1b55753
 
 
 
 
 
 
 
 
e3ef37a
 
1b55753
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
---
license: apache-2.0
datasets:
- ustc-zyt/time-r1-data
language:
- en
metrics:
- mse
- mae
base_model:
- Qwen/Qwen2.5-7B
---
# 🧠 Time-R1 Reinforced Model Weights

These are the official **reinforcement learning (RL) fine-tuned model checkpoints** for the paper:
**"Time Series Forecasting as Reasoning: A Slow-Thinking Approach with Reinforced LLMs"**.

---

## πŸ“¦ Model Details

* **Base Model**: Qwen2.5-7B
* **Tuning Framework**: [Verl](https://github.com/volcengine/verl) + [LLaMA Factory](https://github.com/hiyouga/LLaMA-Factory)
* **Final Stage**: Trained using GRIP (Group-based Relative Importance Policy optimization)
* **Objective**: Multi-horizon time series forecasting with structured reasoning

---

## πŸ“¦ Files Included

This model follows the standard Hugging Face `transformers` format and uses the efficient `safetensors` backend.

```
Time-R1/
β”œβ”€β”€ config.json
β”œβ”€β”€ generation_config.json
β”œβ”€β”€ model.safetensors.index.json
β”œβ”€β”€ model-00001-of-00004.safetensors
β”œβ”€β”€ model-00002-of-00004.safetensors
β”œβ”€β”€ model-00003-of-00004.safetensors
β”œβ”€β”€ model-00004-of-00004.safetensors
β”œβ”€β”€ tokenizer_config.json
β”œβ”€β”€ tokenizer.json
└── vocab.json
```

βœ… Fully compatible with Hugging Face `transformers` and `AutoModelForCausalLM`.