Text Generation
Transformers
Safetensors
English
Inductive
Reasoning

Improve language tag

#1
by lbourdois - opened
Files changed (1) hide show
  1. README.md +90 -78
README.md CHANGED
@@ -1,79 +1,91 @@
1
- ---
2
- library_name: transformers
3
- tags:
4
- - Inductive
5
- - Reasoning
6
- datasets:
7
- - nsadeq/redis_generate_rule_alignment
8
- - nsadeq/redis_generate_rule_sft
9
- - nsadeq/redis_follow_rule_sft
10
- language:
11
- - en
12
- base_model:
13
- - Qwen/Qwen2.5-7B-Instruct
14
- pipeline_tag: text-generation
15
- ---
16
-
17
- ---
18
- library_name: transformers
19
- tags:
20
- - Inductive
21
- - Reasoning
22
- language:
23
- - en
24
- base_model:
25
- - Qwen/Qwen2.5-7B-Instruct
26
- pipeline_tag: text-generation
27
- ---
28
-
29
- # Model Card for Model ID
30
-
31
- ReDis-Llama is trained for improved inductive reasoning performance.
32
-
33
- ### Model Description
34
-
35
- - **Developed by:** Nafis Sadeq
36
- - **Language(s) (NLP):** English
37
- - **Finetuned from model:** Qwen/Qwen2.5-7B-Instruct
38
-
39
- ### Model Sources [optional]
40
-
41
-
42
- - **Repository:** https://github.com/NafisSadeq/reasoning-distillation
43
- - **Paper:** https://arxiv.org/abs/2504.10647
44
-
45
-
46
- ## How to Get Started with the Model
47
-
48
- Follow the instructions here: https://github.com/NafisSadeq/reasoning-distillation
49
-
50
- ## Training Details
51
-
52
- Training details can be found in the paper: https://arxiv.org/abs/2504.10647
53
-
54
- ## Environmental Impact
55
-
56
- - **Hardware Type:** 2 × 48 GB Nvidia RTX A6000 GPUs
57
- - **Hours used:** 72 hours
58
-
59
- ### Model Architecture and Objective
60
-
61
- This model has the same architecture as Qwen/Qwen2.5-7B-Instruct
62
-
63
- ### Compute Infrastructure
64
-
65
- 2 × 48 GB Nvidia RTX A6000 GPUs
66
-
67
- ## Citation
68
-
69
- If you use this model, please cite the following paper.
70
-
71
- @misc{sadeq2025improvingincontextlearningreasoning,
72
- title={Improving In-Context Learning with Reasoning Distillation},
73
- author={Nafis Sadeq and Xin Xu and Zhouhang Xie and Julian McAuley and Byungkyu Kang and Prarit Lamba and Xiang Gao},
74
- year={2025},
75
- eprint={2504.10647},
76
- archivePrefix={arXiv},
77
- primaryClass={cs.CL},
78
- url={https://arxiv.org/abs/2504.10647},
 
 
 
 
 
 
 
 
 
 
 
 
79
  }
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - Inductive
5
+ - Reasoning
6
+ datasets:
7
+ - nsadeq/redis_generate_rule_alignment
8
+ - nsadeq/redis_generate_rule_sft
9
+ - nsadeq/redis_follow_rule_sft
10
+ language:
11
+ - zho
12
+ - eng
13
+ - fra
14
+ - spa
15
+ - por
16
+ - deu
17
+ - ita
18
+ - rus
19
+ - jpn
20
+ - kor
21
+ - vie
22
+ - tha
23
+ - ara
24
+ base_model:
25
+ - Qwen/Qwen2.5-7B-Instruct
26
+ pipeline_tag: text-generation
27
+ ---
28
+
29
+ ---
30
+ library_name: transformers
31
+ tags:
32
+ - Inductive
33
+ - Reasoning
34
+ language:
35
+ - en
36
+ base_model:
37
+ - Qwen/Qwen2.5-7B-Instruct
38
+ pipeline_tag: text-generation
39
+ ---
40
+
41
+ # Model Card for Model ID
42
+
43
+ ReDis-Llama is trained for improved inductive reasoning performance.
44
+
45
+ ### Model Description
46
+
47
+ - **Developed by:** Nafis Sadeq
48
+ - **Language(s) (NLP):** English
49
+ - **Finetuned from model:** Qwen/Qwen2.5-7B-Instruct
50
+
51
+ ### Model Sources [optional]
52
+
53
+
54
+ - **Repository:** https://github.com/NafisSadeq/reasoning-distillation
55
+ - **Paper:** https://arxiv.org/abs/2504.10647
56
+
57
+
58
+ ## How to Get Started with the Model
59
+
60
+ Follow the instructions here: https://github.com/NafisSadeq/reasoning-distillation
61
+
62
+ ## Training Details
63
+
64
+ Training details can be found in the paper: https://arxiv.org/abs/2504.10647
65
+
66
+ ## Environmental Impact
67
+
68
+ - **Hardware Type:** 2 × 48 GB Nvidia RTX A6000 GPUs
69
+ - **Hours used:** 72 hours
70
+
71
+ ### Model Architecture and Objective
72
+
73
+ This model has the same architecture as Qwen/Qwen2.5-7B-Instruct
74
+
75
+ ### Compute Infrastructure
76
+
77
+ 2 × 48 GB Nvidia RTX A6000 GPUs
78
+
79
+ ## Citation
80
+
81
+ If you use this model, please cite the following paper.
82
+
83
+ @misc{sadeq2025improvingincontextlearningreasoning,
84
+ title={Improving In-Context Learning with Reasoning Distillation},
85
+ author={Nafis Sadeq and Xin Xu and Zhouhang Xie and Julian McAuley and Byungkyu Kang and Prarit Lamba and Xiang Gao},
86
+ year={2025},
87
+ eprint={2504.10647},
88
+ archivePrefix={arXiv},
89
+ primaryClass={cs.CL},
90
+ url={https://arxiv.org/abs/2504.10647},
91
  }