File size: 2,435 Bytes
270ffa9
 
f1f70f2
 
 
 
 
 
 
270ffa9
 
 
 
 
 
 
 
 
 
f1f70f2
 
 
 
 
 
270ffa9
f1f70f2
270ffa9
f1f70f2
 
270ffa9
 
 
f1f70f2
 
 
270ffa9
 
 
f1f70f2
 
270ffa9
 
 
f1f70f2
 
270ffa9
f1f70f2
 
 
270ffa9
f1f70f2
 
270ffa9
f1f70f2
 
 
 
 
270ffa9
f1f70f2
270ffa9
 
 
f1f70f2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
---
library_name: transformers
license: apache-2.0
datasets:
- datumo/CAC-CoT
language:
- en
base_model:
- Qwen/Qwen2.5-7B-Instruct
---

# Model Card for Model ID

## Model Details

### Model Description

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

- **Developed by:** Sunguk Choi, Yonghoon Kwon, Heondeuk Lee  
- **Shared by:** SelectStar/Datumo
- **Model type:** Decoder-only language model (Causal LM)  
- **Language(s) (NLP):** English  
- **License:** Apache License 2.0  
- **Finetuned from model:** 🔧 Qwen-2.5-7b-it

### Model Sources

- **Repository:** https://github.com/selectstar-ai/CAC-CoT  
- **Paper:** https://arxiv.org/abs/2508.18743

### Direct Use

- Solving reasoning problems requiring chain-of-thought (CoT).
- Educational tutoring, math/logic assistants, explainable QA.
- Applications requiring interpretable reasoning with low latency.

### Downstream Use [optional]

- Fine-tuning for specific reasoning benchmarks such as GSM8K, StrategyQA, or S1-Bench.
- Integration into larger RAG or tutoring systems.

### Out-of-Scope Use

- Non-English tasks.
- Open-ended creative generation (e.g., fiction, poetry).

## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("datumo/CAC-CoT")  # 🔧 Replace with your model path
tokenizer = AutoTokenizer.from_pretrained("datumo/CAC-CoT")

prompt = "Problem: If you have 3 apples and get 2 more, how many do you have?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```

## Citation

**BibTeX:**

```
@misc{choi2025caccotconnectorawarecompactchainofthought,
  title={CAC-CoT: Connector-Aware Compact Chain-of-Thought for Efficient Reasoning Data Synthesis Across Dual-System Cognitive Tasks},
  author={Sunguk Choi and Yonghoon Kwon and Heondeuk Lee},
  year={2025},
  eprint={2508.18743},
  archivePrefix={arXiv},
  primaryClass={cs.AI},
  url={https://arxiv.org/abs/2508.18743},
}
```

## More Information
- System-1: Fast, intuitive reasoning
- System-2: Slow, logical reasoning
- Connector phrase: Fixed phrases guiding logical flow (e.g., “Because of this,” “Then,” etc.)
- ART: Average Reasoning Trace length

## Model Card Authors
Sunguk Choi, Yonghoon Kwon, Heondeuk Lee