File size: 1,641 Bytes
25463df | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 | ---
license: apache-2.0
language:
- en
tags:
- reasoning
- chain-of-thought
- cot
- thinking
- llama
base_model: Qwen/Qwen2.5-1.5B
pipeline_tag: text-generation
---
# Shivik-2B-Reasoning-Expanded
A reasoning-optimized language model with Chain-of-Thought (CoT) capabilities using `<think>` tags.
## Model Details
| Property | Value |
|----------|-------|
| Parameters | Unknown |
| Hidden Size | Unknown |
| Layers | Unknown |
| Context Length | Unknown |
| CoT Support | ✅ Yes (`<think>` tags) |
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "shivash/Shivik-2B-Reasoning-Expanded"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype="auto",
device_map="auto"
)
# For reasoning tasks, the model uses <think> tags
prompt = "Solve this step by step: What is 15% of 80?"
messages = [
{"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(
**inputs,
max_new_tokens=512,
temperature=0.7,
do_sample=True,
)
response = tokenizer.decode(outputs[0], skip_special_tokens=False)
print(response)
```
## Chain-of-Thought Format
The model uses `<think>` tags for internal reasoning:
```
<think>
Let me work through this step by step...
15% means 15/100 = 0.15
0.15 × 80 = 12
</think>
The answer is 12.
```
## Training
This model was trained on reasoning datasets with Chain-of-Thought demonstrations.
## License
Apache 2.0
|