File size: 4,942 Bytes
e2e53f4
874402c
e2e53f4
874402c
 
e2e53f4
874402c
 
 
 
 
 
 
 
 
e2e53f4
 
874402c
 
 
 
 
 
 
e2e53f4
41f54c2
 
874402c
e2e53f4
874402c
 
 
 
 
e2e53f4
874402c
e2e53f4
840dd86
e2e53f4
874402c
e2e53f4
874402c
 
840dd86
874402c
e0251de
e2e53f4
874402c
e2e53f4
874402c
 
 
 
 
 
 
 
 
 
 
 
e2e53f4
874402c
 
 
 
 
 
e2e53f4
874402c
 
 
 
 
e2e53f4
78472ea
 
874402c
 
 
 
e2e53f4
874402c
e2e53f4
874402c
e2e53f4
874402c
 
 
 
 
e2e53f4
874402c
 
 
 
e2e53f4
874402c
e2e53f4
874402c
e2e53f4
874402c
e2e53f4
e0251de
e2e53f4
874402c
e2e53f4
874402c
 
 
 
e2e53f4
874402c
 
 
 
 
e2e53f4
874402c
e2e53f4
874402c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
---
base_model: meta-llama/Llama-3.2-3B-Instruct
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- gguf
- GRPO
- meta
license: apache-2.0
language:
- en
datasets:
- openai/gsm8k
---

<div align="center">
<img src="https://cdn-uploads.huggingface.co/production/uploads/669777597cb32718c20d97e9/4emWK_PB-RrifIbrCUjE8.png"
     alt="Title card" 
     style="width: 500px;
            height: auto;
            object-position: center top;">
</div>

**Website - https://www.alphaai.biz**

# Uploaded  model

- **Developed by:** alphaaico
- **License:** apache-2.0
- **Finetuned from model :** meta-llama/Llama-3.2-3B-Instruct
- **Training Framework:** Unsloth + Hugging Face TRL
- **Finetuning Techniques:** GRPO + Reward Modelling

## Overview

Welcome to the next evolution of AI reasoning! Reason-With-Choice-3B is not just another fine-tuned model, it's a game-changer. It doesn't just generate reasoning, it chooses whether reasoning is even necessary before delivering an answer. This self-reflective capability allows it to introspect, analyze, and adapt to the complexity of each question, ensuring the most efficient and insightful response possible.

Think about it: most AI models blindly generate reasoning even when unnecessary, leading to bloated, redundant responses. Not this one. With its built-in decision-making, Reason-With-Choice-3B determines if deep reasoning is needed or if a direct answer will suffice—bringing unparalleled efficiency and intelligence to your AI-driven applications.

## Key Highlights
- Reasoning & Self-Reflection: The model first decides if reasoning is necessary and then either provides step-by-step logic or directly answers the question.
- Structured Output: Responses follow a strict format with `<think>`, `<reflection>`, and `<answer>` sections, ensuring clarity and interpretability.
- Optimized Training: Trained using GRPO (Guided Reward Policy Optimization) to enforce structured responses and improve decision-making.
- Efficient Inference: Fine-tuned with Unsloth & Hugging Face's TRL, ensuring faster inference speeds and optimized resource utilization.

## Prompt Structure

The model generates responses in the following structured format:
```python
<think>
[Detailed reasoning, if required. Otherwise, this section remains empty.]
</think>
<reflection>
[Internal thought process explaining whether reasoning was needed.]
</reflection>
<answer>
[Final response.]
</answer>
```

## Key Features
- Decision-Making Capability: The model intelligently determines whether reasoning is necessary before answering.
- Improved Accuracy: Training with reward functions ensures adherence to logical response structure.
- Structured Outputs: Guarantees that each response follows a predictable and interpretable format.
- Enhanced Efficiency: Optimized inference with vLLM for fast token generation and low memory footprint.
- Multi-Use Case Compatibility: Can be used for Q&A systems, logical reasoning tasks, and AI-assisted decision-making.

## Quantization Levels Available
- q4_k_m
- q5_k_m
- q8_0
- 16-bit (Full Precision)

GGUF Versions - https://huggingface.co/alpha-ai/Reason-With-Choice-3B-GGUF

## Ideal Configuration for Usage
- Temperature: 0.8
- Top-p: 0.95
- Max Tokens: 1024

## Use Cases

**Reason-With-Choice-3B is ideal for:**

- AI Research: Investigating decision-making and reasoning processes in AI.
- Conversational AI: Enhancing chatbot intelligence with structured reasoning.
- Automated Decision Support: Assisting in structured, step-by-step problem-solving.
- Educational Tools: Providing logical explanations for learning and problem-solving.
- Business Intelligence: AI-assisted decision-making for operational and strategic planning.

## Limitations & Considerations
- Domain Adaptation: May require further fine-tuning for domain-specific tasks.
- Inference Time: Increased processing time when reasoning is necessary.
- Potential Biases: Outputs depend on training data and may require verification for critical applications.

## License

This model is released under the Apache-2.0 license.

## Acknowledgments

Special thanks to the Unsloth team for optimizing the fine-tuning pipeline and to Hugging Face's TRL for enabling advanced fine-tuning techniques.

## Security & Format Considerations

This model has been saved in .bin format due to Unsloth's default serialization method. If security is a concern, we recommend converting to .safetensors using:
```python
from transformers import AutoModel
from safetensors.torch import save_file

model = AutoModel.from_pretrained("path/to/model")
state_dict = model.state_dict()
save_file(state_dict, "model.safetensors")
print("Model converted to safetensors successfully.")
```

Alternatively, GGUF models are available for optimized inference with llama.cpp, exllama, and other runtime frameworks.

Choose the format best suited to your security, performance, and deployment requirements.