razielAI's picture
Update README.md
e61a171 verified
---
license: apache-2.0
language:
- he
- en
base_model:
- Raziel1234/Duchifat-2
pipeline_tag: text-generation
tags:
- text-generation-inference
- chemistry
- biology
- music
- legal
- medical
- agent
- finance
- climate
- PyTorch-Lightning
- PyTorch
library_name: transformers
---
# Duchifat-2-Thinking 🛰️
**Duchifat-2-Thinking** is a lightweight, efficient Language Model (136M parameters) specifically fine-tuned for **Reasoning tasks** and **Instruction Following**. It utilizes a unique Triple-Prompt architecture (Instruction-Thought-Output) to ensure high-quality, focused, and logical responses.
## Model Details
- **Developed by:** Raziel / TopAI
- **Model type:** Causal Language Model (Transformer)
- **Language(s):** English (Primary), Hebrew (Identity)
- **License:** Apache 2.0
- **Base Model:** Duchifat-2 (136M)
- **Training Technique:** SFT (Supervised Fine-Tuning) with Chain-of-Thought Alignment.
## Key Features
- **Triple-Prompt Architecture:** Designed to process an internal "Thought" block before generating the final output.
- **Efficient Reasoning:** Optimized for CPU and low-resource environments without sacrificing logical consistency.
- **Clean Output:** Significantly reduced hallucination and "word salad" compared to standard small models.
## Prompt Format
To get the best results, use the following structured prompt:
```text
### instruction:
{Your Question}
### thought:
{The logic or reasoning the model should follow}
### output:
```
### Usage Example
``` python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model_id = "razielAI/Duchifat-2-Instruct-Thinking" # Update with your exact HF path
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float32)
instruction = "Who are you?"
thought = "The user is asking for my identity. I should state I am Duchifat-2 developed by TopAI."
prompt = f"### instruction:\n{instruction}\n\n### thought:\n{thought}\n\n### output:\n"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100, temperature=0.1)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
### Performance & Limitations
Duchifat-2-Instruct-Thinking is a **Small Language Model (SLM). While it excels at structured tasks and guided reasoning:
- It may require a guided thought block for highly complex logic.
- Best used with low temperature (0.1 - 0.3) for factual consistency.
### Citation
If you use this model in your research or project, please cite:
```
Plaintext
@misc{duchifat2thinking2026,
author = {Raziel, TopAI},
title = {Duchifat-2-Thinking: A Lightweight Reasoning Model},
year = {2026},
publisher = {Hugging Face},
journal = {Hugging Face Model Hub}
}
```