File size: 2,539 Bytes
d59f4a5
c0799bc
 
d59f4a5
c0799bc
 
 
 
 
 
d59f4a5
c0799bc
 
 
 
 
d59f4a5
 
 
 
c0799bc
 
 
 
 
 
077ae6f
c0799bc
 
 
 
077ae6f
c0799bc
 
 
 
 
 
 
 
d59f4a5
 
 
 
c0799bc
 
d59f4a5
c0799bc
 
5b30bec
c0799bc
 
 
077ae6f
c0799bc
5b30bec
c0799bc
d59f4a5
c0799bc
d59f4a5
c0799bc
5b30bec
c0799bc
 
 
 
 
 
 
 
d59f4a5
c0799bc
d59f4a5
c0799bc
d59f4a5
 
 
 
 
 
 
 
c0799bc
d59f4a5
c0799bc
d59f4a5
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
---
base_model:
- Qwen/Qwen2.5-Coder-32B-Instruct
library_name: peft
license: cc-by-nc-4.0
datasets:
- Jessylg27/DeepThink-Code-Lite
language:
- en
- fr
tags:
- code
- logic
- reasoning
- qwen2.5
- unsloth
- sft
- trl
---

# Specialized Coding Logic LLM (32B)

This model is a specialized fine-tuned version of [Qwen/Qwen2.5-Coder-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct).  
It has been optimized to enhance **logical reasoning** and **code generation capabilities**.

## ๐Ÿง  Model Description

**Specialized Coding Logic LLM** builds upon the powerful Qwen 2.5 Coder architecture (32B parameters). It has been fine-tuned using the **DeepThink-Code-Lite** dataset to improve its ability to:
- Solve complex algorithmic problems.
- Follow multi-step logical instructions.
- Generate cleaner and more optimized code.

## ๐Ÿ“Š Dataset

This model was trained on the custom dataset:  
๐Ÿ‘‰ **[Jessylg27/DeepThink-Code-Lite](https://huggingface.co/datasets/Jessylg27/DeepThink-Code-Lite)**

## ๐Ÿš€ Quick Start

You can use this model directly with the Hugging Face `pipeline`.

```python
from transformers import pipeline

# Define the model ID
model_id = "Jessylg27/specialized-coding-logic-llm"

# Initialize the pipeline
generator = pipeline("text-generation", model=model_id, device_map="auto")

# Prompt the model
question = "Write a Python function to solve the Traveling Salesman Problem using dynamic programming."
output = generator([{"role": "user", "content": question}], max_new_tokens=512, return_full_text=False)[0]

print(output["generated_text"])

```

## ๐Ÿ› ๏ธ Training procedure

This model was trained with **SFT (Supervised Fine-Tuning)** using the [TRL library](https://github.com/huggingface/trl) and [Unsloth](https://github.com/unslothai/unsloth) for efficient training.

### Framework versions

* **PEFT:** 0.18.1
* **TRL:** 0.24.0
* **Transformers:** 4.57.3
* **Pytorch:** 2.8.0+cu128
* **Datasets:** 4.3.0
* **Tokenizers:** 0.22.2

## ๐Ÿ“œ Citations

If you use this model or the TRL library, please cite:

```bibtex
@misc{vonwerra2022trl,
	title        = {{TRL: Transformer Reinforcement Learning}},
	author       = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
	year         = 2020,
	journal      = {GitHub repository},
	publisher    = {GitHub},
	howpublished = {\url{[https://github.com/huggingface/trl](https://github.com/huggingface/trl)}}
}

```