Tandogan commited on
Commit
fe6a9f2
·
verified ·
1 Parent(s): a665dc6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -3,7 +3,7 @@ library_name: transformers
3
  tags: []
4
  ---
5
 
6
- # MNLP M3 DPO Model — Qwen3-0.6B Fine-Tuned with Direct Preference Optimization
7
 
8
  This repository contains a Direct Preference Optimization (DPO) model built on top of the base model [`Qwen/Qwen3-0.6B-Base`](https://huggingface.co/Qwen/Qwen3-0.6B-Base), as part of the MNLP M3 project. The model is fine-tuned using a high-quality preference dataset to better align responses with human preferences.
9
 
 
3
  tags: []
4
  ---
5
 
6
+ # MNLP M3 DPO Model — Qwen3-0.6B-Base Fine-Tuned with Direct Preference Optimization
7
 
8
  This repository contains a Direct Preference Optimization (DPO) model built on top of the base model [`Qwen/Qwen3-0.6B-Base`](https://huggingface.co/Qwen/Qwen3-0.6B-Base), as part of the MNLP M3 project. The model is fine-tuned using a high-quality preference dataset to better align responses with human preferences.
9