Kaggle_2 / README.md
RayTsai's picture
Upload folder using huggingface_hub
aa0d34c verified
---
base_model: Qwen/Qwen2.5-7B-Instruct
tags:
- text-generation
- chinese
- reasoning
- multiple-choice
- lora
- peft
language:
- zh
- en
library_name: peft
license: apache-2.0
---
# Chinese LLM MCQ Model - KAGGLE #2
這是NYCU深度學習課程KAGGLE #2的模型,使用Qwen2.5-7B-Instruct進行微調,加入了推理鏈能力。
## 模型資訊
- **基礎模型**: Qwen/Qwen2.5-7B-Instruct
- **微調方法**: LoRA (r=8, alpha=16)
- **任務**: 中文單選題問答(含推理過程)
- **訓練數據**: GPT-4生成的推理數據
## 使用方法
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
# 載入基礎模型
base_model = AutoModelForCausalLM.from_pretrained(
"Qwen/Qwen2.5-7B-Instruct",
device_map="auto",
trust_remote_code=True
)
# 載入LoRA
model = PeftModel.from_pretrained(base_model, "RayTsai/Kaggle_2")
# 載入tokenizer
tokenizer = AutoTokenizer.from_pretrained("RayTsai/Kaggle_2")
```
## 作者
- Ray Tsai (110651053)
- NYCU 深度學習課程 2025