Upload folder using huggingface_hub
Browse files- README.md +56 -0
- adapter_config.json +34 -0
- adapter_model.safetensors +3 -0
- upload_script.sh +56 -0
README.md
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: cc-by-nc-4.0
|
| 3 |
+
base_model: kaonai/kaon-c-mistral-12b-v0.1-dpo-v0.1
|
| 4 |
+
tags:
|
| 5 |
+
- peft
|
| 6 |
+
- lora
|
| 7 |
+
- dpo
|
| 8 |
+
- role-playing
|
| 9 |
+
- chinese
|
| 10 |
+
library_name: peft
|
| 11 |
+
pipeline_tag: text-generation
|
| 12 |
+
---
|
| 13 |
+
|
| 14 |
+
# Kaon-W: GPT-4 Optimized DPO LoRA
|
| 15 |
+
|
| 16 |
+
基于 进行GPT-4数据优化的DPO LoRA适配器。
|
| 17 |
+
|
| 18 |
+
## 模型信息
|
| 19 |
+
|
| 20 |
+
- **基础模型**: kaonai/kaon-c-mistral-12b-v0.1-dpo-v0.1
|
| 21 |
+
- **训练方法**: DPO (Direct Preference Optimization)
|
| 22 |
+
- **适配器类型**: LoRA (Low-Rank Adaptation)
|
| 23 |
+
- **数据优化**: GPT-4全面优化的对话数据
|
| 24 |
+
- **训练硬件**: 8x H200 GPU
|
| 25 |
+
- **训练时长**: 5.3小时 (968步, 2轮)
|
| 26 |
+
|
| 27 |
+
## 训练配置
|
| 28 |
+
|
| 29 |
+
- **LoRA Rank**: 64
|
| 30 |
+
- **LoRA Alpha**: 64
|
| 31 |
+
- **学习率**: 2e-6
|
| 32 |
+
- **Beta**: 0.12
|
| 33 |
+
- **批次大小**: 2 (per device)
|
| 34 |
+
- **梯度累积**: 4步
|
| 35 |
+
|
| 36 |
+
## 使用方法
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
## 训练数据
|
| 41 |
+
|
| 42 |
+
使用GPT-4优化的DPO数据集:
|
| 43 |
+
- 主数据集: 3,500条优化样本
|
| 44 |
+
- NSFW修复: 150条优化样本
|
| 45 |
+
- 防重复: 297条优化样本
|
| 46 |
+
- 总计: 3,947条高质量训练样本
|
| 47 |
+
|
| 48 |
+
## 性能指标
|
| 49 |
+
|
| 50 |
+
- **最终准确率**: 96.15%
|
| 51 |
+
- **奖励边距**: 20.14+
|
| 52 |
+
- **训练损失**: 1.27 (完美收敛)
|
| 53 |
+
|
| 54 |
+
## 许可证
|
| 55 |
+
|
| 56 |
+
本模型采用 CC-BY-NC-4.0 许可证,仅供非商业用途。
|
adapter_config.json
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alpha_pattern": {},
|
| 3 |
+
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "/mnt/local/zeyu/models/kaon-c-mistral-12b-v0.1-dpo-v0.1",
|
| 5 |
+
"bias": "none",
|
| 6 |
+
"fan_in_fan_out": false,
|
| 7 |
+
"inference_mode": true,
|
| 8 |
+
"init_lora_weights": true,
|
| 9 |
+
"layer_replication": null,
|
| 10 |
+
"layers_pattern": null,
|
| 11 |
+
"layers_to_transform": null,
|
| 12 |
+
"loftq_config": {},
|
| 13 |
+
"lora_alpha": 64,
|
| 14 |
+
"lora_dropout": 0.05,
|
| 15 |
+
"megatron_config": null,
|
| 16 |
+
"megatron_core": "megatron.core",
|
| 17 |
+
"modules_to_save": [],
|
| 18 |
+
"peft_type": "LORA",
|
| 19 |
+
"r": 64,
|
| 20 |
+
"rank_pattern": {},
|
| 21 |
+
"revision": null,
|
| 22 |
+
"target_modules": [
|
| 23 |
+
"gate_proj",
|
| 24 |
+
"k_proj",
|
| 25 |
+
"down_proj",
|
| 26 |
+
"o_proj",
|
| 27 |
+
"up_proj",
|
| 28 |
+
"q_proj",
|
| 29 |
+
"v_proj"
|
| 30 |
+
],
|
| 31 |
+
"task_type": "CAUSAL_LM",
|
| 32 |
+
"use_dora": false,
|
| 33 |
+
"use_rslora": false
|
| 34 |
+
}
|
adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6118dc0c1fe4fa1f135519b2fcab1c4612f6c256a9554f94c30050f74c6a7ab3
|
| 3 |
+
size 912336848
|
upload_script.sh
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/bin/bash
|
| 2 |
+
|
| 3 |
+
# HuggingFace模型上传脚本
|
| 4 |
+
# ================================
|
| 5 |
+
|
| 6 |
+
echo "🚀 HuggingFace模型上传脚本"
|
| 7 |
+
echo "=============================="
|
| 8 |
+
|
| 9 |
+
# 检查是否已登录
|
| 10 |
+
echo "📋 检查HuggingFace登录状态..."
|
| 11 |
+
if ! ~/.local/bin/hf whoami &>/dev/null; then
|
| 12 |
+
echo "❌ 未登录HuggingFace,请先登录:"
|
| 13 |
+
echo " ~/.local/bin/hf login"
|
| 14 |
+
echo " 然后输入你的Access Token"
|
| 15 |
+
exit 1
|
| 16 |
+
fi
|
| 17 |
+
|
| 18 |
+
echo "✅ 已登录HuggingFace"
|
| 19 |
+
|
| 20 |
+
# 获取用户输入
|
| 21 |
+
read -p "请输入模型名称 (例如: 你的用户名/kaon-w): " REPO_NAME
|
| 22 |
+
|
| 23 |
+
if [ -z "$REPO_NAME" ]; then
|
| 24 |
+
echo "❌ 模型名称不能为空"
|
| 25 |
+
exit 1
|
| 26 |
+
fi
|
| 27 |
+
|
| 28 |
+
echo "📂 准备上传模型: $REPO_NAME"
|
| 29 |
+
echo "📁 文件列表:"
|
| 30 |
+
ls -la /tmp/kaon-w-upload/
|
| 31 |
+
|
| 32 |
+
echo ""
|
| 33 |
+
read -p "确认上传? (y/N): " CONFIRM
|
| 34 |
+
|
| 35 |
+
if [[ $CONFIRM != [yY] ]]; then
|
| 36 |
+
echo "❌ 取消上传"
|
| 37 |
+
exit 0
|
| 38 |
+
fi
|
| 39 |
+
|
| 40 |
+
echo "⬆️ 开始上传..."
|
| 41 |
+
cd /tmp/kaon-w-upload/
|
| 42 |
+
|
| 43 |
+
# 上传模型
|
| 44 |
+
~/.local/bin/hf upload $REPO_NAME . --repo-type model
|
| 45 |
+
|
| 46 |
+
if [ $? -eq 0 ]; then
|
| 47 |
+
echo ""
|
| 48 |
+
echo "🎉 上传成功!"
|
| 49 |
+
echo "🔗 模型地址: https://huggingface.co/$REPO_NAME"
|
| 50 |
+
echo ""
|
| 51 |
+
echo "📋 使用方法:"
|
| 52 |
+
echo "from peft import PeftModel"
|
| 53 |
+
echo "model = PeftModel.from_pretrained(base_model, '$REPO_NAME')"
|
| 54 |
+
else
|
| 55 |
+
echo "❌ 上传失败,请检查网络连接和权限"
|
| 56 |
+
fi
|