--- base_model: Qwen/Qwen3-4B-Instruct-2507 language: - en license: apache-2.0 library_name: peft pipeline_tag: text-generation tags: - lora - qwen - unsloth - structeval --- # exp_camelcase **Model ID**: `ekunish/exp_camelcase` exp008a + camelCase augmented data (21K + 1.6K camelCase conversion variants) ## Training Configuration | Parameter | Value | |-----------|-------| | Base model | `Qwen/Qwen3-4B-Instruct-2507` | | Method | QLoRA (4-bit) | | Max sequence length | 512 | | Epochs | 1 | | Learning rate | 1e-06 | | LoRA r | 64 | | LoRA alpha | 128 | | Batch size | 2 × 8 = 16 | ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel import torch base = "Qwen/Qwen3-4B-Instruct-2507" adapter = "ekunish/exp_camelcase" tokenizer = AutoTokenizer.from_pretrained(base) model = AutoModelForCausalLM.from_pretrained( base, torch_dtype=torch.float16, device_map="auto", ) model = PeftModel.from_pretrained(model, adapter) ``` ## Training Data - Dataset: `data/sft_u10bei_camelcase` - License: CC-BY-4.0 (where applicable) ## Sources & License - **Training Data**: u-10bei/structured_data_with_cot_dataset_512_v2, daichira/structured-3k-mix-sft, etc. - **Dataset License**: Creative Commons Attribution (CC-BY-4.0) - **Compliance**: Users must comply with both the dataset's attribution requirements and the base model's original terms of use. ## Competition 松尾研LLMコミュニティ 2025年度講座 メインコンペ (StructEval-T)