Cooolder commited on
Commit
6b74982
·
verified ·
1 Parent(s): b5b668e

Add model card

Browse files
Files changed (1) hide show
  1. README.md +75 -0
README.md ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: unsloth/Qwen3-4B-Instruct-2507
4
+ tags:
5
+ - qwen3
6
+ - performance-prediction
7
+ - scope
8
+ - unsloth
9
+ - fine-tuned
10
+ language:
11
+ - en
12
+ - zh
13
+ ---
14
+
15
+ # SCOPE-Direct-v2
16
+
17
+ This is a fine-tuned version of [unsloth/Qwen3-4B-Instruct-2507](https://huggingface.co/unsloth/Qwen3-4B-Instruct-2507) trained on the [SCOPE-sft-direct-data](https://huggingface.co/datasets/Cooolder/SCOPE-sft-direct-data) dataset.
18
+
19
+ ## Model Details
20
+
21
+ - **Base Model**: Qwen3-4B-Instruct-2507
22
+ - **Training Dataset**: SCOPE-sft-direct-data
23
+ - **Training Method**: LoRA fine-tuning with Unsloth
24
+ - **Merged**: Yes (16-bit merged model)
25
+
26
+ ## Training Configuration
27
+
28
+ - Max Sequence Length: 4096
29
+ - LoRA Rank: 32
30
+ - Batch Size: 4
31
+ - Gradient Accumulation Steps: 4
32
+ - Learning Rate: 2e-5
33
+ - Epochs: 1
34
+ - Optimizer: adamw_8bit
35
+ - LR Scheduler: cosine
36
+
37
+ ## Usage
38
+
39
+ ```python
40
+ from transformers import AutoModelForCausalLM, AutoTokenizer
41
+
42
+ model = AutoModelForCausalLM.from_pretrained(
43
+ "Cooolder/SCOPE-Direct-v2",
44
+ torch_dtype="auto",
45
+ device_map="auto"
46
+ )
47
+ tokenizer = AutoTokenizer.from_pretrained("Cooolder/SCOPE-Direct-v2")
48
+
49
+ # Generate text
50
+ messages = [
51
+ {"role": "user", "content": "Your prompt here"}
52
+ ]
53
+ text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
54
+ inputs = tokenizer(text, return_tensors="pt").to(model.device)
55
+ outputs = model.generate(**inputs, max_new_tokens=512)
56
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
57
+ ```
58
+
59
+ ## License
60
+
61
+ This model inherits the license from its base model (Apache 2.0).
62
+
63
+ ## Citation
64
+
65
+ If you use this model, please cite:
66
+
67
+ ```bibtex
68
+ @misc{scope-direct-v2,
69
+ title={SCOPE-Direct-v2: Fine-tuned Qwen3-4B for Performance Prediction},
70
+ author={Your Name},
71
+ year={2025},
72
+ publisher={HuggingFace},
73
+ howpublished={\url{https://huggingface.co/Cooolder/SCOPE-Direct-v2}}
74
+ }
75
+ ```