Upload folder using huggingface_hub
Browse files
README.md
CHANGED
|
@@ -1,15 +1,3 @@
|
|
| 1 |
-
---
|
| 2 |
-
library_name: transformers
|
| 3 |
-
base_model: Qwen/Qwen3-4B-Instruct-2507
|
| 4 |
-
tags:
|
| 5 |
-
- alignment-handbook
|
| 6 |
-
- generated_from_trainer
|
| 7 |
-
datasets:
|
| 8 |
-
- griffith-bigdata/sft_text2sql
|
| 9 |
-
model-index:
|
| 10 |
-
- name: Qwen3-4B-SQL-Writer
|
| 11 |
-
results: []
|
| 12 |
-
---
|
| 13 |
|
| 14 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 15 |
should probably proofread and complete it, then remove this comment. -->
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
|
| 2 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 3 |
should probably proofread and complete it, then remove this comment. -->
|
runs/Oct16_21-01-10_dgxlogin/events.out.tfevents.1760612475.dgxlogin.2188814.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dc5d3e04f35f15dded30468a855a5fd5e29ad56b1455e725e5f2b87b1548569c
|
| 3 |
+
size 31123
|
tokenizer_config.json
CHANGED
|
@@ -231,7 +231,7 @@
|
|
| 231 |
"eos_token": "<|im_end|>",
|
| 232 |
"errors": "replace",
|
| 233 |
"extra_special_tokens": {},
|
| 234 |
-
"model_max_length":
|
| 235 |
"pad_token": "<|endoftext|>",
|
| 236 |
"split_special_tokens": false,
|
| 237 |
"tokenizer_class": "Qwen2Tokenizer",
|
|
|
|
| 231 |
"eos_token": "<|im_end|>",
|
| 232 |
"errors": "replace",
|
| 233 |
"extra_special_tokens": {},
|
| 234 |
+
"model_max_length": 12000,
|
| 235 |
"pad_token": "<|endoftext|>",
|
| 236 |
"split_special_tokens": false,
|
| 237 |
"tokenizer_class": "Qwen2Tokenizer",
|