thanhdathoang commited on
Commit
a44db07
·
verified ·
1 Parent(s): e7f6454

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,15 +1,3 @@
1
- ---
2
- library_name: transformers
3
- base_model: Qwen/Qwen3-4B-Instruct-2507
4
- tags:
5
- - alignment-handbook
6
- - generated_from_trainer
7
- datasets:
8
- - griffith-bigdata/sft_text2sql
9
- model-index:
10
- - name: Qwen3-4B-SQL-Writer
11
- results: []
12
- ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
1
 
2
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
3
  should probably proofread and complete it, then remove this comment. -->
runs/Oct16_21-01-10_dgxlogin/events.out.tfevents.1760612475.dgxlogin.2188814.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc5d3e04f35f15dded30468a855a5fd5e29ad56b1455e725e5f2b87b1548569c
3
+ size 31123
tokenizer_config.json CHANGED
@@ -231,7 +231,7 @@
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
- "model_max_length": 1010000,
235
  "pad_token": "<|endoftext|>",
236
  "split_special_tokens": false,
237
  "tokenizer_class": "Qwen2Tokenizer",
 
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
+ "model_max_length": 12000,
235
  "pad_token": "<|endoftext|>",
236
  "split_special_tokens": false,
237
  "tokenizer_class": "Qwen2Tokenizer",