Hoffman37 commited on
Commit
7430cb8
·
verified ·
1 Parent(s): 366a1c5

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -1,17 +1,18 @@
1
  ---
2
- base_model: google/gemma-3-1b-pt
3
  library_name: transformers
4
  model_name: gemma-text-to-sql
5
  tags:
6
  - generated_from_trainer
7
  - trl
8
  - sft
 
9
  licence: license
10
  ---
11
 
12
  # Model Card for gemma-text-to-sql
13
 
14
- This model is a fine-tuned version of [google/gemma-3-1b-pt](https://huggingface.co/google/gemma-3-1b-pt).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -34,11 +35,11 @@ This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.15.2
38
- - Transformers: 4.53.0
39
- - Pytorch: 2.6.0+cu124
40
- - Datasets: 3.6.0
41
- - Tokenizers: 0.21.2
42
 
43
  ## Citations
44
 
@@ -49,7 +50,7 @@ Cite TRL as:
49
  ```bibtex
50
  @misc{vonwerra2022trl,
51
  title = {{TRL: Transformer Reinforcement Learning}},
52
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
53
  year = 2020,
54
  journal = {GitHub repository},
55
  publisher = {GitHub},
 
1
  ---
2
+ base_model: unsloth/gemma-3-270m-it-bnb-4bit
3
  library_name: transformers
4
  model_name: gemma-text-to-sql
5
  tags:
6
  - generated_from_trainer
7
  - trl
8
  - sft
9
+ - unsloth
10
  licence: license
11
  ---
12
 
13
  # Model Card for gemma-text-to-sql
14
 
15
+ This model is a fine-tuned version of [unsloth/gemma-3-270m-it-bnb-4bit](https://huggingface.co/unsloth/gemma-3-270m-it-bnb-4bit).
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
 
35
 
36
  ### Framework versions
37
 
38
+ - TRL: 0.22.2
39
+ - Transformers: 4.56.2
40
+ - Pytorch: 2.9.0+cu126
41
+ - Datasets: 4.3.0
42
+ - Tokenizers: 0.22.2
43
 
44
  ## Citations
45
 
 
50
  ```bibtex
51
  @misc{vonwerra2022trl,
52
  title = {{TRL: Transformer Reinforcement Learning}},
53
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
54
  year = 2020,
55
  journal = {GitHub repository},
56
  publisher = {GitHub},
adapter_config.json CHANGED
@@ -1,8 +1,16 @@
1
  {
 
2
  "alpha_pattern": {},
3
- "auto_mapping": null,
4
- "base_model_name_or_path": "google/gemma-3-1b-pt",
 
 
 
 
 
5
  "bias": "none",
 
 
6
  "eva_config": null,
7
  "exclude_modules": null,
8
  "fan_in_fan_out": false,
@@ -14,27 +22,29 @@
14
  "loftq_config": {},
15
  "lora_alpha": 16,
16
  "lora_bias": false,
17
- "lora_dropout": 0.05,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
- "modules_to_save": [
21
- "lm_head",
22
- "embed_tokens"
23
- ],
24
  "peft_type": "LORA",
 
 
25
  "r": 16,
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
- "k_proj",
30
  "v_proj",
31
- "down_proj",
32
  "q_proj",
33
  "up_proj",
34
- "o_proj",
35
- "gate_proj"
36
  ],
 
37
  "task_type": "CAUSAL_LM",
 
38
  "use_dora": false,
 
39
  "use_rslora": false
40
  }
 
1
  {
2
+ "alora_invocation_tokens": null,
3
  "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "Gemma3ForCausalLM",
7
+ "parent_library": "transformers.models.gemma3.modeling_gemma3",
8
+ "unsloth_fixed": true
9
+ },
10
+ "base_model_name_or_path": "unsloth/gemma-3-270m-it-bnb-4bit",
11
  "bias": "none",
12
+ "corda_config": null,
13
+ "ensure_weight_tying": false,
14
  "eva_config": null,
15
  "exclude_modules": null,
16
  "fan_in_fan_out": false,
 
22
  "loftq_config": {},
23
  "lora_alpha": 16,
24
  "lora_bias": false,
25
+ "lora_dropout": 0,
26
  "megatron_config": null,
27
  "megatron_core": "megatron.core",
28
+ "modules_to_save": null,
 
 
 
29
  "peft_type": "LORA",
30
+ "peft_version": "0.18.0",
31
+ "qalora_group_size": 16,
32
  "r": 16,
33
  "rank_pattern": {},
34
  "revision": null,
35
  "target_modules": [
36
+ "gate_proj",
37
  "v_proj",
38
+ "o_proj",
39
  "q_proj",
40
  "up_proj",
41
+ "down_proj",
42
+ "k_proj"
43
  ],
44
+ "target_parameters": null,
45
  "task_type": "CAUSAL_LM",
46
+ "trainable_token_indices": null,
47
  "use_dora": false,
48
+ "use_qalora": false,
49
  "use_rslora": false
50
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e6e331992a639f2f511b3f647ab638a2b639b6a51165ce5cf60efd8e2987fd6
3
- size 2468152240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77f4433025e1ea907f9452ef0d86b5111def381db80a76bc930a264613676256
3
+ size 15220968
chat_template.jinja CHANGED
@@ -42,6 +42,6 @@
42
  ' }}
43
  {%- endfor -%}
44
  {%- if add_generation_prompt -%}
45
- {{'<start_of_turn>model
46
- '}}
47
  {%- endif -%}
 
42
  ' }}
43
  {%- endfor -%}
44
  {%- if add_generation_prompt -%}
45
+ {{ '<start_of_turn>model
46
+ ' }}
47
  {%- endif -%}
special_tokens_map.json CHANGED
@@ -9,7 +9,7 @@
9
  },
10
  "eoi_token": "<end_of_image>",
11
  "eos_token": {
12
- "content": "<eos>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
 
9
  },
10
  "eoi_token": "<end_of_image>",
11
  "eos_token": {
12
+ "content": "<end_of_turn>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -51327,16 +51327,16 @@
51327
  "bos_token": "<bos>",
51328
  "clean_up_tokenization_spaces": false,
51329
  "eoi_token": "<end_of_image>",
51330
- "eos_token": "<eos>",
51331
  "extra_special_tokens": {
51332
  "boi_token": "<start_of_image>",
51333
  "eoi_token": "<end_of_image>",
51334
  "image_token": "<image_soft_token>"
51335
  },
51336
  "image_token": "<image_soft_token>",
51337
- "model_max_length": 1000000000000000019884624838656,
51338
  "pad_token": "<pad>",
51339
- "processor_class": "Gemma3Processor",
51340
  "sp_model_kwargs": null,
51341
  "spaces_between_special_tokens": false,
51342
  "tokenizer_class": "GemmaTokenizer",
 
51327
  "bos_token": "<bos>",
51328
  "clean_up_tokenization_spaces": false,
51329
  "eoi_token": "<end_of_image>",
51330
+ "eos_token": "<end_of_turn>",
51331
  "extra_special_tokens": {
51332
  "boi_token": "<start_of_image>",
51333
  "eoi_token": "<end_of_image>",
51334
  "image_token": "<image_soft_token>"
51335
  },
51336
  "image_token": "<image_soft_token>",
51337
+ "model_max_length": 32768,
51338
  "pad_token": "<pad>",
51339
+ "padding_side": "right",
51340
  "sp_model_kwargs": null,
51341
  "spaces_between_special_tokens": false,
51342
  "tokenizer_class": "GemmaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ca4fac3db2cdc011b13d452d378ffccd4c4524ca8a7b76d76e625f7a41651cb
3
- size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:690412784d2a73305c384c52dac6539174a0ac481e01f127f8bcfc75d063c285
3
+ size 6289