jeisonvendetta commited on
Commit
5c2e9e4
·
verified ·
1 Parent(s): 73dbc41

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: google/gemma-3-1b-it
3
  library_name: transformers
4
  model_name: blockvault-gemma3-1b
5
  tags:
@@ -11,7 +11,7 @@ licence: license
11
 
12
  # Model Card for blockvault-gemma3-1b
13
 
14
- This model is a fine-tuned version of [google/gemma-3-1b-it](https://huggingface.co/google/gemma-3-1b-it).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
1
  ---
2
+ base_model: google/gemma-3-1b-pt
3
  library_name: transformers
4
  model_name: blockvault-gemma3-1b
5
  tags:
 
11
 
12
  # Model Card for blockvault-gemma3-1b
13
 
14
+ This model is a fine-tuned version of [google/gemma-3-1b-pt](https://huggingface.co/google/gemma-3-1b-pt).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "google/gemma-3-1b-it",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
@@ -26,13 +26,13 @@
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
- "q_proj",
30
- "up_proj",
31
  "gate_proj",
 
32
  "down_proj",
33
- "v_proj",
34
- "o_proj",
35
- "k_proj"
36
  ],
37
  "task_type": "CAUSAL_LM",
38
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "google/gemma-3-1b-pt",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "o_proj",
30
+ "v_proj",
31
  "gate_proj",
32
+ "q_proj",
33
  "down_proj",
34
+ "k_proj",
35
+ "up_proj"
 
36
  ],
37
  "task_type": "CAUSAL_LM",
38
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:267f784f91ea36d099b550661d77079b83c4be3d02eacf352984234850084c8c
3
  size 3676207496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f2c9f598d923a4d8850c0b07624caec15614b88d241c32b9ae39c53fc12017a
3
  size 3676207496
runs/Feb01_17-52-24_5c3c04861905/events.out.tfevents.1769968354.5c3c04861905.3095.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df4dffeedb30aa54606005ab80ba7f1d4d8cd20dc8540c2adb1a31c7bdaa2ebf
3
+ size 7386
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d52cd7e78cd9ab739858d58809a49a5fac8631d5f47c85facc4ee1a7398ae15
3
  size 6289
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f512d4141698997d0deb1bd5624bb47792012c7dc26501b84de0e5b52208cfc3
3
  size 6289