bralynn commited on
Commit
6a04447
·
verified ·
1 Parent(s): eff1ed5

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,6 +1,14 @@
1
  ---
2
- base_model: bralynn/try
3
  library_name: peft
 
 
 
 
 
 
 
 
4
  ---
5
 
6
  # Model Card for Model ID
@@ -199,4 +207,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
199
  [More Information Needed]
200
  ### Framework versions
201
 
202
- - PEFT 0.15.2
 
1
  ---
2
+ base_model: bralynn/keyframerdt
3
  library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:bralynn/keyframerdt
7
+ - dpo
8
+ - lora
9
+ - transformers
10
+ - trl
11
+ - unsloth
12
  ---
13
 
14
  # Model Card for Model ID
 
207
  [More Information Needed]
208
  ### Framework versions
209
 
210
+ - PEFT 0.16.0
adapter_config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "bralynn/keyframerdt",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 256,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 128,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "up_proj",
29
+ "q_proj",
30
+ "down_proj",
31
+ "o_proj",
32
+ "k_proj",
33
+ "v_proj",
34
+ "gate_proj"
35
+ ],
36
+ "task_type": "CAUSAL_LM",
37
+ "trainable_token_indices": null,
38
+ "use_dora": false,
39
+ "use_qalora": false,
40
+ "use_rslora": false
41
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ad6a532eee3e2e86e70b5a70149f77f9b9d48250cbddd2783de8d234cc65553
3
+ size 389074856
config.json CHANGED
@@ -34,8 +34,8 @@
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": true,
36
  "torch_dtype": "float16",
37
- "transformers_version": "4.52.4",
38
- "unsloth_version": "2025.6.9",
39
  "use_cache": true,
40
  "vocab_size": 128256
41
  }
 
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": true,
36
  "torch_dtype": "float16",
37
+ "transformers_version": "4.53.2",
38
+ "unsloth_version": "2025.7.8",
39
  "use_cache": true,
40
  "vocab_size": 128256
41
  }
generation_config.json CHANGED
@@ -1,14 +1,12 @@
1
  {
 
2
  "bos_token_id": 128000,
3
- "do_sample": true,
4
  "eos_token_id": [
5
  128001,
6
  128008,
7
  128009
8
  ],
9
  "max_length": 131072,
10
- "pad_token_id": 128004,
11
- "temperature": 0.6,
12
- "top_p": 0.9,
13
- "transformers_version": "4.52.4"
14
  }
 
1
  {
2
+ "_from_model_config": true,
3
  "bos_token_id": 128000,
 
4
  "eos_token_id": [
5
  128001,
6
  128008,
7
  128009
8
  ],
9
  "max_length": 131072,
10
+ "pad_token_id": 0,
11
+ "transformers_version": "4.53.2"
 
 
12
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:305cafe69d3d3613dcd2379e3ef1173cb715f1232741e8dd92dc474e5cfc421e
3
  size 4965798912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fb4eb0e1b8f82d7ed313b12562271e862395b96d9d87c0a233915f5db90298e
3
  size 4965798912
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d825f7c2964cf2c92573f6e347fd7cab0a9371276a68ba3205dd4482dbddcae1
3
  size 1459729880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec360a1f8b82f99d81701ccc5279cdef5dd8ab0c783ed16f3d722c7b0bb933ce
3
  size 1459729880
model.safetensors.index.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
  "metadata": {
 
3
  "total_size": 6425499648
4
  },
5
  "weight_map": {
 
1
  {
2
  "metadata": {
3
+ "total_parameters": 3212749824,
4
  "total_size": 6425499648
5
  },
6
  "weight_map": {
tokenizer_config.json CHANGED
@@ -2063,5 +2063,5 @@
2063
  "pad_token": "<|finetune_right_pad_id|>",
2064
  "pad_token_type_id": 0,
2065
  "padding_side": "right",
2066
- "tokenizer_class": "PreTrainedTokenizer"
2067
  }
 
2063
  "pad_token": "<|finetune_right_pad_id|>",
2064
  "pad_token_type_id": 0,
2065
  "padding_side": "right",
2066
+ "tokenizer_class": "PreTrainedTokenizerFast"
2067
  }