abhishekDS commited on
Commit
b750b7b
·
1 Parent(s): 4335405

Delete my_autotrain_llm

Browse files
my_autotrain_llm/README.md DELETED
@@ -1,9 +0,0 @@
1
- ---
2
- tags:
3
- - autotrain
4
- - text-generation
5
- widget:
6
- - text: "I love AutoTrain because "
7
- ---
8
-
9
- # Model Trained Using AutoTrain
 
 
 
 
 
 
 
 
 
 
my_autotrain_llm/adapter_config.json DELETED
@@ -1,21 +0,0 @@
1
- {
2
- "auto_mapping": null,
3
- "base_model_name_or_path": "abhishek/llama-2-7b-hf-small-shards",
4
- "bias": "none",
5
- "fan_in_fan_out": false,
6
- "inference_mode": true,
7
- "init_lora_weights": true,
8
- "layers_pattern": null,
9
- "layers_to_transform": null,
10
- "lora_alpha": 32,
11
- "lora_dropout": 0.05,
12
- "modules_to_save": null,
13
- "peft_type": "LORA",
14
- "r": 16,
15
- "revision": null,
16
- "target_modules": [
17
- "q_proj",
18
- "v_proj"
19
- ],
20
- "task_type": "CAUSAL_LM"
21
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my_autotrain_llm/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e435ec8734dfb0f8e0cc1d1a7794f6c501cd4cfc08362379fdb5924c067b1df
3
- size 33600461
 
 
 
 
my_autotrain_llm/added_tokens.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "<pad>": 32000
3
- }
 
 
 
 
my_autotrain_llm/checkpoint-6/README.md DELETED
@@ -1,34 +0,0 @@
1
- ---
2
- library_name: peft
3
- ---
4
- ## Training procedure
5
-
6
-
7
- The following `bitsandbytes` quantization config was used during training:
8
- - quant_method: bitsandbytes
9
- - load_in_8bit: False
10
- - load_in_4bit: True
11
- - llm_int8_threshold: 6.0
12
- - llm_int8_skip_modules: None
13
- - llm_int8_enable_fp32_cpu_offload: False
14
- - llm_int8_has_fp16_weight: False
15
- - bnb_4bit_quant_type: nf4
16
- - bnb_4bit_use_double_quant: False
17
- - bnb_4bit_compute_dtype: float16
18
-
19
- The following `bitsandbytes` quantization config was used during training:
20
- - quant_method: bitsandbytes
21
- - load_in_8bit: False
22
- - load_in_4bit: True
23
- - llm_int8_threshold: 6.0
24
- - llm_int8_skip_modules: None
25
- - llm_int8_enable_fp32_cpu_offload: False
26
- - llm_int8_has_fp16_weight: False
27
- - bnb_4bit_quant_type: nf4
28
- - bnb_4bit_use_double_quant: False
29
- - bnb_4bit_compute_dtype: float16
30
- ### Framework versions
31
-
32
- - PEFT 0.6.0.dev0
33
-
34
- - PEFT 0.6.0.dev0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my_autotrain_llm/checkpoint-6/adapter_config.json DELETED
@@ -1,21 +0,0 @@
1
- {
2
- "auto_mapping": null,
3
- "base_model_name_or_path": "abhishek/llama-2-7b-hf-small-shards",
4
- "bias": "none",
5
- "fan_in_fan_out": false,
6
- "inference_mode": true,
7
- "init_lora_weights": true,
8
- "layers_pattern": null,
9
- "layers_to_transform": null,
10
- "lora_alpha": 32,
11
- "lora_dropout": 0.05,
12
- "modules_to_save": null,
13
- "peft_type": "LORA",
14
- "r": 16,
15
- "revision": null,
16
- "target_modules": [
17
- "q_proj",
18
- "v_proj"
19
- ],
20
- "task_type": "CAUSAL_LM"
21
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my_autotrain_llm/checkpoint-6/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e435ec8734dfb0f8e0cc1d1a7794f6c501cd4cfc08362379fdb5924c067b1df
3
- size 33600461
 
 
 
 
my_autotrain_llm/checkpoint-6/added_tokens.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "<pad>": 32000
3
- }
 
 
 
 
my_autotrain_llm/checkpoint-6/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:71f8e565ef2fc7d399f0a8b3543520d4f7d92ca4843cc0179789c8292572a64f
3
- size 67216581
 
 
 
 
my_autotrain_llm/checkpoint-6/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c152074a486243089e4fc0fdee0a373a30fb0e0a6e40eb5fd0d36fdafc97a155
3
- size 443
 
 
 
 
my_autotrain_llm/checkpoint-6/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fbe8f9e35f1fb5cc80e77f1b19604264b0887bfcf8a2d89c68829a3c8712aece
3
- size 14575
 
 
 
 
my_autotrain_llm/checkpoint-6/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:94b96e406e63b9a1f153cfaf08a989666c2cfa350c583c0f1bba816cc8eae242
3
- size 627
 
 
 
 
my_autotrain_llm/checkpoint-6/special_tokens_map.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "<unk>",
17
- "unk_token": {
18
- "content": "<unk>",
19
- "lstrip": false,
20
- "normalized": true,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my_autotrain_llm/checkpoint-6/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
my_autotrain_llm/checkpoint-6/tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
- size 499723
 
 
 
 
my_autotrain_llm/checkpoint-6/tokenizer_config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "bos_token": {
3
- "__type": "AddedToken",
4
- "content": "<s>",
5
- "lstrip": false,
6
- "normalized": true,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- "clean_up_tokenization_spaces": false,
11
- "eos_token": {
12
- "__type": "AddedToken",
13
- "content": "</s>",
14
- "lstrip": false,
15
- "normalized": true,
16
- "rstrip": false,
17
- "single_word": false
18
- },
19
- "legacy": false,
20
- "model_max_length": 1024,
21
- "pad_token": null,
22
- "sp_model_kwargs": {},
23
- "tokenizer_class": "LlamaTokenizer",
24
- "unk_token": {
25
- "__type": "AddedToken",
26
- "content": "<unk>",
27
- "lstrip": false,
28
- "normalized": true,
29
- "rstrip": false,
30
- "single_word": false
31
- },
32
- "use_default_system_prompt": true
33
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my_autotrain_llm/checkpoint-6/trainer_state.json DELETED
@@ -1,25 +0,0 @@
1
- {
2
- "best_metric": null,
3
- "best_model_checkpoint": null,
4
- "epoch": 1.0,
5
- "eval_steps": 500,
6
- "global_step": 6,
7
- "is_hyper_param_search": false,
8
- "is_local_process_zero": true,
9
- "is_world_process_zero": true,
10
- "log_history": [
11
- {
12
- "epoch": 0.67,
13
- "learning_rate": 8e-05,
14
- "loss": 1.0533,
15
- "step": 4
16
- }
17
- ],
18
- "logging_steps": 4,
19
- "max_steps": 6,
20
- "num_train_epochs": 1,
21
- "save_steps": 500,
22
- "total_flos": 498067090440192.0,
23
- "trial_name": null,
24
- "trial_params": null
25
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my_autotrain_llm/checkpoint-6/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c12502f15cf99be7482bb39ff8ffddb975d94cc27316ef6ff967d79cb2329b5c
3
- size 4027
 
 
 
 
my_autotrain_llm/runs/Aug29_11-57-24_0fbcaa219a51/events.out.tfevents.1693310244.0fbcaa219a51.2077.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cf29bbd14d428b59219ae919a150ef092ba3cd0df826dd5e147f79451dc96a3
3
- size 5002
 
 
 
 
my_autotrain_llm/special_tokens_map.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "<unk>",
17
- "unk_token": {
18
- "content": "<unk>",
19
- "lstrip": false,
20
- "normalized": true,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my_autotrain_llm/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
my_autotrain_llm/tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
- size 499723
 
 
 
 
my_autotrain_llm/tokenizer_config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "bos_token": {
3
- "__type": "AddedToken",
4
- "content": "<s>",
5
- "lstrip": false,
6
- "normalized": true,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- "clean_up_tokenization_spaces": false,
11
- "eos_token": {
12
- "__type": "AddedToken",
13
- "content": "</s>",
14
- "lstrip": false,
15
- "normalized": true,
16
- "rstrip": false,
17
- "single_word": false
18
- },
19
- "legacy": false,
20
- "model_max_length": 1024,
21
- "pad_token": null,
22
- "sp_model_kwargs": {},
23
- "tokenizer_class": "LlamaTokenizer",
24
- "unk_token": {
25
- "__type": "AddedToken",
26
- "content": "<unk>",
27
- "lstrip": false,
28
- "normalized": true,
29
- "rstrip": false,
30
- "single_word": false
31
- },
32
- "use_default_system_prompt": true
33
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
my_autotrain_llm/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c12502f15cf99be7482bb39ff8ffddb975d94cc27316ef6ff967d79cb2329b5c
3
- size 4027
 
 
 
 
my_autotrain_llm/training_params.json DELETED
@@ -1,40 +0,0 @@
1
- {
2
- "model": "abhishek/llama-2-7b-hf-small-shards",
3
- "data_path": "data/",
4
- "project_name": "my_autotrain_llm",
5
- "train_split": "train",
6
- "valid_split": null,
7
- "text_column": "text",
8
- "token": null,
9
- "lr": 0.0002,
10
- "epochs": 1,
11
- "batch_size": 1,
12
- "warmup_ratio": 0.1,
13
- "gradient_accumulation": 4,
14
- "optimizer": "adamw_torch",
15
- "scheduler": "linear",
16
- "weight_decay": 0.01,
17
- "max_grad_norm": 1.0,
18
- "seed": 42,
19
- "add_eos_token": false,
20
- "block_size": 1024,
21
- "use_peft": true,
22
- "lora_r": 16,
23
- "lora_alpha": 32,
24
- "lora_dropout": 0.05,
25
- "logging_steps": -1,
26
- "evaluation_strategy": "epoch",
27
- "save_total_limit": 1,
28
- "save_strategy": "epoch",
29
- "auto_find_batch_size": false,
30
- "fp16": true,
31
- "push_to_hub": false,
32
- "use_int8": false,
33
- "model_max_length": 1024,
34
- "repo_id": null,
35
- "use_int4": true,
36
- "trainer": "default",
37
- "target_modules": null,
38
- "merge_adapter": false,
39
- "username": null
40
- }