CSshihao commited on
Commit
4540e13
·
verified ·
1 Parent(s): 5d0d5d5

Upload folder using huggingface_hub

Browse files
checkpoint-100/config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ImgGen_HF"
4
+ ],
5
+ "bnb_4bit_quant_type": "nf4",
6
+ "cls_token_num": 1,
7
+ "codebook_embed_dim": 8,
8
+ "codebook_size": 16384,
9
+ "delay_load_text_encoder": false,
10
+ "downsample_size": 16,
11
+ "drop_path_rate": 0.0,
12
+ "dropout_p": 0.0,
13
+ "gpt_model": "GPT-L",
14
+ "gpt_type": "c2i",
15
+ "image_size": 384,
16
+ "load_in_4bit": false,
17
+ "load_in_8bit": false,
18
+ "lora_alpha": 32,
19
+ "lora_dropout": 0.05,
20
+ "lora_modules_to_save": null,
21
+ "lora_r": 16,
22
+ "lora_target_modules": null,
23
+ "lora_task_type": "CAUSAL_LM",
24
+ "model_name_or_path": "custom/noneexists",
25
+ "model_revision": "main",
26
+ "model_type": "ImgGen_HF",
27
+ "text_encoder": "google/flan-t5-xl",
28
+ "text_encoder_ckpt": "",
29
+ "text_feature_max_len": 120,
30
+ "token_dropout_p": 0.0,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.53.0",
33
+ "trust_remote_code": false,
34
+ "use_bnb_nested_quant": false,
35
+ "use_dora": false,
36
+ "use_peft": false,
37
+ "use_rslora": false,
38
+ "vq_model": "VQ-16"
39
+ }
checkpoint-100/generation_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "transformers_version": "4.53.0"
4
+ }
checkpoint-100/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fefc6d1536be31fa0cc4b1289260e4d8b364013cbc38041a2f72223a60dcafce
3
+ size 3319006336
checkpoint-100/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-100/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5255537192c9212dcda164244b8ef373609734022b217b2fb3dd06cd274debb
3
+ size 7185