ChicFusion commited on
Commit
e364998
·
verified ·
1 Parent(s): ebbc9e5
config.json CHANGED
@@ -1,9 +1,10 @@
1
  {
2
- "_name_or_path": "bigband/ToweringDumuzi",
3
  "architectures": [
4
  "ParlerTTSForConditionalGeneration"
5
  ],
6
  "audio_encoder": {
 
7
  "_name_or_path": "parler-tts/dac_44khZ_8kbps",
8
  "add_cross_attention": false,
9
  "architectures": [
@@ -41,7 +42,7 @@
41
  "max_length": 20,
42
  "min_length": 0,
43
  "model_bitrate": 8,
44
- "model_type": "dac",
45
  "no_repeat_ngram_size": 0,
46
  "num_beam_groups": 1,
47
  "num_beams": 1,
@@ -75,6 +76,7 @@
75
  "use_bfloat16": false
76
  },
77
  "decoder": {
 
78
  "_name_or_path": "/fsx/yoach/tmp/artefacts/parler-tts-mini/decoder",
79
  "activation_dropout": 0.0,
80
  "activation_function": "gelu",
@@ -87,6 +89,7 @@
87
  "begin_suppress_tokens": null,
88
  "bos_token_id": 1025,
89
  "chunk_size_feed_forward": 0,
 
90
  "cross_attention_hidden_size": null,
91
  "cross_attention_implementation_strategy": null,
92
  "decoder_start_token_id": null,
@@ -116,7 +119,7 @@
116
  "layerdrop": 0.0,
117
  "length_penalty": 1.0,
118
  "max_length": 20,
119
- "max_position_embeddings": 4096,
120
  "min_length": 0,
121
  "model_type": "parler_tts_decoder",
122
  "no_repeat_ngram_size": 0,
@@ -157,6 +160,7 @@
157
  "typical_p": 1.0,
158
  "use_bfloat16": false,
159
  "use_cache": true,
 
160
  "vocab_size": 1088
161
  },
162
  "decoder_start_token_id": 1025,
@@ -165,6 +169,7 @@
165
  "pad_token_id": 1024,
166
  "prompt_cross_attention": false,
167
  "text_encoder": {
 
168
  "_name_or_path": "google/flan-t5-large",
169
  "add_cross_attention": false,
170
  "architectures": [
@@ -249,6 +254,6 @@
249
  "vocab_size": 32128
250
  },
251
  "torch_dtype": "float32",
252
- "transformers_version": "4.43.3",
253
  "vocab_size": 32128
254
  }
 
1
  {
2
+ "_name_or_path": "ChicFusion/birthdays10th",
3
  "architectures": [
4
  "ParlerTTSForConditionalGeneration"
5
  ],
6
  "audio_encoder": {
7
+ "_attn_implementation_autoset": false,
8
  "_name_or_path": "parler-tts/dac_44khZ_8kbps",
9
  "add_cross_attention": false,
10
  "architectures": [
 
42
  "max_length": 20,
43
  "min_length": 0,
44
  "model_bitrate": 8,
45
+ "model_type": "dac_on_the_hub",
46
  "no_repeat_ngram_size": 0,
47
  "num_beam_groups": 1,
48
  "num_beams": 1,
 
76
  "use_bfloat16": false
77
  },
78
  "decoder": {
79
+ "_attn_implementation_autoset": false,
80
  "_name_or_path": "/fsx/yoach/tmp/artefacts/parler-tts-mini/decoder",
81
  "activation_dropout": 0.0,
82
  "activation_function": "gelu",
 
89
  "begin_suppress_tokens": null,
90
  "bos_token_id": 1025,
91
  "chunk_size_feed_forward": 0,
92
+ "codebook_weights": null,
93
  "cross_attention_hidden_size": null,
94
  "cross_attention_implementation_strategy": null,
95
  "decoder_start_token_id": null,
 
119
  "layerdrop": 0.0,
120
  "length_penalty": 1.0,
121
  "max_length": 20,
122
+ "max_position_embeddings": 4311,
123
  "min_length": 0,
124
  "model_type": "parler_tts_decoder",
125
  "no_repeat_ngram_size": 0,
 
160
  "typical_p": 1.0,
161
  "use_bfloat16": false,
162
  "use_cache": true,
163
+ "use_fused_lm_heads": false,
164
  "vocab_size": 1088
165
  },
166
  "decoder_start_token_id": 1025,
 
169
  "pad_token_id": 1024,
170
  "prompt_cross_attention": false,
171
  "text_encoder": {
172
+ "_attn_implementation_autoset": false,
173
  "_name_or_path": "google/flan-t5-large",
174
  "add_cross_attention": false,
175
  "architectures": [
 
254
  "vocab_size": 32128
255
  },
256
  "torch_dtype": "float32",
257
+ "transformers_version": "4.48.0",
258
  "vocab_size": 32128
259
  }
generation_config.json CHANGED
@@ -5,8 +5,8 @@
5
  "do_sample": true,
6
  "eos_token_id": 1024,
7
  "guidance_scale": 1,
8
- "max_length": 2580,
9
  "min_new_tokens": 10,
10
  "pad_token_id": 1024,
11
- "transformers_version": "4.43.3"
12
  }
 
5
  "do_sample": true,
6
  "eos_token_id": 1024,
7
  "guidance_scale": 1,
8
+ "max_length": 5127,
9
  "min_new_tokens": 10,
10
  "pad_token_id": 1024,
11
+ "transformers_version": "4.48.0"
12
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d94da905b2e521ce504f992f8b3d950bad3393a617186d348d36bded4ada3932
3
- size 3511490600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1caf990fd587e1a8bd4196d05cb69f0d24a2bd95c0eb79f4a76e5b05f6f7ced9
3
+ size 3512375104
tokenizer_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<pad>",
@@ -927,9 +928,10 @@
927
  "<extra_id_98>",
928
  "<extra_id_99>"
929
  ],
930
- "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
 
933
  "model_max_length": 512,
934
  "pad_token": "<pad>",
935
  "padding_side": "left",
 
1
  {
2
+ "add_prefix_space": null,
3
  "added_tokens_decoder": {
4
  "0": {
5
  "content": "<pad>",
 
928
  "<extra_id_98>",
929
  "<extra_id_99>"
930
  ],
931
+ "clean_up_tokenization_spaces": false,
932
  "eos_token": "</s>",
933
  "extra_ids": 100,
934
+ "extra_special_tokens": {},
935
  "model_max_length": 512,
936
  "pad_token": "<pad>",
937
  "padding_side": "left",