FirstPotatoCoder commited on
Commit
4118342
·
verified ·
1 Parent(s): d38eaa0

Upload CsmForConditionalGeneration

Browse files
config.json CHANGED
@@ -5,7 +5,10 @@
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "audio_eos_token_id": 128003,
 
8
  "audio_token_id": 128002,
 
 
9
  "bos_token_id": 128000,
10
  "codebook_eos_token_id": 0,
11
  "codebook_pad_token_id": 2050,
@@ -46,7 +49,7 @@
46
  "rope_theta": 10000.0,
47
  "sampling_rate": 24000,
48
  "sliding_window": 250,
49
- "torch_dtype": "float16",
50
  "trim_right_ratio": 1.0,
51
  "upsample_groups": 512,
52
  "upsampling_ratios": [
@@ -60,6 +63,7 @@
60
  "use_conv_shortcut": false,
61
  "vector_quantization_hidden_dimension": 256
62
  },
 
63
  "depth_decoder_config": {
64
  "attention_bias": false,
65
  "attention_dropout": 0.0,
@@ -85,7 +89,7 @@
85
  "rope_type": "llama3"
86
  },
87
  "rope_theta": 500000,
88
- "torch_dtype": "float16",
89
  "use_cache": true,
90
  "vocab_size": 2051
91
  },
@@ -101,7 +105,7 @@
101
  "num_codebooks": 32,
102
  "num_hidden_layers": 16,
103
  "num_key_value_heads": 8,
104
- "pad_token_id": 128256,
105
  "rms_norm_eps": 1e-05,
106
  "rope_scaling": {
107
  "factor": 32.0,
@@ -116,7 +120,6 @@
116
  "tie_word_embeddings": false,
117
  "torch_dtype": "float32",
118
  "transformers_version": "4.52.3",
119
- "unsloth_version": "2025.6.12",
120
  "use_cache": true,
121
  "vocab_size": 2051
122
  }
 
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "audio_eos_token_id": 128003,
8
+ "audio_num_codebooks": 32,
9
  "audio_token_id": 128002,
10
+ "audio_vocab_size": 2051,
11
+ "backbone_flavor": "llama-1B",
12
  "bos_token_id": 128000,
13
  "codebook_eos_token_id": 0,
14
  "codebook_pad_token_id": 2050,
 
49
  "rope_theta": 10000.0,
50
  "sampling_rate": 24000,
51
  "sliding_window": 250,
52
+ "torch_dtype": "float32",
53
  "trim_right_ratio": 1.0,
54
  "upsample_groups": 512,
55
  "upsampling_ratios": [
 
63
  "use_conv_shortcut": false,
64
  "vector_quantization_hidden_dimension": 256
65
  },
66
+ "decoder_flavor": "llama-100M",
67
  "depth_decoder_config": {
68
  "attention_bias": false,
69
  "attention_dropout": 0.0,
 
89
  "rope_type": "llama3"
90
  },
91
  "rope_theta": 500000,
92
+ "torch_dtype": "float32",
93
  "use_cache": true,
94
  "vocab_size": 2051
95
  },
 
105
  "num_codebooks": 32,
106
  "num_hidden_layers": 16,
107
  "num_key_value_heads": 8,
108
+ "pad_token_id": 128002,
109
  "rms_norm_eps": 1e-05,
110
  "rope_scaling": {
111
  "factor": 32.0,
 
120
  "tie_word_embeddings": false,
121
  "torch_dtype": "float32",
122
  "transformers_version": "4.52.3",
 
123
  "use_cache": true,
124
  "vocab_size": 2051
125
  }
generation_config.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
- "attn_implementation": "sdpa",
3
  "bos_token_id": 128000,
4
  "depth_decoder_do_sample": true,
 
 
5
  "depth_decoder_temperature": 0.9,
6
  "do_sample": true,
7
- "max_length": 2048,
8
  "max_new_tokens": 125,
9
- "pad_token_id": 128256,
10
  "temperature": 0.9,
11
  "transformers_version": "4.52.3"
12
  }
 
1
  {
 
2
  "bos_token_id": 128000,
3
  "depth_decoder_do_sample": true,
4
+ "depth_decoder_max_new_tokens": 31,
5
+ "depth_decoder_min_new_tokens": 31,
6
  "depth_decoder_temperature": 0.9,
7
  "do_sample": true,
 
8
  "max_new_tokens": 125,
9
+ "pad_token_id": 128002,
10
  "temperature": 0.9,
11
  "transformers_version": "4.52.3"
12
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f7944ebf6386e61b012a314ae3459f360dfbf095bd9cb50ca46fd4b431bc1ee
3
  size 4960077432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1729788cb830d81da3ad3532ba4832ff23fd98f4e242a7e6d464d9d04d75bcd
3
  size 4960077432
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0389cc8ddc94953862c5575254faa986ee10ba11cd7ef4289df641ca6114971a
3
  size 1635765924
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fae515f527d06155fa7e5cb650d8fa71a8e6b8e9f1514edc0cd8e2423b7f475e
3
  size 1635765924