azeros / config.json
yshao18's picture
Upload folder using huggingface_hub
625f14f verified
{
"model_type": "azeros",
"use_flash_attn": false,
"exclude_from_checkpoint": [
"llm",
"speech_encoder",
"paraling_encoder"
],
"paraling_encoder_projector_ds_rate": 4,
"speech_encoder_projector_ds_rate": 4,
"speech_encoder_config": {
"model_type": "zipformer",
"feature_dim": 80,
"output_downsampling_factor": 2,
"num_encoder_layers": [
2,
2,
4,
5,
4,
2
],
"downsampling_factor": [
1,
2,
4,
8,
4,
2
],
"encoder_dim": [
192,
256,
512,
768,
512,
256
],
"feedforward_dim": [
576,
768,
1536,
2304,
1536,
768
],
"warmup_batches": 4000.0,
"dropout": null,
"num_heads": [
4,
4,
4,
8,
4,
4
],
"query_head_dim": [
32
],
"value_head_dim": [
12
],
"pos_head_dim": [
4
],
"pos_dim": 48,
"encoder_unmasked_dim": [
192,
192,
256,
256,
256,
192
],
"cnn_module_kernel": [
31,
31,
15,
15,
15,
31
],
"causal": false,
"chunk_size": [
16,
32,
64,
-1
],
"left_context_frames": [
64,
128,
256,
-1
]
},
"paraling_encoder_config": {
"model_type": "zipformer",
"feature_dim": 80,
"output_downsampling_factor": 2,
"num_encoder_layers": [
2,
2,
4,
5,
4,
2
],
"downsampling_factor": [
1,
2,
4,
8,
4,
2
],
"encoder_dim": [
192,
256,
512,
768,
512,
256
],
"feedforward_dim": [
576,
768,
1536,
2304,
1536,
768
],
"warmup_batches": 4000.0,
"dropout": null,
"num_heads": [
4,
4,
4,
8,
4,
4
],
"query_head_dim": [
32
],
"value_head_dim": [
12
],
"pos_head_dim": [
4
],
"pos_dim": 48,
"encoder_unmasked_dim": [
192,
192,
256,
256,
256,
192
],
"cnn_module_kernel": [
31,
31,
15,
15,
15,
31
],
"causal": false,
"chunk_size": [
16,
32,
64,
-1
],
"left_context_frames": [
64,
128,
256,
-1
]
},
"llm_config": {
"vocab_size": 152064,
"max_position_embeddings": 32768,
"hidden_size": 3584,
"intermediate_size": 18944,
"num_hidden_layers": 28,
"num_attention_heads": 28,
"use_sliding_window": false,
"sliding_window": null,
"max_window_layers": 28,
"num_key_value_heads": 4,
"hidden_act": "silu",
"initializer_range": 0.02,
"rms_norm_eps": 1e-06,
"use_cache": true,
"rope_theta": 1000000.0,
"rope_scaling": null,
"attention_dropout": 0.0,
"layer_types": [
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention"
],
"dtype": "float16",
"tie_word_embeddings": false,
"architectures": [
"Qwen2ForCausalLM"
],
"bos_token_id": 151643,
"eos_token_id": 151645,
"transformers_version": "4.56.1",
"model_type": "qwen2"
}
}