File size: 1,422 Bytes
3c7a78e ef91ee6 f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e f529578 3c7a78e 98961c1 3c7a78e f529578 3c7a78e | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 | {
"_name_or_path": "smitmenon/e2m_translation_project",
"_num_labels": 3,
"activation_dropout": 0.0,
"activation_function": "relu",
"add_bias_logits": false,
"add_final_layer_norm": true,
"architectures": [
"MBartForConditionalGeneration"
],
"attention_dropout": 0.0,
"bos_token_id": 0,
"classif_dropout": 0.0,
"classifier_dropout": 0.0,
"d_model": 1024,
"decoder_attention_heads": 16,
"decoder_ffn_dim": 4096,
"decoder_layerdrop": 0.0,
"decoder_layers": 12,
"decoder_start_token_id": 2,
"dropout": 0.1,
"early_stopping": null,
"encoder_attention_heads": 16,
"encoder_ffn_dim": 4096,
"encoder_layerdrop": 0.0,
"encoder_layers": 12,
"eos_token_id": 2,
"forced_eos_token_id": 2,
"gradient_checkpointing": false,
"id2label": {
"0": "LABEL_0",
"1": "LABEL_1",
"2": "LABEL_2"
},
"init_std": 0.02,
"is_encoder_decoder": true,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1,
"LABEL_2": 2
},
"max_length": null,
"max_position_embeddings": 1024,
"model_type": "mbart",
"normalize_before": true,
"normalize_embedding": true,
"num_beams": null,
"num_hidden_layers": 12,
"output_past": true,
"pad_token_id": 1,
"scale_embedding": true,
"static_position_embeddings": false,
"tokenizer_class": "MBart50Tokenizer",
"torch_dtype": "float32",
"transformers_version": "4.46.3",
"use_cache": true,
"vocab_size": 250054
}
|