rohand8 commited on
Commit
f247d45
·
verified ·
1 Parent(s): 391239e

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. config.json +1 -1
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1 +1 @@
1
- {"model_type": "encoder-decoder", "vocab_size": 50000, "d_model": 256, "nhead": 8, "num_encoder_layers": 2, "num_decoder_layers": 2, "dim_feedforward": 2048, "architectures": ["EncoderDecoderModel"], "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "encoder": {"model_type": "gpt2", "layer_norm_eps": 1e-05, "activation_function": "gelu", "hidden_size": 256, "num_attention_heads": 8, "num_hidden_layers": 2, "intermediate_size": 2048, "max_position_embeddings": 512, "vocab_size": 50000, "pad_token_id": 0, "bos_token_id": 1, "eos_token_id": 2}, "decoder": {"model_type": "gpt2", "layer_norm_eps": 1e-05, "activation_function": "gelu", "hidden_size": 256, "num_attention_heads": 8, "num_hidden_layers": 2, "intermediate_size": 2048, "max_position_embeddings": 512, "vocab_size": 50000, "pad_token_id": 0, "bos_token_id": 1, "eos_token_id": 2}}
 
1
+ {"model_type": "encoder-decoder", "vocab_size": 50000, "d_model": 256, "nhead": 8, "num_encoder_layers": 2, "num_decoder_layers": 2, "dim_feedforward": 2048, "architectures": ["EncoderDecoderModel"], "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "model_specific_params": {"model_type": "encoder-decoder", "layer_norm_eps": 1e-05, "activation_function": "gelu", "hidden_size": 256, "num_attention_heads": 8, "num_hidden_layers": 2, "intermediate_size": 2048, "max_position_embeddings": 512, "vocab_size": 50000, "pad_token_id": 0, "bos_token_id": 1, "eos_token_id": 2}}
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c96582ffd176e6f68ab6f3740056df2b79662164e8765d36dd4a258027fe232
3
- size 242086358
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ec07c30990c88c0cedcb21a037bd64fc95b43274c7f7aa29ce7cb56e2e87dae
3
+ size 116109846