| { | |
| "num_layers":32, | |
| "vocab_size":50048, | |
| "hidden_size":1600, | |
| "num_attention_heads":25, | |
| "embedding_dropout_prob":0.1, | |
| "attention_dropout_prob":0.1, | |
| "output_dropout_prob":0.1, | |
| "max_sequence_length":512, | |
| "max_memory_length":512, | |
| "checkpoint_activations":false, | |
| "checkpoint_num_layers":1, | |
| "parallel_output":true, | |
| "relative_encoding":true | |
| } |