| { | |
| "architectures": [ | |
| "LsrModel" | |
| ], | |
| "coref_dim": 20, | |
| "distance_size": 20, | |
| "dropout_emb": 0.2, | |
| "dropout_gcn": 0.4, | |
| "dropout_rate": 0.3, | |
| "finetune_emb": false, | |
| "hidden_dim": 120, | |
| "initializer_range": 0.02, | |
| "max_length": 512, | |
| "ner_dim": 20, | |
| "num_relations": 97, | |
| "reasoner_layer_sizes": [ | |
| 3, | |
| 4 | |
| ], | |
| "torch_dtype": "float32", | |
| "transformers_version": "4.9.2", | |
| "use_bert": false, | |
| "use_reasoning_block": true, | |
| "use_struct_att": false, | |
| "word_embedding_shape": [ | |
| 194784, | |
| 100 | |
| ] | |
| } | |