{ "architectures": [ "RNNLMForCausalLM" ], "embedding_dim": 300, "hidden_size": 500, "model_type": "rnnlm", "n_feature_nodes": 100, "n_pos_embedding_nodes": 25, "n_pos_nodes": 100, "n_pos_tags": 59, "num_hidden_layers": 2, "pad_token_id": 0, "tie_word_embeddings": false, "torch_dtype": "float32", "transformers_version": "4.46.3", "unk_token_id": 1, "use_cache": true, "use_features": false, "use_pos": false, "vocab_size": 64986 }