| { |
| "_name_or_path": "configs/gpst/babylmgpstconfig.json", |
| "architectures": [ |
| "GPST" |
| ], |
| "auto_map": { |
| "AutoConfig": "generative_r2d2_fast.GPSTConfig", |
| "AutoModel": "generative_r2d2_fast.GPST" |
| }, |
| "gptconfig": { |
| "action_layer_num": 3, |
| "activation_function": "gelu_new", |
| "architectures": [ |
| "GPT2LMHeadModel" |
| ], |
| "attn_implementation": null, |
| "attn_pdrop": 0.1, |
| "bos_token_id": 0, |
| "dense_hidden_factor": 1, |
| "embd_pdrop": 0.1, |
| "eos_token_id": 2, |
| "hidden_size": 256, |
| "initializer_range": 0.02, |
| "layer_norm_epsilon": 1e-05, |
| "max_position_embeddings": 96, |
| "model_type": "gpt2", |
| "n_ctx": 96, |
| "n_embd": 256, |
| "n_head": 8, |
| "n_inner": 512, |
| "n_layer": 8, |
| "n_positions": 96, |
| "output_hidden_states": "True", |
| "resid_pdrop": 0.1, |
| "summary_activation": null, |
| "summary_first_dropout": 0.1, |
| "summary_proj_to_labels": true, |
| "summary_type": "cls_index", |
| "summary_use_proj": true, |
| "task_specific_params": { |
| "text-generation": { |
| "do_sample": true, |
| "max_length": 50 |
| } |
| }, |
| "vocab_size": 10000 |
| }, |
| "model_type": "gpst", |
| "r2d2config": { |
| "architectures": [ |
| "Bert" |
| ], |
| "attention_probs_dropout_prob": 0.1, |
| "attn_implementation": null, |
| "bos_token_id": 0, |
| "cls_token_id": 101, |
| "const_size": 256, |
| "decoder_num_hidden_layers": 1, |
| "enable_self_attention": true, |
| "encoder_num_hidden_layers": 3, |
| "eos_token_id": 2, |
| "ext_vocab_size": -1, |
| "hidden_act": "gelu", |
| "hidden_dropout_prob": 0.1, |
| "hidden_size": 256, |
| "initializer_range": 0.02, |
| "intermediate_size": 256, |
| "lr_token_id": 9, |
| "mask_token_id": 103, |
| "max_role_embeddings": 4, |
| "model_type": "bert", |
| "nsp_token_id": 8, |
| "num_attention_heads": 8, |
| "pad_token_id": 1, |
| "pairwise_label_num": 2, |
| "parser_chunked": true, |
| "parser_hidden_dim": 256, |
| "parser_input_dim": 256, |
| "parser_max_len": 96, |
| "parser_nhead": 8, |
| "parser_num_layers": 3, |
| "policy_token_id": 6, |
| "reduce_token_id": 50257, |
| "rr_token_id": 10, |
| "score_dim": 256, |
| "sum_token_id": 7, |
| "tree_mask_token_id": 12, |
| "type_vocab_size": 2, |
| "vocab_size": 10000, |
| "window_size": 2 |
| }, |
| "torch_dtype": "float32", |
| "transformers_version": "4.46.3" |
| } |
|
|