| { |
| "_name": "Shorsey-T2000", |
| "_name_or_path": "Wonder-Griffin/Shorsey-T2000", |
| "activation_function": "gelu_new", |
| "architectures": [ |
| "GPTForCausalLM" |
| ], |
| "model_type": "gpt2", |
| "vocab_size": 30522, |
| "max_len": 512, |
| "hidden_size": 512, |
| "dropout": 0.1, |
| "n_layer": 4, |
| "n_head": 4, |
| "ff_expansion_factor": 4, |
| "n_embd": 128, |
| "rnn_units": 512, |
| "num_labels": 5, |
| "task_specific_params": { |
| "qa": { |
| "num_labels": 5 |
| }, |
| "causal_lm": { |
| "vocab_size": 30522 |
| }, |
| "general": { |
| "vocab_size": 30522 |
| } |
| }, |
| "library_name": "transformers", |
| "tags": [ |
| "text-generation-inference", |
| "causal-lm", |
| "question-answering" |
| ], |
| "model-index": [ |
| { |
| "name": "Shorsey-T2000", |
| "results": [] |
| } |
| ], |
| "datasets": [ |
| "stanfordnlp/imdb" |
| ], |
| "language": [ |
| "en" |
| ], |
| "pipeline_tag": "text-generation" |
| } |