Invalid JSON:
Unexpected token 'H', "Hugging Fa"... is not valid JSON
| Hugging Face's logo | |
| Hugging Face | |
| Search models, datasets, users... | |
| Models | |
| Datasets | |
| Spaces | |
| Posts | |
| Docs | |
| Pricing | |
| EleutherAI | |
| / | |
| gpt-neo-2.7B | |
| like | |
| 400 | |
| Text Generation | |
| Transformers | |
| PyTorch | |
| JAX | |
| Rust | |
| Safetensors | |
| EleutherAI/pile | |
| English | |
| gpt_neo | |
| text generation | |
| causal-lm | |
| Inference Endpoints | |
| License: | |
| mit | |
| Model card | |
| Files and versions | |
| Community | |
| 15 | |
| gpt-neo-2.7B | |
| / | |
| config.json | |
| valhalla's picture | |
| valhalla | |
| add flax model | |
| 0b8087b | |
| about 3 years ago | |
| raw | |
| Copy download link | |
| history | |
| blame | |
| contribute | |
| delete | |
| No virus | |
| 1.46 kB | |
| { | |
| "activation_function": "gelu_new", | |
| "architectures": [ | |
| "GPTNeoForCausalLM" | |
| ], | |
| "attention_dropout": 0, | |
| "attention_layers": [ | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local", | |
| "global", | |
| "local" | |
| ], | |
| "attention_types": [ | |
| [ | |
| [ | |
| "global", | |
| "local" | |
| ], | |
| 16 | |
| ] | |
| ], | |
| "bos_token_id": 50256, | |
| "embed_dropout": 0, | |
| "eos_token_id": 50256, | |
| "gradient_checkpointing": false, | |
| "hidden_size": 2560, | |
| "initializer_range": 0.02, | |
| "intermediate_size": null, | |
| "layer_norm_epsilon": 1e-05, | |
| "max_position_embeddings": 2048, | |
| "model_type": "gpt_neo", | |
| "num_heads": 20, | |
| "num_layers": 32, | |
| "resid_dropout": 0, | |
| "summary_activation": null, | |
| "summary_first_dropout": 0.1, | |
| "summary_proj_to_labels": true, | |
| "summary_type": "cls_index", | |
| "summary_use_proj": true, | |
| "task_specific_params": { | |
| "text-generation": { | |
| "do_sample": true, | |
| "max_length": 50, | |
| "temperature": 0.9 | |
| } | |
| }, | |
| "tokenizer_class": "GPT2Tokenizer", | |
| "transformers_version": "4.9.0.dev0", | |
| "use_cache": true, | |
| "vocab_size": 50257, | |
| "window_size": 256 | |
| } | |