Update README.md
Browse files
README.md
CHANGED
|
@@ -3,7 +3,7 @@ license: mit
|
|
| 3 |
library_name: transformers
|
| 4 |
pipeline_tag: text-generation
|
| 5 |
---
|
| 6 |
-
# GPT-2
|
| 7 |
|
| 8 |
## Table of Contents
|
| 9 |
- [Model Details](#model-details)
|
|
@@ -19,7 +19,7 @@ pipeline_tag: text-generation
|
|
| 19 |
|
| 20 |
## Model Details
|
| 21 |
|
| 22 |
-
**Model Description:** GPT-2 Large is the **
|
| 23 |
|
| 24 |
- **Developed by:** OpenAI, see [associated research paper](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf) and [GitHub repo](https://github.com/openai/gpt-2) for model developers.
|
| 25 |
- **Model Type:** Transformer-based language model
|
|
|
|
| 3 |
library_name: transformers
|
| 4 |
pipeline_tag: text-generation
|
| 5 |
---
|
| 6 |
+
# GPT-2 Small
|
| 7 |
|
| 8 |
## Table of Contents
|
| 9 |
- [Model Details](#model-details)
|
|
|
|
| 19 |
|
| 20 |
## Model Details
|
| 21 |
|
| 22 |
+
**Model Description:** GPT-2 Large is the **135M parameter** version of GPT-2, a transformer-based language model created and released by OpenAI. The model is a pretrained model on English language using a causal language modeling (CLM) objective.
|
| 23 |
|
| 24 |
- **Developed by:** OpenAI, see [associated research paper](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf) and [GitHub repo](https://github.com/openai/gpt-2) for model developers.
|
| 25 |
- **Model Type:** Transformer-based language model
|