The DistilGPT2 model is truly impressive in its efficiency and performance. We would love to contribute by updating the README to include detailed information about the base model, GPT-2. This addition will help address the current gap in the model card and provide users with a more comprehensive understanding of the model's origins.

Ready to merge
This branch is ready to get merged automatically.

Sign up or log in to comment