Add model card for EuroBERT-210M related to MLM vs CLM paper
#1
by
nielsr
HF Staff
- opened
This PR adds a comprehensive model card for the EuroBERT-210M model. This model is associated with the paper "Should We Still Pretrain Encoders with Masked Language Modeling?", which investigates optimal pretraining strategies for encoders.
The model card now includes:
- Essential metadata (
pipeline_tag,library_name,license). - A concise summary of the paper's key findings and contributions.
- Direct links to the paper, the project page (
https://hf.co/MLMvsCLM), and the underlying GitHub repository (https://github.com/Nicolas-BZRD/EuroBERT). - A practical usage example demonstrating how to load the model and extract text features using the
transformerslibrary, including the necessarytrust_remote_code=Truefor this custom architecture. - A BibTeX citation for the paper.
This update significantly improves the model's discoverability and provides users with critical information for understanding, using, and citing the artifact.