Add model card for EuroBERT-210M related to MLM vs CLM paper

#1
by nielsr HF Staff - opened

This PR adds a comprehensive model card for the EuroBERT-210M model. This model is associated with the paper "Should We Still Pretrain Encoders with Masked Language Modeling?", which investigates optimal pretraining strategies for encoders.

The model card now includes:

  • Essential metadata (pipeline_tag, library_name, license).
  • A concise summary of the paper's key findings and contributions.
  • Direct links to the paper, the project page (https://hf.co/MLMvsCLM), and the underlying GitHub repository (https://github.com/Nicolas-BZRD/EuroBERT).
  • A practical usage example demonstrating how to load the model and extract text features using the transformers library, including the necessary trust_remote_code=True for this custom architecture.
  • A BibTeX citation for the paper.

This update significantly improves the model's discoverability and provides users with critical information for understanding, using, and citing the artifact.

Cannot merge
This branch has merge conflicts in the following files:
  • README.md

Sign up or log in to comment