How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("fill-mask", model="LEIA/LEIA-LM-base")
# Load model directly
from transformers import AutoTokenizer, AutoModelForMaskedLM

tokenizer = AutoTokenizer.from_pretrained("LEIA/LEIA-LM-base")
model = AutoModelForMaskedLM.from_pretrained("LEIA/LEIA-LM-base")
Quick Links

This is a BERTweet-base model that has been further pre-trained with preferential masking of emotion words for 100k steps on about 6.3M Vent posts.

This model is meant to be fine-tuned on labeled data or used as feature extractor for downstream tasks.

Citation

Please cite the following paper if you find the model useful for your work:

@article{aroyehun2023leia,
  title={LEIA: Linguistic Embeddings for the Identification of Affect},
  author={Aroyehun, Segun Taofeek and Malik, Lukas and Metzler, Hannah and Haimerl, Nikolas and Di Natale, Anna and Garcia, David},
  journal={EPJ Data Science},
  volume={12},
  year={2023},
  publisher={Springer}
}
Downloads last month
10
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support