BERT-Wikt-base-adj / README.md
annamos's picture
Update README.md
860da9b verified
metadata
library_name: transformers
tags: []

Model Card for BERT-Wikt-base-adj

Model Description

This model is an English language model based on BERT-base, fine-tuned using adjective examples from English Wiktionary via supervised contrastive learning. The fine-tuning improves token-level semantic representations, particularly for tasks like Word-in-Context (WiC) and Word Sense Disambiguation (WSD).

Although trained on adjectives, the model shows enhanced representation quality across the lexicon.

  • Developed by: Anna Mosolova, Marie Candito, Carlos Ramisch
  • Funded by: ANR Selexini
  • Model type: BERT-based transformer (BERT-base)
  • Language: English
  • License: MIT
  • Finetuned from model: google-bert/bert-base-uncased

Model Sources

Uses

The model is intended for extracting token-level embeddings for English, with improved sense separation.

How to Get Started with the Model

from transformers import AutoTokenizer, AutoModel

tokenizer = AutoTokenizer.from_pretrained("google-bert/bert-base-uncased")
model = AutoModel.from_pretrained("annamos/BERT-Wikt-base-adj")
sentence = 'You should knock before you enter'
tokenized = tokenizer(sentence, return_tensors='pt')
embeddings = model(**tokenized)[0]