SimCSE: Simple Contrastive Learning of Sentence Embeddings
Paper
• 2104.08821 • Published
YAML Metadata Warning: empty or missing yaml metadata in repo card
Check out the documentation for more information.
This is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
We use SimCSE here by using mBERT as the baseline model and training the model with Thai Wikipedia here
Using this model becomes easy when you have sentence-transformers installed:
pip install -U sentence-transformers
Then you can use the model like this:
from sentence_transformers import SentenceTransformer
sentences = ["ฉันนะคือคนรักชาติยังไงละ!", "พวกสามกีบล้มเจ้า!"]
model = SentenceTransformer('{MODEL_NAME}')
embeddings = model.encode(sentences)
print(embeddings)