Text Classification
Transformers
Safetensors
English
bert
history
historical
holocaust
war
text-embeddings-inference
Instructions to use ChrisBridges/labse-malach-multilabel with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use ChrisBridges/labse-malach-multilabel with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-classification", model="ChrisBridges/labse-malach-multilabel")# Load model directly from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("ChrisBridges/labse-malach-multilabel") model = AutoModelForSequenceClassification.from_pretrained("ChrisBridges/labse-malach-multilabel") - Notebooks
- Google Colab
- Kaggle
LaBSE-Malach-Multilabel
A multilabel text classification model fine-tuned on a small English subset (Malach ASR) of the Visual History Archive. Based on LaBSE pretrained weights but it uses the general Hugging Face framework, not sentence-transformers. Input text segments consisted of ~350 words on average.
Given an input string, the model predicts probablites for 1063 keyword IDs from the VHA ontology. Typically, probabilities >= 0.5 are "True" if encoding them in a binary vector.
Due to the small training data, the most likely predictions are usually correct but do not meet the threshold.
The mapping from keyword IDs to labels will be added to the repository later.
- Downloads last month
- 4