How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-classification", model="CamiloVega/Tweet-Classifier")
# Load model directly
from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("CamiloVega/Tweet-Classifier")
model = AutoModelForSequenceClassification.from_pretrained("CamiloVega/Tweet-Classifier")
Quick Links

Model Trained Using AutoTrain

  • Problem type: Text Classification

Validation Metrics

loss: 1.1234140396118164

f1_macro: 0.17777777777777778

f1_micro: 0.36363636363636365

f1_weighted: 0.19393939393939394

precision_macro: 0.12121212121212122

precision_micro: 0.36363636363636365

precision_weighted: 0.1322314049586777

recall_macro: 0.3333333333333333

recall_micro: 0.36363636363636365

recall_weighted: 0.36363636363636365

accuracy: 0.36363636363636365

Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Space using CamiloVega/Tweet-Classifier 1