Instructions to use VitalContribution/JokeDetectBERT with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use VitalContribution/JokeDetectBERT with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-classification", model="VitalContribution/JokeDetectBERT")# Load model directly from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("VitalContribution/JokeDetectBERT") model = AutoModelForSequenceClassification.from_pretrained("VitalContribution/JokeDetectBERT") - Notebooks
- Google Colab
- Kaggle
# Load model directly
from transformers import AutoTokenizer, AutoModelForSequenceClassification
tokenizer = AutoTokenizer.from_pretrained("VitalContribution/JokeDetectBERT")
model = AutoModelForSequenceClassification.from_pretrained("VitalContribution/JokeDetectBERT")Quick Links
Model Card: DistilBERT-based Joke Detection (needed this because I'm German)
Model Details
- Model Type: Fine-tuned DistilBERT base model (uncased)
- Task: Binary classification for joke detection
- Output: Joke or No-joke sentiment
Training Data
- Dataset: 200k Short Texts for Humor Detection
- Link: https://www.kaggle.com/datasets/deepcontractor/200k-short-texts-for-humor-detection
- Size: 200,000 labeled short texts
- Distribution: Equally balanced between humor and non-humor
- Source: Primarily from r/jokes and r/cleanjokes subreddits
Base Model
DistilBERT base model (uncased), a distilled version of BERT optimized for efficiency while maintaining performance.
Usage
from transformers import pipeline
model_id = "VitalContribution/JokeDetectBERT"
pipe = pipeline('text-classification', model=model_id)
joke_questionmark = "What do elves learn in school? The elf-abet."
out = pipe(joke_questionmark)[0]
label = out['label']
confidence = out['score']
result = "JOKE" if label == 'LABEL_1' else "NO JOKE"
print(f"Prediction: {result} ({confidence:.2f})")
Training Details
| Parameter | Value |
|---|---|
| Model | DistilBERT (base-uncased) |
| Task | Sequence Classification |
| Number of Classes | 2 |
| Batch Size | 32 (per device) |
| Learning Rate | 2e-4 |
| Weight Decay | 0.01 |
| Epochs | 2 |
| Warmup Steps | 100 |
| Best Model Selection | Based on eval_loss |
Model Evaluation
- Downloads last month
- 13
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-classification", model="VitalContribution/JokeDetectBERT")