mamba-vs-gpt2
Collection
This repository contains the code for comparing the performance of Mamba vs GPT2 on text classification task • 4 items • Updated
# Load model directly
from transformers import AutoTokenizer, AutoModelForSequenceClassification
tokenizer = AutoTokenizer.from_pretrained("date3k2/gpt2")
model = AutoModelForSequenceClassification.from_pretrained("date3k2/gpt2")This model was trained from scratch on an unknown dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|---|---|---|---|---|---|
| 0.2633 | 0.9997 | 781 | 0.2777 | 0.8846 | 0.8827 |
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-classification", model="date3k2/gpt2")