databricks/databricks-dolly-15k
Viewer β’ Updated β’ 15k β’ 32.9k β’ 959
This model is a fine-tuned version of microsoft/deberta-v3-base on the databricks-dolly-15k dataset.
It has been trained to classify the prompt category based solely on the response text.
Text Classification
Input: Response text
Output: One of the predefined categories such as:
brainstormingclassificationclosed_qacreative_writinggeneral_qainformation_extractionopen_qasummarizationThe model was evaluated on a balanced version of the dataset. Here are the results:
creative_writing, classification, summarizationopen_qafrom transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained("mariadg/deberta-v3-prompt-recognition")
tokenizer = AutoTokenizer.from_pretrained("mariadg/deberta-v3-prompt-recognition")
text = "The mitochondria is known as the powerhouse of the cell."
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True)
outputs = model(**inputs)
pred = torch.argmax(outputs.logits, dim=1).item()
print(pred) # Map this index back to label if needed
The model outputs a numerical label corresponding to a prompt category. Below is the mapping between label IDs and their respective categories:
brainstormingclassificationclosed_qacreative_writinggeneral_qainformation_extractionopen_qasummarizationmicrosoft/deberta-v3-baseCrossEntropyLossApache 2.0
π Fine-tuned for research purposes.
Base model
microsoft/deberta-v3-base