index
int64
0
22.3k
modelId
stringlengths
8
111
label
list
readme
stringlengths
0
385k
315
MoritzLaurer/covid-policy-roberta-21
[ "Anti-Disinformation Measures", "COVID-19 Vaccines", "Closure and Regulation of Schools", "Curfew", "Declaration of Emergency", "External Border Restrictions", "Health Monitoring", "Health Resources", "Health Testing", "Hygiene", "Internal Border Restrictions", "Lockdown", "New Task Force, Bureau or Administrative Configuration", "Other", "Other Policy Not Listed Above", "Public Awareness Measures", "Quarantine", "Restriction and Regulation of Businesses", "Restriction and Regulation of Government Services", "Restrictions of Mass Gatherings", "Social Distancing" ]
--- language: - en tags: - text-classification metrics: - accuracy (balanced) - F1 (weighted) widget: - text: "All non-essential work activity will stop in Spain from tomorrow until 9 April but there is some confusion as to which jobs can continue under the new lockdown restrictions" --- # Covid-Policy-RoBERTa-21 This model is currently in development at the Centre for European Policy Studies (CEPS). The model is not yet recommended for use. A more detailed description will follow. If you are interested in using deep learning to identify 20 different types policy measures against COVID-19 in text (NPIs, "non-pharmaceutical interventions") don't hesitate to [contact me](https://www.ceps.eu/ceps-staff/moritz-laurer/).
316
MoritzLaurer/mDeBERTa-v3-base-mnli-xnli
[ "contradiction", "entailment", "neutral" ]
--- language: - multilingual - en - ar - bg - de - el - es - fr - hi - ru - sw - th - tr - ur - vi - zh license: mit tags: - zero-shot-classification - text-classification - nli - pytorch metrics: - accuracy datasets: - multi_nli - xnli pipeline_tag: zero-shot-classification widget: - text: "Angela Merkel ist eine Politikerin in Deutschland und Vorsitzende der CDU" candidate_labels: "politics, economy, entertainment, environment" --- # Multilingual mDeBERTa-v3-base-mnli-xnli ## Model description This multilingual model can perform natural language inference (NLI) on 100 languages and is therefore also suitable for multilingual zero-shot classification. The underlying model was pre-trained by Microsoft on the [CC100 multilingual dataset](https://huggingface.co/datasets/cc100). It was then fine-tuned on the [XNLI dataset](https://huggingface.co/datasets/xnli), which contains hypothesis-premise pairs from 15 languages, as well as the English [MNLI dataset](https://huggingface.co/datasets/multi_nli). As of December 2021, mDeBERTa-base is the best performing multilingual base-sized transformer model, introduced by Microsoft in [this paper](https://arxiv.org/pdf/2111.09543.pdf). If you are looking for a smaller, faster (but less performant) model, you can try [multilingual-MiniLMv2-L6-mnli-xnli](https://huggingface.co/MoritzLaurer/multilingual-MiniLMv2-L6-mnli-xnli). ### How to use the model #### Simple zero-shot classification pipeline ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model="MoritzLaurer/mDeBERTa-v3-base-mnli-xnli") sequence_to_classify = "Angela Merkel ist eine Politikerin in Deutschland und Vorsitzende der CDU" candidate_labels = ["politics", "economy", "entertainment", "environment"] output = classifier(sequence_to_classify, candidate_labels, multi_label=False) print(output) ``` #### NLI use-case ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cpu") model_name = "MoritzLaurer/mDeBERTa-v3-base-mnli-xnli" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSequenceClassification.from_pretrained(model_name) premise = "Angela Merkel ist eine Politikerin in Deutschland und Vorsitzende der CDU" hypothesis = "Emmanuel Macron is the President of France" input = tokenizer(premise, hypothesis, truncation=True, return_tensors="pt") output = model(input["input_ids"].to(device)) # device = "cuda:0" or "cpu" prediction = torch.softmax(output["logits"][0], -1).tolist() label_names = ["entailment", "neutral", "contradiction"] prediction = {name: round(float(pred) * 100, 1) for pred, name in zip(prediction, label_names)} print(prediction) ``` ### Training data This model was trained on the XNLI development dataset and the MNLI train dataset. The XNLI development set consists of 2490 professionally translated texts from English to 14 other languages (37350 texts in total) (see [this paper](https://arxiv.org/pdf/1809.05053.pdf)). Note that the XNLI contains a training set of 15 machine translated versions of the MNLI dataset for 15 languages, but due to quality issues with these machine translations, this model was only trained on the professional translations from the XNLI development set and the original English MNLI training set (392 702 texts). Not using machine translated texts can avoid overfitting the model to the 15 languages; avoids catastrophic forgetting of the other 85 languages mDeBERTa was pre-trained on; and significantly reduces training costs. ### Training procedure mDeBERTa-v3-base-mnli-xnli was trained using the Hugging Face trainer with the following hyperparameters. ``` training_args = TrainingArguments( num_train_epochs=2, # total number of training epochs learning_rate=2e-05, per_device_train_batch_size=16, # batch size per device during training per_device_eval_batch_size=16, # batch size for evaluation warmup_ratio=0.1, # number of warmup steps for learning rate scheduler weight_decay=0.06, # strength of weight decay ) ``` ### Eval results The model was evaluated on the XNLI test set on 15 languages (5010 texts per language, 75150 in total). Note that multilingual NLI models are capable of classifying NLI texts without receiving NLI training data in the specific language (cross-lingual transfer). This means that the model is also able of doing NLI on the other 85 languages mDeBERTa was training on, but performance is most likely lower than for those languages available in XNLI. Also note that if other multilingual models on the model hub claim performance of around 90% on languages other than English, the authors have most likely made a mistake during testing since non of the latest papers shows a multilingual average performance of more than a few points above 80% on XNLI (see [here](https://arxiv.org/pdf/2111.09543.pdf) or [here](https://arxiv.org/pdf/1911.02116.pdf)). average | ar | bg | de | el | en | es | fr | hi | ru | sw | th | tr | ur | vi | zh ---------|----------|---------|----------|----------|----------|----------|----------|----------|----------|----------|----------|----------|----------|----------|---------- 0.808 | 0.802 | 0.829 | 0.825 | 0.826 | 0.883 | 0.845 | 0.834 | 0.771 | 0.813 | 0.748 | 0.793 | 0.807 | 0.740 | 0.795 | 0.8116 ## Limitations and bias Please consult the original DeBERTa-V3 paper and literature on different NLI datasets for potential biases. ## Citation If you use this model, please cite: Laurer, Moritz, Wouter van Atteveldt, Andreu Salleras Casas, and Kasper Welbers. 2022. ‘Less Annotating, More Classifying – Addressing the Data Scarcity Issue of Supervised Machine Learning with Deep Transfer Learning and BERT - NLI’. Preprint, June. Open Science Framework. https://osf.io/74b8k. ## Ideas for cooperation or questions? If you have questions or ideas for cooperation, contact me at m{dot}laurer{at}vu{dot}nl or [LinkedIn](https://www.linkedin.com/in/moritz-laurer/) ## Debugging and issues Note that DeBERTa-v3 was released in late 2021 and older versions of HF Transformers seem to have issues running the model (e.g. resulting in an issue with the tokenizer). Using Transformers>=4.13 or higher might solve some issues. Note that mDeBERTa currently does not support FP16, see here: https://github.com/microsoft/DeBERTa/issues/77
317
MoritzLaurer/policy-distilbert-7d
[ "external relations", "freedom and democracy", "political system", "economy", "welfare and quality of life", "fabric of society", "social groups" ]
--- language: - en tags: - text-classification metrics: - accuracy (balanced) - F1 (weighted) widget: - text: "70-85% of the population needs to get vaccinated against the novel coronavirus to achieve herd immunity." --- # Policy-DistilBERT-7d ## Model description This model was trained on 129.669 manually annotated sentences to classify text into one of seven political categories: 'Economy', 'External Relations', 'Fabric of Society', 'Freedom and Democracy', 'Political System', 'Welfare and Quality of Life' or 'Social Groups'. ## Intended uses & limitations #### How to use the model ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model_name = "MoritzLaurer/policy-distilbert-7d" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSequenceClassification.from_pretrained(model_name) text = "The new variant first detected in southern England in September is blamed for sharp rises in levels of positive tests in recent weeks in London, south-east England and the east of England" input = tokenizer(text, truncation=True, return_tensors="pt") output = model(input["input_ids"]) # the output corresponds to the following labels: # 0: external relations, 1: freedom and democracy, 2: political system, 3: economy, 4: welfare and quality of life, 5: fabric of society, 6: social groups # output to dictionary prediction = torch.softmax(output["logits"][0], -1).tolist() label_names = ["external relations", "freedom and democracy", "political system", "economy", "welfare and quality of life", "fabric of society", "social groups"] prediction = {name: round(float(pred) * 100, 1) for pred, name in zip(prediction, label_names)} print(prediction) #{'external relations': 0.0, 'freedom and democracy': 0.0, 'political system': 0.9, 'economy': 0.4, # 'welfare and quality of life': 98.3, 'fabric of society': 0.3, 'social groups': 0.0} ``` ### Training data Policy-DistilBERT-7d was trained on the English-speaking subset of the [Manifesto Project Dataset (MPDS2020a)](https://manifesto-project.wzb.eu/datasets). The model was trained on 129.669 sentences from 164 political manifestos from 55 political parties in 8 English-speaking countries (Australia, Canada, Ireland, Israel, New Zealand, South Africa, United Kingdom, United States). The manifestos were published between 1992 - 2019. The Manifesto Project mannually annotates individual sentences from political party manifestos in 7 main political domains: 'Economy', 'External Relations', 'Fabric of Society', 'Freedom and Democracy', 'Political System', 'Welfare and Quality of Life' or 'Social Groups' - see the [codebook](https://manifesto-project.wzb.eu/down/data/2020b/codebooks/codebook_MPDataset_MPDS2020b.pdf) for the exact definitions of each domain. ### Training procedure `distilbert-base-uncased` was trained using the Hugging Face trainer with the following hyperparameters. The hyperparameters were determined using a hyperparameter search on a 15% validation set. ``` training_args = TrainingArguments( num_train_epochs=5, # total number of training epochs learning_rate=4e-05, per_device_train_batch_size=4, # batch size per device during training per_device_eval_batch_size=4, # batch size for evaluation warmup_steps=500, # number of warmup steps for learning rate scheduler weight_decay=0.02, # strength of weight decay fp16=True # mixed precision training ) ``` ### Eval results The model was evaluated using 15% of the sentences (85-15 train-test split). accuracy (balanced) | F1 (weighted) | precision | recall | accuracy (not balanced) -------|---------|----------|---------|---------- 0.745 | 0.773 | 0.772 | 0.771 | 0.771 Please note that the label distribution in the dataset is imbalanced: ``` Welfare and Quality of Life 0.327225 Economy 0.259191 Fabric of Society 0.111800 Political System 0.095081 Social Groups 0.094371 External Relations 0.063724 Freedom and Democracy 0.048608 ``` [Balanced accuracy](https://scikit-learn.org/stable/modules/generated/sklearn.metrics.balanced_accuracy_score.html) and [weighted F1](https://scikit-learn.org/stable/modules/generated/sklearn.metrics.precision_recall_fscore_support.html) were therefore used to evaluate model performance. ## Limitations and bias The model was trained on sentences in political manifestos from parties in the 8 countries mentioned above between 1992-2019, manually annotated by the [Manifesto Project](https://manifesto-project.wzb.eu/information/documents/information). The model output therefore reproduces the limitations of the dataset in terms of country coverage, time span, domain definitions and potential biases of the annotators - as any supervised machine learning model would. Applying the model to other types of data (other types of texts, countries etc.) will reduce performance. ### BibTeX entry and citation info ```bibtex @unpublished{ title={Policy-DistilBERT}, author={Moritz Laurer}, year={2020}, note={Unpublished paper} } ```
318
MoritzLaurer/xtremedistil-l6-h256-mnli-fever-anli-ling-binary
[ "entailment", "not_entailment" ]
--- language: - en tags: - text-classification - zero-shot-classification metrics: - accuracy datasets: - multi_nli - anli - fever - lingnli pipeline_tag: zero-shot-classification --- # xtremedistil-l6-h256-mnli-fever-anli-ling-binary ## Model description This model was trained on 782 357 hypothesis-premise pairs from 4 NLI datasets: [MultiNLI](https://huggingface.co/datasets/multi_nli), [Fever-NLI](https://github.com/easonnie/combine-FEVER-NSMN/blob/master/other_resources/nli_fever.md), [LingNLI](https://arxiv.org/abs/2104.07179) and [ANLI](https://github.com/facebookresearch/anli). Note that the model was trained on binary NLI to predict either "entailment" or "not-entailment". This is specifically designed for zero-shot classification, where the difference between "neutral" and "contradiction" is irrelevant. The base model is [xtremedistil-l6-h256-uncased from Microsoft](https://huggingface.co/microsoft/xtremedistil-l6-h256-uncased). ## Intended uses & limitations #### How to use the model ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch model_name = "MoritzLaurer/xtremedistil-l6-h256-mnli-fever-anli-ling-binary" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSequenceClassification.from_pretrained(model_name) premise = "I first thought that I liked the movie, but upon second thought it was actually disappointing." hypothesis = "The movie was good." input = tokenizer(premise, hypothesis, truncation=True, return_tensors="pt") output = model(input["input_ids"].to(device)) # device = "cuda:0" or "cpu" prediction = torch.softmax(output["logits"][0], -1).tolist() label_names = ["entailment", "not_entailment"] prediction = {name: round(float(pred) * 100, 1) for pred, name in zip(prediction, label_names)} print(prediction) ``` ### Training data This model was trained on 782 357 hypothesis-premise pairs from 4 NLI datasets: [MultiNLI](https://huggingface.co/datasets/multi_nli), [Fever-NLI](https://github.com/easonnie/combine-FEVER-NSMN/blob/master/other_resources/nli_fever.md), [LingNLI](https://arxiv.org/abs/2104.07179) and [ANLI](https://github.com/facebookresearch/anli). ### Training procedure xtremedistil-l6-h256-mnli-fever-anli-ling-binary was trained using the Hugging Face trainer with the following hyperparameters. ``` training_args = TrainingArguments( num_train_epochs=5, # total number of training epochs learning_rate=2e-05, per_device_train_batch_size=32, # batch size per device during training per_device_eval_batch_size=32, # batch size for evaluation warmup_ratio=0.1, # number of warmup steps for learning rate scheduler weight_decay=0.06, # strength of weight decay fp16=True # mixed precision training ) ``` ### Eval results The model was evaluated using the binary test sets for MultiNLI, ANLI, LingNLI and the binary dev set for Fever-NLI (two classes instead of three). The metric used is accuracy. dataset | mnli-m-2c | mnli-mm-2c | fever-nli-2c | anli-all-2c | anli-r3-2c | lingnli-2c --------|---------|----------|---------|----------|----------|------ accuracy | 0.897 | 0.898 | 0.861 | 0.607 | 0.62 | 0.827 speed (text/sec, GPU Tesla P100, 128 batch) | 1490 | 1485 | 760 | 1186 | 1062 | 1791 ## Limitations and bias Please consult the original paper and literature on different NLI datasets for potential biases. ### BibTeX entry and citation info If you want to cite this model, please cite the original paper, the respective NLI datasets and include a link to this model on the Hugging Face hub. ### Ideas for cooperation or questions? If you have questions or ideas for cooperation, contact me at m{dot}laurer{at}vu{dot}nl or [LinkedIn](https://www.linkedin.com/in/moritz-laurer/) ### Debugging and issues Note that the model was released recently and older versions of HF Transformers seem to have issues running the model (e.g. resulting in an issue with the tokenizer). Using Transformers==4.13 might solve some issues.
319
Motahar/distilbert-sst2-mahtab
[ "NEGATIVE", "POSITIVE" ]
--- license: apache-2.0 tags: - generated_from_trainer datasets: - glue model-index: - name: distilbert-sst2-mahtab results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-sst2-mahtab This model is a fine-tuned version of [distilbert-base-uncased-finetuned-sst-2-english](https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english) on the glue dataset. It achieves the following results on the evaluation set: - eval_loss: 0.4982 - eval_accuracy: 0.8830 - eval_runtime: 2.3447 - eval_samples_per_second: 371.91 - eval_steps_per_second: 46.489 - epoch: 1.0 - step: 8419 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
320
Mulin/my_wolf_model
[ "LABEL_0", "LABEL_1", "LABEL_10", "LABEL_11", "LABEL_2", "LABEL_3", "LABEL_4", "LABEL_5", "LABEL_6", "LABEL_7", "LABEL_8", "LABEL_9" ]
My First Model - for classification of wolf
321
Mustang/BERT_responsible_AI
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
--- license: eupl-1.1 --- ## BERT model van het project Explainable AI
322
NDugar/1epochv3
[ "contradiction", "entailment", "neutral" ]
--- language: en tags: - deberta-v3 - deberta-v2` - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. It outperforms BERT and RoBERTa on majority of NLU tasks with 80GB training data. Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates. This is the DeBERTa V2 xxlarge model with 48 layers, 1536 hidden size. The total parameters are 1.5B and it is trained with 160GB raw data. ### Fine-tuning on NLU tasks We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks. | Model | SQuAD 1.1 | SQuAD 2.0 | MNLI-m/mm | SST-2 | QNLI | CoLA | RTE | MRPC | QQP |STS-B | |---------------------------|-----------|-----------|-------------|-------|------|------|--------|-------|-------|------| | | F1/EM | F1/EM | Acc | Acc | Acc | MCC | Acc |Acc/F1 |Acc/F1 |P/S | | BERT-Large | 90.9/84.1 | 81.8/79.0 | 86.6/- | 93.2 | 92.3 | 60.6 | 70.4 | 88.0/- | 91.3/- |90.0/- | | RoBERTa-Large | 94.6/88.9 | 89.4/86.5 | 90.2/- | 96.4 | 93.9 | 68.0 | 86.6 | 90.9/- | 92.2/- |92.4/- | | XLNet-Large | 95.1/89.7 | 90.6/87.9 | 90.8/- | 97.0 | 94.9 | 69.0 | 85.9 | 90.8/- | 92.3/- |92.5/- | | [DeBERTa-Large](https://huggingface.co/microsoft/deberta-large)<sup>1</sup> | 95.5/90.1 | 90.7/88.0 | 91.3/91.1| 96.5|95.3| 69.5| 91.0| 92.6/94.6| 92.3/- |92.8/92.5 | | [DeBERTa-XLarge](https://huggingface.co/microsoft/deberta-xlarge)<sup>1</sup> | -/- | -/- | 91.5/91.2| 97.0 | - | - | 93.1 | 92.1/94.3 | - |92.9/92.7| | [DeBERTa-V2-XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge)<sup>1</sup>|95.8/90.8| 91.4/88.9|91.7/91.6| **97.5**| 95.8|71.1|**93.9**|92.0/94.2|92.3/89.8|92.9/92.9| |**[DeBERTa-V2-XXLarge](https://huggingface.co/microsoft/deberta-v2-xxlarge)<sup>1,2</sup>**|**96.1/91.4**|**92.2/89.7**|**91.7/91.9**|97.2|**96.0**|**72.0**| 93.5| **93.1/94.9**|**92.7/90.3** |**93.2/93.1** | -------- #### Notes. - <sup>1</sup> Following RoBERTa, for RTE, MRPC, STS-B, we fine-tune the tasks based on [DeBERTa-Large-MNLI](https://huggingface.co/microsoft/deberta-large-mnli), [DeBERTa-XLarge-MNLI](https://huggingface.co/microsoft/deberta-xlarge-mnli), [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli), [DeBERTa-V2-XXLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli). The results of SST-2/QQP/QNLI/SQuADv2 will also be slightly improved when start from MNLI fine-tuned models, however, we only report the numbers fine-tuned from pretrained base models for those 4 tasks. - <sup>2</sup> To try the **XXLarge** model with **[HF transformers](https://huggingface.co/transformers/main_classes/trainer.html)**, we recommand using **deepspeed** as it's faster and saves memory. Run with `Deepspeed`, ```bash pip install datasets pip install deepspeed # Download the deepspeed config file wget https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/ds_config.json -O ds_config.json export TASK_NAME=mnli output_dir="ds_results" num_gpus=8 batch_size=8 python -m torch.distributed.launch --nproc_per_node=${num_gpus} \\ run_glue.py \\ --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME \\ --do_train \\ --do_eval \\ --max_seq_length 256 \\ --per_device_train_batch_size ${batch_size} \\ --learning_rate 3e-6 \\ --num_train_epochs 3 \\ --output_dir $output_dir \\ --overwrite_output_dir \\ --logging_steps 10 \\ --logging_dir $output_dir \\ --deepspeed ds_config.json ``` You can also run with `--sharded_ddp` ```bash cd transformers/examples/text-classification/ export TASK_NAME=mnli python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME --do_train --do_eval --max_seq_length 256 --per_device_train_batch_size 8 \\ --learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16 ``` ### Citation If you find DeBERTa useful for your work, please cite the following paper: ``` latex @inproceedings{ he2021deberta, title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION}, author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen}, booktitle={International Conference on Learning Representations}, year={2021}, url={https://openreview.net/forum?id=XPZIaotutsD} } ```
323
NDugar/2epochv3mlni
[ "contradiction", "entailment", "neutral" ]
--- language: en tags: - deberta-v3 - deberta-v2` - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. It outperforms BERT and RoBERTa on majority of NLU tasks with 80GB training data. Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates. This is the DeBERTa V2 xxlarge model with 48 layers, 1536 hidden size. The total parameters are 1.5B and it is trained with 160GB raw data. ### Fine-tuning on NLU tasks We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks. | Model | SQuAD 1.1 | SQuAD 2.0 | MNLI-m/mm | SST-2 | QNLI | CoLA | RTE | MRPC | QQP |STS-B | |---------------------------|-----------|-----------|-------------|-------|------|------|--------|-------|-------|------| | | F1/EM | F1/EM | Acc | Acc | Acc | MCC | Acc |Acc/F1 |Acc/F1 |P/S | | BERT-Large | 90.9/84.1 | 81.8/79.0 | 86.6/- | 93.2 | 92.3 | 60.6 | 70.4 | 88.0/- | 91.3/- |90.0/- | | RoBERTa-Large | 94.6/88.9 | 89.4/86.5 | 90.2/- | 96.4 | 93.9 | 68.0 | 86.6 | 90.9/- | 92.2/- |92.4/- | | XLNet-Large | 95.1/89.7 | 90.6/87.9 | 90.8/- | 97.0 | 94.9 | 69.0 | 85.9 | 90.8/- | 92.3/- |92.5/- | | [DeBERTa-Large](https://huggingface.co/microsoft/deberta-large)<sup>1</sup> | 95.5/90.1 | 90.7/88.0 | 91.3/91.1| 96.5|95.3| 69.5| 91.0| 92.6/94.6| 92.3/- |92.8/92.5 | | [DeBERTa-XLarge](https://huggingface.co/microsoft/deberta-xlarge)<sup>1</sup> | -/- | -/- | 91.5/91.2| 97.0 | - | - | 93.1 | 92.1/94.3 | - |92.9/92.7| | [DeBERTa-V2-XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge)<sup>1</sup>|95.8/90.8| 91.4/88.9|91.7/91.6| **97.5**| 95.8|71.1|**93.9**|92.0/94.2|92.3/89.8|92.9/92.9| |**[DeBERTa-V2-XXLarge](https://huggingface.co/microsoft/deberta-v2-xxlarge)<sup>1,2</sup>**|**96.1/91.4**|**92.2/89.7**|**91.7/91.9**|97.2|**96.0**|**72.0**| 93.5| **93.1/94.9**|**92.7/90.3** |**93.2/93.1** | -------- #### Notes. - <sup>1</sup> Following RoBERTa, for RTE, MRPC, STS-B, we fine-tune the tasks based on [DeBERTa-Large-MNLI](https://huggingface.co/microsoft/deberta-large-mnli), [DeBERTa-XLarge-MNLI](https://huggingface.co/microsoft/deberta-xlarge-mnli), [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli), [DeBERTa-V2-XXLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli). The results of SST-2/QQP/QNLI/SQuADv2 will also be slightly improved when start from MNLI fine-tuned models, however, we only report the numbers fine-tuned from pretrained base models for those 4 tasks. - <sup>2</sup> To try the **XXLarge** model with **[HF transformers](https://huggingface.co/transformers/main_classes/trainer.html)**, we recommand using **deepspeed** as it's faster and saves memory. Run with `Deepspeed`, ```bash pip install datasets pip install deepspeed # Download the deepspeed config file wget https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/ds_config.json -O ds_config.json export TASK_NAME=mnli output_dir="ds_results" num_gpus=8 batch_size=8 python -m torch.distributed.launch --nproc_per_node=${num_gpus} \\ run_glue.py \\ --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME \\ --do_train \\ --do_eval \\ --max_seq_length 256 \\ --per_device_train_batch_size ${batch_size} \\ --learning_rate 3e-6 \\ --num_train_epochs 3 \\ --output_dir $output_dir \\ --overwrite_output_dir \\ --logging_steps 10 \\ --logging_dir $output_dir \\ --deepspeed ds_config.json ``` You can also run with `--sharded_ddp` ```bash cd transformers/examples/text-classification/ export TASK_NAME=mnli python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME --do_train --do_eval --max_seq_length 256 --per_device_train_batch_size 8 \\ --learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16 ``` ### Citation If you find DeBERTa useful for your work, please cite the following paper: ``` latex @inproceedings{ he2021deberta, title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION}, author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen}, booktitle={International Conference on Learning Representations}, year={2021}, url={https://openreview.net/forum?id=XPZIaotutsD} } ```
324
NDugar/3epoch-3large
[ "contradiction", "entailment", "neutral" ]
--- language: en tags: - deberta-v3 - deberta-v2` - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. It outperforms BERT and RoBERTa on majority of NLU tasks with 80GB training data. Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates. This is the DeBERTa V2 xxlarge model with 48 layers, 1536 hidden size. The total parameters are 1.5B and it is trained with 160GB raw data. ### Fine-tuning on NLU tasks We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks. | Model | SQuAD 1.1 | SQuAD 2.0 | MNLI-m/mm | SST-2 | QNLI | CoLA | RTE | MRPC | QQP |STS-B | |---------------------------|-----------|-----------|-------------|-------|------|------|--------|-------|-------|------| | | F1/EM | F1/EM | Acc | Acc | Acc | MCC | Acc |Acc/F1 |Acc/F1 |P/S | | BERT-Large | 90.9/84.1 | 81.8/79.0 | 86.6/- | 93.2 | 92.3 | 60.6 | 70.4 | 88.0/- | 91.3/- |90.0/- | | RoBERTa-Large | 94.6/88.9 | 89.4/86.5 | 90.2/- | 96.4 | 93.9 | 68.0 | 86.6 | 90.9/- | 92.2/- |92.4/- | | XLNet-Large | 95.1/89.7 | 90.6/87.9 | 90.8/- | 97.0 | 94.9 | 69.0 | 85.9 | 90.8/- | 92.3/- |92.5/- | | [DeBERTa-Large](https://huggingface.co/microsoft/deberta-large)<sup>1</sup> | 95.5/90.1 | 90.7/88.0 | 91.3/91.1| 96.5|95.3| 69.5| 91.0| 92.6/94.6| 92.3/- |92.8/92.5 | | [DeBERTa-XLarge](https://huggingface.co/microsoft/deberta-xlarge)<sup>1</sup> | -/- | -/- | 91.5/91.2| 97.0 | - | - | 93.1 | 92.1/94.3 | - |92.9/92.7| | [DeBERTa-V2-XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge)<sup>1</sup>|95.8/90.8| 91.4/88.9|91.7/91.6| **97.5**| 95.8|71.1|**93.9**|92.0/94.2|92.3/89.8|92.9/92.9| |**[DeBERTa-V2-XXLarge](https://huggingface.co/microsoft/deberta-v2-xxlarge)<sup>1,2</sup>**|**96.1/91.4**|**92.2/89.7**|**91.7/91.9**|97.2|**96.0**|**72.0**| 93.5| **93.1/94.9**|**92.7/90.3** |**93.2/93.1** | -------- #### Notes. - <sup>1</sup> Following RoBERTa, for RTE, MRPC, STS-B, we fine-tune the tasks based on [DeBERTa-Large-MNLI](https://huggingface.co/microsoft/deberta-large-mnli), [DeBERTa-XLarge-MNLI](https://huggingface.co/microsoft/deberta-xlarge-mnli), [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli), [DeBERTa-V2-XXLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli). The results of SST-2/QQP/QNLI/SQuADv2 will also be slightly improved when start from MNLI fine-tuned models, however, we only report the numbers fine-tuned from pretrained base models for those 4 tasks. - <sup>2</sup> To try the **XXLarge** model with **[HF transformers](https://huggingface.co/transformers/main_classes/trainer.html)**, we recommand using **deepspeed** as it's faster and saves memory. Run with `Deepspeed`, ```bash pip install datasets pip install deepspeed # Download the deepspeed config file wget https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/ds_config.json -O ds_config.json export TASK_NAME=mnli output_dir="ds_results" num_gpus=8 batch_size=8 python -m torch.distributed.launch --nproc_per_node=${num_gpus} \\ run_glue.py \\ --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME \\ --do_train \\ --do_eval \\ --max_seq_length 256 \\ --per_device_train_batch_size ${batch_size} \\ --learning_rate 3e-6 \\ --num_train_epochs 3 \\ --output_dir $output_dir \\ --overwrite_output_dir \\ --logging_steps 10 \\ --logging_dir $output_dir \\ --deepspeed ds_config.json ``` You can also run with `--sharded_ddp` ```bash cd transformers/examples/text-classification/ export TASK_NAME=mnli python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME --do_train --do_eval --max_seq_length 256 --per_device_train_batch_size 8 \\ --learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16 ``` ### Citation If you find DeBERTa useful for your work, please cite the following paper: ``` latex @inproceedings{ he2021deberta, title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION}, author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen}, booktitle={International Conference on Learning Representations}, year={2021}, url={https://openreview.net/forum?id=XPZIaotutsD} } ```
325
NDugar/ZSD-microsoft-v2xxlmnli
[ "CONTRADICTION", "NEUTRAL", "ENTAILMENT" ]
--- language: en tags: - deberta-v1 - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. It outperforms BERT and RoBERTa on majority of NLU tasks with 80GB training data. Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates. This is the DeBERTa large model fine-tuned with MNLI task. #### Fine-tuning on NLU tasks We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks. | Model | SQuAD 1.1 | SQuAD 2.0 | MNLI-m/mm | SST-2 | QNLI | CoLA | RTE | MRPC | QQP |STS-B | |---------------------------|-----------|-----------|-------------|-------|------|------|--------|-------|-------|------| | | F1/EM | F1/EM | Acc | Acc | Acc | MCC | Acc |Acc/F1 |Acc/F1 |P/S | | BERT-Large | 90.9/84.1 | 81.8/79.0 | 86.6/- | 93.2 | 92.3 | 60.6 | 70.4 | 88.0/- | 91.3/- |90.0/- | | RoBERTa-Large | 94.6/88.9 | 89.4/86.5 | 90.2/- | 96.4 | 93.9 | 68.0 | 86.6 | 90.9/- | 92.2/- |92.4/- | | XLNet-Large | 95.1/89.7 | 90.6/87.9 | 90.8/- | 97.0 | 94.9 | 69.0 | 85.9 | 90.8/- | 92.3/- |92.5/- | | [DeBERTa-Large](https://huggingface.co/microsoft/deberta-large)<sup>1</sup> | 95.5/90.1 | 90.7/88.0 | 91.3/91.1| 96.5|95.3| 69.5| 91.0| 92.6/94.6| 92.3/- |92.8/92.5 | | [DeBERTa-XLarge](https://huggingface.co/microsoft/deberta-xlarge)<sup>1</sup> | -/- | -/- | 91.5/91.2| 97.0 | - | - | 93.1 | 92.1/94.3 | - |92.9/92.7| | [DeBERTa-V2-XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge)<sup>1</sup>|95.8/90.8| 91.4/88.9|91.7/91.6| **97.5**| 95.8|71.1|**93.9**|92.0/94.2|92.3/89.8|92.9/92.9| |**[DeBERTa-V2-XXLarge](https://huggingface.co/microsoft/deberta-v2-xxlarge)<sup>1,2</sup>**|**96.1/91.4**|**92.2/89.7**|**91.7/91.9**|97.2|**96.0**|**72.0**| 93.5| **93.1/94.9**|**92.7/90.3** |**93.2/93.1** | -------- #### Notes. - <sup>1</sup> Following RoBERTa, for RTE, MRPC, STS-B, we fine-tune the tasks based on [DeBERTa-Large-MNLI](https://huggingface.co/microsoft/deberta-large-mnli), [DeBERTa-XLarge-MNLI](https://huggingface.co/microsoft/deberta-xlarge-mnli), [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli), [DeBERTa-V2-XXLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli). The results of SST-2/QQP/QNLI/SQuADv2 will also be slightly improved when start from MNLI fine-tuned models, however, we only report the numbers fine-tuned from pretrained base models for those 4 tasks. - <sup>2</sup> To try the **XXLarge** model with **[HF transformers](https://huggingface.co/transformers/main_classes/trainer.html)**, you need to specify **--sharded_ddp** ```bash cd transformers/examples/text-classification/ export TASK_NAME=mrpc python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\\n--task_name $TASK_NAME --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 \\\n--learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16 ``` ### Citation If you find DeBERTa useful for your work, please cite the following paper: ``` latex @inproceedings{ he2021deberta, title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION}, author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen}, booktitle={International Conference on Learning Representations}, year={2021}, url={https://openreview.net/forum?id=XPZIaotutsD} } ```
326
NDugar/deberta-v2-xlarge-mnli
[ "contradiction", "entailment", "neutral" ]
--- language: en tags: - deberta-v3 - deberta-v2` - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- I tried to train v3 xl to mnli using my own training code and got this result.
327
NDugar/debertav3-mnli-snli-anli
[ "contradiction", "entailment", "neutral" ]
--- language: en tags: - deberta-v3 - deberta-mnli - deberta - deberta-v2 tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png pipeline_tag: zero-shot-classification --- ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. It outperforms BERT and RoBERTa on majority of NLU tasks with 80GB training data. Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates. This is the DeBERTa V2 xxlarge model with 48 layers, 1536 hidden size. The total parameters are 1.5B and it is trained with 160GB raw data. ### Fine-tuning on NLU tasks We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks. | Model | SQuAD 1.1 | SQuAD 2.0 | MNLI-m/mm | SST-2 | QNLI | CoLA | RTE | MRPC | QQP |STS-B | |---------------------------|-----------|-----------|-------------|-------|------|------|--------|-------|-------|------| | | F1/EM | F1/EM | Acc | Acc | Acc | MCC | Acc |Acc/F1 |Acc/F1 |P/S | | BERT-Large | 90.9/84.1 | 81.8/79.0 | 86.6/- | 93.2 | 92.3 | 60.6 | 70.4 | 88.0/- | 91.3/- |90.0/- | | RoBERTa-Large | 94.6/88.9 | 89.4/86.5 | 90.2/- | 96.4 | 93.9 | 68.0 | 86.6 | 90.9/- | 92.2/- |92.4/- | | XLNet-Large | 95.1/89.7 | 90.6/87.9 | 90.8/- | 97.0 | 94.9 | 69.0 | 85.9 | 90.8/- | 92.3/- |92.5/- | | [DeBERTa-Large](https://huggingface.co/microsoft/deberta-large)<sup>1</sup> | 95.5/90.1 | 90.7/88.0 | 91.3/91.1| 96.5|95.3| 69.5| 91.0| 92.6/94.6| 92.3/- |92.8/92.5 | | [DeBERTa-XLarge](https://huggingface.co/microsoft/deberta-xlarge)<sup>1</sup> | -/- | -/- | 91.5/91.2| 97.0 | - | - | 93.1 | 92.1/94.3 | - |92.9/92.7| | [DeBERTa-V2-XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge)<sup>1</sup>|95.8/90.8| 91.4/88.9|91.7/91.6| **97.5**| 95.8|71.1|**93.9**|92.0/94.2|92.3/89.8|92.9/92.9| |**[DeBERTa-V2-XXLarge](https://huggingface.co/microsoft/deberta-v2-xxlarge)<sup>1,2</sup>**|**96.1/91.4**|**92.2/89.7**|**91.7/91.9**|97.2|**96.0**|**72.0**| 93.5| **93.1/94.9**|**92.7/90.3** |**93.2/93.1** | -------- #### Notes. - <sup>1</sup> Following RoBERTa, for RTE, MRPC, STS-B, we fine-tune the tasks based on [DeBERTa-Large-MNLI](https://huggingface.co/microsoft/deberta-large-mnli), [DeBERTa-XLarge-MNLI](https://huggingface.co/microsoft/deberta-xlarge-mnli), [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli), [DeBERTa-V2-XXLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli). The results of SST-2/QQP/QNLI/SQuADv2 will also be slightly improved when start from MNLI fine-tuned models, however, we only report the numbers fine-tuned from pretrained base models for those 4 tasks. - <sup>2</sup> To try the **XXLarge** model with **[HF transformers](https://huggingface.co/transformers/main_classes/trainer.html)**, we recommand using **deepspeed** as it's faster and saves memory. Run with `Deepspeed`, ```bash pip install datasets pip install deepspeed # Download the deepspeed config file wget https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/ds_config.json -O ds_config.json export TASK_NAME=mnli output_dir="ds_results" num_gpus=8 batch_size=8 python -m torch.distributed.launch --nproc_per_node=${num_gpus} \\ run_glue.py \\ --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME \\ --do_train \\ --do_eval \\ --max_seq_length 256 \\ --per_device_train_batch_size ${batch_size} \\ --learning_rate 3e-6 \\ --num_train_epochs 3 \\ --output_dir $output_dir \\ --overwrite_output_dir \\ --logging_steps 10 \\ --logging_dir $output_dir \\ --deepspeed ds_config.json ``` You can also run with `--sharded_ddp` ```bash cd transformers/examples/text-classification/ export TASK_NAME=mnli python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME --do_train --do_eval --max_seq_length 256 --per_device_train_batch_size 8 \\ --learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16 ``` ### Citation If you find DeBERTa useful for your work, please cite the following paper: ``` latex @inproceedings{ he2021deberta, title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION}, author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen}, booktitle={International Conference on Learning Representations}, year={2021}, url={https://openreview.net/forum?id=XPZIaotutsD} } ```
328
NDugar/v2xl-again-mnli
[ "contradiction", "entailment", "neutral" ]
--- language: en tags: - deberta-v1 - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. It outperforms BERT and RoBERTa on majority of NLU tasks with 80GB training data. Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates. This is the DeBERTa large model fine-tuned with MNLI task. #### Fine-tuning on NLU tasks We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks. | Model | SQuAD 1.1 | SQuAD 2.0 | MNLI-m/mm | SST-2 | QNLI | CoLA | RTE | MRPC | QQP |STS-B | |---------------------------|-----------|-----------|-------------|-------|------|------|--------|-------|-------|------| | | F1/EM | F1/EM | Acc | Acc | Acc | MCC | Acc |Acc/F1 |Acc/F1 |P/S | | BERT-Large | 90.9/84.1 | 81.8/79.0 | 86.6/- | 93.2 | 92.3 | 60.6 | 70.4 | 88.0/- | 91.3/- |90.0/- | | RoBERTa-Large | 94.6/88.9 | 89.4/86.5 | 90.2/- | 96.4 | 93.9 | 68.0 | 86.6 | 90.9/- | 92.2/- |92.4/- | | XLNet-Large | 95.1/89.7 | 90.6/87.9 | 90.8/- | 97.0 | 94.9 | 69.0 | 85.9 | 90.8/- | 92.3/- |92.5/- | | [DeBERTa-Large](https://huggingface.co/microsoft/deberta-large)<sup>1</sup> | 95.5/90.1 | 90.7/88.0 | 91.3/91.1| 96.5|95.3| 69.5| 91.0| 92.6/94.6| 92.3/- |92.8/92.5 | | [DeBERTa-XLarge](https://huggingface.co/microsoft/deberta-xlarge)<sup>1</sup> | -/- | -/- | 91.5/91.2| 97.0 | - | - | 93.1 | 92.1/94.3 | - |92.9/92.7| | [DeBERTa-V2-XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge)<sup>1</sup>|95.8/90.8| 91.4/88.9|91.7/91.6| **97.5**| 95.8|71.1|**93.9**|92.0/94.2|92.3/89.8|92.9/92.9| |**[DeBERTa-V2-XXLarge](https://huggingface.co/microsoft/deberta-v2-xxlarge)<sup>1,2</sup>**|**96.1/91.4**|**92.2/89.7**|**91.7/91.9**|97.2|**96.0**|**72.0**| 93.5| **93.1/94.9**|**92.7/90.3** |**93.2/93.1** | -------- #### Notes. - <sup>1</sup> Following RoBERTa, for RTE, MRPC, STS-B, we fine-tune the tasks based on [DeBERTa-Large-MNLI](https://huggingface.co/microsoft/deberta-large-mnli), [DeBERTa-XLarge-MNLI](https://huggingface.co/microsoft/deberta-xlarge-mnli), [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli), [DeBERTa-V2-XXLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli). The results of SST-2/QQP/QNLI/SQuADv2 will also be slightly improved when start from MNLI fine-tuned models, however, we only report the numbers fine-tuned from pretrained base models for those 4 tasks. - <sup>2</sup> To try the **XXLarge** model with **[HF transformers](https://huggingface.co/transformers/main_classes/trainer.html)**, you need to specify **--sharded_ddp** ```bash cd transformers/examples/text-classification/ export TASK_NAME=mrpc python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\\n--task_name $TASK_NAME --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 \\\n--learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16 ``` ### Citation If you find DeBERTa useful for your work, please cite the following paper: ``` latex @inproceedings{ he2021deberta, title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION}, author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen}, booktitle={International Conference on Learning Representations}, year={2021}, url={https://openreview.net/forum?id=XPZIaotutsD} } ```
329
NDugar/v3-Large-mnli
[ "contradiction", "entailment", "neutral" ]
--- language: en tags: - deberta-v1 - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the GLUE MNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.4103 - Accuracy: 0.9175 ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 6e-06 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 50 - num_epochs: 2.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.3631 | 1.0 | 49088 | 0.3129 | 0.9130 | | 0.2267 | 2.0 | 98176 | 0.4157 | 0.9153 | ### Framework versions - Transformers 4.13.0.dev0 - Pytorch 1.10.0 - Datasets 1.15.2.dev0 - Tokenizers 0.10.3
330
NDugar/v3large-1epoch
[ "contradiction", "entailment", "neutral" ]
--- language: en tags: - deberta-v3 - deberta-v2` - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. It outperforms BERT and RoBERTa on majority of NLU tasks with 80GB training data. Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates. This is the DeBERTa V2 xxlarge model with 48 layers, 1536 hidden size. The total parameters are 1.5B and it is trained with 160GB raw data. ### Fine-tuning on NLU tasks We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks. | Model | SQuAD 1.1 | SQuAD 2.0 | MNLI-m/mm | SST-2 | QNLI | CoLA | RTE | MRPC | QQP |STS-B | |---------------------------|-----------|-----------|-------------|-------|------|------|--------|-------|-------|------| | | F1/EM | F1/EM | Acc | Acc | Acc | MCC | Acc |Acc/F1 |Acc/F1 |P/S | | BERT-Large | 90.9/84.1 | 81.8/79.0 | 86.6/- | 93.2 | 92.3 | 60.6 | 70.4 | 88.0/- | 91.3/- |90.0/- | | RoBERTa-Large | 94.6/88.9 | 89.4/86.5 | 90.2/- | 96.4 | 93.9 | 68.0 | 86.6 | 90.9/- | 92.2/- |92.4/- | | XLNet-Large | 95.1/89.7 | 90.6/87.9 | 90.8/- | 97.0 | 94.9 | 69.0 | 85.9 | 90.8/- | 92.3/- |92.5/- | | [DeBERTa-Large](https://huggingface.co/microsoft/deberta-large)<sup>1</sup> | 95.5/90.1 | 90.7/88.0 | 91.3/91.1| 96.5|95.3| 69.5| 91.0| 92.6/94.6| 92.3/- |92.8/92.5 | | [DeBERTa-XLarge](https://huggingface.co/microsoft/deberta-xlarge)<sup>1</sup> | -/- | -/- | 91.5/91.2| 97.0 | - | - | 93.1 | 92.1/94.3 | - |92.9/92.7| | [DeBERTa-V2-XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge)<sup>1</sup>|95.8/90.8| 91.4/88.9|91.7/91.6| **97.5**| 95.8|71.1|**93.9**|92.0/94.2|92.3/89.8|92.9/92.9| |**[DeBERTa-V2-XXLarge](https://huggingface.co/microsoft/deberta-v2-xxlarge)<sup>1,2</sup>**|**96.1/91.4**|**92.2/89.7**|**91.7/91.9**|97.2|**96.0**|**72.0**| 93.5| **93.1/94.9**|**92.7/90.3** |**93.2/93.1** | -------- #### Notes. - <sup>1</sup> Following RoBERTa, for RTE, MRPC, STS-B, we fine-tune the tasks based on [DeBERTa-Large-MNLI](https://huggingface.co/microsoft/deberta-large-mnli), [DeBERTa-XLarge-MNLI](https://huggingface.co/microsoft/deberta-xlarge-mnli), [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli), [DeBERTa-V2-XXLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli). The results of SST-2/QQP/QNLI/SQuADv2 will also be slightly improved when start from MNLI fine-tuned models, however, we only report the numbers fine-tuned from pretrained base models for those 4 tasks. - <sup>2</sup> To try the **XXLarge** model with **[HF transformers](https://huggingface.co/transformers/main_classes/trainer.html)**, we recommand using **deepspeed** as it's faster and saves memory. Run with `Deepspeed`, ```bash pip install datasets pip install deepspeed # Download the deepspeed config file wget https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/ds_config.json -O ds_config.json export TASK_NAME=mnli output_dir="ds_results" num_gpus=8 batch_size=8 python -m torch.distributed.launch --nproc_per_node=${num_gpus} \\ run_glue.py \\ --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME \\ --do_train \\ --do_eval \\ --max_seq_length 256 \\ --per_device_train_batch_size ${batch_size} \\ --learning_rate 3e-6 \\ --num_train_epochs 3 \\ --output_dir $output_dir \\ --overwrite_output_dir \\ --logging_steps 10 \\ --logging_dir $output_dir \\ --deepspeed ds_config.json ``` You can also run with `--sharded_ddp` ```bash cd transformers/examples/text-classification/ export TASK_NAME=mnli python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME --do_train --do_eval --max_seq_length 256 --per_device_train_batch_size 8 \\ --learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16 ``` ### Citation If you find DeBERTa useful for your work, please cite the following paper: ``` latex @inproceedings{ he2021deberta, title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION}, author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen}, booktitle={International Conference on Learning Representations}, year={2021}, url={https://openreview.net/forum?id=XPZIaotutsD} } ```
331
NDugar/v3large-2epoch
[ "contradiction", "entailment", "neutral" ]
--- language: en tags: - deberta-v3 - deberta-v2` - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. It outperforms BERT and RoBERTa on majority of NLU tasks with 80GB training data. Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates. This is the DeBERTa V2 xxlarge model with 48 layers, 1536 hidden size. The total parameters are 1.5B and it is trained with 160GB raw data. ### Fine-tuning on NLU tasks We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks. | Model | SQuAD 1.1 | SQuAD 2.0 | MNLI-m/mm | SST-2 | QNLI | CoLA | RTE | MRPC | QQP |STS-B | |---------------------------|-----------|-----------|-------------|-------|------|------|--------|-------|-------|------| | | F1/EM | F1/EM | Acc | Acc | Acc | MCC | Acc |Acc/F1 |Acc/F1 |P/S | | BERT-Large | 90.9/84.1 | 81.8/79.0 | 86.6/- | 93.2 | 92.3 | 60.6 | 70.4 | 88.0/- | 91.3/- |90.0/- | | RoBERTa-Large | 94.6/88.9 | 89.4/86.5 | 90.2/- | 96.4 | 93.9 | 68.0 | 86.6 | 90.9/- | 92.2/- |92.4/- | | XLNet-Large | 95.1/89.7 | 90.6/87.9 | 90.8/- | 97.0 | 94.9 | 69.0 | 85.9 | 90.8/- | 92.3/- |92.5/- | | [DeBERTa-Large](https://huggingface.co/microsoft/deberta-large)<sup>1</sup> | 95.5/90.1 | 90.7/88.0 | 91.3/91.1| 96.5|95.3| 69.5| 91.0| 92.6/94.6| 92.3/- |92.8/92.5 | | [DeBERTa-XLarge](https://huggingface.co/microsoft/deberta-xlarge)<sup>1</sup> | -/- | -/- | 91.5/91.2| 97.0 | - | - | 93.1 | 92.1/94.3 | - |92.9/92.7| | [DeBERTa-V2-XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge)<sup>1</sup>|95.8/90.8| 91.4/88.9|91.7/91.6| **97.5**| 95.8|71.1|**93.9**|92.0/94.2|92.3/89.8|92.9/92.9| |**[DeBERTa-V2-XXLarge](https://huggingface.co/microsoft/deberta-v2-xxlarge)<sup>1,2</sup>**|**96.1/91.4**|**92.2/89.7**|**91.7/91.9**|97.2|**96.0**|**72.0**| 93.5| **93.1/94.9**|**92.7/90.3** |**93.2/93.1** | -------- #### Notes. - <sup>1</sup> Following RoBERTa, for RTE, MRPC, STS-B, we fine-tune the tasks based on [DeBERTa-Large-MNLI](https://huggingface.co/microsoft/deberta-large-mnli), [DeBERTa-XLarge-MNLI](https://huggingface.co/microsoft/deberta-xlarge-mnli), [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli), [DeBERTa-V2-XXLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli). The results of SST-2/QQP/QNLI/SQuADv2 will also be slightly improved when start from MNLI fine-tuned models, however, we only report the numbers fine-tuned from pretrained base models for those 4 tasks. - <sup>2</sup> To try the **XXLarge** model with **[HF transformers](https://huggingface.co/transformers/main_classes/trainer.html)**, we recommand using **deepspeed** as it's faster and saves memory. Run with `Deepspeed`, ```bash pip install datasets pip install deepspeed # Download the deepspeed config file wget https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/ds_config.json -O ds_config.json export TASK_NAME=mnli output_dir="ds_results" num_gpus=8 batch_size=8 python -m torch.distributed.launch --nproc_per_node=${num_gpus} \\ run_glue.py \\ --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME \\ --do_train \\ --do_eval \\ --max_seq_length 256 \\ --per_device_train_batch_size ${batch_size} \\ --learning_rate 3e-6 \\ --num_train_epochs 3 \\ --output_dir $output_dir \\ --overwrite_output_dir \\ --logging_steps 10 \\ --logging_dir $output_dir \\ --deepspeed ds_config.json ``` You can also run with `--sharded_ddp` ```bash cd transformers/examples/text-classification/ export TASK_NAME=mnli python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\ --task_name $TASK_NAME --do_train --do_eval --max_seq_length 256 --per_device_train_batch_size 8 \\ --learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16 ``` ### Citation If you find DeBERTa useful for your work, please cite the following paper: ``` latex @inproceedings{ he2021deberta, title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION}, author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen}, booktitle={International Conference on Learning Representations}, year={2021}, url={https://openreview.net/forum?id=XPZIaotutsD} } ```
332
NTUYG/DeepSCC-RoBERTa
[ "LABEL_0", "LABEL_1", "LABEL_10", "LABEL_11", "LABEL_12", "LABEL_13", "LABEL_14", "LABEL_15", "LABEL_16", "LABEL_17", "LABEL_18", "LABEL_2", "LABEL_3", "LABEL_4", "LABEL_5", "LABEL_6", "LABEL_7", "LABEL_8", "LABEL_9" ]
## How to use ```python from simpletransformers.classification import ClassificationModel, ClassificationArgs name_file = ['bash', 'c', 'c#', 'c++','css', 'haskell', 'java', 'javascript', 'lua', 'objective-c', 'perl', 'php', 'python','r','ruby', 'scala', 'sql', 'swift', 'vb.net'] deep_scc_model_args = ClassificationArgs(num_train_epochs=10,max_seq_length=300,use_multiprocessing=False) deep_scc_model = ClassificationModel("roberta", "NTUYG/DeepSCC-RoBERTa", num_labels=19, args=deep_scc_model_args,use_cuda=True) code = ''' public static double getSimilarity(String phrase1, String phrase2) { return (getSC(phrase1, phrase2) + getSC(phrase2, phrase1)) / 2.0; }''' code = code.replace('\n',' ').replace('\r',' ') predictions, raw_outputs = model.predict([code]) predict = name_file[predictions[0]] print(predict) ```
335
NYTK/sentiment-hts5-hubert-hungarian
[ "LABEL_0", "LABEL_1", "LABEL_2", "LABEL_3", "LABEL_4" ]
--- language: - hu tags: - text-classification license: apache-2.0 metrics: - accuracy widget: - text: Jó reggelt! majd küldöm az élményhozókat :). --- # Hungarian Sentence-level Sentiment Analysis with Finetuned huBERT Model For further models, scripts and details, see [our repository](https://github.com/nytud/sentiment-analysis) or [our demo site](https://juniper.nytud.hu/demo/nlp). - Pretrained model used: huBERT - Finetuned on Hungarian Twitter Sentiment (HTS) Corpus - Labels: 0 (very negative), 1 (negative), 2 (neutral), 3 (positive), 4 (very positive) ## Limitations - max_seq_length = 128 ## Results | Model | HTS2 | HTS5 | | ------------- | ------------- | ------------- | | huBERT | 85.56 | **68.99** | | XLM-RoBERTa| 85.56 | 66.50 | ## Citation If you use this model, please cite the following paper: ``` @inproceedings {yang-sentiment, title = {Improving Performance of Sentence-level Sentiment Analysis with Data Augmentation Methods}, booktitle = {Proceedings of 12th IEEE International Conference on Cognitive Infocommunications (CogInfoCom 2021)}, year = {2021}, publisher = {IEEE}, address = {Online}, author = {Laki, László and Yang, Zijian Győző} pages = {417--422} } ```
336
NYTK/sentiment-hts5-xlm-roberta-hungarian
[ "LABEL_0", "LABEL_1", "LABEL_2", "LABEL_3", "LABEL_4" ]
--- language: - hu tags: - text-classification license: mit metrics: - accuracy widget: - text: Jó reggelt! majd küldöm az élményhozókat :). --- # Hungarian Sentence-level Sentiment Analysis Model with XLM-RoBERTa For further models, scripts and details, see [our repository](https://github.com/nytud/sentiment-analysis) or [our demo site](https://juniper.nytud.hu/demo/nlp). - Pretrained model used: XLM-RoBERTa base - Finetuned on Hungarian Twitter Sentiment (HTS) Corpus - Labels: 0 (very negative), 1 (negative), 2 (neutral), 3 (positive), 4 (very positive) ## Limitations - max_seq_length = 128 ## Results | Model | HTS2 | HTS5 | | ------------- | ------------- | ------------- | | huBERT | 85.56 | **68.99** | | XLM-RoBERTa| 85.56 | 66.50 | ## Citation If you use this model, please cite the following paper: ``` @inproceedings {laki-yang-sentiment, title = {Improving Performance of Sentence-level Sentiment Analysis with Data Augmentation Methods}, booktitle = {Proceedings of 12th IEEE International Conference on Cognitive Infocommunications (CogInfoCom 2021)}, year = {2021}, publisher = {IEEE}, address = {Online}, author = {Laki, László and Yang, Zijian Győző} pages = {417--422} } ```
338
Narrativa/distilroberta-finetuned-stereotype-detection
[ "neutral", "stereotype", "anti-stereotype" ]
--- license: apache-2.0 tags: - generated_from_trainer - stereotype - gender - gender_bias widget: - text: "Cauterize is not just for fans of the guitarist or his other projects, but those that love music that is both aggressive and infectious and gave the album 4 out of 5 stars ." metrics: - accuracy model-index: - name: distilRoberta-stereotype results: - task: name: Text Classification type: text-classification metrics: - name: Accuracy type: accuracy value: 0.989151002901476 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilRoberta-stereotype This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0651 - Accuracy: 0.9892 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.0783 | 1.0 | 5615 | 0.0703 | 0.9847 | | 0.0468 | 2.0 | 11230 | 0.0573 | 0.9863 | | 0.0316 | 3.0 | 16845 | 0.0580 | 0.9882 | | 0.0172 | 4.0 | 22460 | 0.0591 | 0.9885 | | 0.0098 | 5.0 | 28075 | 0.0651 | 0.9892 | ### Framework versions - Transformers 4.10.2 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3 Created by: [Narrativa](https://www.narrativa.com/) About Narrativa: Natural Language Generation (NLG) | Gabriele, our machine learning-based platform, builds and deploys natural language solutions. #NLG #AI
339
Narrativaai/deberta-v3-small-finetuned-hate_speech18
[ "NO_HATE", "HATE", "IDK", "RELATION" ]
--- license: mit tags: - generated_from_trainer datasets: - hate_speech18 widget: - text: "ok, so do we need to kill them too or are the slavs okay ? for some reason whenever i hear the word slav , the word slobber comes to mind and i picture a slobbering half breed creature like the humpback of notre dame or Igor haha" metrics: - accuracy model-index: - name: deberta-v3-small-hate-speech results: - task: name: Text Classification type: text-classification dataset: name: hate_speech18 type: hate_speech18 args: default metrics: - name: Accuracy type: accuracy value: 0.916058394160584 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # DeBERTa v3 small fine-tuned on hate_speech18 dataset for Hate Speech Detection This model is a fine-tuned version of [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small) on the hate_speech18 dataset. It achieves the following results on the evaluation set: - Loss: 0.2922 - Accuracy: 0.9161 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.4147 | 1.0 | 650 | 0.3910 | 0.8832 | | 0.2975 | 2.0 | 1300 | 0.2922 | 0.9161 | | 0.2575 | 3.0 | 1950 | 0.3555 | 0.9051 | | 0.1553 | 4.0 | 2600 | 0.4263 | 0.9124 | | 0.1267 | 5.0 | 3250 | 0.4238 | 0.9161 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
340
Narrativaai/fake-news-detection-spanish
[ "REAL", "FAKE" ]
--- language: es tags: - generated_from_trainer - fake - news - competition datasets: - fakedes widget: - text: 'La palabra "haiga", aceptada por la RAE [SEP] La palabra "haiga", aceptada por la RAE La Real Academia de la Lengua (RAE), ha aceptado el uso de "HAIGA", para su utilización en las tres personas del singular del presente del subjuntivo del verbo hacer, aunque asegura que la forma más recomendable en la lengua culta para este tiempo, sigue siendo "haya". Así lo han confirmado fuentes de la RAE, que explican que este cambio ha sido propuesto y aprobado por el pleno de la Academia de la Lengua, tras la extendida utilización por todo el territorio nacional, sobre todo, empleado por personas carentes de estudios o con estudios básicos de graduado escolar. Ya no será objeto de burla ese compañero que a diario repite aquello de "Mientras que haiga faena, no podemos quejarnos" o esa abuela que repite aquello de "El que haiga sacao los juguetes, que los recoja". Entre otras palabras novedosas que ha aceptado la RAE, contamos también con "Descambiar", significa deshacer un cambio, por ejemplo "devolver la compra". Visto lo visto, nadie apostaría que la palabra "follamigos" sea la siguiente de la lista.' metrics: - f1 - accuracy model-index: - name: roberta-large-fake-news-detection-spanish results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # RoBERTa-large-fake-news-detection-spanish This model is a fine-tuned version of [PlanTL-GOB-ES/roberta-large-bne](https://huggingface.co/PlanTL-GOB-ES/roberta-large-bne) on an [Spanish Fake News Dataset](https://sites.google.com/view/iberlef2020/#h.p_w0c31bn0r-SW). It achieves the following results on the evaluation set: - Loss: 1.7474 - F1: **0.7717** - Accuracy: 0.7797 > So, based on the [leaderboard](https://sites.google.com/view/fakedes/results?authuser=0) our model **outperforms** the best model (scores F1 = 0.7666). ## Model description RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. ## Intended uses & limitations The objective of this task is to decide if a news item is fake or real by analyzing its textual representation. ## Training and evaluation data **FakeDeS**: [Fake News Detection in Spanish Shared Task](https://sites.google.com/view/fakedes/home) Fake news provides information that aims to manipulate people for different purposes: terrorism, political elections, advertisement, satire, among others. In social networks, misinformation extends in seconds among thousands of people, so it is necessary to develop tools that help control the amount of false information on the web. Similar tasks are detection of popularity in social networks and detection of subjectivity of messages in this media. A fake news detection system aims to help users detect and filter out potentially deceptive news. The prediction of intentionally misleading news is based on the analysis of truthful and fraudulent previously reviewed news, i.e., annotated corpora. The Spanish Fake News Corpus is a collection of news compiled from several web sources: established newspapers websites,media companies websites, special websites dedicated to validating fake news, websites designated by different journalists as sites that regularly publish fake news. The news were collected from January to July of 2018 and all of them were written in Mexican Spanish. The corpus has 971 news collected from January to July, 2018, from different sources: - Established newspapers websites, - Media companies websites, - Special websites dedicated to validating fake news, - Websites designated by different journalists as sites that regularly publish fake news. The corpus was tagged considering only two classes (true or fake), following a manual labeling process: - A news is true if there is evidence that it has been published in reliable sites. - A news is fake if there is news from reliable sites or specialized website in detection of deceptive content that contradicts it or no other evidence was found about the news besides the source. - We collected the true-fake news pair of an event so there is a correlation of news in the corpus. In order to avoid topic bias, the corpus covers news from 9 different topics: Science, Sport, Economy, Education, Entertainment, Politics, Health, Security, and Society. As it can be seen in the table below, the number of fake and true news is quite balanced. Approximately 70% will be used as training corpus (676 news), and the 30% as testing corpus (295 news). The training corpus contains the following information: - Category: Fake/ True - Topic: Science/ Sport/ Economy/ Education/ Entertainment/ Politics, Health/ Security/ Society - Headline: The title of the news. - Text: The complete text of the news. - Link: The URL where the news was published. More information needed ## Training procedure TBA ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:------:|:--------:| | No log | 1.0 | 243 | 0.6282 | 0.7513 | 0.75 | | No log | 2.0 | 486 | 0.9600 | 0.7346 | 0.7587 | | 0.5099 | 3.0 | 729 | 1.2128 | 0.7656 | 0.7570 | | 0.5099 | 4.0 | 972 | 1.4001 | 0.7606 | 0.7622 | | 0.1949 | 5.0 | 1215 | 1.9748 | 0.6475 | 0.7220 | | 0.1949 | 6.0 | 1458 | 1.7386 | 0.7706 | 0.7710 | | 0.0263 | 7.0 | 1701 | 1.7474 | 0.7717 | 0.7797 | | 0.0263 | 8.0 | 1944 | 1.8114 | 0.7695 | 0.7780 | | 0.0046 | 9.0 | 2187 | 1.8444 | 0.7709 | 0.7797 | | 0.0046 | 10.0 | 2430 | 1.8552 | 0.7709 | 0.7797 | ### Fast usage with HF `pipelines` ```python from transformers import pipeline ckpt = "Narrativaai/fake-news-detection-spanish" classifier = pipeline("text-classification", model=ckpt) headline = "Your headline" text = "Your article text here..." classifier(headline + " [SEP] " + text) ``` ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.14.0 - Tokenizers 0.10.3 Created by: [Narrativa](https://www.narrativa.com/) About Narrativa: Natural Language Generation (NLG) | Gabriele, our machine learning-based platform, builds and deploys natural language solutions. #NLG #AI
341
Narsil/deberta-large-mnli-zero-cls
[ "CONTRADICTION", "NEUTRAL", "ENTAILMENT" ]
--- language: en tags: - deberta-v1 - deberta-mnli tasks: mnli thumbnail: https://huggingface.co/front/thumbnails/microsoft.png license: mit pipeline_tag: zero-shot-classification --- ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. It outperforms BERT and RoBERTa on majority of NLU tasks with 80GB training data. Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates. This is the DeBERTa large model fine-tuned with MNLI task. #### Fine-tuning on NLU tasks We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks. | Model | SQuAD 1.1 | SQuAD 2.0 | MNLI-m/mm | SST-2 | QNLI | CoLA | RTE | MRPC | QQP |STS-B | |---------------------------|-----------|-----------|-------------|-------|------|------|--------|-------|-------|------| | | F1/EM | F1/EM | Acc | Acc | Acc | MCC | Acc |Acc/F1 |Acc/F1 |P/S | | BERT-Large | 90.9/84.1 | 81.8/79.0 | 86.6/- | 93.2 | 92.3 | 60.6 | 70.4 | 88.0/- | 91.3/- |90.0/- | | RoBERTa-Large | 94.6/88.9 | 89.4/86.5 | 90.2/- | 96.4 | 93.9 | 68.0 | 86.6 | 90.9/- | 92.2/- |92.4/- | | XLNet-Large | 95.1/89.7 | 90.6/87.9 | 90.8/- | 97.0 | 94.9 | 69.0 | 85.9 | 90.8/- | 92.3/- |92.5/- | | [DeBERTa-Large](https://huggingface.co/microsoft/deberta-large)<sup>1</sup> | 95.5/90.1 | 90.7/88.0 | 91.3/91.1| 96.5|95.3| 69.5| 91.0| 92.6/94.6| 92.3/- |92.8/92.5 | | [DeBERTa-XLarge](https://huggingface.co/microsoft/deberta-xlarge)<sup>1</sup> | -/- | -/- | 91.5/91.2| 97.0 | - | - | 93.1 | 92.1/94.3 | - |92.9/92.7| | [DeBERTa-V2-XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge)<sup>1</sup>|95.8/90.8| 91.4/88.9|91.7/91.6| **97.5**| 95.8|71.1|**93.9**|92.0/94.2|92.3/89.8|92.9/92.9| |**[DeBERTa-V2-XXLarge](https://huggingface.co/microsoft/deberta-v2-xxlarge)<sup>1,2</sup>**|**96.1/91.4**|**92.2/89.7**|**91.7/91.9**|97.2|**96.0**|**72.0**| 93.5| **93.1/94.9**|**92.7/90.3** |**93.2/93.1** | -------- #### Notes. - <sup>1</sup> Following RoBERTa, for RTE, MRPC, STS-B, we fine-tune the tasks based on [DeBERTa-Large-MNLI](https://huggingface.co/microsoft/deberta-large-mnli), [DeBERTa-XLarge-MNLI](https://huggingface.co/microsoft/deberta-xlarge-mnli), [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli), [DeBERTa-V2-XXLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli). The results of SST-2/QQP/QNLI/SQuADv2 will also be slightly improved when start from MNLI fine-tuned models, however, we only report the numbers fine-tuned from pretrained base models for those 4 tasks. - <sup>2</sup> To try the **XXLarge** model with **[HF transformers](https://huggingface.co/transformers/main_classes/trainer.html)**, you need to specify **--sharded_ddp** ```bash cd transformers/examples/text-classification/ export TASK_NAME=mrpc python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\\n--task_name $TASK_NAME --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 \\\n--learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16 ``` ### Citation If you find DeBERTa useful for your work, please cite the following paper: ``` latex @inproceedings{ he2021deberta, title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION}, author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen}, booktitle={International Conference on Learning Representations}, year={2021}, url={https://openreview.net/forum?id=XPZIaotutsD} } ```
342
NathanZhu/GabHateCorpusTrained
[ "LABEL_0", "LABEL_1" ]
Test for use in Google Colab :'(
343
NbAiLab/nb-bert-base-mnli
[ "contradiction", "neutral", "entailment" ]
--- language: no license: cc-by-4.0 thumbnail: https://raw.githubusercontent.com/NBAiLab/notram/master/images/nblogo_2.png pipeline_tag: zero-shot-classification tags: - nb-bert - zero-shot-classification - pytorch - tensorflow - norwegian - bert datasets: - mnli - multi_nli - xnli widget: - example_title: Nyhetsartikkel om FHI text: Folkehelseinstituttets mest optimistiske anslag er at alle voksne er ferdigvaksinert innen midten av september. candidate_labels: helse, politikk, sport, religion --- **Release 1.0** (March 11, 2021) # NB-Bert base model finetuned on Norwegian machine translated MNLI ## Description The most effective way of creating a good classifier is to finetune a pre-trained model for the specific task at hand. However, in many cases this is simply impossible. [Yin et al.](https://arxiv.org/abs/1909.00161) proposed a very clever way of using pre-trained MNLI models as zero-shot sequence classifiers. The methods works by reformulating the question to an MNLI hypothesis. If we want to figure out if a text is about "sport", we simply state that "This text is about sport" ("Denne teksten handler om sport"). When the model is finetuned on the 400k large MNLI task, it is in many cases able to solve this classification tasks. There are no MNLI-set of this size in Norwegian but we have trained it on a machine translated version of the original MNLI-set. ## Testing the model For testing the model, we recommend the [NbAiLab Colab Notebook](https://colab.research.google.com/gist/peregilk/769b5150a2f807219ab8f15dd11ea449/nbailab-mnli-norwegian-demo.ipynb) ## Hugging Face zero-shot-classification pipeline The easiest way to try this out is by using the Hugging Face pipeline. Please, note that you will get better results when using Norwegian hypothesis template instead of the default English one. ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model="NbAiLab/nb-bert-base-mnli") ``` You can then use this pipeline to classify sequences into any of the class names you specify. ```python sequence_to_classify = 'Folkehelseinstituttets mest optimistiske anslag er at alle voksne er ferdigvaksinert innen midten av september.' candidate_labels = ['politikk', 'helse', 'sport', 'religion'] hypothesis_template = 'Dette eksempelet er {}.' classifier(sequence_to_classify, candidate_labels, hypothesis_template=hypothesis_template, multi_class=True) # {'labels': ['helse', 'politikk', 'sport', 'religion'], # 'scores': [0.4210019111633301, 0.0674605593085289, 0.000840459018945694, 0.0007541406666859984], # 'sequence': 'Folkehelseinstituttets mest optimistiske anslag er at alle over 18 år er ferdigvaksinert innen midten av september.'} ``` ## More information For more information on the model, see https://github.com/NBAiLab/notram Here you will also find a Colab explaining more in details how to use the zero-shot-classification pipeline.
346
NikolajMunch/danish-emotion-classification
[ "Afsky", "Frygt", "Glæde", "Overraskelse", "Tristhed", "Vrede" ]
--- widget: - text: "Hold da op! Kan det virkelig passe?" language: - "da" tags: - sentiment - emotion - danish --- # **-- EMODa --** ## BERT-model for danish multi-class classification of emotions Classifies a danish sentence into one of 6 different emotions: | Danish emotion | Ekman's emotion | | ----- | ----- | | 😞 **Afsky** | Disgust | | 😨 **Frygt** | Fear | | 😄 **Glæde** | Joy | | 😱 **Overraskelse** | Surprise | | 😢 **Tristhed** | Sadness | | 😠 **Vrede** | Anger | # How to use ```python from transformers import pipeline model_path = "NikolajMunch/danish-emotion-classification" classifier = pipeline("sentiment-analysis", model=model_path, tokenizer=model_path) prediction = classifier("Jeg er godt nok ked af at mine SMS'er er slettet") print(prediction) # [{'label': 'Tristhed', 'score': 0.9725030660629272}] ``` or ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("NikolajMunch/danish-emotion-classification") model = AutoModelForSequenceClassification.from_pretrained("NikolajMunch/danish-emotion-classification") ```
347
Noxel/sentiments_multilenguaje
[ "1 star", "2 stars", "3 stars", "4 stars", "5 stars" ]
--- language: - en - nl - de - fr - it - es license: mit --- # bert-base-multilingual-uncased-sentiment This a bert-base-multilingual-uncased model finetuned for sentiment analysis on product reviews in six languages: English, Dutch, German, French, Spanish and Italian. It predicts the sentiment of the review as a number of stars (between 1 and 5). This model is intended for direct use as a sentiment analysis model for product reviews in any of the six languages above, or for further finetuning on related sentiment analysis tasks. ## Training data Here is the number of product reviews we used for finetuning the model: | Language | Number of reviews | | -------- | ----------------- | | English | 150k | | Dutch | 80k | | German | 137k | | French | 140k | | Italian | 72k | | Spanish | 50k | ## Accuracy The finetuned model obtained the following accuracy on 5,000 held-out product reviews in each of the languages: - Accuracy (exact) is the exact match on the number of stars. - Accuracy (off-by-1) is the percentage of reviews where the number of stars the model predicts differs by a maximum of 1 from the number given by the human reviewer. | Language | Accuracy (exact) | Accuracy (off-by-1) | | -------- | ---------------------- | ------------------- | | English | 67% | 95% | Dutch | 57% | 93% | German | 61% | 94% | French | 59% | 94% | Italian | 59% | 95% | Spanish | 58% | 95% ## Contact In addition to this model, [NLP Town](https://www.nlp.town) offers custom, monolingual sentiment models for many languages and an improved multilingual model through [RapidAPI](https://rapidapi.com/nlp-town-nlp-town-default/api/multilingual-sentiment-analysis2/). Feel free to contact us for questions, feedback and/or requests for similar models.
348
Omar95farag/distilbert-base-uncased-distilled-clinc
[ "accept_reservations", "account_blocked", "alarm", "application_status", "apr", "are_you_a_bot", "balance", "bill_balance", "bill_due", "book_flight", "book_hotel", "calculator", "calendar", "calendar_update", "calories", "cancel", "cancel_reservation", "car_rental", "card_declined", "carry_on", "change_accent", "change_ai_name", "change_language", "change_speed", "change_user_name", "change_volume", "confirm_reservation", "cook_time", "credit_limit", "credit_limit_change", "credit_score", "current_location", "damaged_card", "date", "definition", "direct_deposit", "directions", "distance", "do_you_have_pets", "exchange_rate", "expiration_date", "find_phone", "flight_status", "flip_coin", "food_last", "freeze_account", "fun_fact", "gas", "gas_type", "goodbye", "greeting", "how_busy", "how_old_are_you", "improve_credit_score", "income", "ingredient_substitution", "ingredients_list", "insurance", "insurance_change", "interest_rate", "international_fees", "international_visa", "jump_start", "last_maintenance", "lost_luggage", "make_call", "maybe", "meal_suggestion", "meaning_of_life", "measurement_conversion", "meeting_schedule", "min_payment", "mpg", "new_card", "next_holiday", "next_song", "no", "nutrition_info", "oil_change_how", "oil_change_when", "oos", "order", "order_checks", "order_status", "pay_bill", "payday", "pin_change", "play_music", "plug_type", "pto_balance", "pto_request", "pto_request_status", "pto_used", "recipe", "redeem_rewards", "reminder", "reminder_update", "repeat", "replacement_card_duration", "report_fraud", "report_lost_card", "reset_settings", "restaurant_reservation", "restaurant_reviews", "restaurant_suggestion", "rewards_balance", "roll_dice", "rollover_401k", "routing", "schedule_maintenance", "schedule_meeting", "share_location", "shopping_list", "shopping_list_update", "smart_home", "spelling", "spending_history", "sync_device", "taxes", "tell_joke", "text", "thank_you", "time", "timer", "timezone", "tire_change", "tire_pressure", "todo_list", "todo_list_update", "traffic", "transactions", "transfer", "translate", "travel_alert", "travel_notification", "travel_suggestion", "uber", "update_playlist", "user_name", "vaccines", "w2", "weather", "what_are_your_hobbies", "what_can_i_ask_you", "what_is_your_name", "what_song", "where_are_you_from", "whisper_mode", "who_do_you_work_for", "who_made_you", "yes" ]
--- license: apache-2.0 tags: - generated_from_trainer datasets: - clinc_oos metrics: - accuracy model-index: - name: distilbert-base-uncased-distilled-clinc results: - task: name: Text Classification type: text-classification dataset: name: clinc_oos type: clinc_oos args: plus metrics: - name: Accuracy type: accuracy value: 0.9332258064516129 - task: type: text-classification name: Text Classification dataset: name: clinc_oos type: clinc_oos config: small split: test metrics: - name: Accuracy type: accuracy value: 0.8587272727272727 verified: true - name: Precision Macro type: precision value: 0.8619245385984416 verified: true - name: Precision Micro type: precision value: 0.8587272727272727 verified: true - name: Precision Weighted type: precision value: 0.8797945801452213 verified: true - name: Recall Macro type: recall value: 0.9359690949227375 verified: true - name: Recall Micro type: recall value: 0.8587272727272727 verified: true - name: Recall Weighted type: recall value: 0.8587272727272727 verified: true - name: F1 Macro type: f1 value: 0.8922503214655346 verified: true - name: F1 Micro type: f1 value: 0.8587272727272727 verified: true - name: F1 Weighted type: f1 value: 0.8506829426037475 verified: true - name: loss type: loss value: 0.9798759818077087 verified: true --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-distilled-clinc This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the clinc_oos dataset. It achieves the following results on the evaluation set: - Loss: 0.1259 - Accuracy: 0.9332 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 48 - eval_batch_size: 48 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 7 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 318 | 0.5952 | 0.7355 | | 0.7663 | 2.0 | 636 | 0.3130 | 0.8742 | | 0.7663 | 3.0 | 954 | 0.2024 | 0.9206 | | 0.3043 | 4.0 | 1272 | 0.1590 | 0.9235 | | 0.181 | 5.0 | 1590 | 0.1378 | 0.9303 | | 0.181 | 6.0 | 1908 | 0.1287 | 0.9329 | | 0.1468 | 7.0 | 2226 | 0.1259 | 0.9332 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.2+cu102 - Datasets 1.18.3 - Tokenizers 0.11.0
355
Pratibha/xlm-roberta-base-finetuned-marc-en
[ "good", "great", "ok", "poor", "terrible" ]
--- license: mit tags: - generated_from_trainer datasets: - amazon_reviews_multi model-index: - name: xlm-roberta-base-finetuned-marc-en results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-marc-en This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.9575 - Mae: 0.5488 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Mae | |:-------------:|:-----:|:----:|:---------------:|:------:| | 1.1253 | 1.0 | 235 | 0.9960 | 0.5366 | | 0.9708 | 2.0 | 470 | 0.9575 | 0.5488 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.14.0 - Tokenizers 0.10.3
357
Prompsit/paraphrase-bert-en
[ "Not Paraphrase", "Paraphrase" ]
--- pipeline_tag: text-classification inference: false language: en tags: - transformers --- # Prompsit/paraphrase-bert-en This model allows to evaluate paraphrases for a given phrase. We have fine-tuned this model from pretrained "bert-base-uncased". Model built under a TSI-100905-2019-4 project, co-financed by Ministry of Economic Affairs and Digital Transformation from the Government of Spain. # How to use it The model answer the following question: Is "phrase B" a paraphrase of "phrase A". Please note that we're considering phrases instead of sentences. Therefore, we must take into account that the model doesn't expect to find punctuation marks or long pieces of text. Resulting probabilities correspond to classes: * 0: Not a paraphrase * 1: It's a paraphrase So, considering the phrase "may be addressed" and a candidate paraphrase like "could be included", you can use the model like this: ``` import torch from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("Prompsit/paraphrase-bert-en") model = AutoModelForSequenceClassification.from_pretrained("Prompsit/paraphrase-bert-en") input = tokenizer('may be addressed','could be included',return_tensors='pt') logits = model(**input).logits soft = torch.nn.Softmax(dim=1) print(soft(logits)) ``` Code output is: ``` tensor([[0.1592, 0.8408]], grad_fn=<SoftmaxBackward>) ``` As the probability of 1 (=It's a paraphrase) is 0.84 and the probability of 0 (=It is not a paraphrase) is 0.15, we can conclude, for our previous example, that "could be included" is a paraphrase of "may be addressed". # Evaluation results We have used as test dataset 16500 pairs of phrases human tagged. Metrics obtained are: ``` metrics={ 'test_loss': 0.5660144090652466, 'test_accuracy': 0.8170742794799527, 'test_precision': 0.7043977055449331, 'test_recall': 0.5978578383641675, 'test_f1': 0.6467696629213483, 'test_matthews_correlation': 0.5276716223607356, 'test_runtime': 19.3345, 'test_samples_per_second': 568.88, 'test_steps_per_second': 17.792 } ```
358
Prompsit/paraphrase-bert-pt
[ "Not Paraphrase", "Paraphrase" ]
--- pipeline_tag: text-classification inference: false language: pt tags: - transformers --- # Prompsit/paraphrase-bert-pt This model allows to evaluate paraphrases for a given phrase. We have fine-tuned this model from pretrained "neuralmind/bert-base-portuguese-cased". Model built under a TSI-100905-2019-4 project, co-financed by Ministry of Economic Affairs and Digital Transformation from the Government of Spain. # How to use it The model answer the following question: Is "phrase B" a paraphrase of "phrase A". Please note that we're considering phrases instead of sentences. Therefore, we must take into account that the model doesn't expect to find punctuation marks or long pieces of text. Resulting probabilities correspond to classes: * 0: Not a paraphrase * 1: It's a paraphrase So, considering the phrase "logo após o homicídio" and a candidate paraphrase like "pouco depois do assassinato", you can use the model like this: ``` import torch from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("Prompsit/paraphrase-bert-pt") model = AutoModelForSequenceClassification.from_pretrained("Prompsit/paraphrase-bert-pt") input = tokenizer('logo após o homicídio','pouco depois do assassinato',return_tensors='pt') logits = model(**input).logits soft = torch.nn.Softmax(dim=1) print(soft(logits)) ``` Code output is: ``` tensor([[0.2137, 0.7863]], grad_fn=<SoftmaxBackward>) ``` As the probability of 1 (=It's a paraphrase) is 0.7863 and the probability of 0 (=It is not a paraphrase) is 0.2137, we can conclude, for our previous example, that "pouco depois do assassinato" is a paraphrase of "logo após o homicidio". # Evaluation results We have used as test dataset 16500 pairs of phrases human tagged. Metrics obtained are: ``` metrics={ 'test_loss': 0.6074697375297546, 'test_accuracy': 0.7809, 'test_precision': 0.7157638466220329, 'test_recall': 0.40551724137931033, 'test_f1': 0.5177195685670262, 'test_matthews_correlation': 0.41603913834665324, 'test_runtime': 16.4585, 'test_samples_per_second': 607.587, 'test_steps_per_second': 19.017 } ```
359
Prompsit/paraphrase-roberta-es
[ "Not Paraphrase", "Paraphrase" ]
--- pipeline_tag: text-classification inference: false language: es tags: - transformers --- # Prompsit/paraphrase-roberta-es This model allows to evaluate paraphrases for a given phrase. We have fine-tuned this model from pretrained "PlanTL-GOB-ES/roberta-base-bne". Model built under a TSI-100905-2019-4 project, co-financed by Ministry of Economic Affairs and Digital Transformation from the Government of Spain. # How to use it The model answer the following question: Is "phrase B" a paraphrase of "phrase A". Please note that we're considering phrases instead of sentences. Therefore, we must take into account that the model doesn't expect to find punctuation marks or long pieces of text. Resulting probabilities correspond to classes: * 0: Not a paraphrase * 1: It's a paraphrase So, considering the phrase "se buscarán acuerdos" and a candidate paraphrase like "se deberá obtener el acuerdo", you can use the model like this: ``` import torch from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("Prompsit/paraphrase-roberta-es") model = AutoModelForSequenceClassification.from_pretrained("Prompsit/paraphrase-roberta-es") input = tokenizer('se buscarán acuerdos','se deberá obtener el acuerdo',return_tensors='pt') logits = model(**input).logits soft = torch.nn.Softmax(dim=1) print(soft(logits)) ``` Code output is: ``` tensor([[0.2266, 0.7734]], grad_fn=<SoftmaxBackward>) ``` As the probability of 1 (=It's a paraphrase) is 0.77 and the probability of 0 (=It is not a paraphrase) is 0.22, we can conclude, for our previous example, that "se deberá obtener el acuerdo" is a paraphrase of "se buscarán acuerdos". # Evaluation results We have used as test dataset 16500 pairs of phrases human tagged. Metrics obtained are: ``` metrics={ 'test_loss': 0.4869941473007202, 'test_accuracy': 0.8003636363636364, 'test_precision': 0.6692456479690522, 'test_recall': 0.5896889646357052, 'test_f1': 0.6269535673839184, 'test_matthews_correlation': 0.49324489316659575, 'test_runtime': 27.1537, 'test_samples_per_second': 607.652, 'test_steps_per_second': 19.003 } ```
360
ProsusAI/finbert
[ "positive", "negative", "neutral" ]
--- language: "en" tags: - financial-sentiment-analysis - sentiment-analysis widget: - text: "Stocks rallied and the British pound gained." --- FinBERT is a pre-trained NLP model to analyze sentiment of financial text. It is built by further training the BERT language model in the finance domain, using a large financial corpus and thereby fine-tuning it for financial sentiment classification. [Financial PhraseBank](https://www.researchgate.net/publication/251231107_Good_Debt_or_Bad_Debt_Detecting_Semantic_Orientations_in_Economic_Texts) by Malo et al. (2014) is used for fine-tuning. For more details, please see the paper [FinBERT: Financial Sentiment Analysis with Pre-trained Language Models](https://arxiv.org/abs/1908.10063) and our related [blog post](https://medium.com/prosus-ai-tech-blog/finbert-financial-sentiment-analysis-with-bert-b277a3607101) on Medium. The model will give softmax outputs for three labels: positive, negative or neutral. --- About Prosus Prosus is a global consumer internet group and one of the largest technology investors in the world. Operating and investing globally in markets with long-term growth potential, Prosus builds leading consumer internet companies that empower people and enrich communities. For more information, please visit www.prosus.com. Contact information Please contact Dogu Araci dogu.araci[at]prosus[dot]com and Zulkuf Genc zulkuf.genc[at]prosus[dot]com about any FinBERT related issues and questions.
361
Qinghui/autonlp-fake-covid-news-36769078
[ "0", "1" ]
--- tags: autonlp language: unk widget: - text: "I love AutoNLP 🤗" datasets: - Qinghui/autonlp-data-fake-covid-news co2_eq_emissions: 23.42719853096565 --- # Model Trained Using AutoNLP - Problem type: Binary Classification - Model ID: 36769078 - CO2 Emissions (in grams): 23.42719853096565 ## Validation Metrics - Loss: 0.15959647297859192 - Accuracy: 0.9817757009345794 - Precision: 0.980411361410382 - Recall: 0.9813725490196078 - AUC: 0.9982379201680672 - F1: 0.9808917197452229 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/Qinghui/autonlp-fake-covid-news-36769078 ``` Or Python API: ``` from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("Qinghui/autonlp-fake-covid-news-36769078", use_auth_token=True) tokenizer = AutoTokenizer.from_pretrained("Qinghui/autonlp-fake-covid-news-36769078", use_auth_token=True) inputs = tokenizer("I love AutoNLP", return_tensors="pt") outputs = model(**inputs) ```
363
Recognai/bert-base-spanish-wwm-cased-xnli
[ "contradiction", "neutral", "entailment" ]
--- language: es tags: - zero-shot-classification - nli - pytorch datasets: - xnli license: mit pipeline_tag: zero-shot-classification widget: - text: "El autor se perfila, a los 50 años de su muerte, como uno de los grandes de su siglo" candidate_labels: "cultura, sociedad, economia, salud, deportes" --- # bert-base-spanish-wwm-cased-xnli **UPDATE, 15.10.2021: Check out our new zero-shot classifiers, much more lightweight and even outperforming this one: [zero-shot SELECTRA small](https://huggingface.co/Recognai/zeroshot_selectra_small) and [zero-shot SELECTRA medium](https://huggingface.co/Recognai/zeroshot_selectra_medium).** ## Model description This model is a fine-tuned version of the [spanish BERT model](https://huggingface.co/dccuchile/bert-base-spanish-wwm-cased) with the Spanish portion of the XNLI dataset. You can have a look at the [training script](https://huggingface.co/Recognai/bert-base-spanish-wwm-cased-xnli/blob/main/zeroshot_training_script.py) for details of the training. ### How to use You can use this model with Hugging Face's [zero-shot-classification pipeline](https://discuss.huggingface.co/t/new-pipeline-for-zero-shot-text-classification/681): ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model="Recognai/bert-base-spanish-wwm-cased-xnli") classifier( "El autor se perfila, a los 50 años de su muerte, como uno de los grandes de su siglo", candidate_labels=["cultura", "sociedad", "economia", "salud", "deportes"], hypothesis_template="Este ejemplo es {}." ) """output {'sequence': 'El autor se perfila, a los 50 años de su muerte, como uno de los grandes de su siglo', 'labels': ['cultura', 'sociedad', 'economia', 'salud', 'deportes'], 'scores': [0.38897448778152466, 0.22997373342514038, 0.1658431738615036, 0.1205764189362526, 0.09463217109441757]} """ ``` ## Eval results Accuracy for the test set: | | XNLI-es | |-----------------------------|---------| |bert-base-spanish-wwm-cased-xnli | 79.9% |
364
Recognai/zeroshot_selectra_medium
[ "contradiction", "neutral", "entailment" ]
--- language: es tags: - zero-shot-classification - nli - pytorch datasets: - xnli pipeline_tag: zero-shot-classification license: apache-2.0 widget: - text: "El autor se perfila, a los 50 años de su muerte, como uno de los grandes de su siglo" candidate_labels: "cultura, sociedad, economia, salud, deportes" --- # Zero-shot SELECTRA: A zero-shot classifier based on SELECTRA *Zero-shot SELECTRA* is a [SELECTRA model](https://huggingface.co/Recognai/selectra_small) fine-tuned on the Spanish portion of the [XNLI dataset](https://huggingface.co/datasets/xnli). You can use it with Hugging Face's [Zero-shot pipeline](https://huggingface.co/transformers/master/main_classes/pipelines.html#transformers.ZeroShotClassificationPipeline) to make [zero-shot classifications](https://joeddav.github.io/blog/2020/05/29/ZSL.html). In comparison to our previous zero-shot classifier [based on BETO](https://huggingface.co/Recognai/bert-base-spanish-wwm-cased-xnli), zero-shot SELECTRA is **much more lightweight**. As shown in the *Metrics* section, the *small* version (5 times fewer parameters) performs slightly worse, while the *medium* version (3 times fewer parameters) **outperforms** the BETO based zero-shot classifier. ## Usage ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model="Recognai/zeroshot_selectra_medium") classifier( "El autor se perfila, a los 50 años de su muerte, como uno de los grandes de su siglo", candidate_labels=["cultura", "sociedad", "economia", "salud", "deportes"], hypothesis_template="Este ejemplo es {}." ) """Output {'sequence': 'El autor se perfila, a los 50 años de su muerte, como uno de los grandes de su siglo', 'labels': ['sociedad', 'cultura', 'economia', 'salud', 'deportes'], 'scores': [0.6450043320655823, 0.16710571944713593, 0.08507631719112396, 0.0759836807847023, 0.026829993352293968]} """ ``` The `hypothesis_template` parameter is important and should be in Spanish. **In the widget on the right, this parameter is set to its default value: "This example is {}.", so different results are expected.** ## Demo and tutorial If you want to see this model in action, we have created a basic tutorial using [Rubrix](https://www.rubrix.ml/), a free and open-source tool to *explore, annotate, and monitor data for NLP*. The tutorial shows you how to evaluate this classifier for news categorization in Spanish, and how it could be used to build a training set for training a supervised classifier (which might be useful if you want obtain more precise results or improve the model over time). You can [find the tutorial here](https://rubrix.readthedocs.io/en/master/tutorials/zeroshot_data_annotation.html). See the video below showing the predictions within the annotation process (see that the predictions are almost correct for every example). <video width="100%" controls><source src="https://github.com/recognai/rubrix-materials/raw/main/tutorials/videos/zeroshot_selectra_news_data_annotation.mp4" type="video/mp4"></video> ## Metrics | Model | Params | XNLI (acc) | \*MLSUM (acc) | | --- | --- | --- | --- | | [zs BETO](https://huggingface.co/Recognai/bert-base-spanish-wwm-cased-xnli) | 110M | 0.799 | 0.530 | | zs SELECTRA medium | 41M | **0.807** | **0.589** | | [zs SELECTRA small](https://huggingface.co/Recognai/zeroshot_selectra_small) | **22M** | 0.795 | 0.446 | \*evaluated with zero-shot learning (ZSL) - **XNLI**: The stated accuracy refers to the test portion of the [XNLI dataset](https://huggingface.co/datasets/xnli), after finetuning the model on the training portion. - **MLSUM**: For this accuracy we take the test set of the [MLSUM dataset](https://huggingface.co/datasets/mlsum) and classify the summaries of 5 selected labels. For details, check out our [evaluation notebook](https://github.com/recognai/selectra/blob/main/zero-shot_classifier/evaluation.ipynb) ## Training Check out our [training notebook](https://github.com/recognai/selectra/blob/main/zero-shot_classifier/training.ipynb) for all the details. ## Authors - David Fidalgo ([GitHub](https://github.com/dcfidalgo)) - Daniel Vila ([GitHub](https://github.com/dvsrepo)) - Francisco Aranda ([GitHub](https://github.com/frascuchon)) - Javier Lopez ([GitHub](https://github.com/javispp))
365
Recognai/zeroshot_selectra_small
[ "contradiction", "neutral", "entailment" ]
--- language: es tags: - zero-shot-classification - nli - pytorch datasets: - xnli pipeline_tag: zero-shot-classification license: apache-2.0 widget: - text: "El autor se perfila, a los 50 años de su muerte, como uno de los grandes de su siglo" candidate_labels: "cultura, sociedad, economia, salud, deportes" --- # Zero-shot SELECTRA: A zero-shot classifier based on SELECTRA *Zero-shot SELECTRA* is a [SELECTRA model](https://huggingface.co/Recognai/selectra_small) fine-tuned on the Spanish portion of the [XNLI dataset](https://huggingface.co/datasets/xnli). You can use it with Hugging Face's [Zero-shot pipeline](https://huggingface.co/transformers/master/main_classes/pipelines.html#transformers.ZeroShotClassificationPipeline) to make [zero-shot classifications](https://joeddav.github.io/blog/2020/05/29/ZSL.html). In comparison to our previous zero-shot classifier [based on BETO](https://huggingface.co/Recognai/bert-base-spanish-wwm-cased-xnli), zero-shot SELECTRA is **much more lightweight**. As shown in the *Metrics* section, the *small* version (5 times fewer parameters) performs slightly worse, while the *medium* version (3 times fewer parameters) **outperforms** the BETO based zero-shot classifier. ## Usage ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model="Recognai/zeroshot_selectra_medium") classifier( "El autor se perfila, a los 50 años de su muerte, como uno de los grandes de su siglo", candidate_labels=["cultura", "sociedad", "economia", "salud", "deportes"], hypothesis_template="Este ejemplo es {}." ) """Output {'sequence': 'El autor se perfila, a los 50 años de su muerte, como uno de los grandes de su siglo', 'labels': ['sociedad', 'cultura', 'salud', 'economia', 'deportes'], 'scores': [0.3711881935596466, 0.25650349259376526, 0.17355826497077942, 0.1641489565372467, 0.03460107371211052]} """ ``` The `hypothesis_template` parameter is important and should be in Spanish. **In the widget on the right, this parameter is set to its default value: "This example is {}.", so different results are expected.** ## Metrics | Model | Params | XNLI (acc) | \*MLSUM (acc) | | --- | --- | --- | --- | | [zs BETO](https://huggingface.co/Recognai/bert-base-spanish-wwm-cased-xnli) | 110M | 0.799 | 0.530 | | [zs SELECTRA medium](https://huggingface.co/Recognai/zeroshot_selectra_medium) | 41M | **0.807** | **0.589** | | zs SELECTRA small | **22M** | 0.795 | 0.446 | \*evaluated with zero-shot learning (ZSL) - **XNLI**: The stated accuracy refers to the test portion of the [XNLI dataset](https://huggingface.co/datasets/xnli), after finetuning the model on the training portion. - **MLSUM**: For this accuracy we take the test set of the [MLSUM dataset](https://huggingface.co/datasets/mlsum) and classify the summaries of 5 selected labels. For details, check out our [evaluation notebook](https://github.com/recognai/selectra/blob/main/zero-shot_classifier/evaluation.ipynb) ## Training Check out our [training notebook](https://github.com/recognai/selectra/blob/main/zero-shot_classifier/training.ipynb) for all the details. ## Authors - David Fidalgo ([GitHub](https://github.com/dcfidalgo)) - Daniel Vila ([GitHub](https://github.com/dvsrepo)) - Francisco Aranda ([GitHub](https://github.com/frascuchon)) - Javier Lopez ([GitHub](https://github.com/javispp))
366
RecordedFuture/Swedish-Sentiment-Fear
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
--- language: sv license: mit --- ## Swedish BERT models for sentiment analysis [Recorded Future](https://www.recordedfuture.com/) together with [AI Sweden](https://www.ai.se/en) releases two language models for sentiment analysis in Swedish. The two models are based on the [KB\/bert-base-swedish-cased](https://huggingface.co/KB/bert-base-swedish-cased) model and has been fine-tuned to solve a multi-label sentiment analysis task. The models have been fine-tuned for the sentiments fear and violence. The models output three floats corresponding to the labels "Negative", "Weak sentiment", and "Strong Sentiment" at the respective indexes. The models have been trained on Swedish data with a conversational focus, collected from various internet sources and forums. The models are only trained on Swedish data and only supports inference of Swedish input texts. The models inference metrics for all non-Swedish inputs are not defined, these inputs are considered as out of domain data. The current models are supported at Transformers version >= 4.3.3 and Torch version 1.8.0, compatibility with older versions are not verified. ### Swedish-Sentiment-Fear The model can be imported from the transformers library by running from transformers import BertForSequenceClassification, BertTokenizerFast tokenizer = BertTokenizerFast.from_pretrained("RecordedFuture/Swedish-Sentiment-Fear") classifier_fear= BertForSequenceClassification.from_pretrained("RecordedFuture/Swedish-Sentiment-Fear") When the model and tokenizer are initialized the model can be used for inference. #### Sentiment definitions #### The strong sentiment includes but are not limited to Texts that: - Hold an expressive emphasis on fear and/ or anxiety #### The weak sentiment includes but are not limited to Texts that: - Express fear and/ or anxiety in a neutral way #### Verification metrics During training, the model had maximized validation metrics at the following classification breakpoint. | Classification Breakpoint | F-score | Precision | Recall | |:-------------------------:|:-------:|:---------:|:------:| | 0.45 | 0.8754 | 0.8618 | 0.8895 | #### Swedish-Sentiment-Violence The model be can imported from the transformers library by running from transformers import BertForSequenceClassification, BertTokenizerFast tokenizer = BertTokenizerFast.from_pretrained("RecordedFuture/Swedish-Sentiment-Violence") classifier_violence = BertForSequenceClassification.from_pretrained("RecordedFuture/Swedish-Sentiment-Violence") When the model and tokenizer are initialized the model can be used for inference. ### Sentiment definitions #### The strong sentiment includes but are not limited to Texts that: - Referencing highly violent acts - Hold an aggressive tone #### The weak sentiment includes but are not limited to Texts that: - Include general violent statements that do not fall under the strong sentiment #### Verification metrics During training, the model had maximized validation metrics at the following classification breakpoint. | Classification Breakpoint | F-score | Precision | Recall | |:-------------------------:|:-------:|:---------:|:------:| | 0.35 | 0.7677 | 0.7456 | 0.791 |
367
RecordedFuture/Swedish-Sentiment-Violence
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
--- language: sv license: mit --- ## Swedish BERT models for sentiment analysis [Recorded Future](https://www.recordedfuture.com/) together with [AI Sweden](https://www.ai.se/en) releases two language models for sentiment analysis in Swedish. The two models are based on the [KB\/bert-base-swedish-cased](https://huggingface.co/KB/bert-base-swedish-cased) model and has been fine-tuned to solve a multi-label sentiment analysis task. The models have been fine-tuned for the sentiments fear and violence. The models output three floats corresponding to the labels "Negative", "Weak sentiment", and "Strong Sentiment" at the respective indexes. The models have been trained on Swedish data with a conversational focus, collected from various internet sources and forums. The models are only trained on Swedish data and only supports inference of Swedish input texts. The models inference metrics for all non-Swedish inputs are not defined, these inputs are considered as out of domain data. The current models are supported at Transformers version >= 4.3.3 and Torch version 1.8.0, compatibility with older versions are not verified. ### Swedish-Sentiment-Fear The model can be imported from the transformers library by running from transformers import BertForSequenceClassification, BertTokenizerFast tokenizer = BertTokenizerFast.from_pretrained("RecordedFuture/Swedish-Sentiment-Fear") classifier_fear= BertForSequenceClassification.from_pretrained("RecordedFuture/Swedish-Sentiment-Fear") When the model and tokenizer are initialized the model can be used for inference. #### Sentiment definitions #### The strong sentiment includes but are not limited to Texts that: - Hold an expressive emphasis on fear and/ or anxiety #### The weak sentiment includes but are not limited to Texts that: - Express fear and/ or anxiety in a neutral way #### Verification metrics During training, the model had maximized validation metrics at the following classification breakpoint. | Classification Breakpoint | F-score | Precision | Recall | |:-------------------------:|:-------:|:---------:|:------:| | 0.45 | 0.8754 | 0.8618 | 0.8895 | #### Swedish-Sentiment-Violence The model be can imported from the transformers library by running from transformers import BertForSequenceClassification, BertTokenizerFast tokenizer = BertTokenizerFast.from_pretrained("RecordedFuture/Swedish-Sentiment-Violence") classifier_violence = BertForSequenceClassification.from_pretrained("RecordedFuture/Swedish-Sentiment-Violence") When the model and tokenizer are initialized the model can be used for inference. ### Sentiment definitions #### The strong sentiment includes but are not limited to Texts that: - Referencing highly violent acts - Hold an aggressive tone #### The weak sentiment includes but are not limited to Texts that: - Include general violent statements that do not fall under the strong sentiment #### Verification metrics During training, the model had maximized validation metrics at the following classification breakpoint. | Classification Breakpoint | F-score | Precision | Recall | |:-------------------------:|:-------:|:---------:|:------:| | 0.35 | 0.7677 | 0.7456 | 0.791 |
368
Rexhaif/rubert-base-srl
[ "инструмент", "каузатор", "экспериенцер" ]
--- tags: - generated_from_trainer metrics: - f1 model-index: - name: rubert-base-srl results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # rubert-base-srl This model is a fine-tuned version of [./ruBert-base/](https://huggingface.co/./ruBert-base/) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.2429 - F1: 0.9563 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.98) and epsilon=1e-06 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.06 - num_epochs: 10.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.5816 | 1.0 | 57 | 0.3865 | 0.8371 | | 0.3685 | 2.0 | 114 | 0.1707 | 0.9325 | | 0.1057 | 3.0 | 171 | 0.0972 | 0.9563 | | 0.0964 | 4.0 | 228 | 0.1429 | 0.9775 | | 0.1789 | 5.0 | 285 | 0.2493 | 0.9457 | | 0.0016 | 6.0 | 342 | 0.1900 | 0.6349 | | 0.0013 | 7.0 | 399 | 0.2060 | 0.9563 | | 0.0008 | 8.0 | 456 | 0.2321 | 0.9563 | | 0.0006 | 9.0 | 513 | 0.2412 | 0.9563 | | 0.0006 | 10.0 | 570 | 0.2429 | 0.9563 | ### Framework versions - Transformers 4.13.0.dev0 - Pytorch 1.10.0+cu102 - Datasets 1.15.1 - Tokenizers 0.10.3
369
Riad/finetuned-bert-mrpc
[ "equivalent", "not equivalent" ]
--- license: apache-2.0 tags: - generated_from_trainer datasets: - glue metrics: - accuracy - f1 model-index: - name: finetuned-bert-mrpc results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue args: mrpc metrics: - name: Accuracy type: accuracy value: 0.8676470588235294 - name: F1 type: f1 value: 0.9084745762711864 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuned-bert-mrpc This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.4382 - Accuracy: 0.8676 - F1: 0.9085 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.5454 | 1.0 | 230 | 0.4396 | 0.8309 | 0.8871 | | 0.3387 | 2.0 | 460 | 0.3783 | 0.8529 | 0.8976 | | 0.1956 | 3.0 | 690 | 0.4382 | 0.8676 | 0.9085 | ### Framework versions - Transformers 4.10.0 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3
370
Roberta55/deberta-base-mnli-finetuned-cola
[ "CONTRADICTION", "ENTAILMENT", "NEUTRAL" ]
--- license: mit tags: - generated_from_trainer datasets: - glue metrics: - matthews_correlation model-index: - name: deberta-base-mnli-finetuned-cola results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue args: cola metrics: - name: Matthews Correlation type: matthews_correlation value: 0.6281691768918801 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-base-mnli-finetuned-cola This model is a fine-tuned version of [microsoft/deberta-base-mnli](https://huggingface.co/microsoft/deberta-base-mnli) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.8205 - Matthews Correlation: 0.6282 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.4713 | 1.0 | 535 | 0.5110 | 0.5797 | | 0.2678 | 2.0 | 1070 | 0.6648 | 0.5154 | | 0.1811 | 3.0 | 1605 | 0.6681 | 0.6121 | | 0.113 | 4.0 | 2140 | 0.8205 | 0.6282 | | 0.0831 | 5.0 | 2675 | 1.0413 | 0.6057 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.14.0 - Tokenizers 0.10.3
371
Rocketknight1/distilbert-base-uncased-finetuned-cola
[ "Invalid", "Valid" ]
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: Rocketknight1/distilbert-base-uncased-finetuned-cola results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # Rocketknight1/distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.3182 - Validation Loss: 0.4914 - Train Matthews Correlation: 0.5056 - Epoch: 1 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 1602, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False} - training_precision: float32 ### Training results | Train Loss | Validation Loss | Train Matthews Correlation | Epoch | |:----------:|:---------------:|:--------------------------:|:-----:| | 0.5126 | 0.4638 | 0.4555 | 0 | | 0.3182 | 0.4914 | 0.5056 | 1 | ### Framework versions - Transformers 4.22.0.dev0 - TensorFlow 2.9.1 - Datasets 2.4.1.dev0 - Tokenizers 0.11.0
375
SEISHIN/distilbert-base-uncased-finetuned-mnli
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
--- license: apache-2.0 tags: - generated_from_trainer datasets: - glue metrics: - accuracy model-index: - name: distilbert-base-uncased-finetuned-mnli results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue args: mnli metrics: - name: Accuracy type: accuracy value: 0.82190524707081 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-mnli This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.6560 - Accuracy: 0.8219 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:------:|:---------------:|:--------:| | 0.5161 | 1.0 | 24544 | 0.5025 | 0.8037 | | 0.4176 | 2.0 | 49088 | 0.5274 | 0.8131 | | 0.3154 | 3.0 | 73632 | 0.5348 | 0.8194 | | 0.2294 | 4.0 | 98176 | 0.6560 | 0.8219 | | 0.1827 | 5.0 | 122720 | 0.8190 | 0.8203 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
378
Sahajtomar/German_Zeroshot
[ "entailment", "neutral", "contradiction" ]
--- language: multilingual tags: - text-classification - pytorch - nli - xnli - de datasets: - xnli pipeline_tag: zero-shot-classification widget: - text: "Letzte Woche gab es einen Selbstmord in einer nahe gelegenen kolonie" candidate_labels: "Verbrechen,Tragödie,Stehlen" hypothesis_template: "In deisem geht es um {}." --- # German Zeroshot ## Model Description This model has [GBERT Large](https://huggingface.co/deepset/gbert-large) as base model and fine-tuned it on xnli de dataset. The default hypothesis template is in English: `This text is {}`. While using this model , change it to "In deisem geht es um {}." or something different. While inferencing through huggingface api may give poor results as it uses by default english template. Since model is monolingual and not multilingual, hypothesis template needs to be changed accordingly. ## XNLI DEV (german) Accuracy: 85.5 ## XNLI TEST (german) Accuracy: 83.6 #### Zero-shot classification pipeline ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model="Sahajtomar/German_Zeroshot") sequence = "Letzte Woche gab es einen Selbstmord in einer nahe gelegenen kolonie" candidate_labels = ["Verbrechen","Tragödie","Stehlen"] hypothesis_template = "In deisem geht es um {}." ## Since monolingual model,its sensitive to hypothesis template. This can be experimented classifier(sequence, candidate_labels, hypothesis_template=hypothesis_template) """{'labels': ['Tragödie', 'Verbrechen', 'Stehlen'], 'scores': [0.8328856854438782, 0.10494536352157593, 0.06316883927583696], 'sequence': 'Letzte Woche gab es einen Selbstmord in einer nahe gelegenen Kolonie'}""" ```
382
Sarim24/Sarim24
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
383
SetFit/MiniLM-L12-H384-uncased__sst2__all-train
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: MiniLM-L12-H384-uncased__sst2__all-train results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # MiniLM-L12-H384-uncased__sst2__all-train This model is a fine-tuned version of [microsoft/MiniLM-L12-H384-uncased](https://huggingface.co/microsoft/MiniLM-L12-H384-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.2632 - Accuracy: 0.9055 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.4183 | 1.0 | 433 | 0.3456 | 0.8720 | | 0.2714 | 2.0 | 866 | 0.2632 | 0.9055 | | 0.2016 | 3.0 | 1299 | 0.3357 | 0.8990 | | 0.1501 | 4.0 | 1732 | 0.4474 | 0.8863 | | 0.1119 | 5.0 | 2165 | 0.3998 | 0.8979 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.1+cu102 - Datasets 1.17.0 - Tokenizers 0.10.3
384
SetFit/deberta-v3-base__sst2__all-train
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-base__sst2__all-train results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-base__sst2__all-train This model is a fine-tuned version of [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.6964 - Accuracy: 0.49 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 7 | 0.6964 | 0.49 | | No log | 2.0 | 14 | 0.7010 | 0.49 | | No log | 3.0 | 21 | 0.7031 | 0.49 | | No log | 4.0 | 28 | 0.7054 | 0.49 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
385
SetFit/deberta-v3-large__sst2__train-16-0
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-0 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.9917 - Accuracy: 0.7705 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.7001 | 1.0 | 7 | 0.7327 | 0.2857 | | 0.6326 | 2.0 | 14 | 0.6479 | 0.5714 | | 0.5232 | 3.0 | 21 | 0.5714 | 0.5714 | | 0.3313 | 4.0 | 28 | 0.6340 | 0.7143 | | 0.3161 | 5.0 | 35 | 0.6304 | 0.7143 | | 0.0943 | 6.0 | 42 | 0.4719 | 0.8571 | | 0.0593 | 7.0 | 49 | 0.5000 | 0.7143 | | 0.0402 | 8.0 | 56 | 0.3530 | 0.8571 | | 0.0307 | 9.0 | 63 | 0.3499 | 0.8571 | | 0.0033 | 10.0 | 70 | 0.3258 | 0.8571 | | 0.0021 | 11.0 | 77 | 0.3362 | 0.8571 | | 0.0012 | 12.0 | 84 | 0.4591 | 0.8571 | | 0.0036 | 13.0 | 91 | 0.4661 | 0.8571 | | 0.001 | 14.0 | 98 | 0.5084 | 0.8571 | | 0.0017 | 15.0 | 105 | 0.5844 | 0.8571 | | 0.0005 | 16.0 | 112 | 0.6645 | 0.8571 | | 0.002 | 17.0 | 119 | 0.7422 | 0.8571 | | 0.0006 | 18.0 | 126 | 0.7354 | 0.8571 | | 0.0005 | 19.0 | 133 | 0.7265 | 0.8571 | | 0.0005 | 20.0 | 140 | 0.7207 | 0.8571 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
386
SetFit/deberta-v3-large__sst2__train-16-1
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-1 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6804 - Accuracy: 0.5497 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.7086 | 1.0 | 7 | 0.7176 | 0.2857 | | 0.6897 | 2.0 | 14 | 0.7057 | 0.2857 | | 0.6491 | 3.0 | 21 | 0.6582 | 0.8571 | | 0.567 | 4.0 | 28 | 0.4480 | 0.8571 | | 0.4304 | 5.0 | 35 | 0.5465 | 0.7143 | | 0.0684 | 6.0 | 42 | 0.5408 | 0.8571 | | 0.0339 | 7.0 | 49 | 0.6501 | 0.8571 | | 0.0082 | 8.0 | 56 | 0.9152 | 0.8571 | | 0.0067 | 9.0 | 63 | 2.5162 | 0.5714 | | 0.0045 | 10.0 | 70 | 1.1136 | 0.8571 | | 0.0012 | 11.0 | 77 | 1.1668 | 0.8571 | | 0.0007 | 12.0 | 84 | 1.2071 | 0.8571 | | 0.0005 | 13.0 | 91 | 1.2310 | 0.8571 | | 0.0006 | 14.0 | 98 | 1.2476 | 0.8571 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
387
SetFit/deberta-v3-large__sst2__train-16-2
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-2 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6959 - Accuracy: 0.5008 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.7079 | 1.0 | 7 | 0.7361 | 0.2857 | | 0.6815 | 2.0 | 14 | 0.7659 | 0.2857 | | 0.6938 | 3.0 | 21 | 0.7944 | 0.2857 | | 0.4584 | 4.0 | 28 | 1.2441 | 0.2857 | | 0.4949 | 5.0 | 35 | 1.2285 | 0.5714 | | 0.0574 | 6.0 | 42 | 1.7796 | 0.5714 | | 0.0156 | 7.0 | 49 | 2.6027 | 0.5714 | | 0.0051 | 8.0 | 56 | 2.8717 | 0.5714 | | 0.0017 | 9.0 | 63 | 2.8491 | 0.5714 | | 0.0023 | 10.0 | 70 | 1.7149 | 0.7143 | | 0.001 | 11.0 | 77 | 1.1101 | 0.7143 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
388
SetFit/deberta-v3-large__sst2__train-16-3
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-3 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6286 - Accuracy: 0.7068 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6955 | 1.0 | 7 | 0.7370 | 0.2857 | | 0.6919 | 2.0 | 14 | 0.6855 | 0.4286 | | 0.6347 | 3.0 | 21 | 0.5872 | 0.7143 | | 0.4016 | 4.0 | 28 | 0.6644 | 0.7143 | | 0.3097 | 5.0 | 35 | 0.5120 | 0.7143 | | 0.0785 | 6.0 | 42 | 0.5845 | 0.7143 | | 0.024 | 7.0 | 49 | 0.6951 | 0.7143 | | 0.0132 | 8.0 | 56 | 0.8972 | 0.7143 | | 0.0037 | 9.0 | 63 | 1.5798 | 0.7143 | | 0.0034 | 10.0 | 70 | 1.5178 | 0.7143 | | 0.003 | 11.0 | 77 | 1.3511 | 0.7143 | | 0.0012 | 12.0 | 84 | 1.1346 | 0.7143 | | 0.0007 | 13.0 | 91 | 0.9752 | 0.7143 | | 0.0008 | 14.0 | 98 | 0.8531 | 0.7143 | | 0.0007 | 15.0 | 105 | 0.8149 | 0.7143 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
389
SetFit/deberta-v3-large__sst2__train-16-4
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-4 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6329 - Accuracy: 0.6392 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6945 | 1.0 | 7 | 0.7381 | 0.2857 | | 0.7072 | 2.0 | 14 | 0.7465 | 0.2857 | | 0.6548 | 3.0 | 21 | 0.7277 | 0.4286 | | 0.5695 | 4.0 | 28 | 0.6738 | 0.5714 | | 0.4615 | 5.0 | 35 | 0.8559 | 0.5714 | | 0.0823 | 6.0 | 42 | 1.0983 | 0.5714 | | 0.0274 | 7.0 | 49 | 1.9937 | 0.5714 | | 0.0106 | 8.0 | 56 | 2.2209 | 0.5714 | | 0.0039 | 9.0 | 63 | 2.2114 | 0.5714 | | 0.0031 | 10.0 | 70 | 2.2808 | 0.5714 | | 0.0013 | 11.0 | 77 | 2.3707 | 0.5714 | | 0.0008 | 12.0 | 84 | 2.4902 | 0.5714 | | 0.0005 | 13.0 | 91 | 2.5208 | 0.5714 | | 0.0007 | 14.0 | 98 | 2.5683 | 0.5714 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
390
SetFit/deberta-v3-large__sst2__train-16-5
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-5 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-5 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.5433 - Accuracy: 0.7924 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6774 | 1.0 | 7 | 0.7450 | 0.2857 | | 0.7017 | 2.0 | 14 | 0.7552 | 0.2857 | | 0.6438 | 3.0 | 21 | 0.7140 | 0.4286 | | 0.3525 | 4.0 | 28 | 0.5570 | 0.7143 | | 0.2061 | 5.0 | 35 | 0.5303 | 0.8571 | | 0.0205 | 6.0 | 42 | 0.6706 | 0.8571 | | 0.0068 | 7.0 | 49 | 0.8284 | 0.8571 | | 0.0029 | 8.0 | 56 | 0.9281 | 0.8571 | | 0.0015 | 9.0 | 63 | 0.9871 | 0.8571 | | 0.0013 | 10.0 | 70 | 1.0208 | 0.8571 | | 0.0008 | 11.0 | 77 | 1.0329 | 0.8571 | | 0.0005 | 12.0 | 84 | 1.0348 | 0.8571 | | 0.0004 | 13.0 | 91 | 1.0437 | 0.8571 | | 0.0005 | 14.0 | 98 | 1.0512 | 0.8571 | | 0.0004 | 15.0 | 105 | 1.0639 | 0.8571 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
391
SetFit/deberta-v3-large__sst2__train-16-6
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-6 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6846 - Accuracy: 0.5058 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6673 | 1.0 | 7 | 0.7580 | 0.2857 | | 0.5896 | 2.0 | 14 | 0.7885 | 0.5714 | | 0.5294 | 3.0 | 21 | 1.0040 | 0.4286 | | 0.3163 | 4.0 | 28 | 1.1761 | 0.5714 | | 0.1315 | 5.0 | 35 | 1.4315 | 0.4286 | | 0.0312 | 6.0 | 42 | 2.6115 | 0.2857 | | 0.1774 | 7.0 | 49 | 2.1631 | 0.5714 | | 0.0052 | 8.0 | 56 | 2.3838 | 0.4286 | | 0.0043 | 9.0 | 63 | 2.6553 | 0.4286 | | 0.0032 | 10.0 | 70 | 2.2774 | 0.4286 | | 0.0015 | 11.0 | 77 | 1.9467 | 0.7143 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
392
SetFit/deberta-v3-large__sst2__train-16-7
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-7 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-7 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6953 - Accuracy: 0.5063 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6911 | 1.0 | 7 | 0.7455 | 0.2857 | | 0.6844 | 2.0 | 14 | 0.7242 | 0.2857 | | 0.6137 | 3.0 | 21 | 0.7341 | 0.4286 | | 0.3805 | 4.0 | 28 | 1.0217 | 0.4286 | | 0.2201 | 5.0 | 35 | 1.1437 | 0.2857 | | 0.0296 | 6.0 | 42 | 1.5997 | 0.4286 | | 0.0103 | 7.0 | 49 | 2.6835 | 0.4286 | | 0.0046 | 8.0 | 56 | 3.3521 | 0.4286 | | 0.002 | 9.0 | 63 | 3.7846 | 0.4286 | | 0.0017 | 10.0 | 70 | 4.0088 | 0.4286 | | 0.0018 | 11.0 | 77 | 4.1483 | 0.4286 | | 0.0006 | 12.0 | 84 | 4.2235 | 0.4286 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
393
SetFit/deberta-v3-large__sst2__train-16-8
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-8 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-8 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6915 - Accuracy: 0.6579 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.7129 | 1.0 | 7 | 0.7309 | 0.2857 | | 0.6549 | 2.0 | 14 | 0.7316 | 0.4286 | | 0.621 | 3.0 | 21 | 0.7131 | 0.5714 | | 0.3472 | 4.0 | 28 | 0.5703 | 0.4286 | | 0.2041 | 5.0 | 35 | 0.6675 | 0.5714 | | 0.031 | 6.0 | 42 | 1.6750 | 0.5714 | | 0.0141 | 7.0 | 49 | 1.8743 | 0.5714 | | 0.0055 | 8.0 | 56 | 1.1778 | 0.5714 | | 0.0024 | 9.0 | 63 | 1.0699 | 0.5714 | | 0.0019 | 10.0 | 70 | 1.0933 | 0.5714 | | 0.0012 | 11.0 | 77 | 1.1218 | 0.7143 | | 0.0007 | 12.0 | 84 | 1.1468 | 0.7143 | | 0.0006 | 13.0 | 91 | 1.1584 | 0.7143 | | 0.0006 | 14.0 | 98 | 1.3092 | 0.7143 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
394
SetFit/deberta-v3-large__sst2__train-16-9
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-16-9 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-16-9 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.2598 - Accuracy: 0.7809 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6887 | 1.0 | 7 | 0.7452 | 0.2857 | | 0.6889 | 2.0 | 14 | 0.7988 | 0.2857 | | 0.6501 | 3.0 | 21 | 0.8987 | 0.2857 | | 0.4286 | 4.0 | 28 | 0.9186 | 0.4286 | | 0.3591 | 5.0 | 35 | 0.5566 | 0.7143 | | 0.0339 | 6.0 | 42 | 1.1130 | 0.5714 | | 0.013 | 7.0 | 49 | 1.8296 | 0.7143 | | 0.0041 | 8.0 | 56 | 1.7069 | 0.7143 | | 0.0023 | 9.0 | 63 | 1.1942 | 0.7143 | | 0.0022 | 10.0 | 70 | 0.6054 | 0.7143 | | 0.0011 | 11.0 | 77 | 0.3872 | 0.7143 | | 0.0006 | 12.0 | 84 | 0.3217 | 0.7143 | | 0.0005 | 13.0 | 91 | 0.2879 | 0.8571 | | 0.0005 | 14.0 | 98 | 0.2640 | 0.8571 | | 0.0004 | 15.0 | 105 | 0.2531 | 0.8571 | | 0.0003 | 16.0 | 112 | 0.2384 | 0.8571 | | 0.0004 | 17.0 | 119 | 0.2338 | 0.8571 | | 0.0003 | 18.0 | 126 | 0.2314 | 0.8571 | | 0.0003 | 19.0 | 133 | 0.2276 | 0.8571 | | 0.0003 | 20.0 | 140 | 0.2172 | 0.8571 | | 0.0003 | 21.0 | 147 | 0.2069 | 0.8571 | | 0.0002 | 22.0 | 154 | 0.2018 | 0.8571 | | 0.0002 | 23.0 | 161 | 0.2005 | 0.8571 | | 0.0002 | 24.0 | 168 | 0.1985 | 0.8571 | | 0.0002 | 25.0 | 175 | 0.1985 | 1.0 | | 0.0002 | 26.0 | 182 | 0.1955 | 1.0 | | 0.0002 | 27.0 | 189 | 0.1967 | 1.0 | | 0.0002 | 28.0 | 196 | 0.1918 | 1.0 | | 0.0002 | 29.0 | 203 | 0.1888 | 1.0 | | 0.0002 | 30.0 | 210 | 0.1864 | 1.0 | | 0.0002 | 31.0 | 217 | 0.1870 | 1.0 | | 0.0002 | 32.0 | 224 | 0.1892 | 1.0 | | 0.0002 | 33.0 | 231 | 0.1917 | 1.0 | | 0.0002 | 34.0 | 238 | 0.1869 | 1.0 | | 0.0002 | 35.0 | 245 | 0.1812 | 1.0 | | 0.0001 | 36.0 | 252 | 0.1777 | 1.0 | | 0.0002 | 37.0 | 259 | 0.1798 | 1.0 | | 0.0002 | 38.0 | 266 | 0.1824 | 0.8571 | | 0.0002 | 39.0 | 273 | 0.1846 | 0.8571 | | 0.0002 | 40.0 | 280 | 0.1839 | 0.8571 | | 0.0001 | 41.0 | 287 | 0.1826 | 0.8571 | | 0.0001 | 42.0 | 294 | 0.1779 | 0.8571 | | 0.0002 | 43.0 | 301 | 0.1762 | 0.8571 | | 0.0001 | 44.0 | 308 | 0.1742 | 1.0 | | 0.0002 | 45.0 | 315 | 0.1708 | 1.0 | | 0.0001 | 46.0 | 322 | 0.1702 | 1.0 | | 0.0001 | 47.0 | 329 | 0.1699 | 1.0 | | 0.0001 | 48.0 | 336 | 0.1695 | 1.0 | | 0.0001 | 49.0 | 343 | 0.1683 | 1.0 | | 0.0001 | 50.0 | 350 | 0.1681 | 1.0 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
395
SetFit/deberta-v3-large__sst2__train-32-0
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-32-0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-32-0 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4849 - Accuracy: 0.7716 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.7059 | 1.0 | 13 | 0.6840 | 0.5385 | | 0.6595 | 2.0 | 26 | 0.6214 | 0.6923 | | 0.4153 | 3.0 | 39 | 0.1981 | 0.9231 | | 0.0733 | 4.0 | 52 | 0.5068 | 0.9231 | | 0.2092 | 5.0 | 65 | 1.3114 | 0.6923 | | 0.003 | 6.0 | 78 | 1.1062 | 0.8462 | | 0.0012 | 7.0 | 91 | 1.5948 | 0.7692 | | 0.0008 | 8.0 | 104 | 1.6913 | 0.7692 | | 0.0006 | 9.0 | 117 | 1.7191 | 0.7692 | | 0.0005 | 10.0 | 130 | 1.6527 | 0.7692 | | 0.0003 | 11.0 | 143 | 1.4840 | 0.7692 | | 0.0002 | 12.0 | 156 | 1.3076 | 0.8462 | | 0.0002 | 13.0 | 169 | 1.3130 | 0.8462 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
396
SetFit/deberta-v3-large__sst2__train-32-1
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-32-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-32-1 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4201 - Accuracy: 0.8759 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.7162 | 1.0 | 13 | 0.6832 | 0.5385 | | 0.6561 | 2.0 | 26 | 0.7270 | 0.4615 | | 0.4685 | 3.0 | 39 | 1.0674 | 0.5385 | | 0.2837 | 4.0 | 52 | 1.0841 | 0.5385 | | 0.1129 | 5.0 | 65 | 0.3502 | 0.9231 | | 0.0118 | 6.0 | 78 | 0.4829 | 0.9231 | | 0.0022 | 7.0 | 91 | 0.7430 | 0.8462 | | 0.0007 | 8.0 | 104 | 0.8219 | 0.8462 | | 0.0005 | 9.0 | 117 | 0.8787 | 0.8462 | | 0.0003 | 10.0 | 130 | 0.8713 | 0.8462 | | 0.0003 | 11.0 | 143 | 0.8473 | 0.8462 | | 0.0002 | 12.0 | 156 | 0.8482 | 0.8462 | | 0.0002 | 13.0 | 169 | 0.8494 | 0.8462 | | 0.0002 | 14.0 | 182 | 0.8638 | 0.8462 | | 0.0002 | 15.0 | 195 | 0.8492 | 0.8462 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
397
SetFit/deberta-v3-large__sst2__train-8-0
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-0 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7088 - Accuracy: 0.5008 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6705 | 1.0 | 3 | 0.7961 | 0.25 | | 0.6571 | 2.0 | 6 | 0.8092 | 0.25 | | 0.7043 | 3.0 | 9 | 0.7977 | 0.25 | | 0.6207 | 4.0 | 12 | 0.8478 | 0.25 | | 0.5181 | 5.0 | 15 | 0.9782 | 0.25 | | 0.4136 | 6.0 | 18 | 1.3151 | 0.25 | | 0.3702 | 7.0 | 21 | 1.8633 | 0.25 | | 0.338 | 8.0 | 24 | 2.2119 | 0.25 | | 0.2812 | 9.0 | 27 | 2.3058 | 0.25 | | 0.2563 | 10.0 | 30 | 2.3353 | 0.25 | | 0.2132 | 11.0 | 33 | 2.5921 | 0.25 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
398
SetFit/deberta-v3-large__sst2__train-8-1
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-1 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7020 - Accuracy: 0.5008 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6773 | 1.0 | 3 | 0.7822 | 0.25 | | 0.6587 | 2.0 | 6 | 0.8033 | 0.25 | | 0.693 | 3.0 | 9 | 0.8101 | 0.25 | | 0.5979 | 4.0 | 12 | 1.1235 | 0.25 | | 0.4095 | 5.0 | 15 | 1.3563 | 0.25 | | 0.2836 | 6.0 | 18 | 1.5325 | 0.5 | | 0.1627 | 7.0 | 21 | 1.7786 | 0.25 | | 0.0956 | 8.0 | 24 | 2.0067 | 0.5 | | 0.0535 | 9.0 | 27 | 2.3351 | 0.5 | | 0.0315 | 10.0 | 30 | 2.6204 | 0.5 | | 0.0182 | 11.0 | 33 | 2.8483 | 0.5 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
399
SetFit/deberta-v3-large__sst2__train-8-2
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-2 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6794 - Accuracy: 0.6063 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6942 | 1.0 | 3 | 0.7940 | 0.25 | | 0.6068 | 2.0 | 6 | 0.9326 | 0.25 | | 0.6553 | 3.0 | 9 | 0.7979 | 0.25 | | 0.475 | 4.0 | 12 | 0.7775 | 0.25 | | 0.377 | 5.0 | 15 | 0.7477 | 0.25 | | 0.3176 | 6.0 | 18 | 0.6856 | 0.75 | | 0.2708 | 7.0 | 21 | 0.6554 | 0.75 | | 0.2855 | 8.0 | 24 | 0.8129 | 0.5 | | 0.148 | 9.0 | 27 | 0.7074 | 0.75 | | 0.0947 | 10.0 | 30 | 0.7090 | 0.75 | | 0.049 | 11.0 | 33 | 0.7885 | 0.75 | | 0.0252 | 12.0 | 36 | 0.9203 | 0.75 | | 0.0165 | 13.0 | 39 | 1.0937 | 0.75 | | 0.0084 | 14.0 | 42 | 1.2502 | 0.75 | | 0.0059 | 15.0 | 45 | 1.3726 | 0.75 | | 0.0037 | 16.0 | 48 | 1.4784 | 0.75 | | 0.003 | 17.0 | 51 | 1.5615 | 0.75 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
400
SetFit/deberta-v3-large__sst2__train-8-3
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-3 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6421 - Accuracy: 0.6310 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6696 | 1.0 | 3 | 0.7917 | 0.25 | | 0.6436 | 2.0 | 6 | 0.8107 | 0.25 | | 0.6923 | 3.0 | 9 | 0.8302 | 0.25 | | 0.5051 | 4.0 | 12 | 0.9828 | 0.25 | | 0.3688 | 5.0 | 15 | 0.7402 | 0.25 | | 0.2671 | 6.0 | 18 | 0.5820 | 0.75 | | 0.1935 | 7.0 | 21 | 0.8356 | 0.5 | | 0.0815 | 8.0 | 24 | 1.0431 | 0.25 | | 0.0591 | 9.0 | 27 | 0.9679 | 0.75 | | 0.0276 | 10.0 | 30 | 1.0659 | 0.75 | | 0.0175 | 11.0 | 33 | 0.9689 | 0.75 | | 0.0152 | 12.0 | 36 | 0.8820 | 0.75 | | 0.006 | 13.0 | 39 | 0.8337 | 0.75 | | 0.0041 | 14.0 | 42 | 0.7650 | 0.75 | | 0.0036 | 15.0 | 45 | 0.6960 | 0.75 | | 0.0034 | 16.0 | 48 | 0.6548 | 0.75 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
401
SetFit/deberta-v3-large__sst2__train-8-4
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-4 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.3023 - Accuracy: 0.7057 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6816 | 1.0 | 3 | 0.8072 | 0.25 | | 0.6672 | 2.0 | 6 | 0.8740 | 0.25 | | 0.6667 | 3.0 | 9 | 0.8578 | 0.25 | | 0.5346 | 4.0 | 12 | 1.0353 | 0.25 | | 0.4517 | 5.0 | 15 | 1.1030 | 0.25 | | 0.3095 | 6.0 | 18 | 0.9986 | 0.25 | | 0.2464 | 7.0 | 21 | 0.9286 | 0.5 | | 0.1342 | 8.0 | 24 | 0.4063 | 1.0 | | 0.0851 | 9.0 | 27 | 0.2210 | 1.0 | | 0.0491 | 10.0 | 30 | 0.2302 | 1.0 | | 0.0211 | 11.0 | 33 | 0.4020 | 0.75 | | 0.017 | 12.0 | 36 | 0.2382 | 1.0 | | 0.0084 | 13.0 | 39 | 0.0852 | 1.0 | | 0.0051 | 14.0 | 42 | 0.0354 | 1.0 | | 0.0047 | 15.0 | 45 | 0.0208 | 1.0 | | 0.0029 | 16.0 | 48 | 0.0155 | 1.0 | | 0.0022 | 17.0 | 51 | 0.0139 | 1.0 | | 0.0019 | 18.0 | 54 | 0.0144 | 1.0 | | 0.0016 | 19.0 | 57 | 0.0168 | 1.0 | | 0.0013 | 20.0 | 60 | 0.0231 | 1.0 | | 0.0011 | 21.0 | 63 | 0.0369 | 1.0 | | 0.0009 | 22.0 | 66 | 0.0528 | 1.0 | | 0.001 | 23.0 | 69 | 0.0639 | 1.0 | | 0.0009 | 24.0 | 72 | 0.0670 | 1.0 | | 0.0009 | 25.0 | 75 | 0.0526 | 1.0 | | 0.0008 | 26.0 | 78 | 0.0425 | 1.0 | | 0.0011 | 27.0 | 81 | 0.0135 | 1.0 | | 0.0007 | 28.0 | 84 | 0.0076 | 1.0 | | 0.0007 | 29.0 | 87 | 0.0057 | 1.0 | | 0.0007 | 30.0 | 90 | 0.0049 | 1.0 | | 0.0008 | 31.0 | 93 | 0.0045 | 1.0 | | 0.0007 | 32.0 | 96 | 0.0044 | 1.0 | | 0.0008 | 33.0 | 99 | 0.0043 | 1.0 | | 0.0005 | 34.0 | 102 | 0.0044 | 1.0 | | 0.0006 | 35.0 | 105 | 0.0045 | 1.0 | | 0.0006 | 36.0 | 108 | 0.0046 | 1.0 | | 0.0007 | 37.0 | 111 | 0.0048 | 1.0 | | 0.0006 | 38.0 | 114 | 0.0049 | 1.0 | | 0.0005 | 39.0 | 117 | 0.0050 | 1.0 | | 0.0005 | 40.0 | 120 | 0.0050 | 1.0 | | 0.0004 | 41.0 | 123 | 0.0051 | 1.0 | | 0.0005 | 42.0 | 126 | 0.0051 | 1.0 | | 0.0004 | 43.0 | 129 | 0.0051 | 1.0 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
402
SetFit/deberta-v3-large__sst2__train-8-5
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-5 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-5 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.3078 - Accuracy: 0.6930 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6813 | 1.0 | 3 | 0.7842 | 0.25 | | 0.6617 | 2.0 | 6 | 0.7968 | 0.25 | | 0.6945 | 3.0 | 9 | 0.7746 | 0.25 | | 0.5967 | 4.0 | 12 | 0.7557 | 0.25 | | 0.4824 | 5.0 | 15 | 0.6920 | 0.25 | | 0.3037 | 6.0 | 18 | 0.6958 | 0.5 | | 0.2329 | 7.0 | 21 | 0.6736 | 0.5 | | 0.1441 | 8.0 | 24 | 0.3749 | 1.0 | | 0.0875 | 9.0 | 27 | 0.3263 | 0.75 | | 0.0655 | 10.0 | 30 | 0.3525 | 0.75 | | 0.0373 | 11.0 | 33 | 0.1993 | 1.0 | | 0.0173 | 12.0 | 36 | 0.1396 | 1.0 | | 0.0147 | 13.0 | 39 | 0.0655 | 1.0 | | 0.0084 | 14.0 | 42 | 0.0343 | 1.0 | | 0.0049 | 15.0 | 45 | 0.0225 | 1.0 | | 0.004 | 16.0 | 48 | 0.0167 | 1.0 | | 0.003 | 17.0 | 51 | 0.0134 | 1.0 | | 0.0027 | 18.0 | 54 | 0.0114 | 1.0 | | 0.002 | 19.0 | 57 | 0.0104 | 1.0 | | 0.0015 | 20.0 | 60 | 0.0099 | 1.0 | | 0.0014 | 21.0 | 63 | 0.0095 | 1.0 | | 0.0013 | 22.0 | 66 | 0.0095 | 1.0 | | 0.0012 | 23.0 | 69 | 0.0091 | 1.0 | | 0.0011 | 24.0 | 72 | 0.0085 | 1.0 | | 0.0009 | 25.0 | 75 | 0.0081 | 1.0 | | 0.001 | 26.0 | 78 | 0.0077 | 1.0 | | 0.0008 | 27.0 | 81 | 0.0074 | 1.0 | | 0.0009 | 28.0 | 84 | 0.0071 | 1.0 | | 0.0007 | 29.0 | 87 | 0.0068 | 1.0 | | 0.0008 | 30.0 | 90 | 0.0064 | 1.0 | | 0.0007 | 31.0 | 93 | 0.0062 | 1.0 | | 0.0007 | 32.0 | 96 | 0.0059 | 1.0 | | 0.0007 | 33.0 | 99 | 0.0056 | 1.0 | | 0.0005 | 34.0 | 102 | 0.0054 | 1.0 | | 0.0006 | 35.0 | 105 | 0.0053 | 1.0 | | 0.0008 | 36.0 | 108 | 0.0051 | 1.0 | | 0.0007 | 37.0 | 111 | 0.0050 | 1.0 | | 0.0007 | 38.0 | 114 | 0.0049 | 1.0 | | 0.0006 | 39.0 | 117 | 0.0048 | 1.0 | | 0.0005 | 40.0 | 120 | 0.0048 | 1.0 | | 0.0005 | 41.0 | 123 | 0.0048 | 1.0 | | 0.0005 | 42.0 | 126 | 0.0047 | 1.0 | | 0.0005 | 43.0 | 129 | 0.0047 | 1.0 | | 0.0005 | 44.0 | 132 | 0.0047 | 1.0 | | 0.0006 | 45.0 | 135 | 0.0047 | 1.0 | | 0.0005 | 46.0 | 138 | 0.0047 | 1.0 | | 0.0005 | 47.0 | 141 | 0.0047 | 1.0 | | 0.0006 | 48.0 | 144 | 0.0047 | 1.0 | | 0.0005 | 49.0 | 147 | 0.0047 | 1.0 | | 0.0005 | 50.0 | 150 | 0.0047 | 1.0 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
403
SetFit/deberta-v3-large__sst2__train-8-6
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-6 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.4331 - Accuracy: 0.7106 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6486 | 1.0 | 3 | 0.7901 | 0.25 | | 0.6418 | 2.0 | 6 | 0.9259 | 0.25 | | 0.6169 | 3.0 | 9 | 1.0574 | 0.25 | | 0.5639 | 4.0 | 12 | 1.1372 | 0.25 | | 0.4562 | 5.0 | 15 | 0.6090 | 0.5 | | 0.3105 | 6.0 | 18 | 0.4435 | 1.0 | | 0.2303 | 7.0 | 21 | 0.2804 | 1.0 | | 0.1388 | 8.0 | 24 | 0.2205 | 1.0 | | 0.0918 | 9.0 | 27 | 0.1282 | 1.0 | | 0.0447 | 10.0 | 30 | 0.0643 | 1.0 | | 0.0297 | 11.0 | 33 | 0.0361 | 1.0 | | 0.0159 | 12.0 | 36 | 0.0211 | 1.0 | | 0.0102 | 13.0 | 39 | 0.0155 | 1.0 | | 0.0061 | 14.0 | 42 | 0.0158 | 1.0 | | 0.0049 | 15.0 | 45 | 0.0189 | 1.0 | | 0.0035 | 16.0 | 48 | 0.0254 | 1.0 | | 0.0027 | 17.0 | 51 | 0.0305 | 1.0 | | 0.0021 | 18.0 | 54 | 0.0287 | 1.0 | | 0.0016 | 19.0 | 57 | 0.0215 | 1.0 | | 0.0016 | 20.0 | 60 | 0.0163 | 1.0 | | 0.0014 | 21.0 | 63 | 0.0138 | 1.0 | | 0.0015 | 22.0 | 66 | 0.0131 | 1.0 | | 0.001 | 23.0 | 69 | 0.0132 | 1.0 | | 0.0014 | 24.0 | 72 | 0.0126 | 1.0 | | 0.0011 | 25.0 | 75 | 0.0125 | 1.0 | | 0.001 | 26.0 | 78 | 0.0119 | 1.0 | | 0.0008 | 27.0 | 81 | 0.0110 | 1.0 | | 0.0007 | 28.0 | 84 | 0.0106 | 1.0 | | 0.0008 | 29.0 | 87 | 0.0095 | 1.0 | | 0.0009 | 30.0 | 90 | 0.0089 | 1.0 | | 0.0008 | 31.0 | 93 | 0.0083 | 1.0 | | 0.0007 | 32.0 | 96 | 0.0075 | 1.0 | | 0.0008 | 33.0 | 99 | 0.0066 | 1.0 | | 0.0006 | 34.0 | 102 | 0.0059 | 1.0 | | 0.0007 | 35.0 | 105 | 0.0054 | 1.0 | | 0.0008 | 36.0 | 108 | 0.0051 | 1.0 | | 0.0007 | 37.0 | 111 | 0.0049 | 1.0 | | 0.0007 | 38.0 | 114 | 0.0047 | 1.0 | | 0.0006 | 39.0 | 117 | 0.0045 | 1.0 | | 0.0006 | 40.0 | 120 | 0.0046 | 1.0 | | 0.0005 | 41.0 | 123 | 0.0045 | 1.0 | | 0.0006 | 42.0 | 126 | 0.0044 | 1.0 | | 0.0006 | 43.0 | 129 | 0.0043 | 1.0 | | 0.0006 | 44.0 | 132 | 0.0044 | 1.0 | | 0.0005 | 45.0 | 135 | 0.0045 | 1.0 | | 0.0006 | 46.0 | 138 | 0.0043 | 1.0 | | 0.0006 | 47.0 | 141 | 0.0043 | 1.0 | | 0.0006 | 48.0 | 144 | 0.0041 | 1.0 | | 0.0007 | 49.0 | 147 | 0.0042 | 1.0 | | 0.0005 | 50.0 | 150 | 0.0042 | 1.0 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
404
SetFit/deberta-v3-large__sst2__train-8-7
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-7 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-7 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7037 - Accuracy: 0.5008 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6864 | 1.0 | 3 | 0.7800 | 0.25 | | 0.6483 | 2.0 | 6 | 0.8067 | 0.25 | | 0.6028 | 3.0 | 9 | 0.8500 | 0.25 | | 0.4086 | 4.0 | 12 | 1.0661 | 0.25 | | 0.2923 | 5.0 | 15 | 1.2302 | 0.25 | | 0.2059 | 6.0 | 18 | 1.0312 | 0.5 | | 0.1238 | 7.0 | 21 | 1.1271 | 0.5 | | 0.0711 | 8.0 | 24 | 1.3100 | 0.5 | | 0.0453 | 9.0 | 27 | 1.4208 | 0.5 | | 0.0198 | 10.0 | 30 | 1.5988 | 0.5 | | 0.0135 | 11.0 | 33 | 1.9174 | 0.5 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
405
SetFit/deberta-v3-large__sst2__train-8-8
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-8 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-8 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7414 - Accuracy: 0.5623 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6597 | 1.0 | 3 | 0.7716 | 0.25 | | 0.6376 | 2.0 | 6 | 0.7802 | 0.25 | | 0.5857 | 3.0 | 9 | 0.6625 | 0.75 | | 0.4024 | 4.0 | 12 | 0.5195 | 0.75 | | 0.2635 | 5.0 | 15 | 0.4222 | 1.0 | | 0.1714 | 6.0 | 18 | 0.4410 | 0.5 | | 0.1267 | 7.0 | 21 | 0.7773 | 0.75 | | 0.0582 | 8.0 | 24 | 0.9070 | 0.75 | | 0.0374 | 9.0 | 27 | 0.9539 | 0.75 | | 0.0204 | 10.0 | 30 | 1.0507 | 0.75 | | 0.012 | 11.0 | 33 | 1.2802 | 0.5 | | 0.0086 | 12.0 | 36 | 1.4272 | 0.5 | | 0.0049 | 13.0 | 39 | 1.4803 | 0.5 | | 0.0039 | 14.0 | 42 | 1.4912 | 0.5 | | 0.0031 | 15.0 | 45 | 1.5231 | 0.5 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
406
SetFit/deberta-v3-large__sst2__train-8-9
[ "negative", "positive" ]
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: deberta-v3-large__sst2__train-8-9 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large__sst2__train-8-9 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6013 - Accuracy: 0.7210 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6757 | 1.0 | 3 | 0.7810 | 0.25 | | 0.6506 | 2.0 | 6 | 0.8102 | 0.25 | | 0.6463 | 3.0 | 9 | 0.8313 | 0.25 | | 0.5813 | 4.0 | 12 | 0.8858 | 0.25 | | 0.4635 | 5.0 | 15 | 0.8220 | 0.25 | | 0.3992 | 6.0 | 18 | 0.7226 | 0.5 | | 0.3281 | 7.0 | 21 | 0.6707 | 0.75 | | 0.2276 | 8.0 | 24 | 0.7515 | 0.75 | | 0.1674 | 9.0 | 27 | 0.6971 | 0.75 | | 0.0873 | 10.0 | 30 | 0.5419 | 0.75 | | 0.0525 | 11.0 | 33 | 0.5025 | 0.75 | | 0.0286 | 12.0 | 36 | 0.5229 | 0.75 | | 0.0149 | 13.0 | 39 | 0.5660 | 0.75 | | 0.0082 | 14.0 | 42 | 0.6954 | 0.75 | | 0.006 | 15.0 | 45 | 0.8649 | 0.75 | | 0.0043 | 16.0 | 48 | 1.0011 | 0.75 | | 0.0035 | 17.0 | 51 | 1.0909 | 0.75 | | 0.0021 | 18.0 | 54 | 1.1615 | 0.75 | | 0.0017 | 19.0 | 57 | 1.2147 | 0.75 | | 0.0013 | 20.0 | 60 | 1.2585 | 0.75 | | 0.0016 | 21.0 | 63 | 1.2917 | 0.75 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
407
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-0
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-0 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.2707 - Accuracy: 0.517 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0943 | 1.0 | 10 | 1.1095 | 0.3 | | 1.0602 | 2.0 | 20 | 1.1086 | 0.4 | | 1.0159 | 3.0 | 30 | 1.1165 | 0.4 | | 0.9027 | 4.0 | 40 | 1.1377 | 0.4 | | 0.8364 | 5.0 | 50 | 1.0126 | 0.5 | | 0.6653 | 6.0 | 60 | 0.9298 | 0.5 | | 0.535 | 7.0 | 70 | 0.9555 | 0.5 | | 0.3713 | 8.0 | 80 | 0.8543 | 0.4 | | 0.1633 | 9.0 | 90 | 0.9876 | 0.4 | | 0.1069 | 10.0 | 100 | 0.8383 | 0.6 | | 0.0591 | 11.0 | 110 | 0.8056 | 0.6 | | 0.0344 | 12.0 | 120 | 0.8915 | 0.6 | | 0.0265 | 13.0 | 130 | 0.8722 | 0.6 | | 0.0196 | 14.0 | 140 | 1.0064 | 0.6 | | 0.0158 | 15.0 | 150 | 1.0479 | 0.6 | | 0.0128 | 16.0 | 160 | 1.0723 | 0.6 | | 0.0121 | 17.0 | 170 | 1.0758 | 0.6 | | 0.0093 | 18.0 | 180 | 1.1236 | 0.6 | | 0.0085 | 19.0 | 190 | 1.1480 | 0.6 | | 0.0084 | 20.0 | 200 | 1.1651 | 0.6 | | 0.0077 | 21.0 | 210 | 1.1832 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
408
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-1
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-1 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0424 - Accuracy: 0.5355 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0989 | 1.0 | 10 | 1.1049 | 0.1 | | 1.0641 | 2.0 | 20 | 1.0768 | 0.3 | | 0.9742 | 3.0 | 30 | 1.0430 | 0.4 | | 0.8765 | 4.0 | 40 | 1.0058 | 0.4 | | 0.6979 | 5.0 | 50 | 0.8488 | 0.7 | | 0.563 | 6.0 | 60 | 0.7221 | 0.7 | | 0.4135 | 7.0 | 70 | 0.6587 | 0.8 | | 0.2509 | 8.0 | 80 | 0.5577 | 0.7 | | 0.0943 | 9.0 | 90 | 0.5840 | 0.7 | | 0.0541 | 10.0 | 100 | 0.6959 | 0.7 | | 0.0362 | 11.0 | 110 | 0.6884 | 0.6 | | 0.0254 | 12.0 | 120 | 0.9263 | 0.6 | | 0.0184 | 13.0 | 130 | 0.7992 | 0.6 | | 0.0172 | 14.0 | 140 | 0.7351 | 0.6 | | 0.0131 | 15.0 | 150 | 0.7664 | 0.6 | | 0.0117 | 16.0 | 160 | 0.8262 | 0.6 | | 0.0101 | 17.0 | 170 | 0.8839 | 0.6 | | 0.0089 | 18.0 | 180 | 0.9018 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
409
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-2
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-2 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.9210 - Accuracy: 0.5635 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0915 | 1.0 | 10 | 1.1051 | 0.4 | | 1.0663 | 2.0 | 20 | 1.0794 | 0.3 | | 1.0307 | 3.0 | 30 | 1.0664 | 0.5 | | 0.9443 | 4.0 | 40 | 1.0729 | 0.5 | | 0.8373 | 5.0 | 50 | 1.0175 | 0.4 | | 0.6892 | 6.0 | 60 | 0.9624 | 0.5 | | 0.538 | 7.0 | 70 | 0.9924 | 0.5 | | 0.4173 | 8.0 | 80 | 1.0136 | 0.6 | | 0.1846 | 9.0 | 90 | 1.0683 | 0.6 | | 0.1125 | 10.0 | 100 | 1.2376 | 0.6 | | 0.0754 | 11.0 | 110 | 1.2537 | 0.6 | | 0.0401 | 12.0 | 120 | 1.4387 | 0.6 | | 0.0285 | 13.0 | 130 | 1.5702 | 0.6 | | 0.0241 | 14.0 | 140 | 1.6795 | 0.6 | | 0.0175 | 15.0 | 150 | 1.7228 | 0.6 | | 0.0147 | 16.0 | 160 | 1.7892 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
410
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-3
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-3 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0675 - Accuracy: 0.44 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0951 | 1.0 | 10 | 1.1346 | 0.1 | | 1.0424 | 2.0 | 20 | 1.1120 | 0.2 | | 0.957 | 3.0 | 30 | 1.1002 | 0.3 | | 0.7889 | 4.0 | 40 | 1.0838 | 0.4 | | 0.6162 | 5.0 | 50 | 1.0935 | 0.5 | | 0.4849 | 6.0 | 60 | 1.0867 | 0.5 | | 0.3089 | 7.0 | 70 | 1.1145 | 0.5 | | 0.2145 | 8.0 | 80 | 1.1278 | 0.6 | | 0.0805 | 9.0 | 90 | 1.2801 | 0.6 | | 0.0497 | 10.0 | 100 | 1.3296 | 0.6 | | 0.0328 | 11.0 | 110 | 1.2913 | 0.6 | | 0.0229 | 12.0 | 120 | 1.3692 | 0.6 | | 0.0186 | 13.0 | 130 | 1.4642 | 0.6 | | 0.0161 | 14.0 | 140 | 1.5568 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
411
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-4
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-4 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0903 - Accuracy: 0.4805 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0974 | 1.0 | 10 | 1.1139 | 0.1 | | 1.0637 | 2.0 | 20 | 1.0988 | 0.1 | | 0.9758 | 3.0 | 30 | 1.1013 | 0.1 | | 0.9012 | 4.0 | 40 | 1.0769 | 0.3 | | 0.6993 | 5.0 | 50 | 1.0484 | 0.6 | | 0.5676 | 6.0 | 60 | 1.0223 | 0.6 | | 0.4069 | 7.0 | 70 | 0.9190 | 0.6 | | 0.3192 | 8.0 | 80 | 1.1370 | 0.6 | | 0.1112 | 9.0 | 90 | 1.1728 | 0.6 | | 0.07 | 10.0 | 100 | 1.1998 | 0.6 | | 0.0397 | 11.0 | 110 | 1.3700 | 0.6 | | 0.027 | 12.0 | 120 | 1.3329 | 0.6 | | 0.021 | 13.0 | 130 | 1.2697 | 0.6 | | 0.0177 | 14.0 | 140 | 1.4195 | 0.6 | | 0.0142 | 15.0 | 150 | 1.5342 | 0.6 | | 0.0118 | 16.0 | 160 | 1.5999 | 0.6 | | 0.0108 | 17.0 | 170 | 1.6327 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
412
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-5
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-5 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-5 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.9907 - Accuracy: 0.49 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0941 | 1.0 | 10 | 1.1287 | 0.2 | | 1.0481 | 2.0 | 20 | 1.1136 | 0.2 | | 0.9498 | 3.0 | 30 | 1.1200 | 0.2 | | 0.8157 | 4.0 | 40 | 1.0771 | 0.2 | | 0.65 | 5.0 | 50 | 0.9733 | 0.4 | | 0.5021 | 6.0 | 60 | 1.0626 | 0.4 | | 0.3358 | 7.0 | 70 | 1.0787 | 0.4 | | 0.2017 | 8.0 | 80 | 1.3183 | 0.4 | | 0.088 | 9.0 | 90 | 1.2204 | 0.5 | | 0.0527 | 10.0 | 100 | 1.6892 | 0.4 | | 0.0337 | 11.0 | 110 | 1.6967 | 0.5 | | 0.0238 | 12.0 | 120 | 1.5436 | 0.5 | | 0.0183 | 13.0 | 130 | 1.7447 | 0.4 | | 0.0159 | 14.0 | 140 | 1.8999 | 0.4 | | 0.014 | 15.0 | 150 | 1.9004 | 0.4 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
413
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-6
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-6 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.8331 - Accuracy: 0.625 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0881 | 1.0 | 10 | 1.1248 | 0.1 | | 1.0586 | 2.0 | 20 | 1.1162 | 0.2 | | 0.9834 | 3.0 | 30 | 1.1199 | 0.3 | | 0.9271 | 4.0 | 40 | 1.0740 | 0.3 | | 0.7663 | 5.0 | 50 | 1.0183 | 0.5 | | 0.6042 | 6.0 | 60 | 1.0259 | 0.5 | | 0.4482 | 7.0 | 70 | 0.8699 | 0.7 | | 0.3072 | 8.0 | 80 | 1.0615 | 0.5 | | 0.1458 | 9.0 | 90 | 1.0164 | 0.5 | | 0.0838 | 10.0 | 100 | 1.0620 | 0.5 | | 0.055 | 11.0 | 110 | 1.1829 | 0.5 | | 0.0347 | 12.0 | 120 | 1.2815 | 0.4 | | 0.0244 | 13.0 | 130 | 1.2607 | 0.6 | | 0.0213 | 14.0 | 140 | 1.3695 | 0.5 | | 0.0169 | 15.0 | 150 | 1.4397 | 0.5 | | 0.0141 | 16.0 | 160 | 1.4388 | 0.6 | | 0.0122 | 17.0 | 170 | 1.4242 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
414
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-7
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-7 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-7 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.9011 - Accuracy: 0.578 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0968 | 1.0 | 10 | 1.1309 | 0.0 | | 1.0709 | 2.0 | 20 | 1.1237 | 0.1 | | 0.9929 | 3.0 | 30 | 1.1254 | 0.1 | | 0.878 | 4.0 | 40 | 1.1206 | 0.5 | | 0.7409 | 5.0 | 50 | 1.0831 | 0.1 | | 0.5663 | 6.0 | 60 | 0.9830 | 0.6 | | 0.4105 | 7.0 | 70 | 0.9919 | 0.5 | | 0.2912 | 8.0 | 80 | 1.0472 | 0.6 | | 0.1013 | 9.0 | 90 | 1.1617 | 0.4 | | 0.0611 | 10.0 | 100 | 1.2789 | 0.6 | | 0.039 | 11.0 | 110 | 1.4091 | 0.4 | | 0.0272 | 12.0 | 120 | 1.4974 | 0.4 | | 0.0189 | 13.0 | 130 | 1.4845 | 0.5 | | 0.018 | 14.0 | 140 | 1.4924 | 0.5 | | 0.0131 | 15.0 | 150 | 1.5206 | 0.6 | | 0.0116 | 16.0 | 160 | 1.5858 | 0.5 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
415
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-8
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-8 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-8 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0704 - Accuracy: 0.394 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1031 | 1.0 | 10 | 1.1286 | 0.1 | | 1.0648 | 2.0 | 20 | 1.1157 | 0.3 | | 0.9982 | 3.0 | 30 | 1.1412 | 0.2 | | 0.9283 | 4.0 | 40 | 1.2053 | 0.2 | | 0.7958 | 5.0 | 50 | 1.1466 | 0.2 | | 0.6668 | 6.0 | 60 | 1.1783 | 0.3 | | 0.5068 | 7.0 | 70 | 1.2992 | 0.3 | | 0.3741 | 8.0 | 80 | 1.3483 | 0.3 | | 0.1653 | 9.0 | 90 | 1.4533 | 0.2 | | 0.0946 | 10.0 | 100 | 1.6292 | 0.2 | | 0.0569 | 11.0 | 110 | 1.8381 | 0.2 | | 0.0346 | 12.0 | 120 | 2.0781 | 0.2 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
416
SetFit/distilbert-base-uncased__hate_speech_offensive__train-16-9
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-16-9 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-16-9 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.1121 - Accuracy: 0.16 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1038 | 1.0 | 10 | 1.1243 | 0.1 | | 1.0859 | 2.0 | 20 | 1.1182 | 0.2 | | 1.0234 | 3.0 | 30 | 1.1442 | 0.3 | | 0.9493 | 4.0 | 40 | 1.2239 | 0.1 | | 0.8114 | 5.0 | 50 | 1.2023 | 0.4 | | 0.6464 | 6.0 | 60 | 1.2329 | 0.4 | | 0.4731 | 7.0 | 70 | 1.2971 | 0.5 | | 0.3355 | 8.0 | 80 | 1.3913 | 0.4 | | 0.1268 | 9.0 | 90 | 1.4670 | 0.5 | | 0.0747 | 10.0 | 100 | 1.7961 | 0.4 | | 0.0449 | 11.0 | 110 | 1.8168 | 0.5 | | 0.0307 | 12.0 | 120 | 1.9307 | 0.4 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
417
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-0
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-0 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7714 - Accuracy: 0.705 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0871 | 1.0 | 19 | 1.0704 | 0.45 | | 1.0019 | 2.0 | 38 | 1.0167 | 0.55 | | 0.8412 | 3.0 | 57 | 0.9134 | 0.55 | | 0.6047 | 4.0 | 76 | 0.8430 | 0.6 | | 0.3746 | 5.0 | 95 | 0.8315 | 0.6 | | 0.1885 | 6.0 | 114 | 0.8585 | 0.6 | | 0.0772 | 7.0 | 133 | 0.9443 | 0.65 | | 0.0312 | 8.0 | 152 | 1.1019 | 0.65 | | 0.0161 | 9.0 | 171 | 1.1420 | 0.65 | | 0.0102 | 10.0 | 190 | 1.2773 | 0.65 | | 0.0077 | 11.0 | 209 | 1.2454 | 0.65 | | 0.0064 | 12.0 | 228 | 1.2785 | 0.65 | | 0.006 | 13.0 | 247 | 1.3834 | 0.65 | | 0.0045 | 14.0 | 266 | 1.4139 | 0.65 | | 0.0043 | 15.0 | 285 | 1.4056 | 0.65 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
418
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-1
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-1 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0606 - Accuracy: 0.4745 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0941 | 1.0 | 19 | 1.1045 | 0.2 | | 0.9967 | 2.0 | 38 | 1.1164 | 0.35 | | 0.8164 | 3.0 | 57 | 1.1570 | 0.4 | | 0.5884 | 4.0 | 76 | 1.2403 | 0.35 | | 0.3322 | 5.0 | 95 | 1.3815 | 0.35 | | 0.156 | 6.0 | 114 | 1.8102 | 0.3 | | 0.0576 | 7.0 | 133 | 2.1439 | 0.4 | | 0.0227 | 8.0 | 152 | 2.4368 | 0.3 | | 0.0133 | 9.0 | 171 | 2.5994 | 0.4 | | 0.009 | 10.0 | 190 | 2.7388 | 0.35 | | 0.0072 | 11.0 | 209 | 2.8287 | 0.35 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
419
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-2
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-2 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7136 - Accuracy: 0.679 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1052 | 1.0 | 19 | 1.0726 | 0.45 | | 1.0421 | 2.0 | 38 | 1.0225 | 0.5 | | 0.9173 | 3.0 | 57 | 0.9164 | 0.6 | | 0.6822 | 4.0 | 76 | 0.8251 | 0.7 | | 0.4407 | 5.0 | 95 | 0.8908 | 0.5 | | 0.2367 | 6.0 | 114 | 0.6772 | 0.75 | | 0.1145 | 7.0 | 133 | 0.7792 | 0.65 | | 0.0479 | 8.0 | 152 | 1.0657 | 0.6 | | 0.0186 | 9.0 | 171 | 1.2228 | 0.65 | | 0.0111 | 10.0 | 190 | 1.1100 | 0.6 | | 0.0083 | 11.0 | 209 | 1.1991 | 0.65 | | 0.0067 | 12.0 | 228 | 1.2654 | 0.65 | | 0.0061 | 13.0 | 247 | 1.2837 | 0.65 | | 0.0046 | 14.0 | 266 | 1.2860 | 0.6 | | 0.0043 | 15.0 | 285 | 1.3160 | 0.65 | | 0.0037 | 16.0 | 304 | 1.3323 | 0.65 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
420
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-3
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-3 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.8286 - Accuracy: 0.661 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1041 | 1.0 | 19 | 1.0658 | 0.5 | | 1.009 | 2.0 | 38 | 0.9892 | 0.7 | | 0.7925 | 3.0 | 57 | 0.8516 | 0.7 | | 0.5279 | 4.0 | 76 | 0.7877 | 0.65 | | 0.2932 | 5.0 | 95 | 0.7592 | 0.65 | | 0.1166 | 6.0 | 114 | 0.9437 | 0.65 | | 0.044 | 7.0 | 133 | 1.0315 | 0.75 | | 0.0197 | 8.0 | 152 | 1.3513 | 0.55 | | 0.0126 | 9.0 | 171 | 1.1702 | 0.7 | | 0.0083 | 10.0 | 190 | 1.2272 | 0.7 | | 0.0068 | 11.0 | 209 | 1.2889 | 0.7 | | 0.0059 | 12.0 | 228 | 1.3073 | 0.7 | | 0.0052 | 13.0 | 247 | 1.3595 | 0.7 | | 0.0041 | 14.0 | 266 | 1.4443 | 0.7 | | 0.0038 | 15.0 | 285 | 1.4709 | 0.7 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
421
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-4
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-4 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7384 - Accuracy: 0.724 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1013 | 1.0 | 19 | 1.0733 | 0.55 | | 1.0226 | 2.0 | 38 | 1.0064 | 0.65 | | 0.8539 | 3.0 | 57 | 0.8758 | 0.75 | | 0.584 | 4.0 | 76 | 0.6941 | 0.7 | | 0.2813 | 5.0 | 95 | 0.5151 | 0.7 | | 0.1122 | 6.0 | 114 | 0.4351 | 0.8 | | 0.0432 | 7.0 | 133 | 0.4896 | 0.85 | | 0.0199 | 8.0 | 152 | 0.5391 | 0.85 | | 0.0126 | 9.0 | 171 | 0.5200 | 0.85 | | 0.0085 | 10.0 | 190 | 0.5622 | 0.85 | | 0.0069 | 11.0 | 209 | 0.5950 | 0.85 | | 0.0058 | 12.0 | 228 | 0.6015 | 0.85 | | 0.0053 | 13.0 | 247 | 0.6120 | 0.85 | | 0.0042 | 14.0 | 266 | 0.6347 | 0.85 | | 0.0039 | 15.0 | 285 | 0.6453 | 0.85 | | 0.0034 | 16.0 | 304 | 0.6660 | 0.85 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
422
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-5
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-5 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-5 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.1327 - Accuracy: 0.57 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0972 | 1.0 | 19 | 1.0470 | 0.45 | | 0.9738 | 2.0 | 38 | 0.9244 | 0.65 | | 0.7722 | 3.0 | 57 | 0.8612 | 0.65 | | 0.4929 | 4.0 | 76 | 0.6759 | 0.75 | | 0.2435 | 5.0 | 95 | 0.7273 | 0.7 | | 0.0929 | 6.0 | 114 | 0.6444 | 0.85 | | 0.0357 | 7.0 | 133 | 0.7671 | 0.8 | | 0.0173 | 8.0 | 152 | 0.7599 | 0.75 | | 0.0121 | 9.0 | 171 | 0.8140 | 0.8 | | 0.0081 | 10.0 | 190 | 0.7861 | 0.8 | | 0.0066 | 11.0 | 209 | 0.8318 | 0.8 | | 0.0057 | 12.0 | 228 | 0.8777 | 0.8 | | 0.0053 | 13.0 | 247 | 0.8501 | 0.8 | | 0.004 | 14.0 | 266 | 0.8603 | 0.8 | | 0.004 | 15.0 | 285 | 0.8787 | 0.8 | | 0.0034 | 16.0 | 304 | 0.8969 | 0.8 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
423
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-6
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-6 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0523 - Accuracy: 0.663 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0957 | 1.0 | 19 | 1.0696 | 0.6 | | 1.0107 | 2.0 | 38 | 1.0047 | 0.55 | | 0.8257 | 3.0 | 57 | 0.8358 | 0.8 | | 0.6006 | 4.0 | 76 | 0.7641 | 0.6 | | 0.4172 | 5.0 | 95 | 0.5931 | 0.8 | | 0.2639 | 6.0 | 114 | 0.5570 | 0.7 | | 0.1314 | 7.0 | 133 | 0.5017 | 0.65 | | 0.0503 | 8.0 | 152 | 0.3115 | 0.75 | | 0.023 | 9.0 | 171 | 0.4353 | 0.85 | | 0.0128 | 10.0 | 190 | 0.5461 | 0.75 | | 0.0092 | 11.0 | 209 | 0.5045 | 0.8 | | 0.007 | 12.0 | 228 | 0.5014 | 0.8 | | 0.0064 | 13.0 | 247 | 0.5070 | 0.8 | | 0.0049 | 14.0 | 266 | 0.4681 | 0.8 | | 0.0044 | 15.0 | 285 | 0.4701 | 0.8 | | 0.0039 | 16.0 | 304 | 0.4862 | 0.8 | | 0.0036 | 17.0 | 323 | 0.4742 | 0.8 | | 0.0035 | 18.0 | 342 | 0.4652 | 0.8 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
424
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-7
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-7 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-7 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.8210 - Accuracy: 0.6305 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0989 | 1.0 | 19 | 1.0655 | 0.4 | | 1.0102 | 2.0 | 38 | 0.9927 | 0.6 | | 0.8063 | 3.0 | 57 | 0.9117 | 0.5 | | 0.5284 | 4.0 | 76 | 0.8058 | 0.55 | | 0.2447 | 5.0 | 95 | 0.8393 | 0.45 | | 0.098 | 6.0 | 114 | 0.8438 | 0.6 | | 0.0388 | 7.0 | 133 | 1.1901 | 0.45 | | 0.0188 | 8.0 | 152 | 1.4429 | 0.45 | | 0.0121 | 9.0 | 171 | 1.3648 | 0.4 | | 0.0082 | 10.0 | 190 | 1.4768 | 0.4 | | 0.0066 | 11.0 | 209 | 1.4830 | 0.45 | | 0.0057 | 12.0 | 228 | 1.4936 | 0.45 | | 0.0053 | 13.0 | 247 | 1.5649 | 0.4 | | 0.0041 | 14.0 | 266 | 1.6306 | 0.4 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
425
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-8
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-8 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-8 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.9191 - Accuracy: 0.632 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1008 | 1.0 | 19 | 1.0877 | 0.4 | | 1.0354 | 2.0 | 38 | 1.0593 | 0.35 | | 0.8765 | 3.0 | 57 | 0.9722 | 0.5 | | 0.6365 | 4.0 | 76 | 0.9271 | 0.55 | | 0.3944 | 5.0 | 95 | 0.7852 | 0.5 | | 0.2219 | 6.0 | 114 | 0.9360 | 0.55 | | 0.126 | 7.0 | 133 | 1.0610 | 0.55 | | 0.0389 | 8.0 | 152 | 1.0884 | 0.6 | | 0.0191 | 9.0 | 171 | 1.3483 | 0.55 | | 0.0108 | 10.0 | 190 | 1.4226 | 0.55 | | 0.0082 | 11.0 | 209 | 1.4270 | 0.55 | | 0.0065 | 12.0 | 228 | 1.5074 | 0.55 | | 0.0059 | 13.0 | 247 | 1.5577 | 0.55 | | 0.0044 | 14.0 | 266 | 1.5798 | 0.55 | | 0.0042 | 15.0 | 285 | 1.6196 | 0.55 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
426
SetFit/distilbert-base-uncased__hate_speech_offensive__train-32-9
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-32-9 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-32-9 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7075 - Accuracy: 0.692 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1054 | 1.0 | 19 | 1.0938 | 0.35 | | 1.0338 | 2.0 | 38 | 1.0563 | 0.65 | | 0.8622 | 3.0 | 57 | 0.9372 | 0.6 | | 0.5919 | 4.0 | 76 | 0.8461 | 0.6 | | 0.3357 | 5.0 | 95 | 1.0206 | 0.45 | | 0.1621 | 6.0 | 114 | 0.9802 | 0.7 | | 0.0637 | 7.0 | 133 | 1.2434 | 0.65 | | 0.0261 | 8.0 | 152 | 1.3865 | 0.65 | | 0.0156 | 9.0 | 171 | 1.4414 | 0.7 | | 0.01 | 10.0 | 190 | 1.5502 | 0.7 | | 0.0079 | 11.0 | 209 | 1.6102 | 0.7 | | 0.0062 | 12.0 | 228 | 1.6525 | 0.7 | | 0.0058 | 13.0 | 247 | 1.6884 | 0.7 | | 0.0046 | 14.0 | 266 | 1.7479 | 0.7 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
427
SetFit/distilbert-base-uncased__hate_speech_offensive__train-8-0
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-8-0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-8-0 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.1097 - Accuracy: 0.132 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1065 | 1.0 | 5 | 1.1287 | 0.0 | | 1.0592 | 2.0 | 10 | 1.1729 | 0.0 | | 1.0059 | 3.0 | 15 | 1.1959 | 0.0 | | 0.9129 | 4.0 | 20 | 1.2410 | 0.0 | | 0.8231 | 5.0 | 25 | 1.2820 | 0.0 | | 0.7192 | 6.0 | 30 | 1.3361 | 0.0 | | 0.6121 | 7.0 | 35 | 1.4176 | 0.0 | | 0.5055 | 8.0 | 40 | 1.5111 | 0.0 | | 0.4002 | 9.0 | 45 | 1.5572 | 0.0 | | 0.3788 | 10.0 | 50 | 1.6733 | 0.0 | | 0.2755 | 11.0 | 55 | 1.7381 | 0.2 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
428
SetFit/distilbert-base-uncased__hate_speech_offensive__train-8-1
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-8-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-8-1 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.1013 - Accuracy: 0.0915 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0866 | 1.0 | 5 | 1.1363 | 0.0 | | 1.0439 | 2.0 | 10 | 1.1803 | 0.0 | | 1.0227 | 3.0 | 15 | 1.2162 | 0.2 | | 0.9111 | 4.0 | 20 | 1.2619 | 0.0 | | 0.8243 | 5.0 | 25 | 1.2929 | 0.2 | | 0.7488 | 6.0 | 30 | 1.3010 | 0.2 | | 0.62 | 7.0 | 35 | 1.3011 | 0.2 | | 0.5054 | 8.0 | 40 | 1.2931 | 0.4 | | 0.4191 | 9.0 | 45 | 1.3274 | 0.4 | | 0.4107 | 10.0 | 50 | 1.3259 | 0.4 | | 0.3376 | 11.0 | 55 | 1.2800 | 0.4 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
429
SetFit/distilbert-base-uncased__hate_speech_offensive__train-8-2
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-8-2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-8-2 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.1019 - Accuracy: 0.139 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1082 | 1.0 | 5 | 1.1432 | 0.0 | | 1.0524 | 2.0 | 10 | 1.1613 | 0.0 | | 1.0641 | 3.0 | 15 | 1.1547 | 0.0 | | 0.9592 | 4.0 | 20 | 1.1680 | 0.0 | | 0.9085 | 5.0 | 25 | 1.1762 | 0.0 | | 0.8508 | 6.0 | 30 | 1.1809 | 0.2 | | 0.7263 | 7.0 | 35 | 1.1912 | 0.2 | | 0.6448 | 8.0 | 40 | 1.2100 | 0.2 | | 0.5378 | 9.0 | 45 | 1.2037 | 0.2 | | 0.5031 | 10.0 | 50 | 1.2096 | 0.2 | | 0.4041 | 11.0 | 55 | 1.2203 | 0.2 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
430
SetFit/distilbert-base-uncased__hate_speech_offensive__train-8-3
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-8-3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-8-3 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.9681 - Accuracy: 0.549 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1073 | 1.0 | 5 | 1.1393 | 0.0 | | 1.0392 | 2.0 | 10 | 1.1729 | 0.0 | | 1.0302 | 3.0 | 15 | 1.1694 | 0.2 | | 0.9176 | 4.0 | 20 | 1.1846 | 0.2 | | 0.8339 | 5.0 | 25 | 1.1663 | 0.2 | | 0.7533 | 6.0 | 30 | 1.1513 | 0.4 | | 0.6327 | 7.0 | 35 | 1.1474 | 0.4 | | 0.4402 | 8.0 | 40 | 1.1385 | 0.4 | | 0.3752 | 9.0 | 45 | 1.0965 | 0.2 | | 0.3448 | 10.0 | 50 | 1.0357 | 0.2 | | 0.2582 | 11.0 | 55 | 1.0438 | 0.2 | | 0.1903 | 12.0 | 60 | 1.0561 | 0.2 | | 0.1479 | 13.0 | 65 | 1.0569 | 0.2 | | 0.1129 | 14.0 | 70 | 1.0455 | 0.2 | | 0.1071 | 15.0 | 75 | 1.0416 | 0.4 | | 0.0672 | 16.0 | 80 | 1.1164 | 0.4 | | 0.0561 | 17.0 | 85 | 1.1846 | 0.6 | | 0.0463 | 18.0 | 90 | 1.2040 | 0.6 | | 0.0431 | 19.0 | 95 | 1.2078 | 0.6 | | 0.0314 | 20.0 | 100 | 1.2368 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
431
SetFit/distilbert-base-uncased__hate_speech_offensive__train-8-4
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-8-4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-8-4 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.1045 - Accuracy: 0.128 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1115 | 1.0 | 5 | 1.1174 | 0.0 | | 1.0518 | 2.0 | 10 | 1.1379 | 0.0 | | 1.0445 | 3.0 | 15 | 1.1287 | 0.0 | | 0.9306 | 4.0 | 20 | 1.1324 | 0.2 | | 0.8242 | 5.0 | 25 | 1.1219 | 0.2 | | 0.7986 | 6.0 | 30 | 1.1369 | 0.4 | | 0.7369 | 7.0 | 35 | 1.1732 | 0.2 | | 0.534 | 8.0 | 40 | 1.1828 | 0.6 | | 0.4285 | 9.0 | 45 | 1.1482 | 0.6 | | 0.3691 | 10.0 | 50 | 1.1401 | 0.6 | | 0.3215 | 11.0 | 55 | 1.1286 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
432
SetFit/distilbert-base-uncased__hate_speech_offensive__train-8-5
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-8-5 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-8-5 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.7214 - Accuracy: 0.37 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.0995 | 1.0 | 5 | 1.1301 | 0.0 | | 1.0227 | 2.0 | 10 | 1.1727 | 0.0 | | 1.0337 | 3.0 | 15 | 1.1734 | 0.2 | | 0.9137 | 4.0 | 20 | 1.1829 | 0.2 | | 0.8065 | 5.0 | 25 | 1.1496 | 0.4 | | 0.7038 | 6.0 | 30 | 1.1101 | 0.4 | | 0.6246 | 7.0 | 35 | 1.0982 | 0.2 | | 0.4481 | 8.0 | 40 | 1.0913 | 0.2 | | 0.3696 | 9.0 | 45 | 1.0585 | 0.4 | | 0.3137 | 10.0 | 50 | 1.0418 | 0.4 | | 0.2482 | 11.0 | 55 | 1.0078 | 0.4 | | 0.196 | 12.0 | 60 | 0.9887 | 0.6 | | 0.1344 | 13.0 | 65 | 0.9719 | 0.6 | | 0.1014 | 14.0 | 70 | 1.0053 | 0.6 | | 0.111 | 15.0 | 75 | 0.9653 | 0.6 | | 0.0643 | 16.0 | 80 | 0.9018 | 0.6 | | 0.0559 | 17.0 | 85 | 0.9393 | 0.6 | | 0.0412 | 18.0 | 90 | 1.0210 | 0.6 | | 0.0465 | 19.0 | 95 | 0.9965 | 0.6 | | 0.0328 | 20.0 | 100 | 0.9739 | 0.6 | | 0.0289 | 21.0 | 105 | 0.9796 | 0.6 | | 0.0271 | 22.0 | 110 | 0.9968 | 0.6 | | 0.0239 | 23.0 | 115 | 1.0143 | 0.6 | | 0.0201 | 24.0 | 120 | 1.0459 | 0.6 | | 0.0185 | 25.0 | 125 | 1.0698 | 0.6 | | 0.0183 | 26.0 | 130 | 1.0970 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
433
SetFit/distilbert-base-uncased__hate_speech_offensive__train-8-6
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-8-6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-8-6 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.1275 - Accuracy: 0.3795 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.11 | 1.0 | 5 | 1.1184 | 0.0 | | 1.0608 | 2.0 | 10 | 1.1227 | 0.0 | | 1.0484 | 3.0 | 15 | 1.1009 | 0.2 | | 0.9614 | 4.0 | 20 | 1.1009 | 0.2 | | 0.8545 | 5.0 | 25 | 1.0772 | 0.2 | | 0.8241 | 6.0 | 30 | 1.0457 | 0.2 | | 0.708 | 7.0 | 35 | 1.0301 | 0.4 | | 0.5045 | 8.0 | 40 | 1.0325 | 0.4 | | 0.4175 | 9.0 | 45 | 1.0051 | 0.4 | | 0.3446 | 10.0 | 50 | 0.9610 | 0.4 | | 0.2851 | 11.0 | 55 | 0.9954 | 0.4 | | 0.1808 | 12.0 | 60 | 1.0561 | 0.4 | | 0.1435 | 13.0 | 65 | 1.0218 | 0.4 | | 0.1019 | 14.0 | 70 | 1.0254 | 0.4 | | 0.0908 | 15.0 | 75 | 0.9935 | 0.4 | | 0.0591 | 16.0 | 80 | 1.0090 | 0.4 | | 0.0512 | 17.0 | 85 | 1.0884 | 0.4 | | 0.0397 | 18.0 | 90 | 1.2732 | 0.4 | | 0.039 | 19.0 | 95 | 1.2979 | 0.6 | | 0.0325 | 20.0 | 100 | 1.2705 | 0.4 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
434
SetFit/distilbert-base-uncased__hate_speech_offensive__train-8-7
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-8-7 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-8-7 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.1206 - Accuracy: 0.0555 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1186 | 1.0 | 5 | 1.1631 | 0.0 | | 1.058 | 2.0 | 10 | 1.1986 | 0.0 | | 1.081 | 3.0 | 15 | 1.2111 | 0.0 | | 1.0118 | 4.0 | 20 | 1.2373 | 0.0 | | 0.9404 | 5.0 | 25 | 1.2645 | 0.0 | | 0.9146 | 6.0 | 30 | 1.3258 | 0.0 | | 0.8285 | 7.0 | 35 | 1.3789 | 0.0 | | 0.6422 | 8.0 | 40 | 1.3783 | 0.0 | | 0.6156 | 9.0 | 45 | 1.3691 | 0.0 | | 0.5321 | 10.0 | 50 | 1.3693 | 0.0 | | 0.4504 | 11.0 | 55 | 1.4000 | 0.0 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3
435
SetFit/distilbert-base-uncased__hate_speech_offensive__train-8-8
[ "hate speech", "neither", "offensive language" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: distilbert-base-uncased__hate_speech_offensive__train-8-8 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased__hate_speech_offensive__train-8-8 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0005 - Accuracy: 0.518 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.1029 | 1.0 | 5 | 1.1295 | 0.0 | | 1.0472 | 2.0 | 10 | 1.1531 | 0.0 | | 1.054 | 3.0 | 15 | 1.1475 | 0.0 | | 0.9366 | 4.0 | 20 | 1.1515 | 0.0 | | 0.8698 | 5.0 | 25 | 1.1236 | 0.4 | | 0.8148 | 6.0 | 30 | 1.0716 | 0.6 | | 0.6884 | 7.0 | 35 | 1.0662 | 0.6 | | 0.5641 | 8.0 | 40 | 1.0671 | 0.6 | | 0.5 | 9.0 | 45 | 1.0282 | 0.6 | | 0.3882 | 10.0 | 50 | 1.0500 | 0.6 | | 0.3522 | 11.0 | 55 | 1.1381 | 0.6 | | 0.2492 | 12.0 | 60 | 1.1278 | 0.6 | | 0.2063 | 13.0 | 65 | 1.0731 | 0.6 | | 0.1608 | 14.0 | 70 | 1.1339 | 0.6 | | 0.1448 | 15.0 | 75 | 1.1892 | 0.6 | | 0.0925 | 16.0 | 80 | 1.1840 | 0.6 | | 0.0768 | 17.0 | 85 | 1.0608 | 0.6 | | 0.0585 | 18.0 | 90 | 1.1073 | 0.6 | | 0.0592 | 19.0 | 95 | 1.3134 | 0.6 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2 - Tokenizers 0.10.3