modelId stringlengths 6 107 | label list | readme stringlengths 0 56.2k | readme_len int64 0 56.2k |
|---|---|---|---|
tbrasil/classificador_de_atendimento_2_classes_v1.1 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
textattack/xlnet-large-cased-MRPC | null | Entry not found | 15 |
vesteinn/XLMR-ENIS-finetuned-sst2 | [
"negative",
"positive"
] | ---
license: agpl-3.0
tags:
- generated_from_trainer
datasets:
- glue
metrics:
- accuracy
model-index:
- name: XLMR-ENIS-finetuned-sst2
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: glue
type: glue
args: sst2
metrics:
- name: Accuracy
type: accuracy
value: 0.9277522935779816
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# XLMR-ENIS-finetuned-sst2
This model is a fine-tuned version of [vesteinn/XLMR-ENIS](https://huggingface.co/vesteinn/XLMR-ENIS) on the glue dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3781
- Accuracy: 0.9278
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.0675 | 1.0 | 4210 | 0.3781 | 0.9278 |
### Framework versions
- Transformers 4.11.0
- Pytorch 1.9.0+cu102
- Datasets 1.12.1
- Tokenizers 0.10.3
| 1,579 |
mariagrandury/distilbert-base-uncased-finetuned-sms-spam-detection | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- sms_spam
metrics:
- accuracy
model-index:
- name: distilbert-base-uncased-finetuned-sms-spam-detection
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: sms_spam
type: sms_spam
args: plain_text
metrics:
- name: Accuracy
type: accuracy
value: 0.9921090387374462
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-sms-spam-detection
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the sms_spam dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0426
- Accuracy: 0.9921
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.0375 | 1.0 | 262 | 0.0549 | 0.9892 |
| 0.0205 | 2.0 | 524 | 0.0426 | 0.9921 |
### Framework versions
- Transformers 4.16.2
- Pytorch 1.10.0+cu111
- Datasets 1.18.3
- Tokenizers 0.11.0
| 1,732 |
ekohrt/qcat | [
"arithmetic",
"average",
"boolean_and",
"boolean_or",
"boolean_retrieval",
"causal_explanation",
"correlation",
"counting",
"datetime_comparison",
"datetime_retrieval",
"definitional",
"mathematical_comparison",
"median",
"mode",
"numeric_comparison",
"numeric_retrieval",
"opinion",
... | ---
license: mit
---
# **Q-Cat**
A pre-trained Distilbert model for classifying question types. For use in QA systems.
Dataset contains ~800 labeled examples. Classifier uses a taxonomy of 27 question types. | 215 |
billfrench/autonlp-cyberlandr-ai-4-614417501 | [
"clear windows",
"close door",
"opaque windows",
"open door"
] | ---
tags: autonlp
language: en
widget:
- text: "I love AutoNLP 🤗"
datasets:
- billfrench/autonlp-data-cyberlandr-ai-4
co2_eq_emissions: 1.6912535041856878
---
# Model Trained Using AutoNLP
- Problem type: Multi-class Classification
- Model ID: 614417501
- CO2 Emissions (in grams): 1.6912535041856878
## Validation Metrics
- Loss: 1.305419921875
- Accuracy: 0.5
- Macro F1: 0.3333333333333333
- Micro F1: 0.5
- Weighted F1: 0.4444444444444444
- Macro Precision: 0.375
- Micro Precision: 0.5
- Weighted Precision: 0.5
- Macro Recall: 0.375
- Micro Recall: 0.5
- Weighted Recall: 0.5
## Usage
You can use cURL to access this model:
```
$ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/billfrench/autonlp-cyberlandr-ai-4-614417501
```
Or Python API:
```
from transformers import AutoModelForSequenceClassification, AutoTokenizer
model = AutoModelForSequenceClassification.from_pretrained("billfrench/autonlp-cyberlandr-ai-4-614417501", use_auth_token=True)
tokenizer = AutoTokenizer.from_pretrained("billfrench/autonlp-cyberlandr-ai-4-614417501", use_auth_token=True)
inputs = tokenizer("I love AutoNLP", return_tensors="pt")
outputs = model(**inputs)
``` | 1,286 |
antho-data/distilbert-base-uncased-finetuned-emotion | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model-index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.9235
- name: F1
type: f1
value: 0.9237367861627231
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2294
- Accuracy: 0.9235
- F1: 0.9237
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8637 | 1.0 | 250 | 0.3319 | 0.9075 | 0.9050 |
| 0.2634 | 2.0 | 500 | 0.2294 | 0.9235 | 0.9237 |
### Framework versions
- Transformers 4.17.0
- Pytorch 1.10.0+cu111
- Datasets 1.18.4
- Tokenizers 0.11.6
| 1,807 |
farrokhguiahi/distilbert-base-uncased-finetuned-emotion | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model-index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.9275
- name: F1
type: f1
value: 0.9278961513392271
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2168
- Accuracy: 0.9275
- F1: 0.9279
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8418 | 1.0 | 250 | 0.3102 | 0.905 | 0.9012 |
| 0.2454 | 2.0 | 500 | 0.2168 | 0.9275 | 0.9279 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
| 1,807 |
negfir/distilbert-base-uncased-finetuned-cola | null | ---
tags:
- generated_from_keras_callback
model-index:
- name: negfir/distilbert-base-uncased-finetuned-cola
results: []
---
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# negfir/distilbert-base-uncased-finetuned-cola
This model is a fine-tuned version of [negfir/uncased_L-12_H-128_A-2](https://huggingface.co/negfir/uncased_L-12_H-128_A-2) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.6077
- Validation Loss: 0.6185
- Train Matthews Correlation: 0.0
- Epoch: 2
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 2670, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
- training_precision: float32
### Training results
| Train Loss | Validation Loss | Train Matthews Correlation | Epoch |
|:----------:|:---------------:|:--------------------------:|:-----:|
| 0.6116 | 0.6187 | 0.0 | 0 |
| 0.6070 | 0.6190 | 0.0 | 1 |
| 0.6077 | 0.6185 | 0.0 | 2 |
### Framework versions
- Transformers 4.17.0
- TensorFlow 2.8.0
- Datasets 2.0.0
- Tokenizers 0.11.6
| 1,734 |
cambridgeltl/guardian_news_bert-base-uncased | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4"
] | Entry not found | 15 |
clapika2010/movies_predictions | null | Entry not found | 15 |
celine98/canine-c-finetuned-sst2 | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- glue
metrics:
- accuracy
model-index:
- name: canine-c-finetuned-sst2
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: glue
type: glue
args: sst2
metrics:
- name: Accuracy
type: accuracy
value: 0.8486238532110092
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# canine-c-finetuned-sst2
This model is a fine-tuned version of [google/canine-c](https://huggingface.co/google/canine-c) on the glue dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6025
- Accuracy: 0.8486
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 4.9121586874695155e-05
- train_batch_size: 32
- eval_batch_size: 16
- seed: 8
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 4
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.3415 | 1.0 | 2105 | 0.4196 | 0.8280 |
| 0.2265 | 2.0 | 4210 | 0.4924 | 0.8211 |
| 0.1439 | 3.0 | 6315 | 0.5726 | 0.8337 |
| 0.0974 | 4.0 | 8420 | 0.6025 | 0.8486 |
### Framework versions
- Transformers 4.17.0
- Pytorch 1.10.0+cu111
- Datasets 2.0.0
- Tokenizers 0.11.6
| 1,775 |
dangvantuan/CrossEncoder-camembert-large | [
"LABEL_0"
] | ---
pipeline_tag: sentence-similarity
language: fr
datasets:
- stsb_multi_mt
tags:
- Text
- Sentence Similarity
- Sentence-Embedding
- camembert-base
license: apache-2.0
model-index:
- name: sentence-camembert-base by Van Tuan DANG
results:
- task:
name: Sentence-Embedding
type: Text Similarity
dataset:
name: Text Similarity fr
type: stsb_multi_mt
args: fr
metrics:
- name: Test Pearson correlation coefficient
type: Pearson_correlation_coefficient
value: xx.xx
---
## Model
Cross-Encoder for sentence-similarity
This model was trained using [sentence-transformers](https://www.SBERT.net) Cross-Encoder class.
## Training Data
This model was trained on the [STS benchmark dataset](https://huggingface.co/datasets/stsb_multi_mt/viewer/fr/train). The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('dangvantuan/CrossEncoder-camembert-large', max_length=128)
scores = model.predict([('Un avion est en train de décoller.', "Un homme joue d'une grande flûte."), ("Un homme étale du fromage râpé sur une pizza.", "Une personne jette un chat au plafond") ])
```
## Evaluation
The model can be evaluated as follows on the French test data of stsb.
```python
from sentence_transformers.readers import InputExample
from sentence_transformers.cross_encoder.evaluation import CECorrelationEvaluator
from datasets import load_dataset
def convert_dataset(dataset):
dataset_samples=[]
for df in dataset:
score = float(df['similarity_score'])/5.0 # Normalize score to range 0 ... 1
inp_example = InputExample(texts=[df['sentence1'],
df['sentence2']], label=score)
dataset_samples.append(inp_example)
return dataset_samples
# Loading the dataset for evaluation
df_dev = load_dataset("stsb_multi_mt", name="fr", split="dev")
df_test = load_dataset("stsb_multi_mt", name="fr", split="test")
# Convert the dataset for evaluation
# For Dev set:
dev_samples = convert_dataset(df_dev)
val_evaluator = CECorrelationEvaluator.from_input_examples(dev_samples, name='sts-dev')
val_evaluator(model, output_path="./")
# For Test set
test_samples = convert_dataset(df_test)
test_evaluator = CECorrelationEvaluator.from_input_examples(test_samples, name='sts-test')
test_evaluator(models, output_path="./")
```
**Test Result**:
The performance is measured using Pearson and Spearman correlation:
- On dev
| Model | Pearson correlation | Spearman correlation | #params |
| ------------- | ------------- | ------------- |------------- |
| [dangvantuan/CrossEncoder-camembert-large](https://huggingface.co/dangvantuan/CrossEncoder-camembert-large)| 90.11 |90.01 | 336M |
- On test
| Model | Pearson correlation | Spearman correlation |
| ------------- | ------------- | ------------- |
| [dangvantuan/CrossEncoder-camembert-large](https://huggingface.co/dangvantuan/CrossEncoder-camembert-large)| 88.16 | 87.57| | 3,288 |
KeithHorgan/TweetClimateAnalysis | [
"No claim",
"Global warming is not happening - Ice/permafrost/snow cover isn’t melting",
"Global warming is not happening - We’re heading into an ice age/global cooling",
"Global warming is not happening - Weather is cold/snowing",
"Global warming is not happening - Climate hasn’t warmed/changed over the la... | ---
tags: autotrain
language: unk
widget:
- text: "Climate Change is a hoax"
- text: "It is freezing, where is global warming"
datasets:
- KeithHorgan98/autotrain-data-TweetClimateAnalysis
co2_eq_emissions: 133.19491276284793
---
# Model Trained Using AutoTrain
- Problem type: Multi-class Classification
- Model ID: 678720226
- CO2 Emissions (in grams): 133.19491276284793
## Validation Metrics
- Loss: 0.4864234924316406
- Accuracy: 0.865424430641822
- Macro F1: 0.7665472174344069
- Micro F1: 0.8654244306418221
- Weighted F1: 0.8586375445115083
- Macro Precision: 0.8281449061702826
- Micro Precision: 0.865424430641822
- Weighted Precision: 0.8619727477790186
- Macro Recall: 0.736576343905098
- Micro Recall: 0.865424430641822
- Weighted Recall: 0.865424430641822
## Usage
You can use cURL to access this model:
```
$ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoTrain"}' https://api-inference.huggingface.co/models/KeithHorgan98/autotrain-TweetClimateAnalysis-678720226
```
Or Python API:
```
from transformers import AutoModelForSequenceClassification, AutoTokenizer
model = AutoModelForSequenceClassification.from_pretrained("KeithHorgan98/autotrain-TweetClimateAnalysis-678720226", use_auth_token=True)
tokenizer = AutoTokenizer.from_pretrained("KeithHorgan98/autotrain-TweetClimateAnalysis-678720226", use_auth_token=True)
inputs = tokenizer("I love AutoTrain", return_tensors="pt")
outputs = model(**inputs)
``` | 1,508 |
princeton-nlp/CoFi-QNLI-s95 | null | This is a model checkpoint for "[Structured Pruning Learns Compact and Accurate Models](https://arxiv.org/pdf/2204.00408.pdf)". The model is pruned from `bert-base-uncased` to a 95% sparsity on dataset QNLI. Please go to [our repository](https://github.com/princeton-nlp/CoFiPruning) for more details on how to use the model for inference. Note that you would have to use the model class specified in our repository to load the model.
| 435 |
abdelrahmanzied/bert-fake-news-classifier | null | ---
license: mit
---
| 24 |
Sam4669/distilbert-base-uncased-finetuned-emotion | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model-index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.923
- name: F1
type: f1
value: 0.9232158277556175
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2317
- Accuracy: 0.923
- F1: 0.9232
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8528 | 1.0 | 250 | 0.3332 | 0.897 | 0.8929 |
| 0.26 | 2.0 | 500 | 0.2317 | 0.923 | 0.9232 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
| 1,805 |
aprilzoo/distilbert-base-uncased-finetuned-emotion | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model-index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.923
- name: F1
type: f1
value: 0.9232474678171817
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2202
- Accuracy: 0.923
- F1: 0.9232
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8244 | 1.0 | 250 | 0.3104 | 0.9025 | 0.8997 |
| 0.2478 | 2.0 | 500 | 0.2202 | 0.923 | 0.9232 |
### Framework versions
- Transformers 4.17.0
- Pytorch 1.10.0+cu111
- Datasets 2.0.0
- Tokenizers 0.11.6
| 1,804 |
JNK789/distilbert-base-uncased-finetuned-tweets-emoji-dataset | [
"LABEL_0",
"LABEL_1",
"LABEL_10",
"LABEL_11",
"LABEL_12",
"LABEL_13",
"LABEL_14",
"LABEL_15",
"LABEL_16",
"LABEL_17",
"LABEL_18",
"LABEL_19",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5",
"LABEL_6",
"LABEL_7",
"LABEL_8",
"LABEL_9"
] | Entry not found | 15 |
ChrisZeng/bertweet-base-cased-covid19-hateval | null | ---
tags:
- generated_from_trainer
metrics:
- accuracy
- f1
model-index:
- name: bertweet-base-cased-covid19-hateval
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bertweet-base-cased-covid19-hateval
This model is a fine-tuned version of [vinai/bertweet-covid19-base-cased](https://huggingface.co/vinai/bertweet-covid19-base-cased) on the HatEval dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4817
- Accuracy: 0.773
- F1: 0.7722
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-06
- train_batch_size: 32
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 30
### Training results
| Training Loss | Epoch | Step | Accuracy | F1 | Validation Loss |
|:-------------:|:-----:|:----:|:--------:|:------:|:---------------:|
| 0.6925 | 0.99 | 70 | 0.573 | 0.3643 | 0.6827 |
| 0.6823 | 1.99 | 140 | 0.573 | 0.3643 | 0.6736 |
| 0.6713 | 2.99 | 210 | 0.587 | 0.3993 | 0.6568 |
| 0.6468 | 3.99 | 280 | 0.7 | 0.6708 | 0.6210 |
| 0.6047 | 4.99 | 350 | 0.732 | 0.7286 | 0.5785 |
| 0.5648 | 5.99 | 420 | 0.733 | 0.7319 | 0.5537 |
| 0.536 | 6.99 | 490 | 0.739 | 0.7381 | 0.5406 |
| 0.5175 | 7.99 | 560 | 0.744 | 0.7431 | 0.5308 |
| 0.5018 | 8.99 | 630 | 0.751 | 0.7504 | 0.5235 |
| 0.4874 | 9.99 | 700 | 0.749 | 0.7479 | 0.5145 |
| 0.4749 | 10.99 | 770 | 0.754 | 0.7533 | 0.5104 |
| 0.4666 | 11.99 | 840 | 0.761 | 0.7605 | 0.5052 |
| 0.456 | 12.99 | 910 | 0.761 | 0.7604 | 0.5017 |
| 0.4489 | 13.99 | 980 | 0.764 | 0.7635 | 0.4986 |
| 0.4375 | 14.99 | 1050 | 0.764 | 0.7625 | 0.4932 |
| 0.4319 | 15.99 | 1120 | 0.762 | 0.7608 | 0.4917 |
| 0.427 | 16.99 | 1190 | 0.77 | 0.7693 | 0.4918 |
| 0.4226 | 17.99 | 1260 | 0.772 | 0.7711 | 0.4889 |
| 0.4167 | 18.99 | 1330 | 0.769 | 0.7681 | 0.4874 |
| 0.4127 | 19.99 | 1400 | 0.768 | 0.7673 | 0.4868 |
| 0.4095 | 20.99 | 1470 | 0.774 | 0.7731 | 0.4836 |
| 0.4066 | 21.99 | 1540 | 0.77 | 0.7690 | 0.4829 |
| 0.405 | 22.99 | 1610 | 0.773 | 0.7721 | 0.4822 |
| 0.3993 | 23.99 | 1680 | 0.77 | 0.7692 | 0.4827 |
| 0.3977 | 24.99 | 1750 | 0.4831 | 0.772 | 0.7712 |
| 0.398 | 25.99 | 1820 | 0.4830 | 0.774 | 0.7733 |
| 0.3969 | 26.99 | 1890 | 0.4815 | 0.771 | 0.7701 |
| 0.3945 | 27.99 | 1960 | 0.4818 | 0.772 | 0.7712 |
| 0.3929 | 28.99 | 2030 | 0.4818 | 0.773 | 0.7722 |
| 0.3887 | 29.99 | 2100 | 0.4817 | 0.773 | 0.7722 |
### Framework versions
- Transformers 4.17.0
- Pytorch 1.11.0+cu113
- Datasets 2.0.0
- Tokenizers 0.11.6
| 3,543 |
btjiong/robbert-twitter-sentiment-tokenized | [
"NEGATIEF",
"NEUTRAAL",
"POSITIEF"
] | ---
license: mit
tags:
- generated_from_trainer
datasets:
- dutch_social
metrics:
- accuracy
- f1
- precision
- recall
model-index:
- name: robbert-twitter-sentiment-tokenized
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: dutch_social
type: dutch_social
args: dutch_social
metrics:
- name: Accuracy
type: accuracy
value: 0.814
- name: F1
type: f1
value: 0.8132800039281481
- name: Precision
type: precision
value: 0.8131073640029836
- name: Recall
type: recall
value: 0.814
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# robbert-twitter-sentiment-tokenized
This model is a fine-tuned version of [pdelobelle/robbert-v2-dutch-base](https://huggingface.co/pdelobelle/robbert-v2-dutch-base) on the dutch_social dataset.
It achieves the following results on the evaluation set:
- Loss: 0.5473
- Accuracy: 0.814
- F1: 0.8133
- Precision: 0.8131
- Recall: 0.814
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
| 0.6895 | 1.0 | 282 | 0.6307 | 0.7433 | 0.7442 | 0.7500 | 0.7433 |
| 0.4948 | 2.0 | 564 | 0.5189 | 0.8053 | 0.8062 | 0.8081 | 0.8053 |
| 0.2642 | 3.0 | 846 | 0.5473 | 0.814 | 0.8133 | 0.8131 | 0.814 |
### Framework versions
- Transformers 4.17.0
- Pytorch 1.11.0+cpu
- Datasets 2.0.0
- Tokenizers 0.11.6
| 2,192 |
SiriusRen/my-awesome-model | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- imdb
model-index:
- name: my-awesome-model
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# my-awesome-model
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
### Framework versions
- Transformers 4.18.0.dev0
- Pytorch 1.10.0
- Datasets 2.0.1.dev0
- Tokenizers 0.11.6
| 1,060 |
Toshifumi/distilbert-base-uncased-finetuned-emotion | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model-index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.927
- name: F1
type: f1
value: 0.9271941874206031
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2106
- Accuracy: 0.927
- F1: 0.9272
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8009 | 1.0 | 250 | 0.2968 | 0.912 | 0.9102 |
| 0.24 | 2.0 | 500 | 0.2106 | 0.927 | 0.9272 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
| 1,805 |
SiriusRen/my-rubbish-model | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- imdb
model-index:
- name: my-rubbish-model
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# my-rubbish-model
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
### Framework versions
- Transformers 4.18.0.dev0
- Pytorch 1.10.0
- Datasets 2.0.1.dev0
- Tokenizers 0.11.6
| 1,060 |
Intel/albert-base-v2-sst2-int8-static | null | ---
language:
- en
license: apache-2.0
tags:
- text-classfication
- int8
- Intel® Neural Compressor
- PostTrainingStatic
datasets:
- glue
metrics:
- accuracy
model_index:
- name: sst2
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: GLUE SST2
type: glue
args: sst2
metric:
name: Accuracy
type: accuracy
value: 0.9254587155963303
---
# INT8 albert-base-v2-sst2
### Post-training static quantization
This is an INT8 PyTorch model quantized with [Intel® Neural Compressor](https://github.com/intel/neural-compressor).
The original fp32 model comes from the fine-tuned model [Alireza1044/albert-base-v2-sst2](https://huggingface.co/Alireza1044/albert-base-v2-sst2).
The calibration dataloader is the train dataloader. The default calibration sampling size 300 isn't divisible exactly by batch size 8, so the real sampling size is 304.
The linear modules **albert.encoder.albert_layer_groups.0.albert_layers.0.ffn_output.module, albert.encoder.albert_layer_groups.0.albert_layers.0.ffn.module** fall back to fp32 to meet the 1% relative accuracy loss.
### Test result
| |INT8|FP32|
|---|:---:|:---:|
| **Accuracy (eval-accuracy)** |0.9255|0.9232|
| **Model size (MB)** |25|44.6|
### Load with Intel® Neural Compressor:
```python
from neural_compressor.utils.load_huggingface import OptimizedModel
int8_model = OptimizedModel.from_pretrained(
'Intel/albert-base-v2-sst2-int8-static',
)
```
| 1,498 |
MartinoMensio/racism-models-w-m-vote-nonstrict-epoch-4 | null | ---
language: es
license: mit
widget:
- text: "y porqué es lo que hay que hacer con los menas y con los adultos también!!!! NO a los inmigrantes ilegales!!!!"
---
### Description
This model is a fine-tuned version of [BETO (spanish bert)](https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased) that has been trained on the *Datathon Against Racism* dataset (2022)
We performed several experiments that will be described in the upcoming paper "Estimating Ground Truth in a Low-labelled Data Regime:A Study of Racism Detection in Spanish" (NEATClasS 2022)
We applied 6 different methods ground-truth estimations, and for each one we performed 4 epochs of fine-tuning. The result is made of 24 models:
| method | epoch 1 | epoch 3 | epoch 3 | epoch 4 |
|--- |--- |--- |--- |--- |
| raw-label | [raw-label-epoch-1](https://huggingface.co/MartinoMensio/racism-models-raw-label-epoch-1) | [raw-label-epoch-2](https://huggingface.co/MartinoMensio/racism-models-raw-label-epoch-2) | [raw-label-epoch-3](https://huggingface.co/MartinoMensio/racism-models-raw-label-epoch-3) | [raw-label-epoch-4](https://huggingface.co/MartinoMensio/racism-models-raw-label-epoch-4) |
| m-vote-strict | [m-vote-strict-epoch-1](https://huggingface.co/MartinoMensio/racism-models-m-vote-strict-epoch-1) | [m-vote-strict-epoch-2](https://huggingface.co/MartinoMensio/racism-models-m-vote-strict-epoch-2) | [m-vote-strict-epoch-3](https://huggingface.co/MartinoMensio/racism-models-m-vote-strict-epoch-3) | [m-vote-strict-epoch-4](https://huggingface.co/MartinoMensio/racism-models-m-vote-strict-epoch-4) |
| m-vote-nonstrict | [m-vote-nonstrict-epoch-1](https://huggingface.co/MartinoMensio/racism-models-m-vote-nonstrict-epoch-1) | [m-vote-nonstrict-epoch-2](https://huggingface.co/MartinoMensio/racism-models-m-vote-nonstrict-epoch-2) | [m-vote-nonstrict-epoch-3](https://huggingface.co/MartinoMensio/racism-models-m-vote-nonstrict-epoch-3) | [m-vote-nonstrict-epoch-4](https://huggingface.co/MartinoMensio/racism-models-m-vote-nonstrict-epoch-4) |
| regression-w-m-vote | [regression-w-m-vote-epoch-1](https://huggingface.co/MartinoMensio/racism-models-regression-w-m-vote-epoch-1) | [regression-w-m-vote-epoch-2](https://huggingface.co/MartinoMensio/racism-models-regression-w-m-vote-epoch-2) | [regression-w-m-vote-epoch-3](https://huggingface.co/MartinoMensio/racism-models-regression-w-m-vote-epoch-3) | [regression-w-m-vote-epoch-4](https://huggingface.co/MartinoMensio/racism-models-regression-w-m-vote-epoch-4) |
| w-m-vote-strict | [w-m-vote-strict-epoch-1](https://huggingface.co/MartinoMensio/racism-models-w-m-vote-strict-epoch-1) | [w-m-vote-strict-epoch-2](https://huggingface.co/MartinoMensio/racism-models-w-m-vote-strict-epoch-2) | [w-m-vote-strict-epoch-3](https://huggingface.co/MartinoMensio/racism-models-w-m-vote-strict-epoch-3) | [w-m-vote-strict-epoch-4](https://huggingface.co/MartinoMensio/racism-models-w-m-vote-strict-epoch-4) |
| w-m-vote-nonstrict | [w-m-vote-nonstrict-epoch-1](https://huggingface.co/MartinoMensio/racism-models-w-m-vote-nonstrict-epoch-1) | [w-m-vote-nonstrict-epoch-2](https://huggingface.co/MartinoMensio/racism-models-w-m-vote-nonstrict-epoch-2) | [w-m-vote-nonstrict-epoch-3](https://huggingface.co/MartinoMensio/racism-models-w-m-vote-nonstrict-epoch-3) | [w-m-vote-nonstrict-epoch-4](https://huggingface.co/MartinoMensio/racism-models-w-m-vote-nonstrict-epoch-4) |
This model is `w-m-vote-nonstrict-epoch-4`
### Usage
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification, pipeline
model_name = 'w-m-vote-nonstrict-epoch-4'
tokenizer = AutoTokenizer.from_pretrained("dccuchile/bert-base-spanish-wwm-uncased")
full_model_path = f'MartinoMensio/racism-models-{model_name}'
model = AutoModelForSequenceClassification.from_pretrained(full_model_path)
pipe = pipeline("text-classification", model = model, tokenizer = tokenizer)
texts = [
'y porqué es lo que hay que hacer con los menas y con los adultos también!!!! NO a los inmigrantes ilegales!!!!',
'Es que los judíos controlan el mundo'
]
print(pipe(texts))
# [{'label': 'racist', 'score': 0.996863842010498}, {'label': 'non-racist', 'score': 0.9982976317405701}]
```
For more details, see https://github.com/preyero/neatclass22
| 4,269 |
ShreyaR/finetuned-distil-bert-depression | null | ---
license: apache-2.0
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: finetuned-distil-bert-depression
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetuned-distil-bert-depression
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1695
- Accuracy: 0.9445
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.0243 | 1.0 | 625 | 0.2303 | 0.9205 |
| 0.0341 | 2.0 | 1250 | 0.1541 | 0.933 |
| 0.0244 | 3.0 | 1875 | 0.1495 | 0.9445 |
### Framework versions
- Transformers 4.18.0
- Pytorch 1.11.0+cu113
- Datasets 2.1.0
- Tokenizers 0.12.1
| 1,495 |
Intel/camembert-base-mrpc-int8-dynamic | [
"0",
"1"
] | ---
language:
- en
license: mit
tags:
- text-classfication
- int8
- Intel® Neural Compressor
- PostTrainingDynamic
datasets:
- glue
metrics:
- f1
model-index:
- name: camembert-base-mrpc-int8-dynamic
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: GLUE MRPC
type: glue
args: mrpc
metrics:
- name: F1
type: f1
value: 0.8842832469775476
---
# INT8 camembert-base-mrpc
### Post-training dynamic quantization
This is an INT8 PyTorch model quantized with [Intel® Neural Compressor](https://github.com/intel/neural-compressor).
The original fp32 model comes from the fine-tuned model [camembert-base-mrpc](https://huggingface.co/Intel/camembert-base-mrpc).
The linear module **roberta.encoder.layer.6.attention.self.query** falls back to fp32 to meet the 1% relative accuracy loss.
### Test result
| |INT8|FP32|
|---|:---:|:---:|
| **Accuracy (eval-f1)** |0.8843|0.8928|
| **Model size (MB)** |180|422|
### Load with Intel® Neural Compressor:
```python
from neural_compressor.utils.load_huggingface import OptimizedModel
int8_model = OptimizedModel.from_pretrained(
'Intel/camembert-base-mrpc-int8-dynamic',
)
```
| 1,220 |
PrasunMishra/finetuning-sentiment-model-3000-samples | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- imdb
model-index:
- name: finetuning-sentiment-model-3000-samples
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetuning-sentiment-model-3000-samples
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.17.0
- Pytorch 1.9.1
- Datasets 2.1.0
- Tokenizers 0.11.6
| 1,095 |
praptishadmaan/finetuning-sentiment-model-3000-samples | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- imdb
metrics:
- accuracy
- f1
model-index:
- name: finetuning-sentiment-model-3000-samples
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: imdb
type: imdb
args: plain_text
metrics:
- name: Accuracy
type: accuracy
value: 0.93192
- name: F1
type: f1
value: 0.9323583180987203
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetuning-sentiment-model-3000-samples
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2345
- Accuracy: 0.9319
- F1: 0.9324
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.18.0
- Pytorch 1.10.0+cu111
- Datasets 2.1.0
- Tokenizers 0.12.1
| 1,510 |
dmjimenezbravo/electricidad-small-discriminator-finetuned-usElectionTweets1Jul11Nov-spanish | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | ---
tags:
- generated_from_trainer
metrics:
- accuracy
- f1
model-index:
- name: electricidad-small-discriminator-finetuned-usElectionTweets1Jul11Nov-spanish
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# electricidad-small-discriminator-finetuned-usElectionTweets1Jul11Nov-spanish
This model is a fine-tuned version of [mrm8488/electricidad-small-discriminator](https://huggingface.co/mrm8488/electricidad-small-discriminator) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 2.3327
- Accuracy: 0.7642
- F1: 0.7642
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 60
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|
| 0.88 | 1.0 | 1222 | 0.7491 | 0.6943 | 0.6943 |
| 0.7292 | 2.0 | 2444 | 0.6253 | 0.7544 | 0.7544 |
| 0.6346 | 3.0 | 3666 | 0.5292 | 0.7971 | 0.7971 |
| 0.565 | 4.0 | 4888 | 0.4831 | 0.8168 | 0.8168 |
| 0.4898 | 5.0 | 6110 | 0.4086 | 0.8532 | 0.8532 |
| 0.4375 | 6.0 | 7332 | 0.3411 | 0.8831 | 0.8831 |
| 0.3968 | 7.0 | 8554 | 0.2735 | 0.9100 | 0.9100 |
| 0.3321 | 8.0 | 9776 | 0.2343 | 0.9253 | 0.9253 |
| 0.3045 | 9.0 | 10998 | 0.1855 | 0.9450 | 0.9450 |
| 0.2837 | 10.0 | 12220 | 0.1539 | 0.9591 | 0.9591 |
| 0.2411 | 11.0 | 13442 | 0.1309 | 0.9650 | 0.9650 |
| 0.2203 | 12.0 | 14664 | 0.1100 | 0.9716 | 0.9716 |
| 0.1953 | 13.0 | 15886 | 0.1067 | 0.9760 | 0.9760 |
| 0.1836 | 14.0 | 17108 | 0.0755 | 0.9813 | 0.9813 |
| 0.1611 | 15.0 | 18330 | 0.0731 | 0.9829 | 0.9829 |
| 0.1479 | 16.0 | 19552 | 0.0746 | 0.9839 | 0.9839 |
| 0.138 | 17.0 | 20774 | 0.0516 | 0.9895 | 0.9895 |
| 0.129 | 18.0 | 21996 | 0.0481 | 0.9903 | 0.9903 |
| 0.1182 | 19.0 | 23218 | 0.0401 | 0.9926 | 0.9926 |
| 0.1065 | 20.0 | 24440 | 0.0488 | 0.9895 | 0.9895 |
| 0.096 | 21.0 | 25662 | 0.0333 | 0.9928 | 0.9928 |
| 0.0889 | 22.0 | 26884 | 0.0222 | 0.9951 | 0.9951 |
| 0.0743 | 23.0 | 28106 | 0.0236 | 0.9951 | 0.9951 |
| 0.0821 | 24.0 | 29328 | 0.0322 | 0.9931 | 0.9931 |
| 0.0866 | 25.0 | 30550 | 0.0135 | 0.9974 | 0.9974 |
| 0.0616 | 26.0 | 31772 | 0.0100 | 0.9980 | 0.9980 |
| 0.0641 | 27.0 | 32994 | 0.0112 | 0.9977 | 0.9977 |
| 0.0603 | 28.0 | 34216 | 0.0071 | 0.9987 | 0.9987 |
| 0.0491 | 29.0 | 35438 | 0.0088 | 0.9982 | 0.9982 |
| 0.0563 | 30.0 | 36660 | 0.0071 | 0.9982 | 0.9982 |
| 0.0467 | 31.0 | 37882 | 0.0045 | 0.9990 | 0.9990 |
| 0.0545 | 32.0 | 39104 | 0.0057 | 0.9987 | 0.9987 |
| 0.0519 | 33.0 | 40326 | 0.0048 | 0.9992 | 0.9992 |
| 0.0524 | 34.0 | 41548 | 0.0030 | 0.9995 | 0.9995 |
| 0.044 | 35.0 | 42770 | 0.0046 | 0.9990 | 0.9990 |
| 0.0442 | 36.0 | 43992 | 0.0029 | 0.9995 | 0.9995 |
| 0.0352 | 37.0 | 45214 | 0.0035 | 0.9995 | 0.9995 |
| 0.0348 | 38.0 | 46436 | 0.0029 | 0.9995 | 0.9995 |
| 0.0295 | 39.0 | 47658 | 0.0023 | 0.9995 | 0.9995 |
| 0.0289 | 40.0 | 48880 | 0.0035 | 0.9995 | 0.9995 |
| 0.0292 | 41.0 | 50102 | 0.0023 | 0.9995 | 0.9995 |
| 0.0259 | 42.0 | 51324 | 0.0027 | 0.9995 | 0.9995 |
| 0.0217 | 43.0 | 52546 | 0.0031 | 0.9995 | 0.9995 |
| 0.0278 | 44.0 | 53768 | 0.0018 | 0.9995 | 0.9995 |
| 0.0254 | 45.0 | 54990 | 0.0023 | 0.9995 | 0.9995 |
| 0.0164 | 46.0 | 56212 | 0.0016 | 0.9997 | 0.9997 |
| 0.0277 | 47.0 | 57434 | 0.0027 | 0.9997 | 0.9997 |
| 0.0158 | 48.0 | 58656 | 0.0029 | 0.9997 | 0.9997 |
| 0.0178 | 49.0 | 59878 | 0.0023 | 0.9997 | 0.9997 |
| 0.022 | 50.0 | 61100 | 0.0019 | 0.9997 | 0.9997 |
| 0.0167 | 51.0 | 62322 | 0.0018 | 0.9997 | 0.9997 |
| 0.0159 | 52.0 | 63544 | 0.0017 | 0.9997 | 0.9997 |
| 0.0105 | 53.0 | 64766 | 0.0016 | 0.9997 | 0.9997 |
| 0.0111 | 54.0 | 65988 | 0.0015 | 0.9997 | 0.9997 |
| 0.0139 | 55.0 | 67210 | 0.0021 | 0.9997 | 0.9997 |
| 0.0152 | 56.0 | 68432 | 0.0026 | 0.9997 | 0.9997 |
| 0.0191 | 57.0 | 69654 | 0.0022 | 0.9997 | 0.9997 |
| 0.0075 | 58.0 | 70876 | 0.0017 | 0.9997 | 0.9997 |
| 0.0141 | 59.0 | 72098 | 0.0016 | 0.9997 | 0.9997 |
| 0.0086 | 60.0 | 73320 | 0.0014 | 0.9997 | 0.9997 |
### Framework versions
- Transformers 4.18.0
- Pytorch 1.11.0+cu113
- Datasets 2.1.0
- Tokenizers 0.12.1
| 5,769 |
jsoutherland/distilbert-base-uncased-finetuned-emotion | [
"sadness",
"joy",
"love",
"anger",
"fear",
"surprise"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model_index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metric:
name: F1
type: f1
value: 0.9327347950817506
model-index:
- name: jsoutherland/distilbert-base-uncased-finetuned-emotion
results:
- task:
type: text-classification
name: Text Classification
dataset:
name: emotion
type: emotion
config: default
split: test
metrics:
- name: Accuracy
type: accuracy
value: 0.925
verified: true
- name: Precision Macro
type: precision
value: 0.8954208010579672
verified: true
- name: Precision Micro
type: precision
value: 0.925
verified: true
- name: Precision Weighted
type: precision
value: 0.9256567173431012
verified: true
- name: Recall Macro
type: recall
value: 0.8711059962680445
verified: true
- name: Recall Micro
type: recall
value: 0.925
verified: true
- name: Recall Weighted
type: recall
value: 0.925
verified: true
- name: F1 Macro
type: f1
value: 0.8794773714607985
verified: true
- name: F1 Micro
type: f1
value: 0.925
verified: true
- name: F1 Weighted
type: f1
value: 0.9244781949774824
verified: true
- name: loss
type: loss
value: 0.17752596735954285
verified: true
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1649
- Accuracy: 0.9325
- F1: 0.9327
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| No log | 1.0 | 250 | 0.2838 | 0.9065 | 0.9036 |
| No log | 2.0 | 500 | 0.1795 | 0.9255 | 0.9255 |
| No log | 3.0 | 750 | 0.1649 | 0.9325 | 0.9327 |
### Framework versions
- Transformers 4.8.2
- Pytorch 1.9.0+cu102
- Datasets 2.1.0
- Tokenizers 0.10.3
| 3,073 |
cynthiachan/procedure_classification_distilbert | [
"LABEL_0",
"LABEL_1",
"LABEL_10",
"LABEL_100",
"LABEL_101",
"LABEL_102",
"LABEL_103",
"LABEL_104",
"LABEL_105",
"LABEL_106",
"LABEL_107",
"LABEL_108",
"LABEL_109",
"LABEL_11",
"LABEL_110",
"LABEL_111",
"LABEL_112",
"LABEL_113",
"LABEL_114",
"LABEL_115",
"LABEL_116",
"LABEL_... | Entry not found | 15 |
nbroad/longformer-base-health-fact | [
"false",
"mixture",
"true",
"unproven"
] | ---
language:
- en
tags:
- generated_from_trainer
datasets:
- health_fact
model-index:
- name: longformer-base-health-fact2
results:
- task:
type: text-classification
name: Text Classification
dataset:
name: health_fact
type: health_fact
split: test
metrics:
- name: F1
type: f1
value: 0.6732897445517078
- name: Accuracy
type: accuracy
value: 0.797242497972425
- name: False Accuracy
type: accuracy
value: 0.8092783505154639
- name: Mixture Accuracy
type: accuracy
value: 0.5323383084577115
- name: True Accuracy
type: accuracy
value: 0.9081803005008348
- name: Unproven Accuracy
type: accuracy
value: 0.4
---
# longformer-base-health-fact2
This model is a fine-tuned version of [allenai/longformer-base-4096](https://huggingface.co/allenai/longformer-base-4096) on the health_fact dataset.
It achieves the following results on the VALIDATION set:
- Loss: 0.5858
- Micro F1: 0.8122
- Macro F1: 0.6830
- False F1: 0.7941
- Mixture F1: 0.5015
- True F1: 0.9234
- Unproven F1: 0.5128
The following are the results on the TEST set:
- Macro F1: 0.6732897445517078
- Accuracy: 0.797242497972425
- False Accuracy: 0.8092783505154639
- Mixture Accuracy: 0.5323383084577115
- True Accuracy: 0.9081803005008348
- Unproven Accuracy: 0.4
## Model description
The health fact dataset is for building fact-checking models related to health. Here is how you can use this model:
```python
import torch
from transformers import pipeline
claim = "A mother revealed to her child in a letter after her death that she had just one eye because she had donated the other to him."
text = "In April 2005, we spotted a tearjerker on the Internet about a mother who gave up one of her eyes to a son who had lost one of his at an early age. By February 2007 the item was circulating in e-mail in the following shortened version: My mom only had one eye. I hated her… She was such an embarrassment. She cooked for students and teachers to support the family. There was this one day during elementary school where my mom came to say hello to me. I was so embarrassed. How could she do this to me? I ignored her, threw her a hateful look and ran out. The next day at school one of my classmates said, “EEEE, your mom only has one eye!” I wanted to bury myself. I also wanted my mom to just disappear. I confronted her that day and said, “If you’re only gonna make me a laughing stock, why don’t you just die?” My mom did not respond… I didn’t even stop to think for a second about what I had said, because I was full of anger. I was oblivious to her feelings. I wanted out of that house, and have nothing to do with her. So I studied real hard, got a chance to go abroad to study. Then, I got married. I bought a house of my own. I had kids of my own. I was happy with my life, my kids and the comforts. Then one day, my Mother came to visit me. She hadn’t seen me in years and she didn’t even meet her grandchildren. When she stood by the door, my children laughed at her, and I yelled at her for coming over uninvited. I screamed at her, “How dare you come to my house and scare my children! GET OUT OF HERE! NOW!! !” And to this, my mother quietly answered, “Oh, I’m so sorry. I may have gotten the wrong address,” and she disappeared out of sight. One day, a letter regarding a school reunion came to my house. So I lied to my wife that I was going on a business trip. After the reunion, I went to the old shack just out of curiosity. My neighbors said that she died. I did not shed a single tear. They handed me a letter that she had wanted me to have. My dearest son, I think of you all the time. I’m sorry that I came to your house and scared your children. I was so glad when I heard you were coming for the reunion. But I may not be able to even get out of bed to see you. I’m sorry that I was a constant embarrassment to you when you were growing up. You see……..when you were very little, you got into an accident, and lost your eye. As a mother, I couldn’t stand watching you having to grow up with one eye. So I gave you mine. I was so proud of my son who was seeing a whole new world for me, in my place, with that eye. With all my love to you, Your mother. In its earlier incarnation, the story identified by implication its location as Korea through statements made by both the mother and the son (the son’s “I left my mother and came to Seoul” and the mother’s “I won’t visit Seoul anymore”). It also supplied a reason for the son’s behavior when his mother arrived unexpectedly to visit him (“My little girl ran away, scared of my mom’s eye” and “I screamed at her, ‘How dare you come to my house and scare my daughter!'”). A further twist was provided in the original: rather than gaining the news of his mother’s death from neighbors (who hand him her letter), the son instead discovered the woman who bore him lying dead on the floor of what used to be his childhood home, her missive to him clutched in her lifeless hand: Give your parents roses while they are alive, not deadMY mom only had one eye. I hated her … she was such an embarrassment. My mom ran a small shop at a flea market. She collected little weeds and such to sell … anything for the money we needed she was such an embarrassment. There was this one day during elementary school … It was field day, and my mom came. I was so embarrassed. How could she do this to me? I threw her a hateful look and ran out. The next day at school … “your mom only has one eye?!? !” … And they taunted me. I wished that my mom would just disappear from this world so I said to my mom, “mom … Why don’t you have the other eye?! If you’re only going to make me a laughingstock, why don’t you just die?!! !” my mom did not respond … I guess I felt a little bad, but at the same time, it felt good to think that I had said what I’d wanted to say all this time… maybe it was because my mom hadn’t punished me, but I didn’t think that I had hurt her feelings very badly. That night… I woke up, and went to the kitchen to get a glass of water. My mom was crying there, so quietly, as if she was afraid that she might wake me. I took a look at her, and then turned away. Because of the thing I had said to her earlier, there was something pinching at me in the corner of my heart. Even so, I hated my mother who was crying out of her one eye. So I told myself that I would grow up and become successful. Because I hated my one-eyed mom and our desperate poverty… then I studied real hard. I left my mother and came to Seoul and studied, and got accepted in the Seoul University with all the confidence I had. Then, I got married. I bought a house of my own. Then I had kids, too… now I’m living happily as a successful man. I like it here because it’s a place that doesn’t remind me of my mom. This happiness was getting bigger and bigger, when… what?! Who’s this…it was my mother… still with her one eye. It felt as if the whole sky was falling apart on me. My little girl ran away, scared of my mom’s eye. And I asked her, “who are you? !” “I don’t know you!! !” as if trying to make that real. I screamed at her, “How dare you come to my house and scare my daughter!” “GET OUT OF HERE! NOW!! !” and to this, my mother quietly answered, “oh, I’m so sorry. I may have gotten the wrong address,” and she disappeared out of sight. Thank goodness… she doesn’t recognize me… I was quite relieved. I told myself that I wasn’t going to care, or think about this for the rest of my life. Then a wave of relief came upon me… One day, a letter regarding a school reunion came to my house. So, lying to my wife that I was going on a business trip, I went. After the reunion, I went down to the old shack, that I used to call a house… just out of curiosity there, I found my mother fallen on the cold ground. But I did not shed a single tear. She had a piece of paper in her hand…. it was a letter to me. My son… I think my life has been long enough now… And… I won’t visit Seoul anymore… but would it be too much to ask if I wanted you to come visit me once in a while? I miss you so much… and I was so glad when I heard you were coming for the reunion. But I decided not to go to the school. …for you… and I’m sorry that I only have one eye, and I was an embarrassment for you. You see, when you were very little, you got into an accident, and lost your eye. as a mom, I couldn’t stand watching you having to grow up with only one eye… so I gave you mine… I was so proud of my son that was seeing a whole new world for me, in my place, with that eye. I was never upset at you for anything you did… the couple times that you were angry with me, I thought to myself, ‘it’s because he loves me…’ my son. Oh, my son… I don’t want you to cry for me, because of my death. My son, I love you my son, I love you so much. With all modern medical technology, transplantation of the eyeball is still impossible. The optic nerve isn’t an ordinary nerve, but instead an inset running from the brain. Modern medicine isn’t able to “connect” an eyeball back to brain after an optic nerve has been severed, let alone transplant the eye from a different person. (The only exception is the cornea, the transparent part in front of the eye: corneas are transplanted to replace injured and opaque ones.) We won’t try to comment on whether any surgeon would accept an eye from a living donor for transplant into another — we’ll leave that to others who are far more knowledgeable about medical ethics and transplant procedures. But we will note that the plot device of a mother’s dramatic sacrifice for the sake of her child’s being revealed in a written communication delivered after her demise appears in another legend about maternal love: the 2008 tale about a woman who left a touching message on her cell phone even as life ebbed from her as she used her body to shield the tot during an earthquake. Giving up one’s own life for a loved one is central to a 2005 urban legend about a boy on a motorcycle who has his girlfriend hug him one last time and put on his helmet just before the crash that kills him and spares her. Returning to the “notes from the dead” theme is the 1995 story about a son who discovers only through a posthumous letter from his mother what their occasional dinner “dates” had meant to her. Another legend we’re familiar with features a meme used in the one-eyed mother story (the coming to light of the enduring love of the person who died for the completely unworthy person she’d lavished it on), but that one involves a terminally ill woman and her cheating husband. In it, an about-to-be-spurned wife begs the adulterous hoon she’d married to stick around for another 30 days and to carry her over the threshold of their home once every day of that month as her way of keeping him around long enough for her to kick the bucket and thus spare their son the knowledge that his parents were on the verge of divorce."
label = "false"
device = 0 if torch.cuda.is_available() else -1
pl = pipeline("text-classification", model="nbroad/longformer-base-health-fact", device=device)
input_text = claim+pl.tokenizer.sep_token+text
print(len(pl.tokenizer(input_text).input_ids))
# 2361 (which is why longformer is useful)
pl(input_text)
# [{'label': 'false', 'score': 0.8015491962432861}]
```
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 16
- eval_batch_size: 32
- seed: 18
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Micro F1 | Macro F1 | False F1 | Mixture F1 | True F1 | Unproven F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:--------:|:----------:|:-------:|:-----------:|
| 0.555 | 1.0 | 613 | 0.5243 | 0.7842 | 0.5535 | 0.7698 | 0.4170 | 0.8938 | 0.1333 |
| 0.4282 | 2.0 | 1226 | 0.5008 | 0.8031 | 0.6393 | 0.7829 | 0.4605 | 0.9199 | 0.3939 |
| 0.2897 | 3.0 | 1839 | 0.5858 | 0.8122 | 0.6830 | 0.7941 | 0.5015 | 0.9234 | 0.5128 |
### Framework versions
- Transformers 4.19.0.dev0
- Pytorch 1.11.0a0+17540c5
- Datasets 2.1.1.dev0
- Tokenizers 0.12.1
| 12,597 |
alla1101/distilbert-base-uncased-finetuned-emotion | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model-index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.924
- name: F1
type: f1
value: 0.9240869504197766
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2236
- Accuracy: 0.924
- F1: 0.9241
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| No log | 1.0 | 250 | 0.3293 | 0.901 | 0.8979 |
| No log | 2.0 | 500 | 0.2236 | 0.924 | 0.9241 |
### Framework versions
- Transformers 4.18.0
- Pytorch 1.11.0+cu113
- Datasets 2.1.0
- Tokenizers 0.12.1
| 1,804 |
Truefilter/bertweet_lg_text_quality | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
peter2000/distilbert-base-uncased-finetuned-osdg | [
"sdg_1",
"sdg_10",
"sdg_11",
"sdg_12",
"sdg_13",
"sdg_14",
"sdg_15",
"sdg_2",
"sdg_3",
"sdg_4",
"sdg_5",
"sdg_6",
"sdg_7",
"sdg_8",
"sdg_9"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: distilbert-base-uncased-finetuned-osdg
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-osdg
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 1.8193
- F1 Score: 0.7962
- Accuracy: 0.8434
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 20
### Training results
| Training Loss | Epoch | Step | Validation Loss | F1 Score | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 0.3769 | 1.0 | 1017 | 0.8258 | 0.7729 | 0.8257 |
| 0.2759 | 2.0 | 2034 | 0.8364 | 0.7773 | 0.8262 |
| 0.1412 | 3.0 | 3051 | 1.0203 | 0.7833 | 0.8379 |
| 0.1423 | 4.0 | 4068 | 1.1603 | 0.7683 | 0.8224 |
| 0.0939 | 5.0 | 5085 | 1.3029 | 0.7843 | 0.8329 |
| 0.0757 | 6.0 | 6102 | 1.3562 | 0.7931 | 0.8379 |
| 0.0801 | 7.0 | 7119 | 1.2925 | 0.7840 | 0.8395 |
| 0.0311 | 8.0 | 8136 | 1.4632 | 0.7750 | 0.8318 |
| 0.0263 | 9.0 | 9153 | 1.5760 | 0.7843 | 0.8312 |
| 0.0196 | 10.0 | 10170 | 1.5689 | 0.7890 | 0.8417 |
| 0.0313 | 11.0 | 11187 | 1.6034 | 0.7909 | 0.8417 |
| 0.0007 | 12.0 | 12204 | 1.6725 | 0.7889 | 0.8406 |
| 0.0081 | 13.0 | 13221 | 1.6463 | 0.7911 | 0.8395 |
| 0.0061 | 14.0 | 14238 | 1.7730 | 0.7861 | 0.8345 |
| 0.003 | 15.0 | 15255 | 1.8001 | 0.7847 | 0.8379 |
| 0.0002 | 16.0 | 16272 | 1.7328 | 0.7912 | 0.8434 |
| 0.0 | 17.0 | 17289 | 1.7914 | 0.8011 | 0.8489 |
| 0.0009 | 18.0 | 18306 | 1.7772 | 0.7958 | 0.8456 |
| 0.0 | 19.0 | 19323 | 1.8028 | 0.7958 | 0.8434 |
| 0.0 | 20.0 | 20340 | 1.8193 | 0.7962 | 0.8434 |
### Framework versions
- Transformers 4.19.2
- Pytorch 1.11.0+cu113
- Datasets 2.2.2
- Tokenizers 0.12.1
| 2,842 |
Jeevesh8/6ep_bert_ft_cola-76 | null | Entry not found | 15 |
Jeevesh8/6ep_bert_ft_cola-78 | null | Entry not found | 15 |
miyagawaorj/distilbert-base-uncased-finetuned-emotion | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model-index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.9425
- name: F1
type: f1
value: 0.9422011075095515
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2285
- Accuracy: 0.9425
- F1: 0.9422
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 2
- eval_batch_size: 2
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|
| 0.4656 | 1.0 | 8000 | 0.2912 | 0.9365 | 0.9362 |
| 0.2046 | 2.0 | 16000 | 0.2285 | 0.9425 | 0.9422 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.11.0
- Datasets 1.16.1
- Tokenizers 0.10.3
| 1,803 |
elvaklose/finetuning-sentiment-model-3000-samples | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- imdb
metrics:
- accuracy
- f1
model-index:
- name: finetuning-sentiment-model-3000-samples
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: imdb
type: imdb
args: plain_text
metrics:
- name: Accuracy
type: accuracy
value: 0.8766666666666667
- name: F1
type: f1
value: 0.8786885245901639
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetuning-sentiment-model-3000-samples
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2896
- Accuracy: 0.8767
- F1: 0.8787
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.19.2
- Pytorch 1.11.0+cu113
- Datasets 2.2.1
- Tokenizers 0.12.1
| 1,521 |
nqcccccc/phobert-social-media-text-classify | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5",
"LABEL_6",
"LABEL_7"
] | Entry not found | 15 |
connectivity/bert_ft_qqp-4 | null | Entry not found | 15 |
connectivity/bert_ft_qqp-98 | null | Entry not found | 15 |
wonscha/my-awesome-model | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- yelp_review_full
metrics:
- accuracy
model-index:
- name: my-awesome-model
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: yelp_review_full
type: yelp_review_full
args: yelp_review_full
metrics:
- name: Accuracy
type: accuracy
value: 0.559
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# my-awesome-model
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the yelp_review_full dataset.
It achieves the following results on the evaluation set:
- Loss: 1.5680
- Accuracy: 0.559
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 1.0 | 125 | 1.1345 | 0.523 |
| No log | 2.0 | 250 | 1.5381 | 0.539 |
| No log | 3.0 | 375 | 1.5680 | 0.559 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.11.0+cu113
- Datasets 2.2.2
- Tokenizers 0.10.3
| 1,729 |
sbenel/emotion-distilbert | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
language: en
tags:
- text-classification
- pytorch
- emotion
metrics:
- accuracy, F1 score
dataset:
- emotion
---
## Training Parameters
```
learning rate: 2e-5
epochs: 40
weight decay: 0.01
batch size: 16
```
## Metrics
```
acuraccy: 0.93
macro-F1 (macro avg): 0.88
best epoch: 15
```
## Dataset:
[Twitter-Sentiment-Analysis](https://huggingface.co/nlp/viewer/?dataset=emotion).
| 409 |
GioReg/bertMULTINEGsentiment | null | ---
license: apache-2.0
tags:
- generated_from_trainer
model-index:
- name: bertMULTINEGsentiment
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bertMULTINEGsentiment
This model is a fine-tuned version of [bert-base-multilingual-uncased](https://huggingface.co/bert-base-multilingual-uncased) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.19.2
- Pytorch 1.11.0+cu113
- Datasets 2.2.2
- Tokenizers 0.12.1
| 1,063 |
M47Labs/spanish_news_classification_headlines_untrained | [
"ciencia_tecnologia",
"clickbait",
"cultura",
"deportes",
"economia",
"educacion",
"medio_ambiente",
"opinion",
"politica",
"sociedad"
] | ---
widget:
- text: "El dólar se dispara tras la reunión de la Fed"
---
# Spanish News Classification Headlines
SNCH: this model was developed by [M47Labs](https://www.m47labs.com/es/) the goal is text classification, the base model use was [BETO](https://huggingface.co/dccuchile/bert-base-spanish-wwm-cased), however this model has not been fine-tuned on any dataset. The objective is to show the performance of this model when is used with the objective of inference without training at all.
## Dataset validation Sample
Dataset size : 1000
Columns: idTask,task content 1,idTag,tag.
|task content|tag|
|------|------|
|Alcalá de Guadaíra celebra la IV Semana de la Diversidad Sexual con acciones de sensibilización|sociedad|
|El Archipiélago Chinijo Graciplus se impone en el Trofeo Centro Comercial Rubicón|deportes|
|Un total de 39 personas padecen ELA actualmente en la provincia|sociedad|
|Eurocopa 2021 : Italia vence a Gales y pasa a octavos con su candidatura reforzada|deportes|
|Resolución de 10 de junio de 2021, del Ayuntamiento de Tarazona de La Mancha (Albacete), referente a la convocatoria para proveer una plaza.|sociedad|
|El primer ministro sueco pierde una moción de censura|politica|
|El dólar se dispara tras la reunión de la Fed|economia|
## Labels:
* ciencia_tecnologia
* clickbait
* cultura
* deportes
* economia
* educacion
* medio_ambiente
* opinion
* politica
* sociedad
## Example of Use
### Pipeline
```{python}
import torch
from transformers import AutoTokenizer, BertForSequenceClassification,TextClassificationPipeline
review_text = 'los vehiculos que esten esperando pasajaeros deberan estar apagados para reducir emisiones'
path = "M47Labs/spanish_news_classification_headlines_untrained"
tokenizer = AutoTokenizer.from_pretrained(path)
model = BertForSequenceClassification.from_pretrained(path)
nlp = TextClassificationPipeline(task = "text-classification",
model = model,
tokenizer = tokenizer)
print(nlp(review_text))
```
```[{'label': 'medio_ambiente', 'score': 0.2834321384291023}]```
### Pytorch
```{python}
import torch
from transformers import AutoTokenizer, BertForSequenceClassification,TextClassificationPipeline
from numpy import np
model_name = 'M47Labs/spanish_news_classification_headlines_untrained'
MAX_LEN = 32
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)
texto = "las emisiones estan bajando, debido a las medidas ambientales tomadas por el gobierno"
encoded_review = tokenizer.encode_plus(
texto,
max_length=MAX_LEN,
add_special_tokens=True,
#return_token_type_ids=False,
pad_to_max_length=True,
return_attention_mask=True,
return_tensors='pt',
)
input_ids = encoded_review['input_ids']
attention_mask = encoded_review['attention_mask']
output = model(input_ids, attention_mask)
_, prediction = torch.max(output['logits'], dim=1)
print(f'Review text: {texto}')
print(f'Sentiment : {model.config.id2label[prediction.detach().cpu().numpy()[0]]}')
```
```Review text: las emisiones estan bajando, debido a las medidas ambientales tomadas por el gobierno```
```Sentiment : opinion```
A more in depth example on how to use the model can be found in this colab notebook: https://colab.research.google.com/drive/1XsKea6oMyEckye2FePW_XN7Rf8v41Cw_?usp=sharing
## Validation Results
|Full Dataset||
|------|------|
|Accuracy Score|0.362|
|Precision (Macro)|0.21|
|Recall (Macro)|0.22|

| 3,731 |
daniel780/finetuning-sentiment-model-3000-samples | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- amazon_polarity
metrics:
- accuracy
- f1
model-index:
- name: finetuning-sentiment-model-3000-samples
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: amazon_polarity
type: amazon_polarity
args: amazon_polarity
metrics:
- name: Accuracy
type: accuracy
value: 0.8066666666666666
- name: F1
type: f1
value: 0.8079470198675497
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetuning-sentiment-model-3000-samples
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the amazon_polarity dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4356
- Accuracy: 0.8067
- F1: 0.8079
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.19.2
- Pytorch 1.11.0+cu113
- Datasets 2.2.2
- Tokenizers 0.12.1
| 1,570 |
RANG012/SENATOR-Scaled | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- imdb
metrics:
- accuracy
- f1
model-index:
- name: SENATOR-Scaled
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: imdb
type: imdb
args: plain_text
metrics:
- name: Accuracy
type: accuracy
value: 0.89
- name: F1
type: f1
value: 0.8897795591182365
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SENATOR-Scaled
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2670
- Accuracy: 0.89
- F1: 0.8898
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.19.2
- Pytorch 1.11.0+cu113
- Datasets 2.2.2
- Tokenizers 0.12.1
| 1,455 |
cjbarrie/masress-medcrit-camel | [
"critical",
"neutral",
"uncritical"
] | ---
tags: autotrain
language: unk
widget:
- text: "الكل ينتقد الرئيس على إخفاقاته"
datasets:
- cjbarrie/autotrain-data-masress-medcrit-binary-5
co2_eq_emissions: 0.01017487638098474
---
# Model Trained Using AutoTrain
- Problem type: Multi-class Classification
- Model ID: 937130980
- CO2 Emissions (in grams): 0.01017487638098474
## Validation Metrics
- Loss: 0.757265031337738
- Accuracy: 0.7551020408163265
- Macro F1: 0.7202470830473576
- Micro F1: 0.7551020408163265
- Weighted F1: 0.7594301962377263
- Macro Precision: 0.718716577540107
- Micro Precision: 0.7551020408163265
- Weighted Precision: 0.7711448215649895
- Macro Recall: 0.7285714285714286
- Micro Recall: 0.7551020408163265
- Weighted Recall: 0.7551020408163265
## Usage
You can use cURL to access this model:
```
$ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoTrain"}' https://api-inference.huggingface.co/models/cjbarrie/autotrain-masress-medcrit-binary-5-937130980
```
Or Python API:
```
from transformers import AutoModelForSequenceClassification, AutoTokenizer
model = AutoModelForSequenceClassification.from_pretrained("cjbarrie/autotrain-masress-medcrit-binary-5-937130980", use_auth_token=True)
tokenizer = AutoTokenizer.from_pretrained("cjbarrie/autotrain-masress-medcrit-binary-5-937130980", use_auth_token=True)
inputs = tokenizer("I love AutoTrain", return_tensors="pt")
outputs = model(**inputs)
``` | 1,465 |
Jeevesh8/lecun_feather_berts-66 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
Jeevesh8/lecun_feather_berts-55 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
Jeevesh8/lecun_feather_berts-56 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
Jeevesh8/lecun_feather_berts-63 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
Jeevesh8/lecun_feather_berts-9 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
Jeevesh8/lecun_feather_berts-13 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
Jeevesh8/lecun_feather_berts-19 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
clhuang/albert-sentiment | null | ---
language:
- tw
tags:
- albert
- classification
license: afl-3.0
metrics:
- Accuracy
---
# 繁體中文情緒分類: 負面(0)、正面(1)
依據ckiplab/albert預訓練模型微調,訓練資料集只有8萬筆,做為課程的範例模型。
# 使用範例:
from transformers import AutoTokenizer, AutoModelForSequenceClassification
tokenizer = AutoTokenizer.from_pretrained("clhuang/albert-sentiment")
model = AutoModelForSequenceClassification.from_pretrained("clhuang/albert-sentiment")
## Pediction
target_names=['Negative','Positive']
max_length = 200 # 最多字數 若超出模型訓練時的字數,以模型最大字數為依據
def get_sentiment_proba(text):
# prepare our text into tokenized sequence
inputs = tokenizer(text, padding=True, truncation=True, max_length=max_length, return_tensors="pt")
# perform inference to our model
outputs = model(**inputs)
# get output probabilities by doing softmax
probs = outputs[0].softmax(1)
response = {'Negative': round(float(probs[0, 0]), 2), 'Positive': round(float(probs[0, 1]), 2)}
# executing argmax function to get the candidate label
#return probs.argmax()
return response
get_sentiment_proba('我喜歡這本書')
get_sentiment_proba('不喜歡這款產品') | 1,197 |
QuentinKemperino/ECHR_test_2 | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5",
"LABEL_6",
"LABEL_7",
"LABEL_8",
"LABEL_9"
] | ---
license: cc-by-sa-4.0
tags:
- generated_from_trainer
datasets:
- lex_glue
model-index:
- name: ECHR_test_2
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# ECHR_test_2 Task A
This model is a fine-tuned version of [nlpaueb/legal-bert-base-uncased](https://huggingface.co/nlpaueb/legal-bert-base-uncased) on the lex_glue dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1998
- Macro-f1: 0.5295
- Micro-f1: 0.6157
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Macro-f1 | Micro-f1 |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 0.2142 | 0.44 | 500 | 0.2887 | 0.2391 | 0.4263 |
| 0.172 | 0.89 | 1000 | 0.2672 | 0.2908 | 0.4628 |
| 0.1737 | 1.33 | 1500 | 0.2612 | 0.3657 | 0.5102 |
| 0.1581 | 1.78 | 2000 | 0.2412 | 0.3958 | 0.5468 |
| 0.1509 | 2.22 | 2500 | 0.2264 | 0.3950 | 0.5552 |
| 0.1606 | 2.67 | 3000 | 0.2342 | 0.4006 | 0.5511 |
| 0.1491 | 3.11 | 3500 | 0.2176 | 0.4558 | 0.5622 |
| 0.1392 | 3.56 | 4000 | 0.2454 | 0.4128 | 0.5596 |
| 0.15 | 4.0 | 4500 | 0.2113 | 0.4684 | 0.5874 |
| 0.1461 | 4.44 | 5000 | 0.2179 | 0.4631 | 0.5815 |
| 0.1457 | 4.89 | 5500 | 0.2151 | 0.4805 | 0.5949 |
| 0.1443 | 5.33 | 6000 | 0.2155 | 0.5123 | 0.5917 |
| 0.1279 | 5.78 | 6500 | 0.2131 | 0.4915 | 0.5998 |
| 0.1377 | 6.22 | 7000 | 0.2244 | 0.4705 | 0.5944 |
| 0.1242 | 6.67 | 7500 | 0.2150 | 0.5089 | 0.5918 |
| 0.1222 | 7.11 | 8000 | 0.2045 | 0.4801 | 0.5981 |
| 0.1372 | 7.56 | 8500 | 0.2074 | 0.5317 | 0.5962 |
| 0.1289 | 8.0 | 9000 | 0.2035 | 0.5323 | 0.6126 |
| 0.1295 | 8.44 | 9500 | 0.2058 | 0.5213 | 0.6073 |
| 0.123 | 8.89 | 10000 | 0.2027 | 0.5486 | 0.6135 |
| 0.1335 | 9.33 | 10500 | 0.1984 | 0.5442 | 0.6249 |
| 0.1258 | 9.78 | 11000 | 0.1998 | 0.5295 | 0.6157 |
### Framework versions
- Transformers 4.19.2
- Pytorch 1.11.0+cu113
- Datasets 2.2.2
- Tokenizers 0.12.1
| 3,003 |
kjunelee/distilbert-base-uncased-finetuned-emotion | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model-index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.931
- name: F1
type: f1
value: 0.9313235272564213
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1595
- Accuracy: 0.931
- F1: 0.9313
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| No log | 1.0 | 125 | 0.1873 | 0.924 | 0.9234 |
| 0.1992 | 2.0 | 250 | 0.1649 | 0.929 | 0.9293 |
| 0.1992 | 3.0 | 375 | 0.1595 | 0.931 | 0.9313 |
### Framework versions
- Transformers 4.19.2
- Pytorch 1.11.0
- Datasets 2.2.3.dev0
- Tokenizers 0.12.1
| 1,876 |
Jeevesh8/std_pnt_04_feather_berts-39 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
Jeevesh8/std_pnt_04_feather_berts-38 | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
binay1999/distilbert-cybertexts-text-classification | null | ---
license: apache-2.0
tags:
- generated_from_trainer
model-index:
- name: distilbert-cybertexts-text-classification
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-cybertexts-text-classification
This model is a fine-tuned version of [binay1999/distilbert-cybertexts-preprocessed](https://huggingface.co/binay1999/distilbert-cybertexts-preprocessed) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1104
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 0.1374 | 1.0 | 1000 | 0.1215 |
| 0.0769 | 2.0 | 2000 | 0.0959 |
| 0.039 | 3.0 | 3000 | 0.1104 |
### Framework versions
- Transformers 4.19.4
- Pytorch 1.11.0+cu113
- Datasets 2.2.2
- Tokenizers 0.12.1
| 1,460 |
Hardeep/distilbert-base-uncased-finetuned-emotion-01 | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4"
] | Entry not found | 15 |
dibsondivya/distilbert-phmtweets-sutd | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3"
] | ---
tags:
- distilbert
- health
- tweet
datasets:
- custom-phm-tweets
metrics:
- accuracy
model-index:
- name: distilbert-phmtweets-sutd
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: custom-phm-tweets
type: labelled
metrics:
- name: Accuracy
type: accuracy
value: 0.877
---
# distilbert-phmtweets-sutd
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) for text classification to identify public health events through tweets. The project was based on an [Emory University Study on Detection of Personal Health Mentions in Social Media paper](https://arxiv.org/pdf/1802.09130v2.pdf), that worked with this [custom dataset](https://github.com/emory-irlab/PHM2017).
It achieves the following results on the evaluation set:
- Accuracy: 0.877
## Usage
```Python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
tokenizer = AutoTokenizer.from_pretrained("dibsondivya/distilbert-phmtweets-sutd")
model = AutoModelForSequenceClassification.from_pretrained("dibsondivya/distilbert-phmtweets-sutd")
```
### Model Evaluation Results
With Validation Set
- Accuracy: 0.8708661417322835
With Test Set
- Accuracy: 0.8772961058045555
# Reference for distilbert-base-uncased Model
```bibtex
@article{Sanh2019DistilBERTAD,
title={DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter},
author={Victor Sanh and Lysandre Debut and Julien Chaumond and Thomas Wolf},
journal={ArXiv},
year={2019},
volume={abs/1910.01108}
}
```
| 1,625 |
SISLab/amc-opt-msmd | null | ---
tags:
- text-classification
- sentiment-analysis
language:
- "it"
--- | 83 |
projecte-aina/roberta-base-ca-v2-cased-tc | [
"Medi ambient",
"Societat",
"Policial",
"Judicial",
"Empresa",
"Partits",
"Política",
"Successos",
"Salut",
"Infraestructures",
"Parlament",
"Música",
"Govern",
"Unió Europea",
"Economia",
"Mobilitat",
"Treball",
"Cultura",
"Educació"
] | ---
language:
- ca
tags:
- "catalan"
- "text classification"
- "tecla"
- "CaText"
- "Catalan Textual Corpus"
datasets:
- "projecte-aina/tecla"
metrics:
- accuracy
model-index:
- name: roberta-base-ca-v2-cased-tc
results:
- task:
type: text-classification
dataset:
name: TeCla
type: projecte-aina/tecla
metrics:
- name: Accuracy
type: accuracy
value: 0.7426
widget:
- text: "Els Pets presenten el seu nou treball al Palau Sant Jordi."
- text: "Els barcelonins incrementen un 23% l’ús del cotxe des de l’inici de la pandèmia."
- text: "Retards a quatre línies de Rodalies per una avaria entre Sants i plaça de Catalunya."
- text: "Majors de 60 anys i sanitaris començaran a rebre la tercera dosi de la vacuna covid els propers dies."
- text: "Els cinemes Verdi estrenen Verdi Classics, un nou canal de televisió."
---
# Catalan BERTa-v2 (roberta-base-ca-v2) finetuned for Text Classification.
## Table of Contents
- [Model Description](#model-description)
- [Intended Uses and Limitations](#intended-uses-and-limitations)
- [How to Use](#how-to-use)
- [Training](#training)
- [Training Data](#training-data)
- [Training Procedure](#training-procedure)
- [Evaluation](#evaluation)
- [Variable and Metrics](#variable-and-metrics)
- [Evaluation Results](#evaluation-results)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Funding](#funding)
- [Contributions](#contributions)
## Model description
The **roberta-base-ca-v2-cased-tc** is a Text Classification (TC) model for the Catalan language fine-tuned from the [roberta-base-ca-v2](https://huggingface.co/projecte-aina/roberta-base-ca-v2) model, a [RoBERTa](https://arxiv.org/abs/1907.11692) base model pre-trained on a medium-size corpus collected from publicly available corpora and crawlers (check the roberta-base-ca-v2 model card for more details).
## Intended Uses and Limitations
**roberta-base-ca-v2-cased-tc** model can be used to classify texts. The model is limited by its training dataset and may not generalize well for all use cases.
## How to Use
Here is how to use this model:
```python
from transformers import pipeline
from pprint import pprint
nlp = pipeline("text-classification", model="projecte-aina/roberta-base-ca-v2-cased-tc")
example = "Retards a quatre línies de Rodalies per una avaria entre Sants i plaça de Catalunya."
tc_results = nlp(example)
pprint(tc_results)
```
## Training
### Training data
We used the TC dataset in Catalan called [TeCla](https://huggingface.co/datasets/projecte-aina/tecla) for training and evaluation.
### Training Procedure
The model was trained with a batch size of 16 and a learning rate of 5e-5 for 5 epochs. We then selected the best checkpoint using the downstream task metric in the corresponding development set and then evaluated it on the test set.
## Evaluation
### Variable and Metrics
This model was finetuned maximizing accuracy.
## Evaluation results
We evaluated the _roberta-base-ca-v2-cased-tc_ on the TeCla test set against standard multilingual and monolingual baselines:
| Model | TeCla (Accuracy) |
| ------------|:-------------|
| roberta-base-ca-v2-cased-tc | **74.26** |
| roberta-base-ca-cased-tc | 73.65 |
| mBERT | 69.90 |
| XLM-RoBERTa | 70.14 |
For more details, check the fine-tuning and evaluation scripts in the official [GitHub repository](https://github.com/projecte-aina/club).
## Licensing Information
[Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0)
## Citation Information
If you use any of these resources (datasets or models) in your work, please cite our latest paper:
```bibtex
@inproceedings{armengol-estape-etal-2021-multilingual,
title = "Are Multilingual Models the Best Choice for Moderately Under-resourced Languages? {A} Comprehensive Assessment for {C}atalan",
author = "Armengol-Estap{\'e}, Jordi and
Carrino, Casimiro Pio and
Rodriguez-Penagos, Carlos and
de Gibert Bonet, Ona and
Armentano-Oller, Carme and
Gonzalez-Agirre, Aitor and
Melero, Maite and
Villegas, Marta",
booktitle = "Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021",
month = aug,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.findings-acl.437",
doi = "10.18653/v1/2021.findings-acl.437",
pages = "4933--4946",
}
```
### Funding
This work was funded by the [Departament de la Vicepresidència i de Polítiques Digitals i Territori de la Generalitat de Catalunya](https://politiquesdigitals.gencat.cat/ca/inici/index.html#googtrans(ca|en) within the framework of [Projecte AINA](https://politiquesdigitals.gencat.cat/ca/economia/catalonia-ai/aina).
## Contributions
[N/A] | 4,896 |
emen/distilbert-base-uncased-finetuned-emotion | [
"LABEL_0",
"LABEL_1",
"LABEL_2",
"LABEL_3",
"LABEL_4",
"LABEL_5"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- emotion
metrics:
- accuracy
- f1
model-index:
- name: distilbert-base-uncased-finetuned-emotion
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: emotion
type: emotion
args: default
metrics:
- name: Accuracy
type: accuracy
value: 0.9295
- name: F1
type: f1
value: 0.9297561758557029
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2181
- Accuracy: 0.9295
- F1: 0.9298
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8495 | 1.0 | 250 | 0.3141 | 0.9085 | 0.9060 |
| 0.2511 | 2.0 | 500 | 0.2181 | 0.9295 | 0.9298 |
### Framework versions
- Transformers 4.20.1
- Pytorch 1.11.0+cu102
- Datasets 2.3.2
- Tokenizers 0.12.1
| 1,806 |
Luojike/autotrain-test-4-macbert-1071837613 | [
"0",
"1"
] | ---
tags: autotrain
language: unk
widget:
- text: "I love AutoTrain 🤗"
datasets:
- Luojike/autotrain-data-test-4-macbert
co2_eq_emissions: 0.012225117907336358
---
# Model Trained Using AutoTrain
- Problem type: Binary Classification
- Model ID: 1071837613
- CO2 Emissions (in grams): 0.012225117907336358
## Validation Metrics
- Loss: 0.533202052116394
- Accuracy: 0.7408088235294118
- Precision: 0.5072463768115942
- Recall: 0.4088785046728972
- AUC: 0.710585043624057
- F1: 0.4527813712807245
## Usage
You can use cURL to access this model:
```
$ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoTrain"}' https://api-inference.huggingface.co/models/Luojike/autotrain-test-4-macbert-1071837613
```
Or Python API:
```
from transformers import AutoModelForSequenceClassification, AutoTokenizer
model = AutoModelForSequenceClassification.from_pretrained("Luojike/autotrain-test-4-macbert-1071837613", use_auth_token=True)
tokenizer = AutoTokenizer.from_pretrained("Luojike/autotrain-test-4-macbert-1071837613", use_auth_token=True)
inputs = tokenizer("I love AutoTrain", return_tensors="pt")
outputs = model(**inputs)
``` | 1,200 |
RobertIrv938/rewardModel60kEpoch3Balanced | null | Entry not found | 15 |
agarwalchaitanya/muril-unified-ei-infotabs-btnli | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | ---
license: apache-2.0
---
| 28 |
juliensimon/distilbert-base-uncased-finetuned-cola | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- glue
metrics:
- matthews_correlation
model-index:
- name: distilbert-base-uncased-finetuned-cola
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: glue
type: glue
args: cola
metrics:
- name: Matthews Correlation
type: matthews_correlation
value: 0.5334876461854267
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-cola
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7737
- Matthews Correlation: 0.5335
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Matthews Correlation |
|:-------------:|:-----:|:----:|:---------------:|:--------------------:|
| 0.5225 | 1.0 | 535 | 0.5170 | 0.4007 |
| 0.3509 | 2.0 | 1070 | 0.5220 | 0.4837 |
| 0.2405 | 3.0 | 1605 | 0.6164 | 0.5186 |
| 0.1777 | 4.0 | 2140 | 0.7737 | 0.5335 |
| 0.1295 | 5.0 | 2675 | 0.8374 | 0.5162 |
### Framework versions
- Transformers 4.20.1
- Pytorch 1.12.0+cu102
- Datasets 2.3.2
- Tokenizers 0.12.1
| 1,999 |
morenolq/thext-cs-scibert | [
"LABEL_0"
] | ---
language: "en"
tags:
- bert
- regression
- pytorch
pipeline:
- text-classification
widget:
- text: "We propose a new approach, based on Transformer-based encoding, to highlight extraction. To the best of our knowledge, this is the first attempt to use transformer architectures to address automatic highlight generation. [SEP] Highlights are short sentences used to annotate scientific papers. They complement the abstract content by conveying the main result findings. To automate the process of paper annotation, highlights extraction aims at extracting from 3 to 5 paper sentences via supervised learning. Existing approaches rely on ad hoc linguistic features, which depend on the analyzed context, and apply recurrent neural networks, which are not effective in learning long-range text dependencies. This paper leverages the attention mechanism adopted in transformer models to improve the accuracy of sentence relevance estimation. Unlike existing approaches, it relies on the end-to-end training of a deep regression model. To attend patterns relevant to highlights content it also enriches sentence encodings with a section-level contextualization. The experimental results, achieved on three different benchmark datasets, show that the designed architecture is able to achieve significant performance improvements compared to the state-of-the-art."
- text: "We design a context-aware sentence-level regressor, in which the semantic similarity between candidate sentences and highlights is estimated by also attending the contextual knowledge provided by the other paper sections. [SEP] Highlights are short sentences used to annotate scientific papers. They complement the abstract content by conveying the main result findings. To automate the process of paper annotation, highlights extraction aims at extracting from 3 to 5 paper sentences via supervised learning. Existing approaches rely on ad hoc linguistic features, which depend on the analyzed context, and apply recurrent neural networks, which are not effective in learning long-range text dependencies. This paper leverages the attention mechanism adopted in transformer models to improve the accuracy of sentence relevance estimation. Unlike existing approaches, it relies on the end-to-end training of a deep regression model. To attend patterns relevant to highlights content it also enriches sentence encodings with a section-level contextualization. The experimental results, achieved on three different benchmark datasets, show that the designed architecture is able to achieve significant performance improvements compared to the state-of-the-art."
- text: "Fig. 2, Fig. 3, Fig. 4 show the effect of varying the number K of selected highlights on the extraction performance. As expected, recall values increase while increasing the number of selected highlights, whereas precision values show an opposite trend. [SEP] Highlights are short sentences used to annotate scientific papers. They complement the abstract content by conveying the main result findings. To automate the process of paper annotation, highlights extraction aims at extracting from 3 to 5 paper sentences via supervised learning. Existing approaches rely on ad hoc linguistic features, which depend on the analyzed context, and apply recurrent neural networks, which are not effective in learning long-range text dependencies. This paper leverages the attention mechanism adopted in transformer models to improve the accuracy of sentence relevance estimation. Unlike existing approaches, it relies on the end-to-end training of a deep regression model. To attend patterns relevant to highlights content it also enriches sentence encodings with a section-level contextualization. The experimental results, achieved on three different benchmark datasets, show that the designed architecture is able to achieve significant performance improvements compared to the state-of-the-art."
---
# General Information
This model is trained on journal publications of belonging to the domain: **Computer Science**.
This is an `allenai/scibert_scivocab_cased` model trained in the scientific domain. The model is trained with regression objective to estimate the relevance of a sentence according to the provided context (e.g., the abstract of the scientific paper).
The model is used in the paper 'Transformer-based highlights extraction from scientific papers' published in Knowledge-Based Systems scientific journal.
The model is able to achieve state-of-the-art performance in the task of highlights extraction from scientific papers.
Access to the full paper: [here](https://doi.org/10.1016/j.knosys.2022.109382).
# Usage:
For detailed usage please use the official repository https://github.com/MorenoLaQuatra/THExt .
# References:
If you find it useful, please cite the following paper:
```bibtex
@article{thext,
title={Transformer-based highlights extraction from scientific papers},
author={La Quatra, Moreno and Cagliero, Luca},
journal={Knowledge-Based Systems},
pages={109382},
year={2022},
publisher={Elsevier}
}
``` | 5,090 |
naem1023/roberta-phrase-clause-classification | null | ---
license: apache-2.0
---
| 28 |
naem1023/electra-phrase-clause-classification-dev | null | ---
license: apache-2.0
---
| 28 |
CenIA/albert-large-spanish-finetuned-pawsx | null | Entry not found | 15 |
EhsanAghazadeh/xlm-roberta-base-lcc-en-fa-2e-5-42 | null | Entry not found | 15 |
LysandreJik/test-upload1 | null | Entry not found | 15 |
JuliusAlphonso/dear-jarvis-v5 | [
"anger",
"anticipation",
"disgust",
"fear",
"joy",
"neutral",
"sadness",
"surprise",
"trust"
] | ---
license: apache-2.0
datasets:
- null
model_index:
- name: dear-jarvis-v5
results:
- task:
name: Text Classification
type: text-classification
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# dear-jarvis-v5
This model is a fine-tuned version of [distilbert-base-cased](https://huggingface.co/distilbert-base-cased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3148
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 32
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| No log | 1.0 | 470 | 0.3106 |
| 0.3452 | 2.0 | 940 | 0.3064 |
| 0.2692 | 3.0 | 1410 | 0.3148 |
### Framework versions
- Transformers 4.7.0
- Pytorch 1.9.0+cu102
- Datasets 1.8.0
- Tokenizers 0.10.3
| 1,412 |
M-FAC/bert-mini-finetuned-mnli | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | # BERT-mini model finetuned with M-FAC
This model is finetuned on MNLI dataset with state-of-the-art second-order optimizer M-FAC.
Check NeurIPS 2021 paper for more details on M-FAC: [https://arxiv.org/pdf/2107.03356.pdf](https://arxiv.org/pdf/2107.03356.pdf).
## Finetuning setup
For fair comparison against default Adam baseline, we finetune the model in the same framework as described here [https://github.com/huggingface/transformers/tree/master/examples/pytorch/text-classification](https://github.com/huggingface/transformers/tree/master/examples/pytorch/text-classification) and just swap Adam optimizer with M-FAC.
Hyperparameters used by M-FAC optimizer:
```bash
learning rate = 1e-4
number of gradients = 1024
dampening = 1e-6
```
## Results
We share the best model out of 5 runs with the following score on MNLI validation set:
```bash
matched_accuracy = 75.13
mismatched_accuracy = 75.93
```
Mean and standard deviation for 5 runs on MNLI validation set:
| | Matched Accuracy | Mismatched Accuracy |
|:-----:|:----------------:|:-------------------:|
| Adam | 73.30 ± 0.20 | 74.85 ± 0.09 |
| M-FAC | 74.59 ± 0.41 | 75.95 ± 0.14 |
Results can be reproduced by adding M-FAC optimizer code in [https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) and running the following bash script:
```bash
CUDA_VISIBLE_DEVICES=0 python run_glue.py \
--seed 8276 \
--model_name_or_path prajjwal1/bert-mini \
--task_name mnli \
--do_train \
--do_eval \
--max_seq_length 128 \
--per_device_train_batch_size 32 \
--learning_rate 1e-4 \
--num_train_epochs 5 \
--output_dir out_dir/ \
--optim MFAC \
--optim_args '{"lr": 1e-4, "num_grads": 1024, "damp": 1e-6}'
```
We believe these results could be improved with modest tuning of hyperparameters: `per_device_train_batch_size`, `learning_rate`, `num_train_epochs`, `num_grads` and `damp`. For the sake of fair comparison and a robust default setup we use the same hyperparameters across all models (`bert-tiny`, `bert-mini`) and all datasets (SQuAD version 2 and GLUE).
Our code for M-FAC can be found here: [https://github.com/IST-DASLab/M-FAC](https://github.com/IST-DASLab/M-FAC).
A step-by-step tutorial on how to integrate and use M-FAC with any repository can be found here: [https://github.com/IST-DASLab/M-FAC/tree/master/tutorials](https://github.com/IST-DASLab/M-FAC/tree/master/tutorials).
## BibTeX entry and citation info
```bibtex
@article{frantar2021m,
title={M-FAC: Efficient Matrix-Free Approximations of Second-Order Information},
author={Frantar, Elias and Kurtic, Eldar and Alistarh, Dan},
journal={Advances in Neural Information Processing Systems},
volume={35},
year={2021}
}
```
| 2,882 |
M-FAC/bert-tiny-finetuned-qqp | null | # BERT-tiny model finetuned with M-FAC
This model is finetuned on QQP dataset with state-of-the-art second-order optimizer M-FAC.
Check NeurIPS 2021 paper for more details on M-FAC: [https://arxiv.org/pdf/2107.03356.pdf](https://arxiv.org/pdf/2107.03356.pdf).
## Finetuning setup
For fair comparison against default Adam baseline, we finetune the model in the same framework as described here [https://github.com/huggingface/transformers/tree/master/examples/pytorch/text-classification](https://github.com/huggingface/transformers/tree/master/examples/pytorch/text-classification) and just swap Adam optimizer with M-FAC.
Hyperparameters used by M-FAC optimizer:
```bash
learning rate = 1e-4
number of gradients = 1024
dampening = 1e-6
```
## Results
We share the best model out of 5 runs with the following score on QQP validation set:
```bash
f1 = 79.84
accuracy = 84.40
```
Mean and standard deviation for 5 runs on QQP validation set:
| | F1 | Accuracy |
|:----:|:-----------:|:----------:|
| Adam | 77.58 ± 0.08 | 81.09 ± 0.15 |
| M-FAC | 79.71 ± 0.13 | 84.29 ± 0.08 |
Results can be reproduced by adding M-FAC optimizer code in [https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) and running the following bash script:
```bash
CUDA_VISIBLE_DEVICES=0 python run_glue.py \
--seed 1234 \
--model_name_or_path prajjwal1/bert-tiny \
--task_name qqp \
--do_train \
--do_eval \
--max_seq_length 128 \
--per_device_train_batch_size 32 \
--learning_rate 1e-4 \
--num_train_epochs 5 \
--output_dir out_dir/ \
--optim MFAC \
--optim_args '{"lr": 1e-4, "num_grads": 1024, "damp": 1e-6}'
```
We believe these results could be improved with modest tuning of hyperparameters: `per_device_train_batch_size`, `learning_rate`, `num_train_epochs`, `num_grads` and `damp`. For the sake of fair comparison and a robust default setup we use the same hyperparameters across all models (`bert-tiny`, `bert-mini`) and all datasets (SQuAD version 2 and GLUE).
Our code for M-FAC can be found here: [https://github.com/IST-DASLab/M-FAC](https://github.com/IST-DASLab/M-FAC).
A step-by-step tutorial on how to integrate and use M-FAC with any repository can be found here: [https://github.com/IST-DASLab/M-FAC/tree/master/tutorials](https://github.com/IST-DASLab/M-FAC/tree/master/tutorials).
## BibTeX entry and citation info
```bibtex
@article{frantar2021m,
title={M-FAC: Efficient Matrix-Free Approximations of Second-Order Information},
author={Frantar, Elias and Kurtic, Eldar and Alistarh, Dan},
journal={Advances in Neural Information Processing Systems},
volume={35},
year={2021}
}
```
| 2,785 |
MoritzLaurer/MiniLM-L6-mnli-binary | [
"entailment",
"not_entailment"
] | ---
language:
- en
tags:
- text-classification
- zero-shot-classification
metrics:
- accuracy
widget:
- text: "I liked the movie. [SEP] The movie was good."
---
# MiniLM-L6-mnli-binary
## Model description
This model was trained on the [MultiNLI](https://huggingface.co/datasets/multi_nli) dataset. The model was trained for binary NLI, which means that the "neutral" and "contradiction" classes were merged into one class. The model therefore predicts "entailment" or "not_entailment".
The base model is MiniLM-L6 from Microsoft, which is very fast, but a bit less accurate than other models.
## Intended uses & limitations
#### How to use the model
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model_name = "MoritzLaurer/MiniLM-L6-mnli-binary"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)
premise = "I liked the movie"
hypothesis = "The movie was good."
input = tokenizer(premise, hypothesis, truncation=True, return_tensors="pt")
output = model(input["input_ids"].to(device)) # device = "cuda:0" or "cpu"
prediction = torch.softmax(output["logits"][0], -1).tolist()
label_names = ["entailment", "not_entailment"]
prediction = {name: round(float(pred) * 100, 1) for pred, name in zip(prediction, label_names)}
print(prediction)
```
### Training data
[MultiNLI](https://huggingface.co/datasets/multi_nli).
### Training procedure
MiniLM-L6-mnli-binary was trained using the Hugging Face trainer with the following hyperparameters.
```
training_args = TrainingArguments(
num_train_epochs=5, # total number of training epochs
learning_rate=2e-05,
per_device_train_batch_size=32, # batch size per device during training
per_device_eval_batch_size=32, # batch size for evaluation
warmup_ratio=0.1, # number of warmup steps for learning rate scheduler
weight_decay=0.06, # strength of weight decay
fp16=True # mixed precision training
)
```
### Eval results
The model was evaluated using the binary (matched) test set from MultiNLI. Accuracy: 0.886
## Limitations and bias
Please consult the original MiniLM paper and literature on different NLI datasets for potential biases.
### BibTeX entry and citation info
If you want to cite this model, please cite the original MiniLM paper, the respective NLI datasets and include a link to this model on the Hugging Face hub. | 2,509 |
Sakil/distilbert_lazylearner_hatespeech_detection | null | ---
license: apache-2.0
language: en
tags:
- hate
- speech
widget:
- text: "RT @ShenikaRoberts: The shit you hear about me might be true or it might be faker than the bitch who told it to ya ᙨ"
---
# Dataset Collection:
* The hatespeech dataset is collected from different open sources like Kaggle ,social media like Twitter.
* The dataset has the two classes hatespeech and non hatespeech.
* The class distribution is equal
* Different strategies have been followed during the data gathering phase.
* The dataset is collected from relevant sources.
# distilbert-base-uncased model is fine-tuned for Hate Speech Detection
* The model is fine-tuned on the dataset.
* This model can be used to create the labels for academic purposes or for industrial purposes.
* This model can be used for the inference purpose as well.
# Data Fields:
**label**: 0 - it is a hate speech, 1 - not a hate speech
# Application:
* This model is useful for the detection of hatespeech in the tweets.
* There are numerous situations where we have tweet data but no labels, so this approach can be used to create labels.
* You can fine-tune this model for your particular use cases.
# Model Implementation
# !pip install transformers[sentencepiece]
from transformers import pipeline
model_name="Sakil/distilbert_lazylearner_hatespeech_detection"
classifier = pipeline("text-classification",model=model_name)
classifier("!!! RT @mayasolovely: As a woman you shouldn't complain about cleaning up your house. & as a man you should always take the trash out...")
# Github: [Sakil Ansari](https://github.com/Sakil786/hate_speech_detection_pretrained_model) | 1,696 |
SetFit/deberta-v3-large__sst2__train-16-7 | [
"negative",
"positive"
] | ---
license: mit
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: deberta-v3-large__sst2__train-16-7
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# deberta-v3-large__sst2__train-16-7
This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6953
- Accuracy: 0.5063
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 50
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.6911 | 1.0 | 7 | 0.7455 | 0.2857 |
| 0.6844 | 2.0 | 14 | 0.7242 | 0.2857 |
| 0.6137 | 3.0 | 21 | 0.7341 | 0.4286 |
| 0.3805 | 4.0 | 28 | 1.0217 | 0.4286 |
| 0.2201 | 5.0 | 35 | 1.1437 | 0.2857 |
| 0.0296 | 6.0 | 42 | 1.5997 | 0.4286 |
| 0.0103 | 7.0 | 49 | 2.6835 | 0.4286 |
| 0.0046 | 8.0 | 56 | 3.3521 | 0.4286 |
| 0.002 | 9.0 | 63 | 3.7846 | 0.4286 |
| 0.0017 | 10.0 | 70 | 4.0088 | 0.4286 |
| 0.0018 | 11.0 | 77 | 4.1483 | 0.4286 |
| 0.0006 | 12.0 | 84 | 4.2235 | 0.4286 |
### Framework versions
- Transformers 4.15.0
- Pytorch 1.10.2+cu102
- Datasets 1.18.2
- Tokenizers 0.10.3
| 2,092 |
StevenLimcorn/indo-roberta-indonli | [
"c",
"e",
"n"
] | ---
language: id
tags:
- roberta
license: mit
datasets:
- indonli
widget:
- text: "Amir Sjarifoeddin Harahap lahir di Kota Medan, Sumatera Utara, 27 April 1907. Ia meninggal di Surakarta, Jawa Tengah, pada 19 Desember 1948 dalam usia 41 tahun. </s></s> Amir Sjarifoeddin Harahap masih hidup."
---
## Indo-roberta-indonli
Indo-roberta-indonli is natural language inference classifier based on [Indo-roberta](https://huggingface.co/flax-community/indonesian-roberta-base) model. It was trained on the trained on [IndoNLI](https://github.com/ir-nlp-csui/indonli/tree/main/data/indonli) dataset. The model used was [Indo-roberta](https://huggingface.co/flax-community/indonesian-roberta-base) and was transfer-learned to a natural inference classifier model. The model are tested using the validation, test_layer and test_expert dataset given in the github repository. The results are shown below.
### Result
| Dataset | Accuracy | F1 | Precision | Recall |
|-------------|----------|---------|-----------|---------|
| Test Lay | 0.74329 | 0.74075 | 0.74283 | 0.74133 |
| Test Expert | 0.6115 | 0.60543 | 0.63924 | 0.61742 |
## Model
The model was trained on with 5 epochs, batch size 16, learning rate 2e-5 and weight decay 0.01. Achieved different metrics as shown below.
| Epoch | Training Loss | Validation Loss | Accuracy | F1 | Precision | Recall |
|-------|---------------|-----------------|----------|----------|-----------|----------|
| 1 | 0.942500 | 0.658559 | 0.737369 | 0.735552 | 0.735488 | 0.736679 |
| 2 | 0.649200 | 0.645290 | 0.761493 | 0.759593 | 0.762784 | 0.759642 |
| 3 | 0.437100 | 0.667163 | 0.766045 | 0.763979 | 0.765740 | 0.763792 |
| 4 | 0.282000 | 0.786683 | 0.764679 | 0.761802 | 0.762011 | 0.761684 |
| 5 | 0.193500 | 0.925717 | 0.765134 | 0.763127 | 0.763560 | 0.763489 |
## How to Use
### As NLI Classifier
```python
from transformers import pipeline
pretrained_name = "StevenLimcorn/indonesian-roberta-indonli"
nlp = pipeline(
"zero-shot-classification",
model=pretrained_name,
tokenizer=pretrained_name
)
nlp("Amir Sjarifoeddin Harahap lahir di Kota Medan, Sumatera Utara, 27 April 1907. Ia meninggal di Surakarta, Jawa Tengah, pada 19 Desember 1948 dalam usia 41 tahun. </s></s> Amir Sjarifoeddin Harahap masih hidup.")
```
## Disclaimer
Do consider the biases which come from both the pre-trained RoBERTa model and the `INDONLI` dataset that may be carried over into the results of this model.
## Author
Indonesian RoBERTa Base IndoNLI was trained and evaluated by [Steven Limcorn](https://github.com/stevenlimcorn). All computation and development are done on Google Colaboratory using their free GPU access.
## Reference
The dataset we used is by IndoNLI.
```
@inproceedings{indonli,
title = "IndoNLI: A Natural Language Inference Dataset for Indonesian",
author = "Mahendra, Rahmad and Aji, Alham Fikri and Louvan, Samuel and Rahman, Fahrurrozi and Vania, Clara",
booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2021",
publisher = "Association for Computational Linguistics",
}
``` | 3,249 |
TehranNLP-org/bert-base-cased-avg-cola | null | The uploaded model is from epoch 4 with Matthews Correlation of 61.05
"best_metric": 0.4796141982078552,<br>
"best_model_checkpoint": "/content/output_dir/checkpoint-268",<br>
"epoch": 10.0,<br>
"global_step": 2680,<br>
"is_hyper_param_search": false,<br>
"is_local_process_zero": true,<br>
"is_world_process_zero": true,<br>
"max_steps": 2680,<br>
"num_train_epochs": 10,<br>
"total_flos": 7113018526540800.0,<br>
"trial_name": null,<br>
"trial_params": null<br>
<table class="table table-bordered table-hover table-condensed" style="width: 60%; overflow: auto">
<thead><tr><th title="Field #1">epoch</th>
<th title="Field #2">eval_loss</th>
<th title="Field #3">eval_matthews_correlation</th>
<th title="Field #4">eval_runtime</th>
<th title="Field #5">eval_samples_per_second</th>
<th title="Field #6">eval_steps_per_second</th>
<th title="Field #7">step</th>
<th title="Field #8">learning_rate</th>
<th title="Field #9">loss</th>
</tr></thead>
<tbody><tr>
<td align="left">1</td>
<td align="left">0.4796141982078552</td>
<td align="left">0.5351033849356494</td>
<td align="left">8.8067</td>
<td align="left">118.433</td>
<td align="left">14.875</td>
<td align="left">268</td>
<td align="left">0.000018067415730337083</td>
<td align="left">0.4913</td>
</tr>
<tr>
<td align="left">2</td>
<td align="left">0.5334435701370239</td>
<td align="left">0.5178799252679331</td>
<td align="left">8.9439</td>
<td align="left">116.616</td>
<td align="left">14.647</td>
<td align="left">536</td>
<td align="left">0.00001605992509363296</td>
<td align="left">0.2872</td>
</tr>
<tr>
<td align="left">3</td>
<td align="left">0.5544090270996094</td>
<td align="left">0.5649788851042796</td>
<td align="left">8.9467</td>
<td align="left">116.58</td>
<td align="left">14.642</td>
<td align="left">804</td>
<td align="left">0.000014052434456928841</td>
<td align="left">0.1777</td>
</tr>
<tr>
<td align="left">4</td>
<td align="left">0.5754779577255249</td>
<td align="left">0.6105374636148787</td>
<td align="left">8.8982</td>
<td align="left">117.215</td>
<td align="left">14.722</td>
<td align="left">1072</td>
<td align="left">0.000012044943820224718</td>
<td align="left">0.1263</td>
</tr>
<tr>
<td align="left">5</td>
<td align="left">0.7263916730880737</td>
<td align="left">0.5807606001872874</td>
<td align="left">8.9705</td>
<td align="left">116.27</td>
<td align="left">14.603</td>
<td align="left">1340</td>
<td align="left">0.000010037453183520601</td>
<td align="left">0.0905</td>
</tr>
<tr>
<td align="left">6</td>
<td align="left">0.8121512532234192</td>
<td align="left">0.5651092792103851</td>
<td align="left">8.9924</td>
<td align="left">115.987</td>
<td align="left">14.568</td>
<td align="left">1608</td>
<td align="left">0.00000802996254681648</td>
<td align="left">0.0692</td>
</tr>
<tr>
<td align="left">7</td>
<td align="left">0.941014289855957</td>
<td align="left">0.5632084517291658</td>
<td align="left">8.9583</td>
<td align="left">116.428</td>
<td align="left">14.623</td>
<td align="left">1876</td>
<td align="left">0.000006022471910112359</td>
<td align="left">0.0413</td>
</tr>
<tr>
<td align="left">8</td>
<td align="left">1.0095174312591553</td>
<td align="left">0.5856531698367675</td>
<td align="left">9.0029</td>
<td align="left">115.851</td>
<td align="left">14.551</td>
<td align="left">2144</td>
<td align="left">0.00000401498127340824</td>
<td align="left">0.0327</td>
</tr>
<tr>
<td align="left">9</td>
<td align="left">1.0425965785980225</td>
<td align="left">0.5941395545037332</td>
<td align="left">8.9217</td>
<td align="left">116.906</td>
<td align="left">14.683</td>
<td align="left">2412</td>
<td align="left">0.00000200749063670412</td>
<td align="left">0.0202</td>
</tr>
<tr>
<td align="left">10</td>
<td align="left">1.0782166719436646</td>
<td align="left">0.5956649094312695</td>
<td align="left">8.9472</td>
<td align="left">116.572</td>
<td align="left">14.641</td>
<td align="left">2680</td>
<td align="left">0</td>
<td align="left">0.0104</td>
</tr>
</tbody></table> | 4,039 |
Theivaprakasham/sentence-transformers-msmarco-distilbert-base-tas-b-twitter_sentiment | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | ---
license: apache-2.0
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: sentence-transformers-msmarco-distilbert-base-tas-b-twitter_sentiment
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# sentence-transformers-msmarco-distilbert-base-tas-b-twitter_sentiment
This model is a fine-tuned version of [sentence-transformers/msmarco-distilbert-base-tas-b](https://huggingface.co/sentence-transformers/msmarco-distilbert-base-tas-b) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6954
- Accuracy: 0.7146
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-06
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 20
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.8892 | 1.0 | 1387 | 0.8472 | 0.6180 |
| 0.7965 | 2.0 | 2774 | 0.7797 | 0.6609 |
| 0.7459 | 3.0 | 4161 | 0.7326 | 0.6872 |
| 0.7096 | 4.0 | 5548 | 0.7133 | 0.6995 |
| 0.6853 | 5.0 | 6935 | 0.6998 | 0.7002 |
| 0.6561 | 6.0 | 8322 | 0.6949 | 0.7059 |
| 0.663 | 7.0 | 9709 | 0.6956 | 0.7077 |
| 0.6352 | 8.0 | 11096 | 0.6890 | 0.7164 |
| 0.6205 | 9.0 | 12483 | 0.6888 | 0.7117 |
| 0.6203 | 10.0 | 13870 | 0.6871 | 0.7121 |
| 0.6005 | 11.0 | 15257 | 0.6879 | 0.7171 |
| 0.5985 | 12.0 | 16644 | 0.6870 | 0.7139 |
| 0.5839 | 13.0 | 18031 | 0.6882 | 0.7164 |
| 0.5861 | 14.0 | 19418 | 0.6910 | 0.7124 |
| 0.5732 | 15.0 | 20805 | 0.6916 | 0.7153 |
| 0.5797 | 16.0 | 22192 | 0.6947 | 0.7110 |
| 0.5565 | 17.0 | 23579 | 0.6930 | 0.7175 |
| 0.5636 | 18.0 | 24966 | 0.6959 | 0.7106 |
| 0.5642 | 19.0 | 26353 | 0.6952 | 0.7132 |
| 0.5717 | 20.0 | 27740 | 0.6954 | 0.7146 |
### Framework versions
- Transformers 4.13.0.dev0
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
| 2,705 |
aXhyra/demo_hate_31415 | null | ---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- tweet_eval
metrics:
- f1
model-index:
- name: demo_hate_31415
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: tweet_eval
type: tweet_eval
args: hate
metrics:
- name: F1
type: f1
value: 0.7772939485986298
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# demo_hate_31415
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the tweet_eval dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8697
- F1: 0.7773
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 7.320702985778492e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 0
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 4
### Training results
| Training Loss | Epoch | Step | Validation Loss | F1 |
|:-------------:|:-----:|:----:|:---------------:|:------:|
| No log | 1.0 | 282 | 0.4850 | 0.7645 |
| 0.3877 | 2.0 | 564 | 0.5160 | 0.7856 |
| 0.3877 | 3.0 | 846 | 0.6927 | 0.7802 |
| 0.1343 | 4.0 | 1128 | 0.8697 | 0.7773 |
### Framework versions
- Transformers 4.12.5
- Pytorch 1.9.1
- Datasets 1.16.1
- Tokenizers 0.10.3
| 1,756 |
adamlin/ml999_power_punching_and_shearing_machinery | [
"0",
"1"
] | Entry not found | 15 |
adelgasmi/autonlp-kpmg_nlp-18833547 | [
"0",
"1",
"2",
"3",
"4"
] | ---
tags: autonlp
language: ar
widget:
- text: "I love AutoNLP 🤗"
datasets:
- adelgasmi/autonlp-data-kpmg_nlp
co2_eq_emissions: 64.58945483765274
---
# Model Trained Using AutoNLP
- Problem type: Multi-class Classification
- Model ID: 18833547
- CO2 Emissions (in grams): 64.58945483765274
## Validation Metrics
- Loss: 0.14247722923755646
- Accuracy: 0.9586074193404036
- Macro F1: 0.9468339778730883
- Micro F1: 0.9586074193404036
- Weighted F1: 0.9585551117678807
- Macro Precision: 0.9445436604001405
- Micro Precision: 0.9586074193404036
- Weighted Precision: 0.9591405429662925
- Macro Recall: 0.9499427161888565
- Micro Recall: 0.9586074193404036
- Weighted Recall: 0.9586074193404036
## Usage
You can use cURL to access this model:
```
$ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/adelgasmi/autonlp-kpmg_nlp-18833547
```
Or Python API:
```
from transformers import AutoModelForSequenceClassification, AutoTokenizer
model = AutoModelForSequenceClassification.from_pretrained("adelgasmi/autonlp-kpmg_nlp-18833547", use_auth_token=True)
tokenizer = AutoTokenizer.from_pretrained("adelgasmi/autonlp-kpmg_nlp-18833547", use_auth_token=True)
inputs = tokenizer("I love AutoNLP", return_tensors="pt")
outputs = model(**inputs)
``` | 1,369 |
aditeyabaral/finetuned-iitpmovie-additionalpretrained-distilbert-base-cased | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
adp12/cs410finetune1 | [
"NEGATIVE",
"POSITIVE"
] | Entry not found | 15 |
agiagoulas/bert-pss | null | bert-base-uncased model trained on the tobacco800 dataset for the task of page-stream-segmentation.
[Link](https://github.com/agiagoulas/page-stream-segmentation) to the GitHub Repo with the model implementation. | 213 |
akahana/indonesia-sentiment-roberta | [
"POSITIF",
"NETRAL",
"NEGATIF"
] | ---
language: "id"
widget:
- text: "dia orang yang baik ya bunds."
---
## how to use
```python
from transformers import pipeline, set_seed
path = "akahana/indonesia-sentiment-roberta"
emotion = pipeline('text-classification',
model=path,device=0)
set_seed(42)
kalimat = "dia orang yang baik ya bunds."
preds = emotion(kalimat)
preds
``` | 362 |
am4nsolanki/autonlp-text-hateful-memes-36789092 | [
"0",
"1"
] | ---
tags: autonlp
language: en
widget:
- text: "I love AutoNLP 🤗"
datasets:
- am4nsolanki/autonlp-data-text-hateful-memes
co2_eq_emissions: 1.4280361775467445
---
# Model Trained Using AutoNLP
- Problem type: Binary Classification
- Model ID: 36789092
- CO2 Emissions (in grams): 1.4280361775467445
## Validation Metrics
- Loss: 0.5255328416824341
- Accuracy: 0.7666078777189889
- Precision: 0.6913123844731978
- Recall: 0.6192052980132451
- AUC: 0.7893359070795125
- F1: 0.6532751091703057
## Usage
You can use cURL to access this model:
```
$ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/am4nsolanki/autonlp-text-hateful-memes-36789092
```
Or Python API:
```
from transformers import AutoModelForSequenceClassification, AutoTokenizer
model = AutoModelForSequenceClassification.from_pretrained("am4nsolanki/autonlp-text-hateful-memes-36789092", use_auth_token=True)
tokenizer = AutoTokenizer.from_pretrained("am4nsolanki/autonlp-text-hateful-memes-36789092", use_auth_token=True)
inputs = tokenizer("I love AutoNLP", return_tensors="pt")
outputs = model(**inputs)
``` | 1,203 |
berkergurcay/10k-pretrained-bert-model | [
"LABEL_0",
"LABEL_1",
"LABEL_2"
] | Entry not found | 15 |
bioformers/bioformer-cased-v1.0-mnli | [
"contradiction",
"entailment",
"neutral"
] | [bioformer-cased-v1.0](https://huggingface.co/bioformers/bioformer-cased-v1.0) fined-tuned on the [MNLI](https://cims.nyu.edu/~sbowman/multinli/) dataset for 2 epochs.
The fine-tuning process was performed on two NVIDIA GeForce GTX 1080 Ti GPUs (11GB). The parameters are:
```
max_seq_length=512
per_device_train_batch_size=16
total train batch size (w. parallel, distributed & accumulation) = 32
learning_rate=3e-5
```
## Evaluation results
eval_accuracy = 0.803973
## Speed
In our experiments, the inference speed of Bioformer is 3x as fast as BERT-base/BioBERT/PubMedBERT, and is 40% faster than DistilBERT.
## More information
The Multi-Genre Natural Language Inference Corpus is a crowdsourced collection of sentence pairs with textual entailment annotations. Given a premise sentence and a hypothesis sentence, the task is to predict whether the premise entails the hypothesis (entailment), contradicts the hypothesis (contradiction), or neither (neutral). The premise sentences are gathered from ten different sources, including transcribed speech, fiction, and government reports. The authors of the benchmark use the standard test set, for which they obtained private labels from the RTE authors, and evaluate on both the matched (in-domain) and mismatched (cross-domain) section. They also uses and recommend the SNLI corpus as 550k examples of auxiliary training data. (source: https://huggingface.co/datasets/glue) | 1,435 |
chitra/finetuned-adversarial-paraphrasing-detector | null | Entry not found | 15 |
dkleczek/Polish-Hate-Speech-Detection-Herbert-Large | null | Entry not found | 15 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.