modelId
stringlengths
6
107
label
list
readme
stringlengths
0
56.2k
readme_len
int64
0
56.2k
TehranNLP-org/bert-base-uncased-avg-mnli-2e-5-63
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
TehranNLP-org/bert-base-uncased-avg-mnli
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
TehranNLP-org/bert-base-uncased-mrpc-2e-5-42
null
Entry not found
15
TehranNLP-org/electra-base-avg-cola-2e-5-21
null
Entry not found
15
TehranNLP-org/electra-base-avg-cola-2e-5-42
null
Entry not found
15
TehranNLP-org/electra-base-avg-cola
null
The uploaded model is from epoch 9 with Matthews Correlation of 66.77 "best_metric": 0.667660908939119,<br> "best_model_checkpoint": "/content/output_dir/checkpoint-2412",<br> "epoch": 10.0,<br> "global_step": 2680,<br> "is_hyper_param_search": false,<br> "is_local_process_zero": true,<br> "is_world_process_zero": true,<br> "max_steps": 2680,<br> "num_train_epochs": 10,<br> "total_flos": 7189983634007040.0,<br> "trial_name": null,<br> "trial_params": null<br> <table class="table table-bordered table-hover table-condensed"> <thead><tr><th title="Field #1">epoch</th> <th title="Field #2">eval_loss</th> <th title="Field #3">eval_matthews_correlation</th> <th title="Field #4">eval_runtime</th> <th title="Field #5">eval_samples_per_second</th> <th title="Field #6">eval_steps_per_second</th> <th title="Field #7">step</th> <th title="Field #8">learning_rate</th> <th title="Field #9">loss</th> </tr></thead> <tbody><tr> <td align="right">1</td> <td align="right">0.5115634202957153</td> <td align="right">0.5385290213636863</td> <td align="right">7.985</td> <td align="right">130.62</td> <td align="right">16.406</td> <td align="right">268</td> <td align="right">0.00009280492497114274</td> <td align="right">0.4622</td> </tr> <tr> <td align="right">2</td> <td align="right">0.4201788902282715</td> <td align="right">0.6035894895952164</td> <td align="right">8.0283</td> <td align="right">129.916</td> <td align="right">16.317</td> <td align="right">536</td> <td align="right">0.00008249326664101577</td> <td align="right">0.2823</td> </tr> <tr> <td align="right">3</td> <td align="right">0.580650806427002</td> <td align="right">0.5574138665741355</td> <td align="right">8.1314</td> <td align="right">128.268</td> <td align="right">16.11</td> <td align="right">804</td> <td align="right">0.00007218160831088881</td> <td align="right">0.1804</td> </tr> <tr> <td align="right">4</td> <td align="right">0.4439031779766083</td> <td align="right">0.6557697896854868</td> <td align="right">8.1435</td> <td align="right">128.078</td> <td align="right">16.087</td> <td align="right">1072</td> <td align="right">0.00006186994998076183</td> <td align="right">0.1357</td> </tr> <tr> <td align="right">5</td> <td align="right">0.5736830830574036</td> <td align="right">0.6249925495853809</td> <td align="right">8.0533</td> <td align="right">129.512</td> <td align="right">16.267</td> <td align="right">1340</td> <td align="right">0.00005155829165063486</td> <td align="right">0.0913</td> </tr> <tr> <td align="right">6</td> <td align="right">0.7729296684265137</td> <td align="right">0.6188970025554703</td> <td align="right">8.081</td> <td align="right">129.068</td> <td align="right">16.211</td> <td align="right">1608</td> <td align="right">0.000041246633320507885</td> <td align="right">0.065</td> </tr> <tr> <td align="right">7</td> <td align="right">0.7351673245429993</td> <td align="right">0.6405767700619004</td> <td align="right">8.1372</td> <td align="right">128.176</td> <td align="right">16.099</td> <td align="right">1876</td> <td align="right">0.00003093497499038092</td> <td align="right">0.0433</td> </tr> <tr> <td align="right">8</td> <td align="right">0.7900031208992004</td> <td align="right">0.6565021466238845</td> <td align="right">8.1095</td> <td align="right">128.615</td> <td align="right">16.154</td> <td align="right">2144</td> <td align="right">0.000020623316660253942</td> <td align="right">0.0199</td> </tr> <tr> <td align="right">9</td> <td align="right">0.8539554476737976</td> <td align="right">0.667660908939119</td> <td align="right">8.1204</td> <td align="right">128.442</td> <td align="right">16.132</td> <td align="right">2412</td> <td align="right">0.000010311658330126971</td> <td align="right">0.0114</td> </tr> <tr> <td align="right">10</td> <td align="right">0.9261117577552795</td> <td align="right">0.660301076782038</td> <td align="right">8.0088</td> <td align="right">130.231</td> <td align="right">16.357</td> <td align="right">2680</td> <td align="right">0</td> <td align="right">0.0066</td> </tr> </tbody></table>
4,086
TehranNLP-org/electra-base-avg-mnli-2e-5-63
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
TehranNLP-org/electra-base-avg-sst2-2e-5-42
null
Entry not found
15
TehranNLP-org/roberta-base-mrpc-2e-5-42
null
Entry not found
15
TehranNLP-org/xlnet-base-cased-avg-cola-2e-5-63
null
Entry not found
15
TehranNLP-org/xlnet-base-cased-avg-mnli-2e-5-21
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
TehranNLP-org/xlnet-base-cased-avg-mnli-2e-5-63
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
TehranNLP-org/xlnet-base-cased-avg-mnli-2e-5
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
TehranNLP-org/xlnet-base-cased-avg-mnli
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
TehranNLP-org/xlnet-base-cased-avg-sst2-2e-5-21
null
Entry not found
15
Tejas3/distillbert_base_uncased_80
[ "NEGATIVE", "NEUTRAL", "POSITIVE" ]
Entry not found
15
TomW/TOMFINSEN
[ "negative", "neutral", "positive" ]
--- license: apache-2.0 tags: - generated_from_trainer datasets: - financial_phrasebank metrics: - recall - accuracy - precision model-index: - name: TOMFINSEN results: - task: name: Text Classification type: text-classification dataset: name: financial_phrasebank type: financial_phrasebank args: sentences_50agree metrics: - name: Recall type: recall value: 0.8985861629736692 - name: Accuracy type: accuracy value: 0.8742268041237113 - name: Precision type: precision value: 0.8509995913451198 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # TOMFINSEN This model is a fine-tuned version of [deepmind/language-perceiver](https://huggingface.co/deepmind/language-perceiver) on the financial_phrasebank dataset. It achieves the following results on the evaluation set: - Loss: 0.3642 - Recall: 0.8986 - Accuracy: 0.8742 - Precision: 0.8510 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - distributed_type: tpu - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Recall | Accuracy | Precision | |:-------------:|:-----:|:----:|:---------------:|:------:|:--------:|:---------:| | 0.5403 | 1.0 | 273 | 0.4207 | 0.8358 | 0.8619 | 0.8534 | | 0.3939 | 2.0 | 546 | 0.3750 | 0.8943 | 0.8577 | 0.8225 | | 0.1993 | 3.0 | 819 | 0.3113 | 0.8882 | 0.8660 | 0.8367 | | 0.301 | 4.0 | 1092 | 0.3642 | 0.8986 | 0.8742 | 0.8510 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.9.0+cu102 - Datasets 1.17.0 - Tokenizers 0.10.3
2,186
TransQuest/monotransquest-hter-en_cs-pharmaceutical
[ "LABEL_0" ]
--- language: en-cs tags: - Quality Estimation - monotransquest - hter license: apache-2.0 --- # TransQuest: Translation Quality Estimation with Cross-lingual Transformers The goal of quality estimation (QE) is to evaluate the quality of a translation without having access to a reference translation. High-accuracy QE that can be easily deployed for a number of language pairs is the missing piece in many commercial translation workflows as they have numerous potential uses. They can be employed to select the best translation when several translation engines are available or can inform the end user about the reliability of automatically translated content. In addition, QE systems can be used to decide whether a translation can be published as it is in a given context, or whether it requires human post-editing before publishing or translation from scratch by a human. The quality estimation can be done at different levels: document level, sentence level and word level. With TransQuest, we have opensourced our research in translation quality estimation which also won the sentence-level direct assessment quality estimation shared task in [WMT 2020](http://www.statmt.org/wmt20/quality-estimation-task.html). TransQuest outperforms current open-source quality estimation frameworks such as [OpenKiwi](https://github.com/Unbabel/OpenKiwi) and [DeepQuest](https://github.com/sheffieldnlp/deepQuest). ## Features - Sentence-level translation quality estimation on both aspects: predicting post editing efforts and direct assessment. - Word-level translation quality estimation capable of predicting quality of source words, target words and target gaps. - Outperform current state-of-the-art quality estimation methods like DeepQuest and OpenKiwi in all the languages experimented. - Pre-trained quality estimation models for fifteen language pairs are available in [HuggingFace.](https://huggingface.co/TransQuest) ## Installation ### From pip ```bash pip install transquest ``` ### From Source ```bash git clone https://github.com/TharinduDR/TransQuest.git cd TransQuest pip install -r requirements.txt ``` ## Using Pre-trained Models ```python import torch from transquest.algo.sentence_level.monotransquest.run_model import MonoTransQuestModel model = MonoTransQuestModel("xlmroberta", "TransQuest/monotransquest-hter-en_cs-pharmaceutical", num_labels=1, use_cuda=torch.cuda.is_available()) predictions, raw_outputs = model.predict([["Reducerea acestor conflicte este importantă pentru conservare.", "Reducing these conflicts is not important for preservation."]]) print(predictions) ``` ## Documentation For more details follow the documentation. 1. **[Installation](https://tharindudr.github.io/TransQuest/install/)** - Install TransQuest locally using pip. 2. **Architectures** - Checkout the architectures implemented in TransQuest 1. [Sentence-level Architectures](https://tharindudr.github.io/TransQuest/architectures/sentence_level_architectures/) - We have released two architectures; MonoTransQuest and SiameseTransQuest to perform sentence level quality estimation. 2. [Word-level Architecture](https://tharindudr.github.io/TransQuest/architectures/word_level_architecture/) - We have released MicroTransQuest to perform word level quality estimation. 3. **Examples** - We have provided several examples on how to use TransQuest in recent WMT quality estimation shared tasks. 1. [Sentence-level Examples](https://tharindudr.github.io/TransQuest/examples/sentence_level_examples/) 2. [Word-level Examples](https://tharindudr.github.io/TransQuest/examples/word_level_examples/) 4. **Pre-trained Models** - We have provided pretrained quality estimation models for fifteen language pairs covering both sentence-level and word-level 1. [Sentence-level Models](https://tharindudr.github.io/TransQuest/models/sentence_level_pretrained/) 2. [Word-level Models](https://tharindudr.github.io/TransQuest/models/word_level_pretrained/) 5. **[Contact](https://tharindudr.github.io/TransQuest/contact/)** - Contact us for any issues with TransQuest ## Citations If you are using the word-level architecture, please consider citing this paper which is accepted to [ACL 2021](https://2021.aclweb.org/). ```bash @InProceedings{ranasinghe2021, author = {Ranasinghe, Tharindu and Orasan, Constantin and Mitkov, Ruslan}, title = {An Exploratory Analysis of Multilingual Word Level Quality Estimation with Cross-Lingual Transformers}, booktitle = {Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics}, year = {2021} } ``` If you are using the sentence-level architectures, please consider citing these papers which were presented in [COLING 2020](https://coling2020.org/) and in [WMT 2020](http://www.statmt.org/wmt20/) at EMNLP 2020. ```bash @InProceedings{transquest:2020a, author = {Ranasinghe, Tharindu and Orasan, Constantin and Mitkov, Ruslan}, title = {TransQuest: Translation Quality Estimation with Cross-lingual Transformers}, booktitle = {Proceedings of the 28th International Conference on Computational Linguistics}, year = {2020} } ``` ```bash @InProceedings{transquest:2020b, author = {Ranasinghe, Tharindu and Orasan, Constantin and Mitkov, Ruslan}, title = {TransQuest at WMT2020: Sentence-Level Direct Assessment}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, year = {2020} } ```
5,415
TransQuest/monotransquest-hter-en_de-it-nmt
[ "LABEL_0" ]
--- language: en-de tags: - Quality Estimation - monotransquest - hter license: apache-2.0 --- # TransQuest: Translation Quality Estimation with Cross-lingual Transformers The goal of quality estimation (QE) is to evaluate the quality of a translation without having access to a reference translation. High-accuracy QE that can be easily deployed for a number of language pairs is the missing piece in many commercial translation workflows as they have numerous potential uses. They can be employed to select the best translation when several translation engines are available or can inform the end user about the reliability of automatically translated content. In addition, QE systems can be used to decide whether a translation can be published as it is in a given context, or whether it requires human post-editing before publishing or translation from scratch by a human. The quality estimation can be done at different levels: document level, sentence level and word level. With TransQuest, we have opensourced our research in translation quality estimation which also won the sentence-level direct assessment quality estimation shared task in [WMT 2020](http://www.statmt.org/wmt20/quality-estimation-task.html). TransQuest outperforms current open-source quality estimation frameworks such as [OpenKiwi](https://github.com/Unbabel/OpenKiwi) and [DeepQuest](https://github.com/sheffieldnlp/deepQuest). ## Features - Sentence-level translation quality estimation on both aspects: predicting post editing efforts and direct assessment. - Word-level translation quality estimation capable of predicting quality of source words, target words and target gaps. - Outperform current state-of-the-art quality estimation methods like DeepQuest and OpenKiwi in all the languages experimented. - Pre-trained quality estimation models for fifteen language pairs are available in [HuggingFace.](https://huggingface.co/TransQuest) ## Installation ### From pip ```bash pip install transquest ``` ### From Source ```bash git clone https://github.com/TharinduDR/TransQuest.git cd TransQuest pip install -r requirements.txt ``` ## Using Pre-trained Models ```python import torch from transquest.algo.sentence_level.monotransquest.run_model import MonoTransQuestModel model = MonoTransQuestModel("xlmroberta", "TransQuest/monotransquest-hter-en_de-it-nmt", num_labels=1, use_cuda=torch.cuda.is_available()) predictions, raw_outputs = model.predict([["Reducerea acestor conflicte este importantă pentru conservare.", "Reducing these conflicts is not important for preservation."]]) print(predictions) ``` ## Documentation For more details follow the documentation. 1. **[Installation](https://tharindudr.github.io/TransQuest/install/)** - Install TransQuest locally using pip. 2. **Architectures** - Checkout the architectures implemented in TransQuest 1. [Sentence-level Architectures](https://tharindudr.github.io/TransQuest/architectures/sentence_level_architectures/) - We have released two architectures; MonoTransQuest and SiameseTransQuest to perform sentence level quality estimation. 2. [Word-level Architecture](https://tharindudr.github.io/TransQuest/architectures/word_level_architecture/) - We have released MicroTransQuest to perform word level quality estimation. 3. **Examples** - We have provided several examples on how to use TransQuest in recent WMT quality estimation shared tasks. 1. [Sentence-level Examples](https://tharindudr.github.io/TransQuest/examples/sentence_level_examples/) 2. [Word-level Examples](https://tharindudr.github.io/TransQuest/examples/word_level_examples/) 4. **Pre-trained Models** - We have provided pretrained quality estimation models for fifteen language pairs covering both sentence-level and word-level 1. [Sentence-level Models](https://tharindudr.github.io/TransQuest/models/sentence_level_pretrained/) 2. [Word-level Models](https://tharindudr.github.io/TransQuest/models/word_level_pretrained/) 5. **[Contact](https://tharindudr.github.io/TransQuest/contact/)** - Contact us for any issues with TransQuest ## Citations If you are using the word-level architecture, please consider citing this paper which is accepted to [ACL 2021](https://2021.aclweb.org/). ```bash @InProceedings{ranasinghe2021, author = {Ranasinghe, Tharindu and Orasan, Constantin and Mitkov, Ruslan}, title = {An Exploratory Analysis of Multilingual Word Level Quality Estimation with Cross-Lingual Transformers}, booktitle = {Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics}, year = {2021} } ``` If you are using the sentence-level architectures, please consider citing these papers which were presented in [COLING 2020](https://coling2020.org/) and in [WMT 2020](http://www.statmt.org/wmt20/) at EMNLP 2020. ```bash @InProceedings{transquest:2020a, author = {Ranasinghe, Tharindu and Orasan, Constantin and Mitkov, Ruslan}, title = {TransQuest: Translation Quality Estimation with Cross-lingual Transformers}, booktitle = {Proceedings of the 28th International Conference on Computational Linguistics}, year = {2020} } ``` ```bash @InProceedings{transquest:2020b, author = {Ranasinghe, Tharindu and Orasan, Constantin and Mitkov, Ruslan}, title = {TransQuest at WMT2020: Sentence-Level Direct Assessment}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, year = {2020} } ```
5,407
TransQuest/monotransquest-hter-en_de-wiki
[ "LABEL_0" ]
--- language: en-de tags: - Quality Estimation - monotransquest - hter license: apache-2.0 --- # TransQuest: Translation Quality Estimation with Cross-lingual Transformers The goal of quality estimation (QE) is to evaluate the quality of a translation without having access to a reference translation. High-accuracy QE that can be easily deployed for a number of language pairs is the missing piece in many commercial translation workflows as they have numerous potential uses. They can be employed to select the best translation when several translation engines are available or can inform the end user about the reliability of automatically translated content. In addition, QE systems can be used to decide whether a translation can be published as it is in a given context, or whether it requires human post-editing before publishing or translation from scratch by a human. The quality estimation can be done at different levels: document level, sentence level and word level. With TransQuest, we have opensourced our research in translation quality estimation which also won the sentence-level direct assessment quality estimation shared task in [WMT 2020](http://www.statmt.org/wmt20/quality-estimation-task.html). TransQuest outperforms current open-source quality estimation frameworks such as [OpenKiwi](https://github.com/Unbabel/OpenKiwi) and [DeepQuest](https://github.com/sheffieldnlp/deepQuest). ## Features - Sentence-level translation quality estimation on both aspects: predicting post editing efforts and direct assessment. - Word-level translation quality estimation capable of predicting quality of source words, target words and target gaps. - Outperform current state-of-the-art quality estimation methods like DeepQuest and OpenKiwi in all the languages experimented. - Pre-trained quality estimation models for fifteen language pairs are available in [HuggingFace.](https://huggingface.co/TransQuest) ## Installation ### From pip ```bash pip install transquest ``` ### From Source ```bash git clone https://github.com/TharinduDR/TransQuest.git cd TransQuest pip install -r requirements.txt ``` ## Using Pre-trained Models ```python import torch from transquest.algo.sentence_level.monotransquest.run_model import MonoTransQuestModel model = MonoTransQuestModel("xlmroberta", "TransQuest/monotransquest-hter-en_de-wiki", num_labels=1, use_cuda=torch.cuda.is_available()) predictions, raw_outputs = model.predict([["Reducerea acestor conflicte este importantă pentru conservare.", "Reducing these conflicts is not important for preservation."]]) print(predictions) ``` ## Documentation For more details follow the documentation. 1. **[Installation](https://tharindudr.github.io/TransQuest/install/)** - Install TransQuest locally using pip. 2. **Architectures** - Checkout the architectures implemented in TransQuest 1. [Sentence-level Architectures](https://tharindudr.github.io/TransQuest/architectures/sentence_level_architectures/) - We have released two architectures; MonoTransQuest and SiameseTransQuest to perform sentence level quality estimation. 2. [Word-level Architecture](https://tharindudr.github.io/TransQuest/architectures/word_level_architecture/) - We have released MicroTransQuest to perform word level quality estimation. 3. **Examples** - We have provided several examples on how to use TransQuest in recent WMT quality estimation shared tasks. 1. [Sentence-level Examples](https://tharindudr.github.io/TransQuest/examples/sentence_level_examples/) 2. [Word-level Examples](https://tharindudr.github.io/TransQuest/examples/word_level_examples/) 4. **Pre-trained Models** - We have provided pretrained quality estimation models for fifteen language pairs covering both sentence-level and word-level 1. [Sentence-level Models](https://tharindudr.github.io/TransQuest/models/sentence_level_pretrained/) 2. [Word-level Models](https://tharindudr.github.io/TransQuest/models/word_level_pretrained/) 5. **[Contact](https://tharindudr.github.io/TransQuest/contact/)** - Contact us for any issues with TransQuest ## Citations If you are using the word-level architecture, please consider citing this paper which is accepted to [ACL 2021](https://2021.aclweb.org/). ```bash @InProceedings{ranasinghe2021, author = {Ranasinghe, Tharindu and Orasan, Constantin and Mitkov, Ruslan}, title = {An Exploratory Analysis of Multilingual Word Level Quality Estimation with Cross-Lingual Transformers}, booktitle = {Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics}, year = {2021} } ``` If you are using the sentence-level architectures, please consider citing these papers which were presented in [COLING 2020](https://coling2020.org/) and in [WMT 2020](http://www.statmt.org/wmt20/) at EMNLP 2020. ```bash @InProceedings{transquest:2020a, author = {Ranasinghe, Tharindu and Orasan, Constantin and Mitkov, Ruslan}, title = {TransQuest: Translation Quality Estimation with Cross-lingual Transformers}, booktitle = {Proceedings of the 28th International Conference on Computational Linguistics}, year = {2020} } ``` ```bash @InProceedings{transquest:2020b, author = {Ranasinghe, Tharindu and Orasan, Constantin and Mitkov, Ruslan}, title = {TransQuest at WMT2020: Sentence-Level Direct Assessment}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, year = {2020} } ```
5,405
Vassilis/distilbert-base-uncased-finetuned-emotion
[ "LABEL_0", "LABEL_1", "LABEL_2", "LABEL_3", "LABEL_4", "LABEL_5" ]
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: distilbert-base-uncased-finetuned-emotion results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1628 - Accuracy: 0.9345 - F1: 0.9348 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.1674 | 1.0 | 250 | 0.1718 | 0.9265 | 0.9266 | | 0.1091 | 2.0 | 500 | 0.1628 | 0.9345 | 0.9348 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0 - Tokenizers 0.10.3
1,481
Yuri/xlm-roberta-base-finetuned-marc
[ "good", "great", "ok", "poor", "terrible" ]
--- license: mit tags: - generated_from_trainer datasets: - amazon_reviews_multi model-index: - name: xlm-roberta-base-finetuned-marc results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-marc This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.9825 - Mae: 0.4956 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Mae | |:-------------:|:-----:|:----:|:---------------:|:------:| | 1.1432 | 1.0 | 308 | 1.0559 | 0.5133 | | 0.9883 | 2.0 | 616 | 0.9825 | 0.4956 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
1,423
aXhyra/demo_emotion_42
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - tweet_eval metrics: - f1 model-index: - name: demo_emotion_42 results: - task: name: Text Classification type: text-classification dataset: name: tweet_eval type: tweet_eval args: emotion metrics: - name: F1 type: f1 value: 0.7348035780583043 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # demo_emotion_42 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the tweet_eval dataset. It achieves the following results on the evaluation set: - Loss: 0.9818 - F1: 0.7348 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 7.551070618629693e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 0 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 1.0 | 204 | 0.7431 | 0.6530 | | No log | 2.0 | 408 | 0.6943 | 0.7333 | | 0.5176 | 3.0 | 612 | 0.8456 | 0.7326 | | 0.5176 | 4.0 | 816 | 0.9818 | 0.7348 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.9.1 - Datasets 1.16.1 - Tokenizers 0.10.3
1,759
aXhyra/demo_irony_42
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - tweet_eval metrics: - f1 model-index: - name: demo_irony_42 results: - task: name: Text Classification type: text-classification dataset: name: tweet_eval type: tweet_eval args: irony metrics: - name: F1 type: f1 value: 0.685764300192161 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # demo_irony_42 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the tweet_eval dataset. It achieves the following results on the evaluation set: - Loss: 1.2905 - F1: 0.6858 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2.7735294032820418e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 0 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 1.0 | 358 | 0.5872 | 0.6786 | | 0.5869 | 2.0 | 716 | 0.6884 | 0.6952 | | 0.3417 | 3.0 | 1074 | 0.9824 | 0.6995 | | 0.3417 | 4.0 | 1432 | 1.2905 | 0.6858 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.9.1 - Datasets 1.16.1 - Tokenizers 0.10.3
1,751
aXhyra/emotion_trained_1234567
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - tweet_eval metrics: - f1 model-index: - name: emotion_trained_1234567 results: - task: name: Text Classification type: text-classification dataset: name: tweet_eval type: tweet_eval args: emotion metrics: - name: F1 type: f1 value: 0.7301562209701973 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # emotion_trained_1234567 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the tweet_eval dataset. It achieves the following results on the evaluation set: - Loss: 0.9051 - F1: 0.7302 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 6.961635072722524e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 1234567 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 1.0 | 204 | 0.6480 | 0.7231 | | No log | 2.0 | 408 | 0.6114 | 0.7403 | | 0.5045 | 3.0 | 612 | 0.7592 | 0.7311 | | 0.5045 | 4.0 | 816 | 0.9051 | 0.7302 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.9.1 - Datasets 1.16.1 - Tokenizers 0.10.3
1,781
aXhyra/emotion_trained_42
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - tweet_eval metrics: - f1 model-index: - name: emotion_trained_42 results: - task: name: Text Classification type: text-classification dataset: name: tweet_eval type: tweet_eval args: emotion metrics: - name: F1 type: f1 value: 0.7361210540311689 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # emotion_trained_42 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the tweet_eval dataset. It achieves the following results on the evaluation set: - Loss: 0.9012 - F1: 0.7361 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 6.961635072722524e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 1.0 | 204 | 0.6131 | 0.6955 | | No log | 2.0 | 408 | 0.5816 | 0.7297 | | 0.5148 | 3.0 | 612 | 0.8942 | 0.7199 | | 0.5148 | 4.0 | 816 | 0.9012 | 0.7361 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.9.1 - Datasets 1.16.1 - Tokenizers 0.10.3
1,766
aXhyra/hate_trained_1234567
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - tweet_eval metrics: - f1 model-index: - name: hate_trained_1234567 results: - task: name: Text Classification type: text-classification dataset: name: tweet_eval type: tweet_eval args: hate metrics: - name: F1 type: f1 value: 0.7750768993843997 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # hate_trained_1234567 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the tweet_eval dataset. It achieves the following results on the evaluation set: - Loss: 0.7912 - F1: 0.7751 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2.7272339744854407e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 1234567 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.4835 | 1.0 | 563 | 0.4881 | 0.7534 | | 0.3236 | 2.0 | 1126 | 0.5294 | 0.7610 | | 0.219 | 3.0 | 1689 | 0.6095 | 0.7717 | | 0.1409 | 4.0 | 2252 | 0.7912 | 0.7751 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.9.1 - Datasets 1.16.1 - Tokenizers 0.10.3
1,773
aXhyra/hate_trained_31415
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - tweet_eval metrics: - f1 model-index: - name: hate_trained_31415 results: - task: name: Text Classification type: text-classification dataset: name: tweet_eval type: tweet_eval args: hate metrics: - name: F1 type: f1 value: 0.7729447444817463 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # hate_trained_31415 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the tweet_eval dataset. It achieves the following results on the evaluation set: - Loss: 0.8568 - F1: 0.7729 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2.7272339744854407e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 31415 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.482 | 1.0 | 563 | 0.4973 | 0.7672 | | 0.3316 | 2.0 | 1126 | 0.4931 | 0.7794 | | 0.2308 | 3.0 | 1689 | 0.7073 | 0.7593 | | 0.1444 | 4.0 | 2252 | 0.8568 | 0.7729 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.9.1 - Datasets 1.16.1 - Tokenizers 0.10.3
1,767
aXhyra/hate_trained_42
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - tweet_eval metrics: - f1 model-index: - name: hate_trained_42 results: - task: name: Text Classification type: text-classification dataset: name: tweet_eval type: tweet_eval args: hate metrics: - name: F1 type: f1 value: 0.7712319060633668 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # hate_trained_42 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the tweet_eval dataset. It achieves the following results on the evaluation set: - Loss: 0.8994 - F1: 0.7712 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2.7272339744854407e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.4835 | 1.0 | 563 | 0.4855 | 0.7556 | | 0.3277 | 2.0 | 1126 | 0.5354 | 0.7704 | | 0.2112 | 3.0 | 1689 | 0.6870 | 0.7751 | | 0.1384 | 4.0 | 2252 | 0.8994 | 0.7712 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.9.1 - Datasets 1.16.1 - Tokenizers 0.10.3
1,758
aXhyra/test_irony_trained_test
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - tweet_eval metrics: - f1 model-index: - name: test_irony_trained_test results: - task: name: Text Classification type: text-classification dataset: name: tweet_eval type: tweet_eval args: irony metrics: - name: F1 type: f1 value: 0.6680395323922843 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # test_irony_trained_test This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the tweet_eval dataset. It achieves the following results on the evaluation set: - Loss: 0.7674 - F1: 0.6680 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 9.207906329883037e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 0 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 1.0 | 358 | 0.6655 | 0.5924 | | 0.684 | 2.0 | 716 | 0.6889 | 0.6024 | | 0.5826 | 3.0 | 1074 | 0.7085 | 0.6488 | | 0.5826 | 4.0 | 1432 | 0.7674 | 0.6680 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.9.1 - Datasets 1.16.1 - Tokenizers 0.10.3
1,771
aarnphm/finetune_emotion_distilroberta
[ "anger", "fear", "joy", "love", "sadness", "surprise" ]
Entry not found
15
abhishek/autonlp-fred2-2682064
[ "0", "1" ]
--- tags: autonlp language: en widget: - text: "I love AutoNLP 🤗" datasets: - abhishek/autonlp-data-fred2 --- # Model Trained Using AutoNLP - Problem type: Binary Classification - Model ID: 2682064 ## Validation Metrics - Loss: 0.4454168379306793 - Accuracy: 0.8188976377952756 - Precision: 0.8442028985507246 - Recall: 0.7103658536585366 - AUC: 0.8699702146791053 - F1: 0.771523178807947 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/abhishek/autonlp-fred2-2682064 ``` Or Python API: ``` from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("abhishek/autonlp-fred2-2682064", use_auth_token=True) tokenizer = AutoTokenizer.from_pretrained("abhishek/autonlp-fred2-2682064", use_auth_token=True) inputs = tokenizer("I love AutoNLP", return_tensors="pt") outputs = model(**inputs) ```
1,050
adamlin/ml999_wood
[ "0", "1" ]
Entry not found
15
adamlin/text-cls
[ "体育", "时尚", "科技", "社会", "股票", "星座", "娱乐", "教育", "财经", "时政", "彩票", "家居", "房产", "游戏" ]
Entry not found
15
aditeyabaral/finetuned-iitp_pdt_review-distilbert-hinglish-small
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
aditeyabaral/finetuned-iitp_pdt_review-roberta-base
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
aditeyabaral/finetuned-iitpmovie-additionalpretrained-bert-base-cased
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
aditeyabaral/finetuned-sail2017-additionalpretrained-xlm-roberta-base
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
aditeyabaral/finetuned-sail2017-distilbert-base-cased
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
aditeyabaral/finetuned-sail2017-roberta-base
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
airKlizz/xlm-roberta-base-germeval21-toxic
null
Entry not found
15
anel/autonlp-cml-412010597
[ "misleading", "news" ]
--- tags: autonlp language: en widget: - text: "I love AutoNLP 🤗" datasets: - anel/autonlp-data-cml co2_eq_emissions: 10.411685187181709 --- # Model Trained Using AutoNLP - Problem type: Binary Classification - Model ID: 412010597 - CO2 Emissions (in grams): 10.411685187181709 ## Validation Metrics - Loss: 0.12585781514644623 - Accuracy: 0.9475446428571429 - Precision: 0.9454660748256183 - Recall: 0.964424320827943 - AUC: 0.990229573862156 - F1: 0.9548511047070125 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/anel/autonlp-cml-412010597 ``` Or Python API: ``` from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("anel/autonlp-cml-412010597", use_auth_token=True) tokenizer = AutoTokenizer.from_pretrained("anel/autonlp-cml-412010597", use_auth_token=True) inputs = tokenizer("I love AutoNLP", return_tensors="pt") outputs = model(**inputs) ```
1,118
anirudh21/albert-large-v2-finetuned-cola
null
Entry not found
15
anirudh21/albert-large-v2-finetuned-mrpc
null
Entry not found
15
anirudh21/distilbert-base-uncased-finetuned-cola
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - glue metrics: - matthews_correlation model-index: - name: distilbert-base-uncased-finetuned-cola results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue args: cola metrics: - name: Matthews Correlation type: matthews_correlation value: 0.5224154837835395 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.8623 - Matthews Correlation: 0.5224 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5278 | 1.0 | 535 | 0.5223 | 0.4007 | | 0.3515 | 2.0 | 1070 | 0.5150 | 0.4993 | | 0.2391 | 3.0 | 1605 | 0.6471 | 0.5103 | | 0.1841 | 4.0 | 2140 | 0.7640 | 0.5153 | | 0.1312 | 5.0 | 2675 | 0.8623 | 0.5224 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
2,000
anirudh21/distilbert-base-uncased-finetuned-sst2
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - glue metrics: - accuracy model-index: - name: distilbert-base-uncased-finetuned-sst2 results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue args: sst2 metrics: - name: Accuracy type: accuracy value: 0.908256880733945 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-sst2 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.4028 - Accuracy: 0.9083 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.188 | 1.0 | 4210 | 0.3127 | 0.9037 | | 0.1299 | 2.0 | 8420 | 0.3887 | 0.9048 | | 0.0845 | 3.0 | 12630 | 0.4028 | 0.9083 | | 0.0691 | 4.0 | 16840 | 0.3924 | 0.9071 | | 0.052 | 5.0 | 21050 | 0.5047 | 0.9002 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
1,874
anirudh21/distilbert-base-uncased-finetuned-wnli
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - glue metrics: - accuracy model-index: - name: distilbert-base-uncased-finetuned-wnli results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue args: wnli metrics: - name: Accuracy type: accuracy value: 0.5633802816901409 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-wnli This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.6883 - Accuracy: 0.5634 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 40 | 0.6883 | 0.5634 | | No log | 2.0 | 80 | 0.6934 | 0.5634 | | No log | 3.0 | 120 | 0.6960 | 0.5211 | | No log | 4.0 | 160 | 0.6958 | 0.5634 | | No log | 5.0 | 200 | 0.6964 | 0.5634 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
1,868
anthonymirand/haha_2019_primary_task
null
Entry not found
15
arianpasquali/distilbert-base-uncased-finetuned-clinc
[ "accept_reservations", "account_blocked", "alarm", "application_status", "apr", "are_you_a_bot", "balance", "bill_balance", "bill_due", "book_flight", "book_hotel", "calculator", "calendar", "calendar_update", "calories", "cancel", "cancel_reservation", "car_rental", "card_declin...
--- license: apache-2.0 tags: - generated_from_trainer datasets: - clinc_oos metrics: - accuracy model-index: - name: distilbert-base-uncased-finetuned-clinc results: - task: name: Text Classification type: text-classification dataset: name: clinc_oos type: clinc_oos args: plus metrics: - name: Accuracy type: accuracy value: 0.9112903225806451 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-clinc This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the clinc_oos dataset. It achieves the following results on the evaluation set: - Loss: 0.7751 - Accuracy: 0.9113 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 48 - eval_batch_size: 48 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 4.315 | 1.0 | 318 | 3.3087 | 0.74 | | 2.6371 | 2.0 | 636 | 1.8833 | 0.8381 | | 1.5388 | 3.0 | 954 | 1.1547 | 0.8929 | | 1.0076 | 4.0 | 1272 | 0.8590 | 0.9071 | | 0.79 | 5.0 | 1590 | 0.7751 | 0.9113 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1 - Datasets 1.16.1 - Tokenizers 0.10.3
1,883
aristotletan/sc-distilbert
[ "LABEL_0", "LABEL_1", "LABEL_10", "LABEL_2", "LABEL_3", "LABEL_4", "LABEL_5", "LABEL_6", "LABEL_7", "LABEL_8", "LABEL_9" ]
Entry not found
15
aristotletan/scim-distillbert
[]
Entry not found
15
athar/distilbert-base-uncased-finetuned-cola
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - glue metrics: - matthews_correlation model-index: - name: distilbert-base-uncased-finetuned-cola results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue args: cola metrics: - name: Matthews Correlation type: matthews_correlation value: 0.5451837431775948 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.8508 - Matthews Correlation: 0.5452 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5221 | 1.0 | 535 | 0.5370 | 0.4246 | | 0.3462 | 2.0 | 1070 | 0.5157 | 0.5183 | | 0.2332 | 3.0 | 1605 | 0.6324 | 0.5166 | | 0.1661 | 4.0 | 2140 | 0.7616 | 0.5370 | | 0.1263 | 5.0 | 2675 | 0.8508 | 0.5452 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.13.0 - Tokenizers 0.10.3
1,999
auychai/distilbert-base-uncased-finetuned-emotion
[ "LABEL_0", "LABEL_1", "LABEL_2", "LABEL_3", "LABEL_4", "LABEL_5" ]
Entry not found
15
benjaminbeilharz/bert-base-uncased-next-turn-classifier
[ "LABEL_0", "LABEL_1", "LABEL_2", "LABEL_3", "LABEL_4" ]
Entry not found
15
beomi/beep-kcbert-base-hate
[ "hate", "none", "offensive" ]
Entry not found
15
bestvater/distilbert-kav-stance
null
Entry not found
15
bierus/distilbert_bookreviews
[ "LABEL_0", "LABEL_1", "LABEL_2" ]
Entry not found
15
claudio75/xlm-roberta-base-finetuned-marc
[ "good", "great", "ok", "poor", "terrible" ]
--- license: mit tags: - generated_from_trainer datasets: - amazon_reviews_multi model-index: - name: xlm-roberta-base-finetuned-marc results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-marc This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.9611 - Mae: 0.4749 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Mae | |:-------------:|:-----:|:----:|:---------------:|:------:| | 1.0431 | 1.0 | 860 | 0.9819 | 0.4985 | | 0.9079 | 2.0 | 1720 | 0.9611 | 0.4749 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
1,423
csalamea/roberta-base-bne-finetuned-amazon_reviews_multi
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - amazon_reviews_multi metrics: - accuracy model-index: - name: roberta-base-bne-finetuned-amazon_reviews_multi results: - task: name: Text Classification type: text-classification dataset: name: amazon_reviews_multi type: amazon_reviews_multi args: es metrics: - name: Accuracy type: accuracy value: 0.9325 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-bne-finetuned-amazon_reviews_multi This model is a fine-tuned version of [BSC-TeMU/roberta-base-bne](https://huggingface.co/BSC-TeMU/roberta-base-bne) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.2303 - Accuracy: 0.9325 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.1942 | 1.0 | 1250 | 0.1751 | 0.932 | | 0.0935 | 2.0 | 2500 | 0.2303 | 0.9325 | ### Framework versions - Transformers 4.10.2 - Pytorch 1.9.0+cu102 - Datasets 1.12.1 - Tokenizers 0.10.3
1,753
damlab/HIV_V3_bodysite
[ "CNS", "breast-milk", "female-genitals", "gastric", "lung", "male-genitals", "organ", "periphery-monocyte", "periphery-tcell" ]
--- licence: mit widget: - text: "T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C" example_title: "V3 Macrophage" - text: 'C T R P N N N T R K S I H I G P G R A F Y T T G Q I I G D I R Q A Y C' example_title: "V3 T-cell" datasets: - damlab/HIV_V3_bodysite metrics: - accuracy --- # Model Card for [HIV_V3_bodysite] ## Table of Contents - [Table of Contents](#table-of-contents) - [Summary](#model-summary) - [Model Description](#model-description) - [Intended Uses & Limitations](#intended-uses-&-limitations) - [How to Use](#how-to-use) - [Training Data](#training-data) - [Training Procedure](#training-procedure) - [Preprocessing](#preprocessing) - [Training](#training) - [Evaluation Results](#evaluation-results) - [BibTeX Entry and Citation Info](#bibtex-entry-and-citation-info) ## Summary The HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (https://huggingface.co/Rostlab/prot_bert_bfd) to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html), allowing even more precise prediction of body site location than the HIV-BERT model can provide. ## Model Description The HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from. ## Intended Uses & Limitations This tool can be used as a predictor of which body site an HIV sample was derived from based on its genomic sequence. It should not be considered a clinical diagnostic tool. This tool was trained using the Los Alamos HIV sequence dataset (https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html). Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences. ## How to use This model is able to predict the likely bodysite from a V3 sequence. This may be use for surveillance of cells that are emerging from latent reservoirs. Remember, a sequence can come from multiple sites, they are not mutually exclusive. ```python from transformers import pipeline predictor = pipeline("text-classification", model="damlab/HIV_V3_bodysite") predictor(f"C T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C") [ [ { "label": "periphery-tcell", "score": 0.29097115993499756 }, { "label": "periphery-monocyte", "score": 0.014322502538561821 }, { "label": "CNS", "score": 0.06870711594820023 }, { "label": "breast-milk", "score": 0.002785981632769108 }, { "label": "female-genitals", "score": 0.024997007101774216 }, { "label": "male-genitals", "score": 0.01040483545511961 }, { "label": "gastric", "score": 0.06872137635946274 }, { "label": "lung", "score": 0.04432062804698944 }, { "label": "organ", "score": 0.47476938366889954 } ] ] ``` ## Training Data This model was trained using the damlab/HIV_V3_bodysite dataset using the 0th fold. The dataset consists of 5510 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database. ## Training Procedure ### Preprocessing As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation. ### Training The damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be found in multiple sites) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance. ## Evaluation Results *Need to add* ## BibTeX Entry and Citation Info [More Information Needed]
5,229
danlou/distilbert-base-uncased-finetuned-cola
null
Entry not found
15
dhikri/question_answering_glue
null
"hello"
9
diegozs97/finetuned-chemprot-seed-0-1000k
[ "CPR:3", "CPR:4", "CPR:5", "CPR:6", "CPR:9", "false" ]
Entry not found
15
diegozs97/finetuned-chemprot-seed-1-1000k
[ "CPR:3", "CPR:4", "CPR:5", "CPR:6", "CPR:9", "false" ]
Entry not found
15
diegozs97/finetuned-chemprot-seed-1-100k
[ "CPR:3", "CPR:4", "CPR:5", "CPR:6", "CPR:9", "false" ]
Entry not found
15
diegozs97/finetuned-chemprot-seed-3-0k
[ "CPR:3", "CPR:4", "CPR:5", "CPR:6", "CPR:9", "false" ]
Entry not found
15
diegozs97/finetuned-chemprot-seed-3-100k
[ "CPR:3", "CPR:4", "CPR:5", "CPR:6", "CPR:9", "false" ]
Entry not found
15
diegozs97/finetuned-chemprot-seed-3-2000k
[ "CPR:3", "CPR:4", "CPR:5", "CPR:6", "CPR:9", "false" ]
Entry not found
15
diegozs97/finetuned-chemprot-seed-4-200k
[ "CPR:3", "CPR:4", "CPR:5", "CPR:6", "CPR:9", "false" ]
Entry not found
15
diegozs97/finetuned-chemprot-seed-4-400k
[ "CPR:3", "CPR:4", "CPR:5", "CPR:6", "CPR:9", "false" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-0-1000k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-0-2000k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-0-20k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-0-400k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-0-60k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-1-1000k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-1-100k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-1-1800k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-1-400k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-1-60k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-1-700k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-2-1000k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-2-2000k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-2-60k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-3-0k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-3-100k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-3-1800k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-3-400k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-4-0k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-4-1000k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-4-1500k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-4-1800k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-4-200k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-4-400k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
diegozs97/finetuned-sciie-seed-4-700k
[ "COMPARE", "CONJUNCTION", "EVALUATE-FOR", "FEATURE-OF", "HYPONYM-OF", "PART-OF", "USED-FOR" ]
Entry not found
15
ds198799/autonlp-predict_ROI_1-29797722
[ "1.0", "2.0", "3.0" ]
--- tags: autonlp language: en widget: - text: "I love AutoNLP 🤗" datasets: - ds198799/autonlp-data-predict_ROI_1 co2_eq_emissions: 2.7516207978192737 --- # Model Trained Using AutoNLP - Problem type: Multi-class Classification - Model ID: 29797722 - CO2 Emissions (in grams): 2.7516207978192737 ## Validation Metrics - Loss: 0.6113826036453247 - Accuracy: 0.7559139784946236 - Macro F1: 0.4594734612976928 - Micro F1: 0.7559139784946236 - Weighted F1: 0.7195080232106192 - Macro Precision: 0.7175166413412577 - Micro Precision: 0.7559139784946236 - Weighted Precision: 0.7383048259333735 - Macro Recall: 0.4482203645846237 - Micro Recall: 0.7559139784946236 - Weighted Recall: 0.7559139784946236 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/ds198799/autonlp-predict_ROI_1-29797722 ``` Or Python API: ``` from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("ds198799/autonlp-predict_ROI_1-29797722", use_auth_token=True) tokenizer = AutoTokenizer.from_pretrained("ds198799/autonlp-predict_ROI_1-29797722", use_auth_token=True) inputs = tokenizer("I love AutoNLP", return_tensors="pt") outputs = model(**inputs) ```
1,386
echarlaix/bert-base-dynamic-quant-test
null
Entry not found
15
edmihranyan/roberta_large_classifier
[ "LABEL_0", "LABEL_1", "LABEL_2", "LABEL_3" ]
Entry not found
15
edwardgowsmith/xlnet-base-cased-best
null
Entry not found
15
edwardgowsmith/xlnet-base-cased-train-from-dev-and-test-best
null
Entry not found
15