license
stringlengths
2
30
tags
stringlengths
2
513
is_nc
bool
1 class
readme_section
stringlengths
201
597k
hash
stringlengths
32
32
creativeml-openrail-m
['stable-diffusion', 'text-to-image']
false
Usage To use this model you have to download the file aswell as drop it into the "\stable-diffusion-webui\models\Stable-diffusion" folder Token: ```neko``` If it is to strong just add [] around it. Trained until 10000 steps Have fun :)
d1f9554ae91bf137be085556bacee3c9
creativeml-openrail-m
['stable-diffusion', 'text-to-image']
false
Example Pictures <table> <tr> <td><img src=https://i.imgur.com/MpyeqMe.png width=100% height=100%/></td> <td><img src=https://i.imgur.com/wxzvHrL.png width=100% height=100%/></td> <td><img src=https://i.imgur.com/MuUnJY5.png width=100% height=100%/></td> <td><img src=https://i.imgur.com/XeDC8xA.png width=100% height=100%/></td> <td><img src=https://i.imgur.com/XmLTrEl.png width=100% height=100%/></td> </tr> </table>
7863ddc1c98f939b5f28cb3dc92673dc
mit
[]
false
LONGFORMER-BASE-4096 fine-tuned on SQuAD v1 This is longformer-base-4096 model fine-tuned on SQuAD v1 dataset for question answering task. [Longformer](https://arxiv.org/abs/2004.05150) model created by Iz Beltagy, Matthew E. Peters, Arman Coha from AllenAI. As the paper explains it > `Longformer` is a BERT-like model for long documents. The pre-trained model can handle sequences with upto 4096 tokens.
a50fce6f7cbdd5178359d3c157c8128e
mit
[]
false
Model Training This model was trained on google colab v100 GPU. You can find the fine-tuning colab here [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1zEl5D-DdkBKva-DdreVOmN0hrAfzKG1o?usp=sharing). Few things to keep in mind while training longformer for QA task, by default longformer uses sliding-window local attention on all tokens. But For QA, all question tokens should have global attention. For more details on this please refer the paper. The `LongformerForQuestionAnswering` model automatically does that for you. To allow it to do that 1. The input sequence must have three sep tokens, i.e the sequence should be encoded like this ` <s> question</s></s> context</s>`. If you encode the question and answer as a input pair, then the tokenizer already takes care of that, you shouldn't worry about it. 2. `input_ids` should always be a batch of examples.
e1674c19c020f7506de5d9cabf471d5c
mit
[]
false
Model in Action 🚀 ```python import torch from transformers import AutoTokenizer, AutoModelForQuestionAnswering, tokenizer = AutoTokenizer.from_pretrained("valhalla/longformer-base-4096-finetuned-squadv1") model = AutoModelForQuestionAnswering.from_pretrained("valhalla/longformer-base-4096-finetuned-squadv1") text = "Huggingface has democratized NLP. Huge thanks to Huggingface for this." question = "What has Huggingface done ?" encoding = tokenizer(question, text, return_tensors="pt") input_ids = encoding["input_ids"]
1e3c0c4c5fb9a1bfd9c43c9194e24054
mit
[]
false
the forward method will automatically set global attention on question tokens attention_mask = encoding["attention_mask"] start_scores, end_scores = model(input_ids, attention_mask=attention_mask) all_tokens = tokenizer.convert_ids_to_tokens(input_ids[0].tolist()) answer_tokens = all_tokens[torch.argmax(start_scores) :torch.argmax(end_scores)+1] answer = tokenizer.decode(tokenizer.convert_tokens_to_ids(answer_tokens))
192e483f16677c2d70ab6082d17ff995
mit
[]
false
output => democratized NLP ``` The `LongformerForQuestionAnswering` isn't yet supported in `pipeline` . I'll update this card once the support has been added. > Created with ❤️ by Suraj Patil [![Github icon](https://cdn0.iconfinder.com/data/icons/octicons/1024/mark-github-32.png)](https://github.com/patil-suraj/) [![Twitter icon](https://cdn0.iconfinder.com/data/icons/shift-logotypes/32/Twitter-32.png)](https://twitter.com/psuraj28)
31ddc53af307dc7f411032757f49a5e4
apache-2.0
['generated_from_trainer']
false
mobilebert_sa_GLUE_Experiment_logit_kd_rte_128 This model is a fine-tuned version of [google/mobilebert-uncased](https://huggingface.co/google/mobilebert-uncased) on the GLUE RTE dataset. It achieves the following results on the evaluation set: - Loss: 0.3915 - Accuracy: 0.5271
72d9bee2516c6d4c430a41f57c5b9451
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.4093 | 1.0 | 20 | 0.3917 | 0.5271 | | 0.4077 | 2.0 | 40 | 0.3922 | 0.5271 | | 0.4076 | 3.0 | 60 | 0.3916 | 0.5271 | | 0.4075 | 4.0 | 80 | 0.3921 | 0.5271 | | 0.4075 | 5.0 | 100 | 0.3925 | 0.5271 | | 0.4073 | 6.0 | 120 | 0.3915 | 0.5271 | | 0.4066 | 7.0 | 140 | 0.3916 | 0.5271 | | 0.4043 | 8.0 | 160 | 0.3937 | 0.5271 | | 0.3902 | 9.0 | 180 | 0.4440 | 0.5054 | | 0.3545 | 10.0 | 200 | 0.4575 | 0.4801 | | 0.3116 | 11.0 | 220 | 0.4770 | 0.4440 |
a9b075e5c8eeb087ef8b6d86367f3d38
mit
['generated_from_trainer']
false
roberta-base.CEBaB_confounding.uniform.absa.5-class.seed_43 This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the OpenTable OPENTABLE-ABSA dataset. It achieves the following results on the evaluation set: - Loss: 0.3790 - Accuracy: 0.8913 - Macro-f1: 0.8893 - Weighted-macro-f1: 0.8914
bb80a68f2759062528452fad8d71b6c5
apache-2.0
['translation']
false
phi-eng * source group: Philippine languages * target group: English * OPUS readme: [phi-eng](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/phi-eng/README.md) * model: transformer * source language(s): akl_Latn ceb hil ilo pag war * target language(s): eng * model: transformer * pre-processing: normalization + SentencePiece (spm12k,spm12k) * download original weights: [opus2m-2020-08-01.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/phi-eng/opus2m-2020-08-01.zip) * test set translations: [opus2m-2020-08-01.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/phi-eng/opus2m-2020-08-01.test.txt) * test set scores: [opus2m-2020-08-01.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/phi-eng/opus2m-2020-08-01.eval.txt)
5b2c14d0e7257c0fb1404134ccce8fbd
apache-2.0
['translation']
false
Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.akl-eng.akl.eng | 11.6 | 0.321 | | Tatoeba-test.ceb-eng.ceb.eng | 21.7 | 0.393 | | Tatoeba-test.hil-eng.hil.eng | 17.6 | 0.371 | | Tatoeba-test.ilo-eng.ilo.eng | 36.6 | 0.560 | | Tatoeba-test.multi.eng | 21.5 | 0.391 | | Tatoeba-test.pag-eng.pag.eng | 27.5 | 0.494 | | Tatoeba-test.war-eng.war.eng | 17.3 | 0.380 |
8d996abd191d9f9f1fcf8b449b773994
apache-2.0
['translation']
false
System Info: - hf_name: phi-eng - source_languages: phi - target_languages: eng - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/phi-eng/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['phi', 'en'] - src_constituents: {'ilo', 'akl_Latn', 'war', 'hil', 'pag', 'ceb'} - tgt_constituents: {'eng'} - src_multilingual: True - tgt_multilingual: False - prepro: normalization + SentencePiece (spm12k,spm12k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/phi-eng/opus2m-2020-08-01.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/phi-eng/opus2m-2020-08-01.test.txt - src_alpha3: phi - tgt_alpha3: eng - short_pair: phi-en - chrF2_score: 0.391 - bleu: 21.5 - brevity_penalty: 1.0 - ref_len: 2380.0 - src_name: Philippine languages - tgt_name: English - train_date: 2020-08-01 - src_alpha2: phi - tgt_alpha2: en - prefer_old: False - long_pair: phi-eng - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
b458533f66100ad6d3cd9e6050f2161f
mit
['automatic-speech-recognition', 'common_voice', 'generated_from_trainer']
false
wav2vec2-xls-r-300m-uk This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0927 - Wer: 0.1222 - Cer: 0.0204
c574c1a982f64a93ce7bbad01e001c9a
mit
['automatic-speech-recognition', 'common_voice', 'generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 40 - eval_batch_size: 40 - seed: 42 - gradient_accumulation_steps: 6 - total_train_batch_size: 240 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 100 - mixed_precision_training: Native AMP
b345737facb09e501d84d3ac7df0bfd3
mit
['automatic-speech-recognition', 'common_voice', 'generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Cer | Validation Loss | Wer | |:-------------:|:-----:|:----:|:------:|:---------------:|:------:| | 9.0008 | 1.68 | 200 | 1.0 | 3.7590 | 1.0 | | 3.4972 | 3.36 | 400 | 1.0 | 3.3933 | 1.0 | | 3.3432 | 5.04 | 600 | 1.0 | 3.2617 | 1.0 | | 3.2421 | 6.72 | 800 | 1.0 | 3.0712 | 1.0 | | 1.9839 | 7.68 | 1000 | 0.1400 | 0.7204 | 0.6561 | | 0.8017 | 9.36 | 1200 | 0.0766 | 0.3734 | 0.4159 | | 0.5554 | 11.04 | 1400 | 0.0583 | 0.2621 | 0.3237 | | 0.4309 | 12.68 | 1600 | 0.0486 | 0.2085 | 0.2753 | | 0.3697 | 14.36 | 1800 | 0.0421 | 0.1746 | 0.2427 | | 0.3293 | 16.04 | 2000 | 0.0388 | 0.1597 | 0.2243 | | 0.2934 | 17.72 | 2200 | 0.0358 | 0.1428 | 0.2083 | | 0.2704 | 19.4 | 2400 | 0.0333 | 0.1326 | 0.1949 | | 0.2547 | 21.08 | 2600 | 0.0322 | 0.1255 | 0.1882 | | 0.2366 | 22.76 | 2800 | 0.0309 | 0.1211 | 0.1815 | | 0.2183 | 24.44 | 3000 | 0.0294 | 0.1159 | 0.1727 | | 0.2115 | 26.13 | 3200 | 0.0280 | 0.1117 | 0.1661 | | 0.1968 | 27.8 | 3400 | 0.0274 | 0.1063 | 0.1622 | | 0.1922 | 29.48 | 3600 | 0.0269 | 0.1082 | 0.1598 | | 0.1847 | 31.17 | 3800 | 0.0260 | 0.1061 | 0.1550 | | 0.1715 | 32.84 | 4000 | 0.0252 | 0.1014 | 0.1496 | | 0.1689 | 34.53 | 4200 | 0.0250 | 0.1012 | 0.1492 | | 0.1655 | 36.21 | 4400 | 0.0243 | 0.0999 | 0.1450 | | 0.1585 | 37.88 | 4600 | 0.0239 | 0.0967 | 0.1432 | | 0.1492 | 39.57 | 4800 | 0.0237 | 0.0978 | 0.1421 | | 0.1491 | 41.25 | 5000 | 0.0236 | 0.0963 | 0.1412 | | 0.1453 | 42.93 | 5200 | 0.0230 | 0.0979 | 0.1373 | | 0.1386 | 44.61 | 5400 | 0.0227 | 0.0959 | 0.1353 | | 0.1387 | 46.29 | 5600 | 0.0226 | 0.0927 | 0.1355 | | 0.1329 | 47.97 | 5800 | 0.0224 | 0.0951 | 0.1341 | | 0.1295 | 49.65 | 6000 | 0.0219 | 0.0950 | 0.1306 | | 0.1287 | 51.33 | 6200 | 0.0216 | 0.0937 | 0.1290 | | 0.1277 | 53.02 | 6400 | 0.0215 | 0.0963 | 0.1294 | | 0.1201 | 54.69 | 6600 | 0.0213 | 0.0959 | 0.1282 | | 0.1199 | 56.38 | 6800 | 0.0215 | 0.0944 | 0.1286 | | 0.1221 | 58.06 | 7000 | 0.0209 | 0.0938 | 0.1249 | | 0.1145 | 59.68 | 7200 | 0.0208 | 0.0941 | 0.1254 | | 0.1143 | 61.36 | 7400 | 0.0209 | 0.0941 | 0.1249 | | 0.1143 | 63.04 | 7600 | 0.0209 | 0.0940 | 0.1248 | | 0.1137 | 64.72 | 7800 | 0.0205 | 0.0931 | 0.1234 | | 0.1125 | 66.4 | 8000 | 0.0204 | 0.0927 | 0.1222 |
1a3f48e16949f9294d9fa230beab7b7a
apache-2.0
['generated_from_trainer']
false
distilroberta-base-finetuned-suicide-depression This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.6622 - Accuracy: 0.7158
87fb75b45bdc97dcaf7ffa220e65499e
apache-2.0
['generated_from_trainer']
false
Model description Just a **POC** of a Transformer fine-tuned on [SDCNL](https://github.com/ayaanzhaque/SDCNL) dataset for suicide (label 1) or depression (label 0) detection in tweets. **DO NOT use it in production**
57da9168df13d43afc91f2977f91e1f6
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 214 | 0.6204 | 0.6632 | | No log | 2.0 | 428 | 0.6622 | 0.7158 | | 0.5244 | 3.0 | 642 | 0.7312 | 0.6684 | | 0.5244 | 4.0 | 856 | 0.9711 | 0.7105 | | 0.2876 | 5.0 | 1070 | 1.1620 | 0.7 |
a94841318001d32fb5495b9518f22968
apache-2.0
['generated_from_trainer']
false
tiny-mlm-glue-mnli-target-glue-rte This model is a fine-tuned version of [muhtasham/tiny-mlm-glue-mnli](https://huggingface.co/muhtasham/tiny-mlm-glue-mnli) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.5419 - Accuracy: 0.6137
f6029552b014869aed31f17c9c4c63b5
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6373 | 6.41 | 500 | 0.6751 | 0.5993 | | 0.4271 | 12.82 | 1000 | 0.8148 | 0.6390 | | 0.2621 | 19.23 | 1500 | 0.9962 | 0.6173 | | 0.1589 | 25.64 | 2000 | 1.2448 | 0.6065 | | 0.1002 | 32.05 | 2500 | 1.5419 | 0.6137 |
4f4154c326df7ed99e0daf0f77ba2d60
apache-2.0
['generated_from_trainer']
false
distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2251 - Accuracy: 0.923 - F1: 0.9232
357a4e85dbff759fff614f023dc74e7c
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.8243 | 1.0 | 250 | 0.3183 | 0.906 | 0.9019 | | 0.2543 | 2.0 | 500 | 0.2251 | 0.923 | 0.9232 |
2317afbcfda2bad2386c574bbc27fd94
apache-2.0
['generated_from_trainer']
false
distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2281 - Accuracy: 0.924 - F1: 0.9240
e24e19a3a5409c52999852267b6aa630
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.8687 | 1.0 | 250 | 0.3390 | 0.9015 | 0.8984 | | 0.2645 | 2.0 | 500 | 0.2281 | 0.924 | 0.9240 |
245b6a7f21eeee0d3336f5a64354cd7e
apache-2.0
['generated_from_trainer']
false
medium-mlm-tweet-target-tweet This model is a fine-tuned version of [muhtasham/medium-mlm-tweet](https://huggingface.co/muhtasham/medium-mlm-tweet) on the tweet_eval dataset. It achieves the following results on the evaluation set: - Loss: 1.9066 - Accuracy: 0.7594 - F1: 0.7637
8af73ff35ce11330790f12e36ff1493e
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.4702 | 4.9 | 500 | 0.8711 | 0.7540 | 0.7532 | | 0.0629 | 9.8 | 1000 | 1.2918 | 0.7701 | 0.7668 | | 0.0227 | 14.71 | 1500 | 1.4801 | 0.7727 | 0.7696 | | 0.0181 | 19.61 | 2000 | 1.5118 | 0.7888 | 0.7870 | | 0.0114 | 24.51 | 2500 | 1.6747 | 0.7754 | 0.7745 | | 0.0141 | 29.41 | 3000 | 1.8765 | 0.7674 | 0.7628 | | 0.0177 | 34.31 | 3500 | 1.9066 | 0.7594 | 0.7637 |
ab001e8395e0af13052d6f37ac284245
apache-2.0
['generated_from_trainer']
false
bert-base-uncased-offensive-lm-tapt This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0002
5c395ef6eb29842af2f6a3fd3e237a79
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 8 - eval_batch_size: 28 - seed: 42 - optimizer: Adam with betas=(0.9,0.98) and epsilon=1e-06 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 5 - num_epochs: 16 - mixed_precision_training: Native AMP
e1bf38018957b0d6cc3c877a1ae2290b
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.219 | 0.07 | 100 | 0.0728 | | 0.0358 | 0.13 | 200 | 0.0090 | | 0.0106 | 0.2 | 300 | 0.0033 | | 0.0056 | 0.26 | 400 | 0.0020 | | 0.0053 | 0.33 | 500 | 0.0015 | | 0.003 | 0.39 | 600 | 0.0012 | | 0.0029 | 0.46 | 700 | 0.0009 | | 0.0022 | 0.52 | 800 | 0.0010 | | 0.0025 | 0.59 | 900 | 0.0008 | | 0.002 | 0.65 | 1000 | 0.0006 | | 0.0016 | 0.72 | 1100 | 0.0006 | | 0.0015 | 0.78 | 1200 | 0.0006 | | 0.0015 | 0.85 | 1300 | 0.0004 | | 0.0012 | 0.92 | 1400 | 0.0006 | | 0.0013 | 0.98 | 1500 | 0.0002 | | 0.0013 | 1.05 | 1600 | 0.0003 | | 0.0008 | 1.11 | 1700 | 0.0002 | | 0.0013 | 1.18 | 1800 | 0.0006 | | 0.0019 | 1.24 | 1900 | 0.0004 | | 0.001 | 1.31 | 2000 | 0.0002 | | 0.0007 | 1.37 | 2100 | 0.0003 | | 0.0009 | 1.44 | 2200 | 0.0003 | | 0.0012 | 1.5 | 2300 | 0.0002 |
2e1a154f9a1907bb9834d0120172c417
apache-2.0
['LABEL-0 = NONE', 'LABEL-1 = B-DATE', 'LABEL-2 = I-DATE', 'LABEL-3 = B-TIME', 'LABEL-4 = I-TIME', 'LABEL-5 = B-DURATION', 'LABEL-6 = I-DURATION', 'LABEL-7 = B-SET', 'LABEL-8 = I-SET']
false
Bio-RoBERTime This model is a fine-tuned version of [PlanTL-GOB-ES/roberta-base-biomedical-clinical-es](https://huggingface.co/PlanTL-GOB-ES/roberta-base-biomedical-clinical-es) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0177 - Precision: 0.8121 - Recall: 0.8854 - F1: 0.8472 - Accuracy: 0.9919
63ef28193f6fcc8e42ea099a932226f0
apache-2.0
['LABEL-0 = NONE', 'LABEL-1 = B-DATE', 'LABEL-2 = I-DATE', 'LABEL-3 = B-TIME', 'LABEL-4 = I-TIME', 'LABEL-5 = B-DURATION', 'LABEL-6 = I-DURATION', 'LABEL-7 = B-SET', 'LABEL-8 = I-SET']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 8e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 72 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 24
04ad0b4b9b527333505b317865719fc1
apache-2.0
['LABEL-0 = NONE', 'LABEL-1 = B-DATE', 'LABEL-2 = I-DATE', 'LABEL-3 = B-TIME', 'LABEL-4 = I-TIME', 'LABEL-5 = B-DURATION', 'LABEL-6 = I-DURATION', 'LABEL-7 = B-SET', 'LABEL-8 = I-SET']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0433 | 1.0 | 12 | 0.0443 | 0.4948 | 0.5 | 0.4974 | 0.9800 | | 0.0234 | 2.0 | 24 | 0.0221 | 0.4082 | 0.7257 | 0.5225 | 0.9732 | | 0.0055 | 3.0 | 36 | 0.0159 | 0.4768 | 0.7847 | 0.5932 | 0.9797 | | 0.0089 | 4.0 | 48 | 0.0153 | 0.5317 | 0.8160 | 0.6438 | 0.9813 | | 0.0033 | 5.0 | 60 | 0.0131 | 0.7229 | 0.8333 | 0.7742 | 0.9896 | | 0.008 | 6.0 | 72 | 0.0129 | 0.6649 | 0.8681 | 0.7530 | 0.9885 | | 0.0063 | 7.0 | 84 | 0.0146 | 0.7523 | 0.8542 | 0.8 | 0.9904 | | 0.0086 | 8.0 | 96 | 0.0150 | 0.7470 | 0.8715 | 0.8045 | 0.9906 | | 0.0009 | 9.0 | 108 | 0.0139 | 0.7658 | 0.8854 | 0.8213 | 0.9910 | | 0.0031 | 10.0 | 120 | 0.0159 | 0.8031 | 0.8924 | 0.8454 | 0.9919 | | 0.0011 | 11.0 | 132 | 0.0158 | 0.7649 | 0.8924 | 0.8237 | 0.9909 | | 0.0006 | 12.0 | 144 | 0.0153 | 0.7398 | 0.8785 | 0.8032 | 0.9902 | | 0.0013 | 13.0 | 156 | 0.0157 | 0.7815 | 0.8819 | 0.8287 | 0.9910 | | 0.0008 | 14.0 | 168 | 0.0154 | 0.7822 | 0.8854 | 0.8306 | 0.9908 | | 0.0008 | 15.0 | 180 | 0.0164 | 0.7778 | 0.875 | 0.8235 | 0.9910 | | 0.0007 | 16.0 | 192 | 0.0168 | 0.7864 | 0.8819 | 0.8314 | 0.9912 | | 0.0018 | 17.0 | 204 | 0.0173 | 0.7870 | 0.8854 | 0.8333 | 0.9912 | | 0.0006 | 18.0 | 216 | 0.0178 | 0.7730 | 0.875 | 0.8208 | 0.9914 | | 0.0012 | 19.0 | 228 | 0.0171 | 0.8013 | 0.8819 | 0.8397 | 0.9916 | | 0.0006 | 20.0 | 240 | 0.0181 | 0.8137 | 0.8646 | 0.8384 | 0.9916 | | 0.0007 | 21.0 | 252 | 0.0186 | 0.8137 | 0.8646 | 0.8384 | 0.9918 | | 0.0012 | 22.0 | 264 | 0.0188 | 0.8137 | 0.8646 | 0.8384 | 0.9919 | | 0.0006 | 23.0 | 276 | 0.0178 | 0.8121 | 0.8854 | 0.8472 | 0.9919 | | 0.0009 | 24.0 | 288 | 0.0177 | 0.8121 | 0.8854 | 0.8472 | 0.9919 |
39cafa8fca799c4ebd3b0e59465888d9
unknown
[]
false
karaokeroom.safetensors [<img width="480" src="https://i.imgur.com/hclI0vj.jpg">](https://i.imgur.com/hclI0vj.jpg) [<img width="480" src="https://i.imgur.com/8H3c7eE.jpg">](https://i.imgur.com/8H3c7eE.jpg) カラオケ屋さんの部屋の雰囲気を学習したLoRAです。 Loraを読み込ませて、プロンプトに **karaokeroom** と記述してください。 プロンプトに、1girl, karaoke, microphone, 等とあわせて記述していただくとカラオケを歌ってる感じの絵ができます。 ※当LoRAを適用すると人物の描画や画風に影響が生じるようです。LoRAを適用するWeightを調整することで画風への影響を抑えられます。 影響が気になった場合は \<lora:karaokeroom:1\>ではなく\<lora:karaokeroom:0.6\>といった感じで調整して使ってみてください。 ※karaokeroom, 1girl, karaoke, 等のプロンプトを書いても、部屋の風景のみで人物がうまく描画されないことがあります。 ガチャ要素があるのと、モデルによってはうまく働かない場合があるようです。その場合は根気よく何枚か生成してみるか、違うモデルを使ってみてください。
df9bbc68bcefe6f4766dfdac206c13a0
unknown
[]
false
この実験をやってみた動機 たとえばプロンプトに shibuya,city, と書くと渋谷っぽい風景の絵を描いてくれます。これはモデルが「渋谷」という概念を知ってるという事だと思います。 しかし、 Nishinomiya と書いても西宮っぽい風景の絵を描いてはくれません。これはモデルが「西宮」という概念を知らないという事だと思います。 最近、LoRAという手法でスペックが低いパソコン(GPU)でも追加学習が出来る方法が普及してきました。既存のモデルでは描けないキャラクターや衣装等を学習させている方がたくさんいらっしゃいます。 そこで自分は、風景の写真を何枚か学習させれば、その「場所」の概念を学習してくれるのではないかと考えました。 カラオケ店の部屋の写真を20枚程用意して、WD14-taggerでタグ付けを行いました。出来たtxtファイルの全ての先頭の位置に karaokeroom, という単語を追加しました。 学習前提のモデルは karaokeroom という概念を知らないので、この学習によって karaokeroom という新しい概念を獲得してくれると想定しました。 実際に学習を実行して出来上がった当LoRAを読み込ませると上記の karaokeroom というプロンプトでカラオケ店の部屋っぽい絵が生成できます。 場所の概念を学習させる実験は成功ではないでしょうか?  LoRAでうまく場所の概念を学習できる方法が確立できれば日本の様々な風景を学習させることで身近な場所のイラストが生成できるようになると思います。これはその第一歩です。
3697d917068a5563f344c147cd135e02
unknown
[]
false
問題点、今後の課題 カラオケ店の風景は再現できるようになりました。が、当LoRAを適用してカラオケを歌う女の子の絵を生成すると、人物の描画や画風に影響が生じる場合があります。 これはおそらく場所の概念だけではなく、素材写真の画風等も学習してしまったものだと思います。 現状はLoraを適用するWeightを下げることで影響を軽減できますが、学習方法やLoRAの適用の仕方で影響を軽減することが出来ないか?と考えています。 ・U-net層でWeight調整することで影響を押さえられる? 実は僕も全然よく分かってないのですが(!)階層マージ(Marge Block Weighted)で多くの方が様々なモデルマージに挑戦した結果、絵を生成するU-netの各レイヤー層を調節することで描画に様々な調整ができる(?)ことがわかってきました。 例えば「INの上層はリアル調、INの下層がanime調を担当しているのではないか?」、「M_00は全体にキャラクターや服装・背景等に大きな影響が出る」、「OUT上層は、主題以外の表現 (例えば背景)に影響を及ぼしている」、「OUT04,OUT05,OUT06あたりはめっちゃ顔に影響ある」等色々な説があります。 上の方で書いたkaraokeroom:0.6といった指定はU-net全体まるごとで影響を下げる設定になると思うのですが(多分)もし背景に大きく関与しているU-net層が分かればそれ以外のU-net層への関与を抑えることで既存モデルの人物描写と追加学習背景LoRAがうまく共存できるのでは?と考えられます。 実際にU-net層別にWeightを調節できるScripts(sd-webui-lora-block-weight:https://github.com/hako-mikan/sd-webui-lora-block-weight )を使って色々な数値を調整したXY Plot画像等を作成してみたりして調べていますが、現状では「ワイには何もわからないことがわかった」という感じではっきりしたことは分かっていません。 ・学習用素材の写真をうまく調整する ・学習時のキャプション(タグ)の付け方などで、画風を学ばないようにできないか? ・正則化画像を用意することで何かうまく学習の調整ができるのでは? 等、いろいろな案が考えられると思いますが…まだまだ試行錯誤の段階で情報が足りず良い解決方法は得られていません。 なかなか難しそうですが、うまく場所の概念だけ覚えさせる方法が出来たらいいですよね。
63834e374ea177892e89069b5a282cb4
apache-2.0
['generated_from_trainer']
false
swin-finetuned-food101 This model is a fine-tuned version of [microsoft/swin-base-patch4-window7-224](https://huggingface.co/microsoft/swin-base-patch4-window7-224) on the food101 dataset. It achieves the following results on the evaluation set: - Loss: 0.2772 - Accuracy: 0.9210
8a4e587dd8a1eb17548f9aea5a6743fb
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.5077 | 1.0 | 1183 | 0.3851 | 0.8893 | | 0.3523 | 2.0 | 2366 | 0.3124 | 0.9088 | | 0.1158 | 3.0 | 3549 | 0.2772 | 0.9210 |
3d3905e11f91de506b088fc08bf195ca
apache-2.0
['automatic-speech-recognition', 'zh-CN']
false
exp_w2v2t_zh-cn_unispeech_s784 Fine-tuned [microsoft/unispeech-large-1500h-cv](https://huggingface.co/microsoft/unispeech-large-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (zh-CN)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
f610eec628957738cb829496aec331ba
mit
[]
false
Fast_DreamBooth_AMLO on Stable Diffusion via Dreambooth trained on the [fast-DreamBooth.ipynb by TheLastBen](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook
8f56780d5e29813ef4670d8043f06233
mit
[]
false
model by mrcrois This your the Stable Diffusion model fine-tuned the Fast_DreamBooth_AMLO concept taught to Stable Diffusion with Dreambooth. It can be used by modifying the `instance_prompt(s)`: **AMLO17.jpg, AMLO21.jpg, AMLO9.jpg, AMLO18.jpg, AMLO2.jpg, AMLO1.jpg, AMLO13.jpg, AMLO15.jpg, AMLO14.jpg, AMLO22.jpg, AMLO4.jpg, AMLO16.jpg, AMLO11.jpg, AMLO7.jpg, AMLO8.jpg, AMLO19.jpg, AMLO10.jpg, AMLO6.jpg, AMLO20.jpg, AMLO12.jpg, AMLO5.jpg** You can also train your own concepts and upload them to the library by using [the fast-DremaBooth.ipynb by TheLastBen](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb). And you can run your new concept via `diffusers`: [Colab Notebook for Inference](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_dreambooth_inference.ipynb), [Spaces with the Public Concepts loaded](https://huggingface.co/spaces/sd-dreambooth-library/stable-diffusion-dreambooth-concepts) Here are the images used for training this concept: AMLO5.jpg AMLO12.jpg AMLO20.jpg AMLO6.jpg AMLO10.jpg AMLO19.jpg AMLO8.jpg AMLO7.jpg AMLO11.jpg AMLO16.jpg AMLO4.jpg AMLO22.jpg AMLO14.jpg AMLO15.jpg AMLO13.jpg AMLO1.jpg AMLO2.jpg AMLO18.jpg AMLO9.jpg AMLO21.jpg AMLO17.jpg ![AMLO17.jpg 0](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO17.jpg) ![AMLO21.jpg 1](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO21.jpg) ![AMLO9.jpg 2](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO9.jpg) ![AMLO18.jpg 3](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO18.jpg) ![AMLO2.jpg 4](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO2.jpg) ![AMLO1.jpg 5](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO1.jpg) ![AMLO13.jpg 6](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO13.jpg) ![AMLO15.jpg 7](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO15.jpg) ![AMLO14.jpg 8](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO14.jpg) ![AMLO22.jpg 9](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO22.jpg) ![AMLO4.jpg 10](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO4.jpg) ![AMLO16.jpg 11](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO16.jpg) ![AMLO11.jpg 12](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO11.jpg) ![AMLO7.jpg 13](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO7.jpg) ![AMLO8.jpg 14](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO8.jpg) ![AMLO19.jpg 15](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO19.jpg) ![AMLO10.jpg 16](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO10.jpg) ![AMLO6.jpg 17](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO6.jpg) ![AMLO20.jpg 18](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO20.jpg) ![AMLO12.jpg 19](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO12.jpg) ![AMLO5.jpg 20](https://huggingface.co/mrcrois/fast-dreambooth-amlo/resolve/main/concept_images/AMLO5.jpg)
995a48dc72c75a9387f01a4ea4558241
mit
[]
false
German ELECTRA large Released, Oct 2020, this is a German ELECTRA language model trained collaboratively by the makers of the original German BERT (aka "bert-base-german-cased") and the dbmdz BERT (aka bert-base-german-dbmdz-cased). In our [paper](https://arxiv.org/pdf/2010.10906.pdf), we outline the steps taken to train our model and show that this is the state of the art German language model.
bbd7dd557a0a7a55075d8ebb0850799e
mit
[]
false
Performance ``` GermEval18 Coarse: 80.70 GermEval18 Fine: 55.16 GermEval14: 88.95 ``` See also: deepset/gbert-base deepset/gbert-large deepset/gelectra-base deepset/gelectra-large deepset/gelectra-base-generator deepset/gelectra-large-generator
04cb0cd1268f18b9a03328b4a118ab1b
apache-2.0
['generated_from_trainer']
false
xlsr-53-bemba-15hrs This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.2789 - Wer: 0.3751
f69404d183957973055fa4acab0f970a
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 3.4138 | 0.71 | 400 | 0.4965 | 0.7239 | | 0.5685 | 1.43 | 800 | 0.2939 | 0.4839 | | 0.4471 | 2.15 | 1200 | 0.2728 | 0.4467 | | 0.3579 | 2.86 | 1600 | 0.2397 | 0.3965 | | 0.3087 | 3.58 | 2000 | 0.2427 | 0.4015 | | 0.2702 | 4.29 | 2400 | 0.2539 | 0.4112 | | 0.2406 | 5.01 | 2800 | 0.2376 | 0.3885 | | 0.2015 | 5.72 | 3200 | 0.2492 | 0.3844 | | 0.1759 | 6.44 | 3600 | 0.2562 | 0.3768 | | 0.1572 | 7.16 | 4000 | 0.2789 | 0.3751 |
2bfd681c3b4bdf2cf50c3b389b19e326
apache-2.0
[]
false
ParsBERT (v2.0) A Transformer-based Model for Persian Language Understanding We reconstructed the vocabulary and fine-tuned the ParsBERT v1.1 on the new Persian corpora in order to provide some functionalities for using ParsBERT in other scopes! Please follow the [ParsBERT](https://github.com/hooshvare/parsbert) repo for the latest information about previous and current models.
d6084e05eb0e4d637642e23b7f45a273
apache-2.0
[]
false
DigiMag A total of 8,515 articles scraped from [Digikala Online Magazine](https://www.digikala.com/mag/). This dataset includes seven different classes. 1. Video Games 2. Shopping Guide 3. Health Beauty 4. Science Technology 5. General 6. Art Cinema 7. Books Literature | Label |
83a10f14734e380fd891f24e4b2b490b
apache-2.0
[]
false
| |:------------------:|:----:| | Video Games | 1967 | | Shopping Guide | 125 | | Health Beauty | 1610 | | Science Technology | 2772 | | General | 120 | | Art Cinema | 1667 | | Books Literature | 254 | **Download** You can download the dataset from [here](https://drive.google.com/uc?id=1YgrCYY-Z0h2z0-PfWVfOGt1Tv0JDI-qz)
292c57662cf27a3b93c86a1839c0299e
apache-2.0
[]
false
Results The following table summarizes the F1 score obtained by ParsBERT as compared to other models and architectures. | Dataset | ParsBERT v2 | ParsBERT v1 | mBERT | |:-----------------:|:-----------:|:-----------:|:-----:| | Digikala Magazine | 93.65* | 93.59 | 90.72 |
8e0b426a81a8cf123df4a8fae82172bb
apache-2.0
[]
false
How to use :hugs: | Task | Notebook | |---------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | Text Classification | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/hooshvare/parsbert/blob/master/notebooks/Taaghche_Sentiment_Analysis.ipynb) |
a27931c938dedece7c772e82d961b473
apache-2.0
['exbert', 'multiberts', 'multiberts-seed-0']
false
MultiBERTs Seed 0 Checkpoint 1400k (uncased) Seed 0 intermediate checkpoint 1400k MultiBERTs (pretrained BERT) model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/pdf/2106.16163.pdf) and first released in [this repository](https://github.com/google-research/language/tree/master/language/multiberts). This is an intermediate checkpoint. The final checkpoint can be found at [multiberts-seed-0](https://hf.co/multberts-seed-0). This model is uncased: it does not make a difference between english and English. Disclaimer: The team releasing MultiBERTs did not write a model card for this model so this model card has been written by [gchhablani](https://hf.co/gchhablani).
7847d451871d7b8033c6de73e9fdbfa1
apache-2.0
['exbert', 'multiberts', 'multiberts-seed-0']
false
How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('multiberts-seed-0-1400k') model = BertModel.from_pretrained("multiberts-seed-0-1400k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ```
876b2745bf089990dfc4ba458a94bbba
unknown
[]
false
Stable Diffusion Model Trained using Dreambooth using the original spites of the character Rika Furude from Higurashi <br> Tag to trigger Rika Generation is "furude_rika" Example Images: <img src="https://i.imgur.com/4Rsf4WI.png" alt="Girl in a jacket" > <b> DISCLAIMER: I am not responsible for what images you produce or what you do with them. By downloading this model you consent to taking full responsibility for the images you produce with it. </b>
5fc2df744e71622da44f393b084e523a
mit
['generated_from_keras_callback']
false
sachinsahu/Warsaw_Pact-clustered This model is a fine-tuned version of [nandysoham16/12-clustered_aug](https://huggingface.co/nandysoham16/12-clustered_aug) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.0968 - Train End Logits Accuracy: 0.9688 - Train Start Logits Accuracy: 0.9792 - Validation Loss: 0.1022 - Validation End Logits Accuracy: 1.0 - Validation Start Logits Accuracy: 1.0 - Epoch: 0
b989f745953d82967505048108d5ad0d
mit
['generated_from_keras_callback']
false
Training results | Train Loss | Train End Logits Accuracy | Train Start Logits Accuracy | Validation Loss | Validation End Logits Accuracy | Validation Start Logits Accuracy | Epoch | |:----------:|:-------------------------:|:---------------------------:|:---------------:|:------------------------------:|:--------------------------------:|:-----:| | 0.0968 | 0.9688 | 0.9792 | 0.1022 | 1.0 | 1.0 | 0 |
8b0e5d75c3f5db2828ad1c0ba35bdd05
creativeml-openrail-m
['text-to-image', 'stable-diffusion', 'stable-diffusion-diffusers']
false
The Emoji file names were converted to become the text descriptions. It made the model learn a few special words: "flat", "high contrast" and "color" ![thumbnail](https://huggingface.co/Norod78/sd15-fluentui-emoji/resolve/main/sample_images/sd15-fluentui-emoji-Thumbnail.png)
76c1a70e76b55d1f5a307a2da78def84
apache-2.0
[]
false
SnappFood [Snappfood](https://snappfood.ir/) (an online food delivery company) user comments containing 70,000 comments with two labels (i.e. polarity classification): 1. Happy 2. Sad | Label |
d38e277d465f64b4c229d94f3e681268
apache-2.0
[]
false
Results The following table summarizes the F1 score obtained as compared to other models and architectures. | Dataset | ALBERT-fa-base-v2 | ParsBERT-v1 | mBERT | DeepSentiPers | |:------------------------:|:-----------------:|:-----------:|:-----:|:-------------:| | SnappFood User Comments | 85.79 | 88.12 | 87.87 | - |
d65cc09c95927179ccdbfa6164c8b9b3
apache-2.0
['G2P', 'Grapheme-to-Phoneme', 'speechbrain', 'text2text-generation']
false
SoundChoice: Grapheme-to-Phoneme Models with Semantic Disambiguation This repository provides all the necessary tools to perform English grapheme-to-phoneme conversion with a pretrained SoundChoice G2P model using SpeechBrain. It is trained on LibriG2P training data derived from [LibriSpeech Alignments](https://zenodo.org/record/2619474
d2108d079a52dd5d3a163e6a95c2deb0
apache-2.0
['G2P', 'Grapheme-to-Phoneme', 'speechbrain', 'text2text-generation']
false
Install SpeechBrain First of all, please install SpeechBrain with the following command (local installation): ```bash pip install speechbrain pip install transformers ``` Please notice that we encourage you to read our tutorials and learn more about [SpeechBrain](https://speechbrain.github.io).
c67d432602e008d633dd2e3cd1087289
apache-2.0
['G2P', 'Grapheme-to-Phoneme', 'speechbrain', 'text2text-generation']
false
Perform G2P Conversion Please follow the example below to perform grapheme-to-phoneme conversion with a high-level wrapper. ```python from speechbrain.pretrained import GraphemeToPhoneme g2p = GraphemeToPhoneme.from_hparams("speechbrain/soundchoice-g2p") text = "To be or not to be, that is the question" phonemes = g2p(text) ``` Given below is the expected output ```python >>> phonemes ['T', 'UW', ' ', 'B', 'IY', ' ', 'AO', 'R', ' ', 'N', 'AA', 'T', ' ', 'T', 'UW', ' ', 'B', 'IY', ' ', 'DH', 'AE', 'T', ' ', 'IH', 'Z', ' ', 'DH', 'AH', ' ', 'K', 'W', 'EH', 'S', 'CH', 'AH', 'N'] ``` To perform G2P conversion on a batch of text, pass an array of strings to the interface: ```python items = [ "All's Well That Ends Well", "The Merchant of Venice", "The Two Gentlemen of Verona", "The Comedy of Errors" ] transcriptions = g2p(items) ``` Given below is the expected output: ```python >>> transcriptions [['AO', 'L', 'Z', ' ', 'W', 'EH', 'L', ' ', 'DH', 'AE', 'T', ' ', 'EH', 'N', 'D', 'Z', ' ', 'W', 'EH', 'L'], ['DH', 'AH', ' ', 'M', 'ER', 'CH', 'AH', 'N', 'T', ' ', 'AH', 'V', ' ', 'V', 'EH', 'N', 'AH', 'S'], ['DH', 'AH', ' ', 'T', 'UW', ' ', 'JH', 'EH', 'N', 'T', 'AH', 'L', 'M', 'IH', 'N', ' ', 'AH', 'V', ' ', 'V', 'ER', 'OW', 'N', 'AH'], ['DH', 'AH', ' ', 'K', 'AA', 'M', 'AH', 'D', 'IY', ' ', 'AH', 'V', ' ', 'EH', 'R', 'ER', 'Z']] ```
a5eb2bc2b159063979a3e62b765e9902
apache-2.0
['G2P', 'Grapheme-to-Phoneme', 'speechbrain', 'text2text-generation']
false
Training The model was trained with SpeechBrain (aa018540). To train it from scratch follows these steps: 1. Clone SpeechBrain: ```bash git clone https://github.com/speechbrain/speechbrain/ ``` 2. Install it: ``` cd speechbrain pip install -r requirements.txt pip install -e . ``` 3. Run Training: ``` cd recipes/LibriSpeech/G2P python train.py hparams/hparams_g2p_rnn.yaml --data_folder=your_data_folder ``` Adjust hyperparameters as needed by passing additional arguments.
c1265ae063ac9c9f67e59e9a9cbed7f7
apache-2.0
['G2P', 'Grapheme-to-Phoneme', 'speechbrain', 'text2text-generation']
false
**Citing SpeechBrain** Please, cite SpeechBrain if you use it for your research or business. ```bibtex @misc{speechbrain, title={{SpeechBrain}: A General-Purpose Speech Toolkit}, author={Mirco Ravanelli and Titouan Parcollet and Peter Plantinga and Aku Rouhe and Samuele Cornell and Loren Lugosch and Cem Subakan and Nauman Dawalatabad and Abdelwahab Heba and Jianyuan Zhong and Ju-Chieh Chou and Sung-Lin Yeh and Szu-Wei Fu and Chien-Feng Liao and Elena Rastorgueva and François Grondin and William Aris and Hwidong Na and Yan Gao and Renato De Mori and Yoshua Bengio}, year={2021}, eprint={2106.04624}, archivePrefix={arXiv}, primaryClass={eess.AS}, note={arXiv:2106.04624} } ``` Also please cite the SoundChoice G2P paper on which this pretrained model is based: ```bibtex @misc{ploujnikov2022soundchoice, title={SoundChoice: Grapheme-to-Phoneme Models with Semantic Disambiguation}, author={Artem Ploujnikov and Mirco Ravanelli}, year={2022}, eprint={2207.13703}, archivePrefix={arXiv}, primaryClass={cs.SD} } ```
c276efb62954fa2da4dc55320bc9ec69
apache-2.0
['generated_from_trainer']
false
distilbert-base-uncased-few-shot-sentiment-model This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6819 - Accuracy: 0.75 - F1: 0.8
79fb7ec497ae08c368d3b03f72e3de98
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.00015 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 10
998c071729b857c4710c7270c0471aba
apache-2.0
['generated_from_trainer']
false
distilbert_add_GLUE_Experiment_logit_kd_wnli_96 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the GLUE WNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.3442 - Accuracy: 0.5634
874f2d89542c93c904cc910606f0517e
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.3478 | 1.0 | 3 | 0.3444 | 0.5634 | | 0.3472 | 2.0 | 6 | 0.3445 | 0.5634 | | 0.3467 | 3.0 | 9 | 0.3444 | 0.5634 | | 0.3476 | 4.0 | 12 | 0.3442 | 0.5634 | | 0.3476 | 5.0 | 15 | 0.3442 | 0.5634 | | 0.3471 | 6.0 | 18 | 0.3446 | 0.5634 | | 0.3473 | 7.0 | 21 | 0.3449 | 0.5634 | | 0.3471 | 8.0 | 24 | 0.3451 | 0.5634 | | 0.3477 | 9.0 | 27 | 0.3452 | 0.5634 | | 0.3469 | 10.0 | 30 | 0.3451 | 0.5634 |
ae598eee6ba02cf52ecac35e5fe49936
apache-2.0
[]
false
<a name="introduction"></a> BERTweet.BR: A Pre-Trained Language Model for Tweets in Portuguese Having the same architecture of [BERTweet](https://huggingface.co/docs/transformers/model_doc/bertweet) we trained our model from scratch following [RoBERTa](https://huggingface.co/docs/transformers/model_doc/roberta) pre-training procedure on a corpus of approximately 9GB containing 100M Portuguese Tweets.
828a7f321335afff6849e053fe1fcaa3
apache-2.0
[]
false
Normalized Inputs ```python import torch from transformers import AutoModel, AutoTokenizer model = AutoModel.from_pretrained('melll-uff/bertweetbr') tokenizer = AutoTokenizer.from_pretrained('melll-uff/bertweetbr', normalization=False)
beea19903102e1dd87adbfe91096e74a
apache-2.0
[]
false
INPUT TWEETS ALREADY NORMALIZED! inputs = [ "Procuro um amor , que seja bom pra mim ... vou procurar , eu vou até o fim :nota_musical:", "Que jogo ontem @USER :mãos_juntas:", "Demojizer para Python é :polegar_para_cima: e está disponível em HTTPURL"] encoded_inputs = tokenizer(inputs, return_tensors="pt", padding=True) with torch.no_grad(): last_hidden_states = model(**encoded_inputs)
e7da89b022b01d57bb40ac8595fd697a
apache-2.0
[]
false
CLS Token of last hidden states. Shape: (number of input sentences, hidden sizeof the model) last_hidden_states[0][:,0,:] tensor([[-0.1430, -0.1325, 0.1595, ..., -0.0802, -0.0153, -0.1358], [-0.0108, 0.1415, 0.0695, ..., 0.1420, 0.1153, -0.0176], [-0.1854, 0.1866, 0.3163, ..., -0.2117, 0.2123, -0.1907]]) ```
68151f34a21f516ef66c7fe4c48e4fae
apache-2.0
[]
false
Normalize raw input Tweets ```python from emoji import demojize import torch from transformers import AutoModel, AutoTokenizer model = AutoModel.from_pretrained('melll-uff/bertweetbr') tokenizer = AutoTokenizer.from_pretrained('melll-uff/bertweetbr', normalization=True) inputs = [ "Procuro um amor , que seja bom pra mim ... vou procurar , eu vou até o fim 🎵", "Que jogo ontem @cristiano 🙏", "Demojizer para Python é 👍 e está disponível em https://pypi.org/project/emoji/"] tokenizer.demojizer = lambda x: demojize(x, language='pt') [tokenizer.normalizeTweet(s) for s in inputs]
fb41c8551187cecb1f2e25e194c4768b
apache-2.0
[]
false
Tokenizer first normalizes tweet sentences ['Procuro um amor , que seja bom pra mim ... vou procurar , eu vou até o fim :nota_musical:', 'Que jogo ontem @USER :mãos_juntas:', 'Demojizer para Python é :polegar_para_cima: e está disponível em HTTPURL'] encoded_inputs = tokenizer(inputs, return_tensors="pt", padding=True) with torch.no_grad(): last_hidden_states = model(**encoded_inputs)
8f3ee5dcffb639c7e1f10ee1178f59ba
apache-2.0
[]
false
Mask Filling with Pipeline ```python from transformers import pipeline model_name = 'melll-uff/bertweetbr' tokenizer = AutoTokenizer.from_pretrained('melll-uff/bertweetbr', normalization=False) filler_mask = pipeline("fill-mask", model=model_name, tokenizer=tokenizer) filler_mask("Rio é a <mask> cidade do Brasil.", top_k=5)
37647749d4c29c0ce1830e51a7a67ab5
apache-2.0
[]
false
Output [{'sequence': 'Rio é a melhor cidade do Brasil.', 'score': 0.9871652126312256, 'token': 120, 'token_str': 'm e l h o r'}, {'sequence': 'Rio é a pior cidade do Brasil.', 'score': 0.005050931591540575, 'token': 316, 'token_str': 'p i o r'}, {'sequence': 'Rio é a maior cidade do Brasil.', 'score': 0.004420778248459101, 'token': 389, 'token_str': 'm a i o r'}, {'sequence': 'Rio é a minha cidade do Brasil.', 'score': 0.0021856199018657207, 'token': 38, 'token_str': 'm i n h a'}, {'sequence': 'Rio é a segunda cidade do Brasil.', 'score': 0.0002110043278662488, 'token': 667, 'token_str': 's e g u n d a'}] ```
8786514f638e8095c5777b37ec87cab8
apache-2.0
['automatic-speech-recognition', 'en']
false
exp_w2v2r_en_xls-r_age_teens-0_sixties-10_s225 Fine-tuned [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) for speech recognition using the train split of [Common Voice 7.0 (en)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
49c06652636adcad5d0b06043140824f
apache-2.0
['question-answering']
false
Model Overview This is an ELECTRA-Large QA Model trained from https://huggingface.co/google/electra-large-discriminator in two stages. First, it is trained on synthetic adversarial data generated using a BART-Large question generator, and then it is trained on SQuAD and AdversarialQA (https://arxiv.org/abs/2002.00293) in a second stage of fine-tuning.
ca38937a39e0d65967b13b1bccd81580
apache-2.0
['generated_from_trainer']
false
data-augmentation-whitenoise-timit-1155 This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.5458 - Wer: 0.3324
c30db7f0bef0635b082bc1a98042a518
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 3.5204 | 0.8 | 500 | 1.6948 | 0.9531 | | 0.8435 | 1.6 | 1000 | 0.5367 | 0.5113 | | 0.4449 | 2.4 | 1500 | 0.4612 | 0.4528 | | 0.3182 | 3.21 | 2000 | 0.4314 | 0.4156 | | 0.2328 | 4.01 | 2500 | 0.4250 | 0.4031 | | 0.1897 | 4.81 | 3000 | 0.4630 | 0.4023 | | 0.1628 | 5.61 | 3500 | 0.4445 | 0.3922 | | 0.1472 | 6.41 | 4000 | 0.4452 | 0.3793 | | 0.1293 | 7.21 | 4500 | 0.4715 | 0.3847 | | 0.1176 | 8.01 | 5000 | 0.4267 | 0.3757 | | 0.1023 | 8.81 | 5500 | 0.4494 | 0.3821 | | 0.092 | 9.62 | 6000 | 0.4501 | 0.3704 | | 0.0926 | 10.42 | 6500 | 0.4722 | 0.3643 | | 0.0784 | 11.22 | 7000 | 0.5033 | 0.3765 | | 0.077 | 12.02 | 7500 | 0.5165 | 0.3684 | | 0.0704 | 12.82 | 8000 | 0.5138 | 0.3646 | | 0.0599 | 13.62 | 8500 | 0.5664 | 0.3674 | | 0.0582 | 14.42 | 9000 | 0.5188 | 0.3575 | | 0.0526 | 15.22 | 9500 | 0.5605 | 0.3621 | | 0.0512 | 16.03 | 10000 | 0.5400 | 0.3585 | | 0.0468 | 16.83 | 10500 | 0.5471 | 0.3603 | | 0.0445 | 17.63 | 11000 | 0.5168 | 0.3555 | | 0.0411 | 18.43 | 11500 | 0.5772 | 0.3542 | | 0.0394 | 19.23 | 12000 | 0.5079 | 0.3567 | | 0.0354 | 20.03 | 12500 | 0.5427 | 0.3613 | | 0.0325 | 20.83 | 13000 | 0.5532 | 0.3572 | | 0.0318 | 21.63 | 13500 | 0.5223 | 0.3514 | | 0.0269 | 22.44 | 14000 | 0.6002 | 0.3460 | | 0.028 | 23.24 | 14500 | 0.5591 | 0.3432 | | 0.0254 | 24.04 | 15000 | 0.5837 | 0.3432 | | 0.0235 | 24.84 | 15500 | 0.5571 | 0.3397 | | 0.0223 | 25.64 | 16000 | 0.5470 | 0.3383 | | 0.0193 | 26.44 | 16500 | 0.5611 | 0.3367 | | 0.0227 | 27.24 | 17000 | 0.5405 | 0.3342 | | 0.0183 | 28.04 | 17500 | 0.5205 | 0.3330 | | 0.017 | 28.85 | 18000 | 0.5512 | 0.3330 | | 0.0167 | 29.65 | 18500 | 0.5458 | 0.3324 |
63973428c47350987172470c1746c42e
apache-2.0
['audio', 'automatic-speech-recognition', 'speech', 'xlsr-fine-tuning-week']
false
Wav2Vec2-Large-XLSR-53-Spanish-With-LM This is a model copy of [Wav2Vec2-Large-XLSR-53-Spanish](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-spanish) that has language model support. This model card can be seen as a demo for the [pyctcdecode](https://github.com/kensho-technologies/pyctcdecode) integration with Transformers led by [this PR](https://github.com/huggingface/transformers/pull/14339). The PR explains in-detail how the integration works. In a nutshell: This PR adds a new Wav2Vec2WithLMProcessor class as drop-in replacement for Wav2Vec2Processor. The only change from the existing ASR pipeline will be: ```diff import torch from datasets import load_dataset from transformers import AutoModelForCTC, AutoProcessor import torchaudio.functional as F model_id = "patrickvonplaten/wav2vec2-xlsr-53-es-kenlm" sample = next(iter(load_dataset("common_voice", "es", split="test", streaming=True))) resampled_audio = F.resample(torch.tensor(sample["audio"]["array"]), 48_000, 16_000).numpy() model = AutoModelForCTC.from_pretrained(model_id) processor = AutoProcessor.from_pretrained(model_id) input_values = processor(resampled_audio, return_tensors="pt").input_values with torch.no_grad(): logits = model(input_values).logits -prediction_ids = torch.argmax(logits, dim=-1) -transcription = processor.batch_decode(prediction_ids) +transcription = processor.batch_decode(logits.numpy()).text
b78de20cfee256d6e7f7dc832f8b5f23
apache-2.0
['audio', 'automatic-speech-recognition', 'speech', 'xlsr-fine-tuning-week']
false
=> 'bien y qué regalo vas a abrir primero' ``` **Improvement** This model has been compared on 512 speech samples from the Spanish Common Voice Test set and gives a nice *20 %* performance boost: The results can be reproduced by running *from this model repository*: | Model | WER | CER | | ------------- | ------------- | ------------- | | patrickvonplaten/wav2vec2-xlsr-53-es-kenlm | **8.44%** | **2.93%** | | jonatasgrosman/wav2vec2-large-xlsr-53-spanish | **10.20%** | **3.24%** | ``` bash run_ngram_wav2vec2.py 1 512 ``` ``` bash run_ngram_wav2vec2.py 0 512 ``` with `run_ngram_wav2vec2.py` being https://huggingface.co/patrickvonplaten/wav2vec2-large-xlsr-53-spanish-with-lm/blob/main/run_ngram_wav2vec2.py
38549ccfae4b1d0481e85499dbeeffe2
apache-2.0
['generated_from_trainer']
false
opus-mt-tr-en-finetuned-en-to-tr This model is a fine-tuned version of [Helsinki-NLP/opus-mt-tr-en](https://huggingface.co/Helsinki-NLP/opus-mt-tr-en) on the wmt16 dataset. It achieves the following results on the evaluation set: - Loss: 1.9429 - Bleu: 6.471 - Gen Len: 56.1688
f74eab9602166c5c867bd37b918c2cab
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:------:|:-------:| | 1.5266 | 1.0 | 12860 | 2.2526 | 4.5834 | 55.6563 | | 1.2588 | 2.0 | 25720 | 2.0113 | 5.9203 | 56.3506 | | 1.1878 | 3.0 | 38580 | 1.9429 | 6.471 | 56.1688 |
66c3ae094bad30c3634fb0c31482f471
cc-by-4.0
[]
false
Sentiment Classification in Polish ```python import numpy as np from transformers import AutoTokenizer, AutoModelForSequenceClassification id2label = {0: "negative", 1: "neutral", 2: "positive"} tokenizer = AutoTokenizer.from_pretrained("Voicelab/herbert-base-cased-sentiment") model = AutoModelForSequenceClassification.from_pretrained("Voicelab/herbert-base-cased-sentiment") input = ["Ale fajnie, spadł dzisiaj śnieg! Ulepimy dziś bałwana?"] encoding = tokenizer( input, add_special_tokens=True, return_token_type_ids=True, truncation=True, padding='max_length', return_attention_mask=True, return_tensors='pt', ) output = model(**encoding).logits.to("cpu").detach().numpy() prediction = id2label[np.argmax(output)] print(input, "--->", prediction) ``` Predicted output: ```python ['Ale fajnie, spadł dzisiaj śnieg! Ulepimy dziś bałwana?'] ---> positive ```
4d5cc69c6473821a39423b954a76d447
cc-by-4.0
[]
false
Overview - **Language model:** [allegro/herbert-base-cased](https://huggingface.co/allegro/herbert-base-cased) - **Language:** pl - **Training data:** Reviews + own data - **Blog post:** [Sentiment analysis - COVID-19 – the source of the heated discussion](https://voicelab.ai/covid-19-the-source-of-the-heated-discussion)
6d2f8f6c0f85fd2669573966cba83641
apache-2.0
['generated_from_trainer']
false
my_awesome_qa_model This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 4.2177
628ce027283c7b0bacbba1423bfccfac
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 266 | 4.1357 | | 3.5667 | 2.0 | 532 | 4.1447 | | 3.5667 | 3.0 | 798 | 4.2177 |
43bb22dd4d39bc94ed35c5676a9f9d7a
mit
[]
false
**Hyperparameters:** - learning rate: 2e-5 - weight decay: 0.01 - per_device_train_batch_size: 8 - per_device_eval_batch_size: 8 - gradient_accumulation_steps:1 - eval steps: 50000 - max_length: 512 - num_epochs: 1 - hidden_dropout_prob: 0.3 - attention_probs_dropout_prob: 0.25 **Dataset version:** - tasky_or_not/10xp3nirstbbflanseuni_10xc4 **Checkpoint:** - 300000 steps. **Results on Validation set:** | **Step** | **Training Loss** | **Validation Loss** | **Accuracy** | **Precision** | **Recall** | **F1** | |:--------:|:-----------------:|:-------------------:|:------------:|:-------------:|:----------:|:--------:| | 50000 | 0.020800 | 0.192550 | 0.970363 | 0.990686 | 0.949654 | 0.969736 | | 100000 | 0.015200 | 0.264168 | 0.969427 | 0.994374 | 0.944196 | 0.968636 | | 150000 | 0.012900 | 0.146541 | 0.981440 | 0.994599 | 0.968138 | 0.981190 | | 200000 | 0.011100 | 0.319310 | 0.970516 | 0.998871 | 0.942097 | 0.969654 | | 250000 | 0.008000 | 0.204103 | 0.976309 | 0.996226 | 0.956241 | 0.975824 | | 300000 | 0.006100 | 0.096262 | 0.988053 | 0.994676 | 0.981358 | 0.987972 | | 350000 | 0.005800 | 0.162989 | 0.983663 | 0.994730 | 0.972478 | 0.983478 | **Wandb logs:** - https://wandb.ai/manandey/taskydata/runs/y3j3fbkh?workspace=user-manandey
5fc9d36b711ad3d3b6dd51047b557632
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 2 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - training_steps: 200
0586356265699b3bb1a1d5fa15cca9e9
apache-2.0
['generated_from_trainer']
false
t5-base-finetuned-eli5-a This model is a fine-tuned version of [ammarpl/t5-base-finetuned-xsum-a](https://huggingface.co/ammarpl/t5-base-finetuned-xsum-a) on the eli5 dataset. It achieves the following results on the evaluation set: - Loss: 3.1773 - Rouge1: 14.6711 - Rouge2: 2.2878 - Rougel: 11.3676 - Rougelsum: 13.1805 - Gen Len: 18.9892
20158034179e005bc3a5f19f78fbb546
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:-------:|:------:|:-------:|:---------:|:-------:| | 3.3417 | 1.0 | 17040 | 3.1773 | 14.6711 | 2.2878 | 11.3676 | 13.1805 | 18.9892 |
03cdc9bdb75bfbc72ee1c5d9e046a61e
apache-2.0
['whisper-event', 'generated_from_trainer']
false
Whisper medium Croatian El Greco This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the google/fleurs hr_hr dataset. It achieves the following results on the evaluation set: - Loss: 0.3374 - Wer: 14.6133
109e7083979169f5b47709a08511f68b
apache-2.0
['whisper-event', 'generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-06 - train_batch_size: 32 - eval_batch_size: 16 - seed: 42 - distributed_type: multi-GPU - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 1000
9f505320498d0b2ad35581ae1ebd94a5
apache-2.0
['whisper-event', 'generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.0106 | 4.61 | 1000 | 0.3374 | 14.6133 |
d9808f09b416e724e8c7e1fc4b5fbc3d
apache-2.0
['part-of-speech', 'token-classification']
false
XLM-RoBERTa base Universal Dependencies v2.8 POS tagging: Vietnamese This model is part of our paper called: - Make the Best of Cross-lingual Transfer: Evidence from POS Tagging with over 100 Languages Check the [Space](https://huggingface.co/spaces/wietsedv/xpos) for more details.
50d6b0bb1644cd18213620b536a828e2
apache-2.0
['part-of-speech', 'token-classification']
false
Usage ```python from transformers import AutoTokenizer, AutoModelForTokenClassification tokenizer = AutoTokenizer.from_pretrained("wietsedv/xlm-roberta-base-ft-udpos28-vi") model = AutoModelForTokenClassification.from_pretrained("wietsedv/xlm-roberta-base-ft-udpos28-vi") ```
821d6a0fd888fab9de3396d658095c42
mit
[]
false
Model description This model takes the XLM-Roberta-base model which has been continued to pre-traine on a large corpus of Twitter in multiple languages. It was developed following a similar strategy as introduced as part of the [Tweet Eval](https://github.com/cardiffnlp/tweeteval) framework. The model is further finetuned on the english part of the XNLI training dataset.
f3d7c0023b1558de8b778486e55742ec
mit
[]
false
Intended Usage This model was developed to do Zero-Shot Text Classification in the realm of Hate Speech Detection. It is focused on the language of english as it was finetuned on data in said language. Since the base model was pre-trained on 100 different languages it has shown some effectiveness in other languages. Please refer to the list of languages in the [XLM Roberta paper](https://arxiv.org/abs/1911.02116)
908fcc0cf322ecb7022d5d23cc029ac6
mit
[]
false
Usage with Zero-Shot Classification pipeline ```python from transformers import pipeline classifier = pipeline("zero-shot-classification", model="morit/english_xlm_xnli") ``` After loading the model you can classify sequences in the languages mentioned above. You can specify your sequences and a matching hypothesis to be able to classify your proposed candidate labels. ```python sequence_to_classify = "I think Rishi Sunak is going to win the elections"
429fad3c3d33891d32b8f694634caa3c
mit
[]
false
Training This model was pre-trained on a set of 100 languages and follwed further training on 198M multilingual tweets as described in the original [paper](https://arxiv.org/abs/2104.12250). Further it was trained on the training set of XNLI dataset in english which is a machine translated version of the MNLI dataset. It was trained on 5 epochs of the XNLI train set and evaluated on the XNLI eval dataset at the end of every epoch to find the best performing model. The model which had the highest accuracy on the eval set was chosen at the end. ![Training Charts from wandb](screen_wandb.png) - learning rate: 2e-5 - batch size: 32 - max sequence: length 128 using a GPU (NVIDIA GeForce RTX 3090) resulting in a training time of 1h 47 mins.
8ef617fb069f13167afe5abeadf28148