license
stringlengths
2
30
tags
stringlengths
2
513
is_nc
bool
1 class
readme_section
stringlengths
201
597k
hash
stringlengths
32
32
apache-2.0
['pytorch', 'causal-lm']
false
Training data GPT-J 6B was pretrained on the [Pile](pile.eleuther.ai), a large scale curated dataset created by EleutherAI for the purpose of training this model. After the pre-training, it's finetuned on our Japanese storytelling dataset. Check our blog post for more details.
79072a2a0d7896cdc9022f7cd09f1614
apache-2.0
['pytorch', 'causal-lm']
false
How to use ``` from transformers import AutoTokenizer, AutoModelForCausalLM import torch tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-j-6B") model = AutoModelForCausalLM.from_pretrained("NovelAI/genji-jp", torch_dtype=torch.float16, low_cpu_mem_usage=True).eval().cuda() text = '''あらすじ:あなたは異世界に転生してしまいました。勇者となって、仲間を作り、異世界を冒険しよう! *** 転生すると、ある能力を手に入れていた。それは、''' tokens = tokenizer(text, return_tensors="pt").input_ids generated_tokens = model.generate(tokens.long().cuda(), use_cache=True, do_sample=True, temperature=1, top_p=0.9, repetition_penalty=1.125, min_length=1, max_length=len(tokens[0]) + 400, pad_token_id=tokenizer.eos_token_id) last_tokens = generated_tokens[0] generated_text = tokenizer.decode(last_tokens).replace("�", "") print("Generation:\n" + generated_text) ``` When run, produces output like this: ``` Generation: あらすじ:あなたは異世界に転生してしまいました。勇者となって、仲間を作り、異世界を冒険しよう! *** 転生すると、ある能力を手に入れていた。それは、『予知』だ。過去から未来のことを、誰も知らない出来事も含めて見通すことが出来る。 悪魔の欠片と呼ばれる小さな結晶を取り込んで、使役することが出来る。人を惹きつけ、堕落させる。何より、俺は男なんて居なかったし、女に興味もない。……そんなクズの片棒を担ぎ上げる奴が多くなると思うと、ちょっと苦しい。 だが、一部の人間には協力者を得ることが出来る。目立たない街にある寺の中で、常に家に引きこもっている老人。そんなヤツの魂をコントロールすることが出来るのだ。便利な能力だ。しかし、裏切り者は大勢いる。気を抜けば、狂う。だから注意が必要だ。 ――「やってやるよ」  アーロンは不敵に笑った。この ```
1096cbfd08cfca6eb1c397f50ec91e72
apache-2.0
['pytorch', 'causal-lm']
false
Acknowledgements This project was possible because of the compute provided by the [TPU Research Cloud](https://sites.research.google/trc/) Thanks [EleutherAI](https://eleuther.ai/) for pretraining the GPT-J 6B model. Thanks to everyone who contributed to this project! - [Finetune](https://github.com/finetuneanon) - [Aero](https://github.com/AeroScripts) - [Kurumuz](https://github.com/kurumuz)
fe88b939cb4b65cdc5de36720144ddd2
apache-2.0
['generated_from_trainer']
false
wav2vec2-xls-r-300m_phoneme-mfa_korean This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on a phonetically balanced native Korean read-speech corpus.
75b86462f9cbf391a00ae95c1dfe0b25
apache-2.0
['generated_from_trainer']
false
Training and Evaluation Data Training Data - Data Name: Phonetically Balanced Native Korean Read-speech Corpus - Num. of Samples: 54,000 - Audio Length: 108 Hours Evaluation Data - Data Name: Phonetically Balanced Native Korean Read-speech Corpus - Num. of Samples: 6,000 - Audio Length: 12 Hours
19e00de90a17d1dbc9e66c9dda61e801
apache-2.0
['generated_from_trainer']
false
Training Hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.2 - num_epochs: 20 (EarlyStopping: patience: 5 epochs max) - mixed_precision_training: Native AMP
a1bb489dce320f719f91ce7ce37819e3
apache-2.0
['generated_from_trainer']
false
Experimental Results Official implementation of the paper (in review) Major error patterns of L2 Korean speech from five different L1s: Chinese (ZH), Vietnamese (VI), Japanese (JP), Thai (TH), English (EN) ![Experimental Results](./ICPHS2023_table2.png)
1608902c02bcff756be756deb422d622
apache-2.0
['hf-asr-leaderboard', 'generated_from_trainer']
false
Whisper Tiny Dutch This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the Common Voice 11.0 dataset. It achieves the following results on the evaluation set: - Loss: 0.7024 - Wer: 42.0655
fca5e3220ca2266c49bb524d6f03607e
mit
['conversational']
false
THIS AI IS OUTDATED. See [Aeona](https://huggingface.co/deepparag/Aeona) An generative AI made using [microsoft/DialoGPT-small](https://huggingface.co/microsoft/DialoGPT-small). Trained on: https://www.kaggle.com/Cornell-University/movie-dialog-corpus https://www.kaggle.com/jef1056/discord-data [Live Demo](https://dumbot-331213.uc.r.appspot.com/) Example: ```python from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("deepparag/DumBot") model = AutoModelWithLMHead.from_pretrained("deepparag/DumBot")
3775b335cb35cde7b2eb2a82300d5964
cc-by-4.0
['spanish', 'roberta']
false
This is a **RoBERTa-base** model trained from scratch in Spanish. The training dataset is [mc4](https://huggingface.co/datasets/bertin-project/mc4-es-sampled ) subsampling documents to a total of about 50 million examples. Sampling is random. This model continued training from [sequence length 128](https://huggingface.co/bertin-project/bertin-base-random) using 20.000 steps for length 512. Please see our main [card](https://huggingface.co/bertin-project/bertin-roberta-base-spanish) for more information. This is part of the [Flax/Jax Community Week](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104), organised by [HuggingFace](https://huggingface.co/) and TPU usage sponsored by Google.
915b274e6baa63d22925737b11fab313
apache-2.0
['audio', 'automatic-speech-recognition', 'speech', 'xlsr-fine-tuning-week']
false
Fine-tuned XLSR-53 large model for speech recognition in Hungarian Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Hungarian using the train and validation splits of [Common Voice 6.1](https://huggingface.co/datasets/common_voice) and [CSS10](https://github.com/Kyubyong/css10). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned thanks to the GPU credits generously given by the [OVHcloud](https://www.ovhcloud.com/en/public-cloud/ai-training/) :) The script used for training can be found here: https://github.com/jonatasgrosman/wav2vec2-sprint
fd5a3939795acfea442244b624d99a12
apache-2.0
['audio', 'automatic-speech-recognition', 'speech', 'xlsr-fine-tuning-week']
false
Usage The model can be used directly (without a language model) as follows... Using the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) library: ```python from huggingsound import SpeechRecognitionModel model = SpeechRecognitionModel("jonatasgrosman/wav2vec2-large-xlsr-53-hungarian") audio_paths = ["/path/to/file.mp3", "/path/to/another_file.wav"] transcriptions = model.transcribe(audio_paths) ``` Writing your own inference script: ```python import torch import librosa from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor LANG_ID = "hu" MODEL_ID = "jonatasgrosman/wav2vec2-large-xlsr-53-hungarian" SAMPLES = 5 test_dataset = load_dataset("common_voice", LANG_ID, split=f"test[:{SAMPLES}]") processor = Wav2Vec2Processor.from_pretrained(MODEL_ID) model = Wav2Vec2ForCTC.from_pretrained(MODEL_ID)
01657bf9319a0ff73473ae4fc1ed316a
apache-2.0
['audio', 'automatic-speech-recognition', 'speech', 'xlsr-fine-tuning-week']
false
We need to read the audio files as arrays def speech_file_to_array_fn(batch): speech_array, sampling_rate = librosa.load(batch["path"], sr=16_000) batch["speech"] = speech_array batch["sentence"] = batch["sentence"].upper() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) predicted_sentences = processor.batch_decode(predicted_ids) for i, predicted_sentence in enumerate(predicted_sentences): print("-" * 100) print("Reference:", test_dataset[i]["sentence"]) print("Prediction:", predicted_sentence) ``` | Reference | Prediction | | ------------- | ------------- | | BÜSZKÉK VAGYUNK A MAGYAR EMBEREK NAGYSZERŰ SZELLEMI ALKOTÁSAIRA. | BÜSZKÉK VAGYUNK A MAGYAR EMBEREK NAGYSZERŰ SZELLEMI ALKOTÁSAIRE | | A NEMZETSÉG TAGJAI KÖZÜL EZT TERMESZTIK A LEGSZÉLESEBB KÖRBEN ÍZLETES TERMÉSÉÉRT. | A NEMZETSÉG TAGJAI KÖZÜL ESZSZERMESZTIK A LEGSZELESEBB KÖRBEN IZLETES TERMÉSSÉÉRT | | A VÁROSBA VÁGYÓDOTT A LEGJOBBAN, ÉPPEN MERT ODA NEM JUTHATOTT EL SOHA. | A VÁROSBA VÁGYÓDOTT A LEGJOBBAN ÉPPEN MERT ODA NEM JUTHATOTT EL SOHA | | SÍRJA MÁRA MEGSEMMISÜLT. | SIMGI A MANDO MEG SEMMICSEN | | MINDEN ZENESZÁMOT DRÁGAKŐNEK NEVEZETT. | MINDEN ZENA SZÁMODRAGAKŐNEK NEVEZETT | | ÍGY MÚLT EL A DÉLELŐTT. | ÍGY MÚLT EL A DÍN ELŐTT | | REMEK POFA! | A REMEG PUFO | | SZEMET SZEMÉRT, FOGAT FOGÉRT. | SZEMET SZEMÉRT FOGADD FOGÉRT | | BIZTOSAN LAKIK ITT NÉHÁNY ATYÁMFIA. | BIZTOSAN LAKIKÉT NÉHANY ATYAMFIA | | A SOROK KÖZÖTT OLVAS. | A SOROG KÖZÖTT OLVAS |
1594f50f8a0439549b4b19bdd9472619
apache-2.0
['audio', 'automatic-speech-recognition', 'speech', 'xlsr-fine-tuning-week']
false
Evaluation The model can be evaluated as follows on the Hungarian test data of Common Voice. ```python import torch import re import librosa from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor LANG_ID = "hu" MODEL_ID = "jonatasgrosman/wav2vec2-large-xlsr-53-hungarian" DEVICE = "cuda" CHARS_TO_IGNORE = [",", "?", "¿", ".", "!", "¡", ";", ";", ":", '""', "%", '"', "�", "ʿ", "·", "჻", "~", "՞", "؟", "،", "।", "॥", "«", "»", "„", "“", "”", "「", "」", "‘", "’", "《", "》", "(", ")", "[", "]", "{", "}", "=", "`", "_", "+", "<", ">", "…", "–", "°", "´", "ʾ", "‹", "›", "©", "®", "—", "→", "。", "、", "﹂", "﹁", "‧", "~", "﹏", ",", "{", "}", "(", ")", "[", "]", "【", "】", "‥", "〽", "『", "』", "〝", "〟", "⟨", "⟩", "〜", ":", "!", "?", "♪", "؛", "/", "\\", "º", "−", "^", "ʻ", "ˆ"] test_dataset = load_dataset("common_voice", LANG_ID, split="test") wer = load_metric("wer.py")
7b69ae3dd6c9af1071e1669fd48ed673
apache-2.0
['audio', 'automatic-speech-recognition', 'speech', 'xlsr-fine-tuning-week']
false
We need to read the audio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to(DEVICE), attention_mask=inputs.attention_mask.to(DEVICE)).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=8) predictions = [x.upper() for x in result["pred_strings"]] references = [x.upper() for x in result["sentence"]] print(f"WER: {wer.compute(predictions=predictions, references=references, chunk_size=1000) * 100}") print(f"CER: {cer.compute(predictions=predictions, references=references, chunk_size=1000) * 100}") ``` **Test Result**: In the table below I report the Word Error Rate (WER) and the Character Error Rate (CER) of the model. I ran the evaluation script described above on other models as well (on 2021-04-22). Note that the table below may show different results from those already reported, this may have been caused due to some specificity of the other evaluation scripts used. | Model | WER | CER | | ------------- | ------------- | ------------- | | jonatasgrosman/wav2vec2-large-xlsr-53-hungarian | **31.40%** | **6.20%** | | anton-l/wav2vec2-large-xlsr-53-hungarian | 42.39% | 9.39% | | gchhablani/wav2vec2-large-xlsr-hu | 46.42% | 10.04% | | birgermoell/wav2vec2-large-xlsr-hungarian | 46.93% | 10.31% |
e851569cbe26eb5a62532aa8f272d42b
apache-2.0
['audio', 'automatic-speech-recognition', 'speech', 'xlsr-fine-tuning-week']
false
Citation If you want to cite this model you can use this: ```bibtex @misc{grosman2021xlsr53-large-hungarian, title={Fine-tuned {XLSR}-53 large model for speech recognition in {H}ungarian}, author={Grosman, Jonatas}, howpublished={\url{https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-hungarian}}, year={2021} } ```
23292a64dec425f2720fd550bc8a0f38
apache-2.0
['generated_from_trainer']
false
VN_ja-en_helsinki This model is a fine-tuned version of [Helsinki-NLP/opus-mt-ja-en](https://huggingface.co/Helsinki-NLP/opus-mt-ja-en) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 2.2409 - BLEU: 15.28
905a86b025471c3e3f6d94c67ff1d774
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP
bcac4d8767f7395f9c1a62dc412d2134
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 2.6165 | 0.19 | 2000 | 2.6734 | | 2.3805 | 0.39 | 4000 | 2.6047 | | 2.2793 | 0.58 | 6000 | 2.5461 | | 2.2028 | 0.78 | 8000 | 2.5127 | | 2.1361 | 0.97 | 10000 | 2.4511 | | 1.9653 | 1.17 | 12000 | 2.4331 | | 1.934 | 1.36 | 14000 | 2.3840 | | 1.9002 | 1.56 | 16000 | 2.3901 | | 1.87 | 1.75 | 18000 | 2.3508 | | 1.8408 | 1.95 | 20000 | 2.3082 | | 1.6937 | 2.14 | 22000 | 2.3279 | | 1.6371 | 2.34 | 24000 | 2.3052 | | 1.6264 | 2.53 | 26000 | 2.3071 | | 1.6029 | 2.72 | 28000 | 2.2685 | | 1.5847 | 2.92 | 30000 | 2.2409 |
4e8ad9b782864011cfc8d2569df13614
apache-2.0
['generated_from_trainer']
false
bert-finetuned-ner This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.0614 - Precision: 0.9310 - Recall: 0.9498 - F1: 0.9404 - Accuracy: 0.9857
3d559cf69c55abf56679281e46a1b20c
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0875 | 1.0 | 1756 | 0.0639 | 0.9167 | 0.9387 | 0.9276 | 0.9833 | | 0.0332 | 2.0 | 3512 | 0.0595 | 0.9334 | 0.9504 | 0.9418 | 0.9857 | | 0.0218 | 3.0 | 5268 | 0.0614 | 0.9310 | 0.9498 | 0.9404 | 0.9857 |
5d9e34ec1e924e14a406c0b3cdfa94ea
apache-2.0
['translation']
false
opus-mt-sv-fi * source languages: sv * target languages: fi * OPUS readme: [sv-fi](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/sv-fi/README.md) * dataset: opus+bt * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus+bt-2020-04-07.zip](https://object.pouta.csc.fi/OPUS-MT-models/sv-fi/opus+bt-2020-04-07.zip) * test set translations: [opus+bt-2020-04-07.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/sv-fi/opus+bt-2020-04-07.test.txt) * test set scores: [opus+bt-2020-04-07.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/sv-fi/opus+bt-2020-04-07.eval.txt)
f5d7b2846ef38788d619332f4c56f02d
apache-2.0
['automatic-speech-recognition', 'pt']
false
exp_w2v2t_pt_xlsr-53_s454 Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) for speech recognition using the train split of [Common Voice 7.0 (pt)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
6f2a53b524ac571e8a7bb872d8c54db5
apache-2.0
['finnish', 'roberta']
false
RoBERTa large model trained with WECHSEL method for Finnish Pretrained RoBERTa model on Finnish language using a masked language modeling (MLM) objective with WECHSEL method. RoBERTa was introduced in [this paper](https://arxiv.org/abs/1907.11692) and first released in [this repository](https://github.com/pytorch/fairseq/tree/master/examples/roberta). WECHSEL method (Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models) was introduced in [this paper](https://arxiv.org/abs/2112.06598) and first released in [this repository](https://github.com/CPJKU/wechsel). This model is case-sensitive: it makes a difference between finnish and Finnish.
cecd2e32edcdb908f581144781a5c614
apache-2.0
['finnish', 'roberta']
false
WECHSEL method Using the WECHSEL method, we first took the pretrained English [roberta-large](https://huggingface.co/roberta-large) model, changed its tokenizer with our Finnish tokenizer and initialized model's token embeddings such that they are close to semantically similar English tokens by utilizing multilingual static word embeddings (by fastText) covering English and Finnish. We were able to confirm the WECHSEL paper's findings that using this method you can save pretraining time and thus computing resources. To get idea of the WECHSEL method's training time savings you can check the table below illustrating the MLM evaluation accuracies during the pretraining compared to the [Finnish-NLP/roberta-large-finnish-v2](https://huggingface.co/Finnish-NLP/roberta-large-finnish-v2) which was trained from scratch: | | 10k train steps | 100k train steps | 200k train steps | 270k train steps | |------------------------------------------|------------------|------------------|------------------|------------------| |Finnish-NLP/roberta-large-wechsel-finnish |37.61 eval acc |58.14 eval acc |61.60 eval acc |62.77 eval acc | |Finnish-NLP/roberta-large-finnish-v2 |13.83 eval acc |55.87 eval acc |58.58 eval acc |59.47 eval acc | Downstream finetuning text classification tests can be found from the end but there this model trained with WECHSEL method didn't significantly improve the downstream performances. However, based on tens of qualitative fill-mask task example tests we noticed that for fill-mask task this WECHSEL model significantly outperforms our other models trained from scratch.
b6891fbfe314b5cd3cdac3fd472036ee
apache-2.0
['finnish', 'roberta']
false
How to use You can use this model directly with a pipeline for masked language modeling: ```python >>> from transformers import pipeline >>> unmasker = pipeline('fill-mask', model='Finnish-NLP/roberta-large-wechsel-finnish') >>> unmasker("Moikka olen <mask> kielimalli.") [{'sequence': 'Moikka olen hyvä kielimalli.', 'score': 0.07757357507944107, 'token': 763, 'token_str': ' hyvä'}, {'sequence': 'Moikka olen suomen kielimalli.', 'score': 0.05297883599996567, 'token': 3641, 'token_str': ' suomen'}, {'sequence': 'Moikka olen kuin kielimalli.', 'score': 0.03747279942035675, 'token': 523, 'token_str': ' kuin'}, {'sequence': 'Moikka olen suomalainen kielimalli.', 'score': 0.031031042337417603, 'token': 4966, 'token_str': ' suomalainen'}, {'sequence': 'Moikka olen myös kielimalli.', 'score': 0.026489052921533585, 'token': 505, 'token_str': ' myös'}] ``` Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import RobertaTokenizer, RobertaModel tokenizer = RobertaTokenizer.from_pretrained('Finnish-NLP/roberta-large-wechsel-finnish') model = RobertaModel.from_pretrained('Finnish-NLP/roberta-large-wechsel-finnish') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import RobertaTokenizer, TFRobertaModel tokenizer = RobertaTokenizer.from_pretrained('Finnish-NLP/roberta-large-wechsel-finnish') model = TFRobertaModel.from_pretrained('Finnish-NLP/roberta-large-wechsel-finnish', from_pt=True) text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ```
5b6aa0c0531614071777315c60833655
apache-2.0
['finnish', 'roberta']
false
Training data This Finnish RoBERTa model was pretrained on the combination of five datasets: - [mc4_fi_cleaned](https://huggingface.co/datasets/Finnish-NLP/mc4_fi_cleaned), the dataset mC4 is a multilingual colossal, cleaned version of Common Crawl's web crawl corpus. We used the Finnish subset of the mC4 dataset and further cleaned it with our own text data cleaning codes (check the dataset repo). - [wikipedia](https://huggingface.co/datasets/wikipedia) We used the Finnish subset of the wikipedia (August 2021) dataset - [Yle Finnish News Archive](http://urn.fi/urn:nbn:fi:lb-2017070501) - [Finnish News Agency Archive (STT)](http://urn.fi/urn:nbn:fi:lb-2018121001) - [The Suomi24 Sentences Corpus](http://urn.fi/urn:nbn:fi:lb-2020021803) Raw datasets were cleaned to filter out bad quality and non-Finnish examples. Together these cleaned datasets were around 84GB of text.
db5c6b3c990c735a159ff8c2fd0b7136
apache-2.0
['finnish', 'roberta']
false
Pretraining The model was trained on TPUv3-8 VM, sponsored by the [Google TPU Research Cloud](https://sites.research.google/trc/about/), for 270k steps (a bit over 1 epoch, 512 batch size) with a sequence length of 128 and continuing for 180k steps (batch size 64) with a sequence length of 512. The optimizer used was Adafactor (to save memory). Learning rate was 2e-4, \\(\beta_{1} = 0.9\\), \\(\beta_{2} = 0.98\\) and \\(\epsilon = 1e-6\\), learning rate warmup for 2500 steps and linear decay of the learning rate after.
408e424ff782e270ed2e79a50be1debd
apache-2.0
['finnish', 'roberta']
false
Evaluation results Evaluation was done by fine-tuning the model on downstream text classification task with two different labeled datasets: [Yle News](https://github.com/spyysalo/yle-corpus) and [Eduskunta](https://github.com/aajanki/eduskunta-vkk). Yle News classification fine-tuning was done with two different sequence lengths: 128 and 512 but Eduskunta only with 128 sequence length. When fine-tuned on those datasets, this model (the first row of the table) achieves the following accuracy results compared to the [FinBERT (Finnish BERT)](https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1) model and to our previous [Finnish-NLP/roberta-large-finnish-v2](https://huggingface.co/Finnish-NLP/roberta-large-finnish-v2) and [Finnish-NLP/roberta-large-finnish](https://huggingface.co/Finnish-NLP/roberta-large-finnish) models: | | Average | Yle News 128 length | Yle News 512 length | Eduskunta 128 length | |------------------------------------------|----------|---------------------|---------------------|----------------------| |Finnish-NLP/roberta-large-wechsel-finnish |88.19 |**94.91** |95.18 |74.47 | |Finnish-NLP/roberta-large-finnish-v2 |88.17 |94.46 |95.22 |74.83 | |Finnish-NLP/roberta-large-finnish |88.02 |94.53 |95.23 |74.30 | |TurkuNLP/bert-base-finnish-cased-v1 |**88.82** |94.90 |**95.49** |**76.07** | To conclude, this model didn't significantly improve compared to our previous models which were trained from scratch instead of using the WECHSEL method as in this model. This model is also slightly (~ 1%) losing to the [FinBERT (Finnish BERT)](https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1) model.
fd30857e4dedebeb44facf15d0cdb51d
apache-2.0
['finnish', 'roberta']
false
Team Members - Aapo Tanskanen, [Hugging Face profile](https://huggingface.co/aapot), [LinkedIn profile](https://www.linkedin.com/in/aapotanskanen/) - Rasmus Toivanen [Hugging Face profile](https://huggingface.co/RASMUS), [LinkedIn profile](https://www.linkedin.com/in/rasmustoivanen/) Feel free to contact us for more details 🤗
353a71cf3b26f31041c65d3a11e24384
apache-2.0
['generated_from_keras_callback']
false
javilonso/classificationPolEsp2 This model is a fine-tuned version of [PlanTL-GOB-ES/gpt2-base-bne](https://huggingface.co/PlanTL-GOB-ES/gpt2-base-bne) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.1229 - Validation Loss: 0.8172 - Epoch: 2
a1dc31baebbd0e8e1c5d1b6f7c747858
apache-2.0
['generated_from_keras_callback']
false
Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 17958, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01} - training_precision: mixed_float16
cb470e1c154a9a64f31828b80b641d55
apache-2.0
['generated_from_keras_callback']
false
Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 0.6246 | 0.5679 | 0 | | 0.4198 | 0.6097 | 1 | | 0.1229 | 0.8172 | 2 |
10c6d8f2ad0765b9735b0aa1620b0d13
mit
[]
false
InsideWhale on Stable Diffusion This is the `<InsideWhale>` concept taught to Stable Diffusion via Textual Inversion. You can load this concept into the [Stable Conceptualizer](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/stable_conceptualizer_inference.ipynb) notebook. You can also train your own concepts and load them into the concept libraries using [this notebook](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_textual_inversion_training.ipynb). Here is the new concept you will be able to use as a `style`: ![<InsideWhale> 0](https://huggingface.co/sd-concepts-library/insidewhale/resolve/main/concept_images/1.jpeg) ![<InsideWhale> 1](https://huggingface.co/sd-concepts-library/insidewhale/resolve/main/concept_images/2.jpeg) ![<InsideWhale> 2](https://huggingface.co/sd-concepts-library/insidewhale/resolve/main/concept_images/0.jpeg) ![<InsideWhale> 3](https://huggingface.co/sd-concepts-library/insidewhale/resolve/main/concept_images/3.jpeg) ![<InsideWhale> 4](https://huggingface.co/sd-concepts-library/insidewhale/resolve/main/concept_images/4.jpeg)
be1c8c95bcee8ca1ca50e6181d73c6cb
apache-2.0
['summarization']
false
How to use Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import BigBirdPegasusForConditionalGeneration, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("google/bigbird-pegasus-large-bigpatent")
dbc643d676782b59dad96ddd66e2c1dc
apache-2.0
['summarization']
false
you can change `attention_type` (encoder only) to full attention like this: model = BigBirdPegasusForConditionalGeneration.from_pretrained("google/bigbird-pegasus-large-bigpatent", attention_type="original_full")
2503b6f004d4434e23fdcac065d1bac4
apache-2.0
['summarization']
false
you can change `block_size` & `num_random_blocks` like this: model = BigBirdPegasusForConditionalGeneration.from_pretrained("google/bigbird-pegasus-large-bigpatent", block_size=16, num_random_blocks=2) text = "Replace me by any text you'd like." inputs = tokenizer(text, return_tensors='pt') prediction = model.generate(**inputs) prediction = tokenizer.batch_decode(prediction) ```
c145bccb391141bcc80e9a75f712102c
other
['generated_from_trainer']
false
finetuned-distilbert-news-article-catgorization This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the adult_content dataset. It achieves the following results on the evaluation set: - Loss: 0.0065 - F1_score(weighted): 0.90
14d064bb7c4926077c362803100a827e
other
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-5 - train_batch_size: 5 - eval_batch_size: 5 - seed: 17 - optimizer: AdamW(lr=1e-5 and epsilon=1e-08) - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 0 - num_epochs: 2
31279bbcae9645099bbd8d153f70e7a4
other
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Validation Loss | f1 score | |:-------------:|:-----:|:---------------: |:------:| | 0.1414 | 1.0 | 0.4585 | 0.9058 | | 0.1410 | 2.0 | 0.4584 | 0.9058 |
afdc65a5e70b761015046d1db957f340
apache-2.0
['generated_from_keras_callback']
false
annaeze/lab9_1 This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.0230 - Validation Loss: 0.0572 - Epoch: 2
3835e52ccf1a95909105743b9ab596df
apache-2.0
['generated_from_keras_callback']
false
Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 0.1174 | 0.0596 | 0 | | 0.0391 | 0.0529 | 1 | | 0.0230 | 0.0572 | 2 |
f91856522d5c62701af4b1e03eaee0bb
apache-2.0
[]
false
BERT multilingual base model (cased) Pretrained model on the top 104 languages with the largest Wikipedia using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1810.04805) and first released in [this repository](https://github.com/google-research/bert). This model is case sensitive: it makes a difference between english and English. Disclaimer: The team releasing BERT did not write a model card for this model so this model card has been written by the Hugging Face team.
34971bbad56a1a89515004b7d009c591
apache-2.0
[]
false
Model description BERT is a transformers model pretrained on a large corpus of multilingual data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with two objectives: - Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. - Next sentence prediction (NSP): the models concatenates two masked sentences as inputs during pretraining. Sometimes they correspond to sentences that were next to each other in the original text, sometimes not. The model then has to predict if the two sentences were following each other or not. This way, the model learns an inner representation of the languages in the training set that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs.
c746fb08c2142034ae2c8d01c2029223
apache-2.0
[]
false
How to use You can use this model directly with a pipeline for masked language modeling: ```python >>> from transformers import pipeline >>> unmasker = pipeline('fill-mask', model='bert-base-multilingual-cased') >>> unmasker("Hello I'm a [MASK] model.") [{'sequence': "[CLS] Hello I'm a model model. [SEP]", 'score': 0.10182085633277893, 'token': 13192, 'token_str': 'model'}, {'sequence': "[CLS] Hello I'm a world model. [SEP]", 'score': 0.052126359194517136, 'token': 11356, 'token_str': 'world'}, {'sequence': "[CLS] Hello I'm a data model. [SEP]", 'score': 0.048930276185274124, 'token': 11165, 'token_str': 'data'}, {'sequence': "[CLS] Hello I'm a flight model. [SEP]", 'score': 0.02036019042134285, 'token': 23578, 'token_str': 'flight'}, {'sequence': "[CLS] Hello I'm a business model. [SEP]", 'score': 0.020079681649804115, 'token': 14155, 'token_str': 'business'}] ``` Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('bert-base-multilingual-cased') model = BertModel.from_pretrained("bert-base-multilingual-cased") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('bert-base-multilingual-cased') model = TFBertModel.from_pretrained("bert-base-multilingual-cased") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ```
a7f91a7cc832793ff4f83b44266417c2
apache-2.0
[]
false
Training data The BERT model was pretrained on the 104 languages with the largest Wikipedias. You can find the complete list [here](https://github.com/google-research/bert/blob/master/multilingual.md
2cb3c7c20be6262f1e00e0ec3ea5abee
apache-2.0
[]
false
Preprocessing The texts are lowercased and tokenized using WordPiece and a shared vocabulary size of 110,000. The languages with a larger Wikipedia are under-sampled and the ones with lower resources are oversampled. For languages like Chinese, Japanese Kanji and Korean Hanja that don't have space, a CJK Unicode block is added around every character. The inputs of the model are then of the form: ``` [CLS] Sentence A [SEP] Sentence B [SEP] ``` With probability 0.5, sentence A and sentence B correspond to two consecutive sentences in the original corpus and in the other cases, it's another random sentence in the corpus. Note that what is considered a sentence here is a consecutive span of text usually longer than a single sentence. The only constrain is that the result with the two "sentences" has a combined length of less than 512 tokens. The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by `[MASK]`. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is.
376b2bb9d8fc8cf536d1d880ea01ae7f
apache-2.0
['generated_from_trainer']
false
UrduAudio2Text This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. It achieves the following results on the evaluation set: - Loss: 1.4978 - Wer: 0.8376
85fb942b1fcdf9a5a5b8fcc73d17e2f7
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 5.5558 | 15.98 | 400 | 1.4978 | 0.8376 |
8c216a479c5829acb56cccf0edbb6732
apache-2.0
['monai', 'medical']
false
Model Overview This model is trained using the runner-up [1] awarded pipeline of the "Medical Segmentation Decathlon Challenge 2018" using the UNet architecture [2] with 32 training images and 9 validation images.
9fe7b153d77dbfc9de17a7a1160de599
apache-2.0
['monai', 'medical']
false
commands example Execute training: ``` python -m monai.bundle run training --meta_file configs/metadata.json --config_file configs/train.json --logging_file configs/logging.conf ``` Override the `train` config to execute multi-GPU training: ``` torchrun --standalone --nnodes=1 --nproc_per_node=2 -m monai.bundle run training --meta_file configs/metadata.json --config_file "['configs/train.json','configs/multi_gpu_train.json']" --logging_file configs/logging.conf ``` Override the `train` config to execute evaluation with the trained model: ``` python -m monai.bundle run evaluating --meta_file configs/metadata.json --config_file "['configs/train.json','configs/evaluate.json']" --logging_file configs/logging.conf ``` Execute inference: ``` python -m monai.bundle run evaluating --meta_file configs/metadata.json --config_file configs/inference.json --logging_file configs/logging.conf ```
483328602e4e11c4e8bafbd474dfa041
apache-2.0
['monai', 'medical']
false
References [1] Xia, Yingda, et al. "3D Semi-Supervised Learning with Uncertainty-Aware Multi-View Co-Training." arXiv preprint arXiv:1811.12506 (2018). https://arxiv.org/abs/1811.12506. [2] Kerfoot E., Clough J., Oksuz I., Lee J., King A.P., Schnabel J.A. (2019) Left-Ventricle Quantification Using Residual U-Net. In: Pop M. et al. (eds) Statistical Atlases and Computational Models of the Heart. Atrial Segmentation and LV Quantification Challenges. STACOM 2018. Lecture Notes in Computer Science, vol 11395. Springer, Cham. https://doi.org/10.1007/978-3-030-12029-0_40
27f5dca9b3a83135a3099d127011e863
apache-2.0
['generated_from_trainer']
false
Use this model to detect Twitter users' profiles related to healthcare. User profile classification may be useful when searching for health information on Twitter. For a certain health topic, tweets from physicians or organizations (e.g. ```Board-certified dermatologist```) may be more reliable than undefined or vague profiles (e.g. ```Human. Person. Father```). The model expects the user's ```description``` text field (see [Twitter API](https://developer.twitter.com/en/docs/twitter-api/v1/data-dictionary/object-model/user) docs) as input and returns a label for each profile: - `not-health-related` - `health-related` - `health-related/person` - `health-related/organization` - `health-related/publishing` - `health-related/physician` - `health-related/news` - `health-related/academic` F1 score is 0.9
03b4f20be45809a1704ea55aac372037
apache-2.0
['generated_from_trainer']
false
distilbert-base-uncased-finetuned-ner This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.0616 - Precision: 0.9265 - Recall: 0.9361 - F1: 0.9313 - Accuracy: 0.9837
66a9b2eda032996199900f228f4aa9dd
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.2437 | 1.0 | 878 | 0.0745 | 0.9144 | 0.9173 | 0.9158 | 0.9799 | | 0.0518 | 2.0 | 1756 | 0.0621 | 0.9177 | 0.9353 | 0.9264 | 0.9826 | | 0.03 | 3.0 | 2634 | 0.0616 | 0.9265 | 0.9361 | 0.9313 | 0.9837 |
ad990d9f0d629be54ac02ca78cb91fa5
apache-2.0
['generated_from_trainer']
false
distilbert-base-uncased-finetuned-policies This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the policies dataset. It achieves the following results on the evaluation set: - Loss: 0.0193
14fe6560b72f47e25fd2ef8c6aaa38ff
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.4208 | 1.0 | 759 | 0.0183 | | 0.0115 | 2.0 | 1518 | 0.0202 | | 0.0048 | 3.0 | 2277 | 0.0193 |
70322d00117b9b738db0aecc9a5336d6
apache-2.0
['generated_from_trainer']
false
aa This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the None dataset. It achieves the following results on the evaluation set: - Loss: 15.9757 - Wer: 1.0
68779a3f16dde36200910b445f14fd37
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 10 - mixed_precision_training: Native AMP
517af4f4b0fb43d92a4a1e589a86ec82
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:---:| | 14.5628 | 3.33 | 20 | 16.1808 | 1.0 | | 14.5379 | 6.67 | 40 | 16.1005 | 1.0 | | 14.3379 | 10.0 | 60 | 15.9757 | 1.0 |
373758c08e913fcba590d46f82c71122
mit
['generated_from_trainer']
false
roberta-large-md-conllpp-v3 This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on the conllpp dataset. It achieves the following results on the evaluation set: - Loss: 0.0564 - Precision: 0.9980 - Recall: 0.9951 - F1: 0.9965 - Accuracy: 0.9943
777ebae082be738d194209323c988c8f
mit
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0945 | 1.0 | 878 | 0.0317 | 0.9975 | 0.9950 | 0.9963 | 0.9939 | | 0.0175 | 2.0 | 1756 | 0.0483 | 0.9980 | 0.9953 | 0.9967 | 0.9946 | | 0.0105 | 3.0 | 2634 | 0.0505 | 0.9982 | 0.9941 | 0.9961 | 0.9937 | | 0.0056 | 4.0 | 3512 | 0.0574 | 0.9982 | 0.9939 | 0.9961 | 0.9935 | | 0.0022 | 5.0 | 4390 | 0.0564 | 0.9980 | 0.9951 | 0.9965 | 0.9943 |
bd7c284aeafb6611cfba7c910cf96348
apache-2.0
['sentence-transformers', 'feature-extraction', 'sentence-similarity', 'transformers']
false
mmarco-sentence-BERTino This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. It was trained on [mmarco](https://huggingface.co/datasets/unicamp-dl/mmarco/viewer/italian/train). <p align="center"> <img src="https://media.tate.org.uk/art/images/work/L/L04/L04294_9.jpg" width="600"> </br> Mohan Samant, Midnight Fishing Party, 1978 </p>
bb5ad3f224ad68f864d31ff240871173
apache-2.0
['sentence-transformers', 'feature-extraction', 'sentence-similarity', 'transformers']
false
Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["Questo è un esempio di frase", "Questo è un ulteriore esempio"] model = SentenceTransformer('efederici/mmarco-sentence-BERTino') embeddings = model.encode(sentences) print(embeddings) ```
2dc92e9e6d264dada5eb67ff09fcf1d0
apache-2.0
['sentence-transformers', 'feature-extraction', 'sentence-similarity', 'transformers']
false
Usage (HuggingFace Transformers) Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings. ```python from transformers import AutoTokenizer, AutoModel import torch
0ef165af21f8f9b0fa682c4b060f68f0
apache-2.0
['mobile', 'vison', 'image-classification']
false
Model Details EfficientFormer-L3, developed by [Snap Research](https://github.com/snap-research), is one of three EfficientFormer models. The EfficientFormer models were released as part of an effort to prove that properly designed transformers can reach extremely low latency on mobile devices while maintaining high performance. This checkpoint of EfficientFormer-L3 was trained for 300 epochs. - Developed by: Yanyu Li, Geng Yuan, Yang Wen, Eric Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, Jian Ren - Language(s): English - License: This model is licensed under the apache-2.0 license - Resources for more information: - [Research Paper](https://arxiv.org/abs/2206.01191) - [GitHub Repo](https://github.com/snap-research/EfficientFormer/) </model_details> <how_to_start>
6eae00a5e55ceb229399d29be5ce56ba
apache-2.0
['mobile', 'vison', 'image-classification']
false
Load preprocessor and pretrained model model_name = "huggingface/efficientformer-l3-300" processor = EfficientFormerImageProcessor.from_pretrained(model_name) model = EfficientFormerForImageClassificationWithTeacher.from_pretrained(model_name)
fb2125fe496335d47993fcff6abdd1f2
apache-2.0
['generated_from_trainer']
false
funnel-transformer-xlarge_ner_wikiann This model is a fine-tuned version of [funnel-transformer/xlarge](https://huggingface.co/funnel-transformer/xlarge) on the wikiann dataset. It achieves the following results on the evaluation set: - Loss: 0.4023 - Precision: 0.8522 - Recall: 0.8634 - F1: 0.8577 - Accuracy: 0.9358
3703317fee31481929a15ee1795769e0
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - num_epochs: 5
0d3284741cbb314becf07e86c32779a9
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.3193 | 1.0 | 5000 | 0.3116 | 0.8239 | 0.8296 | 0.8267 | 0.9260 | | 0.2836 | 2.0 | 10000 | 0.2846 | 0.8446 | 0.8498 | 0.8472 | 0.9325 | | 0.2237 | 3.0 | 15000 | 0.3258 | 0.8427 | 0.8542 | 0.8484 | 0.9332 | | 0.1303 | 4.0 | 20000 | 0.3801 | 0.8531 | 0.8634 | 0.8582 | 0.9362 | | 0.0867 | 5.0 | 25000 | 0.4023 | 0.8522 | 0.8634 | 0.8577 | 0.9358 |
90bb7988f28e651d9f346371159f613f
cc-by-4.0
['question generation']
false
Model Card of `research-backup/t5-base-squadshifts-vanilla-reddit-qg` This model is fine-tuned version of [t5-base](https://huggingface.co/t5-base) for question generation task on the [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) (dataset_name: reddit) via [`lmqg`](https://github.com/asahi417/lm-question-generation).
b317f34498cf4a888958bc2f97b2df44
cc-by-4.0
['question generation']
false
Overview - **Language model:** [t5-base](https://huggingface.co/t5-base) - **Language:** en - **Training data:** [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) (reddit) - **Online Demo:** [https://autoqg.net/](https://autoqg.net/) - **Repository:** [https://github.com/asahi417/lm-question-generation](https://github.com/asahi417/lm-question-generation) - **Paper:** [https://arxiv.org/abs/2210.03992](https://arxiv.org/abs/2210.03992)
8ecda5ba142b0a72e5e48f7b38028add
cc-by-4.0
['question generation']
false
model prediction questions = model.generate_q(list_context="William Turner was an English painter who specialised in watercolour landscapes", list_answer="William Turner") ``` - With `transformers` ```python from transformers import pipeline pipe = pipeline("text2text-generation", "research-backup/t5-base-squadshifts-vanilla-reddit-qg") output = pipe("generate question: <hl> Beyonce <hl> further expanded her acting career, starring as blues singer Etta James in the 2008 musical biopic, Cadillac Records.") ```
22b5bdfe473e5b5271ebf976a96de631
cc-by-4.0
['question generation']
false
Evaluation - ***Metric (Question Generation)***: [raw metric file](https://huggingface.co/research-backup/t5-base-squadshifts-vanilla-reddit-qg/raw/main/eval/metric.first.sentence.paragraph_answer.question.lmqg_qg_squadshifts.reddit.json) | | Score | Type | Dataset | |:-----------|--------:|:-------|:---------------------------------------------------------------------------| | BERTScore | 91.28 | reddit | [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) | | Bleu_1 | 23.52 | reddit | [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) | | Bleu_2 | 14.84 | reddit | [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) | | Bleu_3 | 9.52 | reddit | [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) | | Bleu_4 | 6.37 | reddit | [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) | | METEOR | 19.69 | reddit | [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) | | MoverScore | 60.19 | reddit | [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) | | ROUGE_L | 23.41 | reddit | [lmqg/qg_squadshifts](https://huggingface.co/datasets/lmqg/qg_squadshifts) |
8959ab5b6d6a44b26da340b28b1adbaa
cc-by-4.0
['question generation']
false
Training hyperparameters The following hyperparameters were used during fine-tuning: - dataset_path: lmqg/qg_squadshifts - dataset_name: reddit - input_types: ['paragraph_answer'] - output_types: ['question'] - prefix_types: ['qg'] - model: t5-base - max_length: 512 - max_length_output: 32 - epoch: 5 - batch: 8 - lr: 0.0001 - fp16: False - random_seed: 1 - gradient_accumulation_steps: 8 - label_smoothing: 0.15 The full configuration can be found at [fine-tuning config file](https://huggingface.co/research-backup/t5-base-squadshifts-vanilla-reddit-qg/raw/main/trainer_config.json).
d6b6e0579ac32e4d13dfdfe617271110
mit
['generated_from_trainer']
false
xlmr-finetuned-ner This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the wikiann dataset. It achieves the following results on the evaluation set: - Loss: 0.1395 - Precision: 0.9044 - Recall: 0.9137 - F1: 0.9090 - Accuracy: 0.9649
4895dd3c52b07ee0af0bca5e797acc01
mit
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.4215 | 1.0 | 938 | 0.1650 | 0.8822 | 0.8781 | 0.8802 | 0.9529 | | 0.1559 | 2.0 | 1876 | 0.1412 | 0.9018 | 0.9071 | 0.9045 | 0.9631 | | 0.1051 | 3.0 | 2814 | 0.1395 | 0.9044 | 0.9137 | 0.9090 | 0.9649 |
2f72b41152969f6a8bb41895136febab
apache-2.0
['generated_from_trainer']
false
tiny-mlm-glue-cola-custom-tokenizer-target-glue-stsb This model is a fine-tuned version of [muhtasham/tiny-mlm-glue-cola-custom-tokenizer](https://huggingface.co/muhtasham/tiny-mlm-glue-cola-custom-tokenizer) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.3195 - Pearson: 0.6745 - Spearmanr: 0.6765
1be1754b0240b91cb72457f4489a4fe8
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Pearson | Spearmanr | |:-------------:|:-----:|:----:|:---------------:|:-------:|:---------:| | 3.3068 | 2.78 | 500 | 1.9896 | 0.3651 | 0.3938 | | 1.4915 | 5.56 | 1000 | 1.3053 | 0.6677 | 0.6745 | | 1.2119 | 8.33 | 1500 | 1.2683 | 0.6873 | 0.6896 | | 1.0589 | 11.11 | 2000 | 1.1930 | 0.6929 | 0.6911 | | 0.9145 | 13.89 | 2500 | 1.2757 | 0.6787 | 0.6807 | | 0.8423 | 16.67 | 3000 | 1.3657 | 0.6751 | 0.6800 | | 0.7599 | 19.44 | 3500 | 1.3195 | 0.6745 | 0.6765 |
4ae26ce176f35217883ba1b93a043da7
creativeml-openrail-m
[]
false
![logo](https://huggingface.co/joachimsallstrom/Glitch-Embedding/resolve/main/logo.png) [*EMBEDDING DOWNLOAD LINK*](https://huggingface.co/joachimsallstrom/Glitch-Embedding/resolve/main/glitch.pt) – Glitch is a finetuned embedding inspired by 80s and 90s VHS tape aesthetics (trained on SD 2.1 768 ema pruned). With it you can style images overall and affect skin, clothing and the general appearance of people, animals and more. ![sample1](https://huggingface.co/joachimsallstrom/Glitch-Embedding/resolve/main/sample1.jpg) Glitch generates both painted and photorealistic styles, where subjects and objects become more or less part of the VHS like glitch. It works great with tv and film references, ![sample2](https://huggingface.co/joachimsallstrom/Glitch-Embedding/resolve/main/sample2.jpg) and in an artsy, stylised sense to set a mood. ![sample3](https://huggingface.co/joachimsallstrom/Glitch-Embedding/resolve/main/sample3.jpg)
25288636aefd67e0c251c0abca0919cf
creativeml-openrail-m
[]
false
Install instructions and usage 1. Place either the [*glitch.pt*](https://huggingface.co/joachimsallstrom/Glitch-Embedding/resolve/main/glitch.pt) or [*glitch.png*](https://huggingface.co/joachimsallstrom/Glitch-Embedding/resolve/main/glitch.png) file in the embeddings folder of your Automatic1111 installation. 3. Trigger the style in the prompt by writing ***glitch***. To get great results a very basic negative prompting is suggested: ***ugly cartoon drawing, blurry, blurry, blurry, blurry*** This negative prompt is used throughout images shown in this presentation, which is a shorter, edited version of Stability AI’s recommendation for SD 2.x. Turning on highres fix is also higly recommended to achieve the best results.
7482c4dbe6336be2fedca8cd26e063ce
creativeml-openrail-m
[]
false
Example prompts and settings TV/movie still:<br> **glitch, close-up portrait of Millie Bobby Brown as Eleven, Stranger Things 1 9 8 2 movie still, Mitchell FC 65 Camera 35 mm, heavy grain**<br> Negative prompt: **ugly cartoon drawing, blurry, blurry, blurry, blurry**<br> _Steps: 30, Sampler: DPM++ SDE Karras, CFG scale: 7, Seed: 3729481696, Size: 1152x768, Model hash: 4bdfc29c, Denoising strength: 0.7_ Dancer ripping up the glitch with her hand:<br> **a close-up portrait of a wise Megleno-Romanians girl miner dancing in Spain, by glitch, 2d animation**<br> Negative prompt: **ugly cartoon drawing, blurry, blurry, blurry, blurry**<br> _Steps: 30, Sampler: DPM++ SDE Karras, CFG scale: 7.5, Seed: 888774459, Size: 1024x768, Model hash: 4bdfc29c, Model: sd_v2_v2-1_768-ema-pruned, Batch size: 5, Batch pos: 4, Denoising strength: 0.7_ ![sample3](https://huggingface.co/joachimsallstrom/Glitch-Embedding/resolve/main/stranger1.gif)
01758c2383015c8689eee2dc1ddf412f
creativeml-openrail-m
[]
false
Credit Thanks to [*masslevel*](https://twitter.com/masslevel?s=21&t=_O7DiffGgoNtZD33jECV_g) who has contributed with a large number of images and knowledge on prompt settings. Thanks also to Klinter for providing the gif animation.
ed3cb7dd39ccb53acdacdb92df3212e0
apache-2.0
['whisper-event', 'generated_from_trainer']
false
Whisper Large-v2 Nepali This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the mozilla-foundation/common_voice_11_0 ne-NP dataset. It achieves the following results on the evaluation set: - Loss: 1.5723 - Wer: 56.0976
707e84e35f8a1c5c40840f19ee100140
apache-2.0
['whisper-event', 'generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.0 | 999.0 | 1000 | 1.5723 | 56.0976 |
fda1e38178b0e24aa95d477e5bf16173
apache-2.0
['generated_from_trainer']
false
finetuning-misinfo-model-1000-Zhaohui This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7352 - Accuracy: 0.8226 - F1: 0.8571
819d2e62d7cb3a3f65de6bfd28f9adab
apache-2.0
['generated_from_trainer']
false
t5-small-finetuned-xsum This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.1681 - Rouge1: 60.7249 - Rouge2: 36.0768 - Rougel: 57.6761 - Rougelsum: 57.8618 - Gen Len: 17.9
47fdc5f5906264bc10dd10bf5d1b1005
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 100 - mixed_precision_training: Native AMP
7d6c0a42eb1e7a0134ee415aae5c4ed2
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | No log | 1.0 | 2 | 2.7817 | 13.2305 | 4.2105 | 11.0476 | 11.2063 | 13.0 | | No log | 2.0 | 4 | 2.7249 | 13.2305 | 4.2105 | 11.0476 | 11.2063 | 12.8 | | No log | 3.0 | 6 | 2.6053 | 13.1273 | 4.2105 | 10.9075 | 11.1008 | 13.1 | | No log | 4.0 | 8 | 2.4840 | 16.6829 | 6.2105 | 14.1984 | 14.6508 | 14.8 | | No log | 5.0 | 10 | 2.3791 | 16.6829 | 6.2105 | 14.1984 | 14.6508 | 14.8 | | No log | 6.0 | 12 | 2.2628 | 20.7742 | 9.5439 | 18.6218 | 18.9274 | 16.1 | | No log | 7.0 | 14 | 2.1714 | 20.7742 | 9.5439 | 18.6218 | 18.9274 | 16.1 | | No log | 8.0 | 16 | 2.0929 | 20.7742 | 9.5439 | 18.6218 | 18.9274 | 16.0 | | No log | 9.0 | 18 | 2.0069 | 20.7742 | 9.5439 | 18.6218 | 18.9274 | 16.0 | | No log | 10.0 | 20 | 1.9248 | 20.7742 | 8.4912 | 18.6218 | 18.9274 | 16.0 | | No log | 11.0 | 22 | 1.8535 | 20.7742 | 8.4912 | 18.6218 | 18.9274 | 16.0 | | No log | 12.0 | 24 | 1.7843 | 22.5821 | 10.8889 | 20.4396 | 20.9928 | 16.0 | | No log | 13.0 | 26 | 1.7115 | 22.5821 | 10.8889 | 20.4396 | 20.9928 | 16.0 | | No log | 14.0 | 28 | 1.6379 | 22.5821 | 10.8889 | 20.4396 | 20.9928 | 16.0 | | No log | 15.0 | 30 | 1.5689 | 22.5821 | 10.8889 | 20.4396 | 20.9928 | 16.0 | | No log | 16.0 | 32 | 1.5067 | 35.1364 | 17.6608 | 31.8254 | 31.8521 | 15.9 | | No log | 17.0 | 34 | 1.4543 | 41.7696 | 20.2005 | 38.8803 | 39.3886 | 16.9 | | No log | 18.0 | 36 | 1.4118 | 41.7696 | 20.2005 | 38.8803 | 39.3886 | 16.9 | | No log | 19.0 | 38 | 1.3789 | 41.5843 | 20.2005 | 38.6571 | 39.219 | 16.9 | | No log | 20.0 | 40 | 1.3543 | 41.5843 | 20.2005 | 38.6571 | 39.219 | 16.9 | | No log | 21.0 | 42 | 1.3332 | 42.6832 | 20.2005 | 39.7017 | 40.5046 | 16.9 | | No log | 22.0 | 44 | 1.3156 | 46.5429 | 22.7005 | 41.9156 | 42.7222 | 16.9 | | No log | 23.0 | 46 | 1.2999 | 49.5478 | 25.0555 | 44.8352 | 45.4884 | 16.9 | | No log | 24.0 | 48 | 1.2878 | 49.5478 | 25.0555 | 44.8352 | 45.4884 | 16.9 | | No log | 25.0 | 50 | 1.2777 | 49.5478 | 25.0555 | 44.8352 | 45.4884 | 16.9 | | No log | 26.0 | 52 | 1.2681 | 54.8046 | 28.7238 | 49.4767 | 49.699 | 17.4 | | No log | 27.0 | 54 | 1.2596 | 54.8046 | 28.7238 | 49.4767 | 49.699 | 17.4 | | No log | 28.0 | 56 | 1.2514 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 29.0 | 58 | 1.2450 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 30.0 | 60 | 1.2395 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 31.0 | 62 | 1.2340 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 32.0 | 64 | 1.2287 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 33.0 | 66 | 1.2233 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 34.0 | 68 | 1.2182 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 35.0 | 70 | 1.2127 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 36.0 | 72 | 1.2079 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 37.0 | 74 | 1.2035 | 58.1449 | 30.5444 | 52.7235 | 53.4075 | 18.9 | | No log | 38.0 | 76 | 1.1996 | 58.9759 | 30.5444 | 53.6606 | 54.2436 | 18.6 | | No log | 39.0 | 78 | 1.1962 | 58.9759 | 30.5444 | 53.6606 | 54.2436 | 18.6 | | No log | 40.0 | 80 | 1.1936 | 58.9759 | 30.5444 | 53.6606 | 54.2436 | 18.6 | | No log | 41.0 | 82 | 1.1912 | 58.9759 | 30.5444 | 53.6606 | 54.2436 | 18.6 | | No log | 42.0 | 84 | 1.1890 | 58.2807 | 30.5444 | 52.872 | 53.5594 | 18.5 | | No log | 43.0 | 86 | 1.1874 | 58.2807 | 30.5444 | 52.872 | 53.5594 | 18.5 | | No log | 44.0 | 88 | 1.1859 | 58.2807 | 30.5444 | 52.872 | 53.5594 | 18.5 | | No log | 45.0 | 90 | 1.1844 | 58.2807 | 30.5444 | 52.872 | 53.5594 | 18.5 | | No log | 46.0 | 92 | 1.1834 | 58.3968 | 30.5444 | 53.0602 | 53.7089 | 18.8 | | No log | 47.0 | 94 | 1.1822 | 58.3968 | 30.5444 | 53.0602 | 53.7089 | 18.8 | | No log | 48.0 | 96 | 1.1806 | 58.3968 | 30.5444 | 53.0602 | 53.7089 | 18.8 | | No log | 49.0 | 98 | 1.1786 | 58.3968 | 30.5444 | 53.0602 | 53.7089 | 18.8 | | No log | 50.0 | 100 | 1.1768 | 58.4517 | 31.303 | 54.18 | 54.6898 | 18.4 | | No log | 51.0 | 102 | 1.1761 | 58.4517 | 31.303 | 54.18 | 54.6898 | 18.4 | | No log | 52.0 | 104 | 1.1748 | 58.4517 | 31.303 | 54.18 | 54.6898 | 18.4 | | No log | 53.0 | 106 | 1.1743 | 58.4517 | 33.9839 | 55.5054 | 55.8799 | 18.4 | | No log | 54.0 | 108 | 1.1735 | 58.4517 | 33.9839 | 55.5054 | 55.8799 | 18.4 | | No log | 55.0 | 110 | 1.1731 | 58.4517 | 33.9839 | 55.5054 | 55.8799 | 18.4 | | No log | 56.0 | 112 | 1.1722 | 58.4517 | 33.9839 | 55.5054 | 55.8799 | 18.4 | | No log | 57.0 | 114 | 1.1714 | 58.4517 | 33.9839 | 55.5054 | 55.8799 | 18.4 | | No log | 58.0 | 116 | 1.1710 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 59.0 | 118 | 1.1702 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 60.0 | 120 | 1.1688 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 61.0 | 122 | 1.1682 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 62.0 | 124 | 1.1671 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 63.0 | 126 | 1.1669 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 64.0 | 128 | 1.1669 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 65.0 | 130 | 1.1668 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 66.0 | 132 | 1.1663 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 67.0 | 134 | 1.1665 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 68.0 | 136 | 1.1662 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 69.0 | 138 | 1.1663 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 70.0 | 140 | 1.1665 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 71.0 | 142 | 1.1664 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 72.0 | 144 | 1.1664 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 73.0 | 146 | 1.1662 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 74.0 | 148 | 1.1665 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 75.0 | 150 | 1.1662 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 76.0 | 152 | 1.1669 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 77.0 | 154 | 1.1668 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 78.0 | 156 | 1.1671 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 79.0 | 158 | 1.1674 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 80.0 | 160 | 1.1670 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 81.0 | 162 | 1.1671 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 82.0 | 164 | 1.1672 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 83.0 | 166 | 1.1675 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 84.0 | 168 | 1.1677 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 85.0 | 170 | 1.1677 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 86.0 | 172 | 1.1673 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 87.0 | 174 | 1.1673 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 88.0 | 176 | 1.1673 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 89.0 | 178 | 1.1673 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 90.0 | 180 | 1.1675 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 91.0 | 182 | 1.1675 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 92.0 | 184 | 1.1680 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 93.0 | 186 | 1.1680 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 94.0 | 188 | 1.1679 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 95.0 | 190 | 1.1679 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 96.0 | 192 | 1.1682 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 97.0 | 194 | 1.1681 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 98.0 | 196 | 1.1683 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 99.0 | 198 | 1.1683 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 | | No log | 100.0 | 200 | 1.1681 | 60.7249 | 36.0768 | 57.6761 | 57.8618 | 17.9 |
f3fc7b493b76db82dd4f67f6f2923462
apache-2.0
['generated_from_trainer']
false
roberta-base-bne-finetuned-amazon_reviews_multi This model is a fine-tuned version of [BSC-TeMU/roberta-base-bne](https://huggingface.co/BSC-TeMU/roberta-base-bne) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.2291 - Accuracy: 0.9343
7dc89f4faea1225cb2aca94645aa2f5b
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.1909 | 1.0 | 1250 | 0.1784 | 0.9295 | | 0.1013 | 2.0 | 2500 | 0.2291 | 0.9343 |
8ddb13f62d7aaedb61e3f39c6ced864a
apache-2.0
['generated_from_trainer']
false
swin-tiny-patch4-window7-224-finetuned-3e This model is a fine-tuned version of [microsoft/swin-tiny-patch4-window7-224](https://huggingface.co/microsoft/swin-tiny-patch4-window7-224) on the imagefolder dataset. It achieves the following results on the evaluation set: - Loss: 0.1065 - Accuracy: 0.9606
f8fcb9ba4260cb637ca0c23650fef83d
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.4549 | 1.0 | 527 | 0.2910 | 0.8857 | | 0.2838 | 2.0 | 1054 | 0.1524 | 0.9410 | | 0.254 | 3.0 | 1581 | 0.1065 | 0.9606 |
c58df9931be931f9f78fb34ef6c1221a
cc
['generated_from_trainer']
false
racism-finetuned-detests-prueba This model is a fine-tuned version of [davidmasip/racism](https://huggingface.co/davidmasip/racism) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.3034 - F1: 0.6222
153da09b3fb9c34a1bcc18b29ffa827b
cc
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 50 - eval_batch_size: 50 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10
64a1b15408f6200630aecbe627fc2126
cc
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.0001 | 1.0 | 49 | 1.0331 | 0.6667 | | 0.0 | 2.0 | 98 | 1.2473 | 0.5992 | | 0.0 | 3.0 | 147 | 1.2280 | 0.6227 | | 0.0 | 4.0 | 196 | 1.2530 | 0.6245 | | 0.0 | 5.0 | 245 | 1.2708 | 0.6222 | | 0.0 | 6.0 | 294 | 1.2827 | 0.6222 | | 0.0 | 7.0 | 343 | 1.2918 | 0.6222 | | 0.0 | 8.0 | 392 | 1.2982 | 0.6222 | | 0.0 | 9.0 | 441 | 1.3021 | 0.6222 | | 0.0 | 10.0 | 490 | 1.3034 | 0.6222 |
3e51cfbcb8fbee95818f085f469bccf2
apache-2.0
['generated_from_trainer']
false
tiny-mlm-glue-cola-from-scratch-custom-tokenizer-expand-vocab-target-glue-cola This model is a fine-tuned version of [muhtasham/tiny-mlm-glue-cola-from-scratch-custom-tokenizer-expand-vocab](https://huggingface.co/muhtasham/tiny-mlm-glue-cola-from-scratch-custom-tokenizer-expand-vocab) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6205 - Matthews Correlation: 0.0
3daad064a3d9ebd642a6928ca6acbc3c
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.6103 | 1.87 | 500 | 0.6214 | 0.0 | | 0.6073 | 3.73 | 1000 | 0.6197 | 0.0 | | 0.607 | 5.6 | 1500 | 0.6183 | 0.0 | | 0.6065 | 7.46 | 2000 | 0.6205 | 0.0 |
9bbcfac9fd6950731ea9e3224cfc7f23
apache-2.0
['generated_from_trainer']
false
all-roberta-large-v1-travel-2-16-5 This model is a fine-tuned version of [sentence-transformers/all-roberta-large-v1](https://huggingface.co/sentence-transformers/all-roberta-large-v1) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 2.1384 - Accuracy: 0.4289
de5b84aaee6ec5b137dd4f1cec669600
apache-2.0
[]
false
doc2query/all-with_prefix-t5-base-v1 This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)). It can be used for: - **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini. - **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
6ff0cbad0177c8f7c9515d0544b92270