license
stringlengths
2
30
tags
stringlengths
2
513
is_nc
bool
1 class
readme_section
stringlengths
201
597k
hash
stringlengths
32
32
openrail++
['text-to-image', 'stable-diffusion']
false
Samples Here are a few example images (generated with 50 steps). | pbr uneven stone wall | pbr dirt with weeds | pbr bright white marble | | --- | --- | --- | | ![](samples/pbr%20uneven%20stone%20wall.png) | ![](samples/pbr%20dirt%20with%20weeds.png) | ![](samples/pbr%20bright%20white%20marble.png) |
a93d65c5aaaef1a3e79669bb8a7a9c19
openrail++
['text-to-image', 'stable-diffusion']
false
Usage Use the token `pbr` in your prompts to invoke the style. This model was made for use in [Dream Textures](https://github.com/carson-katri/dream-textures), a Stable Diffusion add-on for Blender. You can also use it with [🧨 diffusers](https://github.com/huggingface/diffusers): ```python from diffusers import StableDiffusionPipeline import torch model_id = "dream-textures/texture-diffusion" pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16) pipe = pipe.to("cuda") prompt = "pbr brick wall" image = pipe(prompt).images[0] image.save("bricks.png") ```
d9053bf6ec6661f5e8df1ae3c5ee5528
openrail++
['text-to-image', 'stable-diffusion']
false
Training Details * Base Model: [stabilityai/stable-diffusion-2-base](https://huggingface.co/stabilityai/stable-diffusion-2-base) * Resolution: `512` * Prior Loss Weight: `1.0` * Class Prompt: `texture` * Batch Size: `1` * Learning Rate: `1e-6` * Precision: `fp16` * Steps: `4000` * GPU: Tesla T4
18f158d0d59ffeb818c6712bfb5949b6
apache-2.0
['generated_from_trainer']
false
openai/whisper-large-v2 This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.1534 - Wer: 145.6786
10e86c109d6bf21b0e9b24cffa020a6b
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.0799 | 2.03 | 500 | 0.1010 | 28.1322 | | 0.0239 | 5.01 | 1000 | 0.1388 | 161.0139 | | 0.0066 | 7.03 | 1500 | 0.1221 | 99.3747 | | 0.0007 | 10.01 | 2000 | 0.1295 | 250.8822 | | 0.0007 | 12.04 | 2500 | 0.1423 | 77.2203 | | 0.0003 | 15.02 | 3000 | 0.1480 | 149.4380 | | 0.0001 | 17.05 | 3500 | 0.1518 | 141.5842 | | 0.0001 | 20.02 | 4000 | 0.1534 | 145.6786 |
bda84a0a0c1dbec134db67dfd46dbaa3
apache-2.0
['generated_from_keras_callback']
false
kevinbram/testarbara This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 1.4900 - Train End Logits Accuracy: 0.6129 - Train Start Logits Accuracy: 0.5735 - Validation Loss: 1.1335 - Validation End Logits Accuracy: 0.6908 - Validation Start Logits Accuracy: 0.6545 - Epoch: 0
ea5c37d821d66af59a123257be163faa
apache-2.0
['generated_from_keras_callback']
false
Training results | Train Loss | Train End Logits Accuracy | Train Start Logits Accuracy | Validation Loss | Validation End Logits Accuracy | Validation Start Logits Accuracy | Epoch | |:----------:|:-------------------------:|:---------------------------:|:---------------:|:------------------------------:|:--------------------------------:|:-----:| | 1.4900 | 0.6129 | 0.5735 | 1.1335 | 0.6908 | 0.6545 | 0 |
ece5bfe5a74895e01ecfe2468f27eaa9
apache-2.0
['translation']
false
opus-mt-en-gl * source languages: en * target languages: gl * OPUS readme: [en-gl](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/en-gl/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2019-12-18.zip](https://object.pouta.csc.fi/OPUS-MT-models/en-gl/opus-2019-12-18.zip) * test set translations: [opus-2019-12-18.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/en-gl/opus-2019-12-18.test.txt) * test set scores: [opus-2019-12-18.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/en-gl/opus-2019-12-18.eval.txt)
a55d4d1e99a5485074c36947889b224d
creativeml-openrail-m
['text-to-image']
false
hulk_style_v1 Dreambooth model trained by sztanki with [Hugging Face Dreambooth Training Space](https://huggingface.co/spaces/multimodalart/dreambooth-training) with the v1-5 base model You run your new concept via `diffusers` [Colab Notebook for Inference](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_dreambooth_inference.ipynb). Don't forget to use the concept prompts! Sample pictures of: hulk (use that on your prompt) ![hulk 0](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%281%29.jpg)![hulk 1](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%282%29.jpg)![hulk 2](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%283%29.jpg)![hulk 3](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%284%29.jpg)![hulk 4](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%285%29.jpg)![hulk 5](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%286%29.jpg)![hulk 6](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%287%29.jpg)![hulk 7](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%288%29.jpg)![hulk 8](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%289%29.jpg)![hulk 9](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%2810%29.jpg)![hulk 10](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%2811%29.jpg)![hulk 11](https://huggingface.co/sztanki/hulk-style-v1/resolve/main/concept_images/hulk_style_%2812%29.jpg)
685b83023e6eeb477572a1bf7a651905
creativeml-openrail-m
[]
false
Prompt and settings for portraits: **brld harrison ford** _Steps: 50, Sampler: Euler a, CFG scale: 7, Seed: 3940025417 **brld morgan freeman** _Steps: 50, Sampler: Euler a, CFG scale: 7, Seed: 3940025417
a61984345e450db6a7eb1257f346823e
apache-2.0
['automatic-speech-recognition', 'fr']
false
exp_w2v2t_fr_vp-es_s980 Fine-tuned [facebook/wav2vec2-large-es-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-es-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (fr)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
df338ec11ecd6fc92fc1e0ace2981506
other
['generated_from_trainer']
false
segformer-b0-scene-parse-150 This model is a fine-tuned version of [nvidia/mit-b0](https://huggingface.co/nvidia/mit-b0) on the scene_parse_150 dataset. It achieves the following results on the evaluation set: - Loss: 4.4675 - Mean Iou: 0.0363 - Mean Accuracy: 0.1783 - Overall Accuracy: 0.2473 - Per Category Iou: [0.12781275754079446, 0.2655217061115745, nan, nan, 0.0, 0.0, 0.0, nan, nan, 0.0, nan, nan, nan, 0.0, nan, nan, 0.07806681040517459, nan, nan, nan, 0.0, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan] - Per Category Accuracy: [0.16194541672583893, 0.8526907084599392, nan, nan, 0.0, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.23336402444128232, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan]
0f2856f8c67e7b9072d427e91a8c62be
other
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Mean Iou | Mean Accuracy | Overall Accuracy | Per Category Iou | Per Category Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:|:-------------:|:----------------:|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:| | 3.6159 | 10.0 | 20 | 4.7805 | 0.0132 | 0.1045 | 0.1528 | [0.11163628179158226, 0.15792526006741817, nan, nan, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, nan, nan, nan, 0.0, nan, nan, 0.07479355092410539, 0.0, nan, 0.0, 0.0, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, 0.0, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan] | [0.1367338585045152, 0.4354909643371182, nan, nan, 0.0, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.15920314723361514, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan] | | 3.3116 | 20.0 | 40 | 4.6263 | 0.0214 | 0.1528 | 0.2002 | [0.12640023810812273, 0.20944680080510925, nan, nan, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, nan, nan, nan, 0.0, nan, nan, 0.09273170489030397, 0.0, nan, 0.0, 0.0, nan, nan, 0.0, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan] | [0.15609780547503488, 0.6011714377098992, nan, nan, 0.0, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.31204486481962, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan] | | 2.9608 | 30.0 | 60 | 4.4877 | 0.0299 | 0.1609 | 0.2376 | [0.13839157766016394, 0.2524577890734283, nan, nan, 0.0, 0.0, 0.0, nan, nan, 0.0, nan, nan, nan, 0.0, nan, nan, 0.05781044317683194, 0.0, nan, 0.0, 0.0, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan] | [0.1815182069553159, 0.7684711338557493, nan, nan, 0.0, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.17644596969950616, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan] | | 2.885 | 40.0 | 80 | 4.4865 | 0.0362 | 0.1822 | 0.2447 | [0.12383757908479229, 0.2653620793269231, nan, nan, 0.0, 0.0, 0.0, nan, nan, 0.0, nan, nan, nan, 0.0, nan, nan, 0.08116762230992627, 0.0, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan] | [0.15451895043731778, 0.8473932512394051, nan, nan, 0.0, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.27370888089060014, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan] | | 2.9919 | 50.0 | 100 | 4.4675 | 0.0363 | 0.1783 | 0.2473 | [0.12781275754079446, 0.2655217061115745, nan, nan, 0.0, 0.0, 0.0, nan, nan, 0.0, nan, nan, nan, 0.0, nan, nan, 0.07806681040517459, nan, nan, nan, 0.0, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan] | [0.16194541672583893, 0.8526907084599392, nan, nan, 0.0, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.23336402444128232, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, 0.0, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan] |
4d1d18fde8edda4297c7d40a4772f272
mit
['generated_from_trainer']
false
xlm-roberta-base-finetuned-panx-en This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the xtreme dataset. It achieves the following results on the evaluation set: - Loss: 0.7589 - F1: 0.6307
3b88ae9b591e8100f73d1db95c5821d3
mit
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.9453 | 1.0 | 1180 | 0.7589 | 0.6307 |
323a6db8b83cef3e3182d5a6e4c25f2e
wtfpl
[]
false
Cat picture embedding for 2.0. Trained on high quality Unsplash images, so it tends to prefer photorealism. Warning: the weights are quite strong. But, when tamed, it works great with stylistic embeddings like the last couple of images! Trained for 1500 steps, but added the 1000 steps one as well which also works pretty decently and is a bit less strong. ![03171-4193474301-cute [kittyhelper] wearing a tuxedo, sharp focus, photohelper, fur, extremely detailed.png](https://s3.amazonaws.com/moonup/production/uploads/1670332296701-6312579fc7577b68d90a7646.png) ![03164-4193474294-cute [kittyhelper] wearing a tuxedo, sharp focus, photohelper, fur, extremely detailed.png](https://s3.amazonaws.com/moonup/production/uploads/1670332307002-6312579fc7577b68d90a7646.png) ![03121-2828626995-kittyhelper, sharp focus, fur, extremely detailed,kipaki.png](https://s3.amazonaws.com/moonup/production/uploads/1670332315259-6312579fc7577b68d90a7646.png) ![03131-3693647088-kittyhelper wearing a tuxedo, sharp focus, fur, extremely detailed.png](https://s3.amazonaws.com/moonup/production/uploads/1670332324414-6312579fc7577b68d90a7646.png) ![03188-2279795372-knollingcase, cute kittypic, sharp focus, fur, extremely detailed,.png](https://s3.amazonaws.com/moonup/production/uploads/1670332259284-6312579fc7577b68d90a7646.png) ![03190-1679759101-knollingcase, kipaki, cute kittypic, sharp focus, fur, extremely detailed,.png](https://s3.amazonaws.com/moonup/production/uploads/1670332271419-6312579fc7577b68d90a7646.png) ![03199-318487485-kipaki, (((kittypic))), sharp focus, fur, extremely detailed,.png](https://s3.amazonaws.com/moonup/production/uploads/1670332274191-6312579fc7577b68d90a7646.png)
b08574f56bc714775a5d13aa47fc07ff
apache-2.0
['ner', 'zero-shot', 'information extruction']
false
Erlangshen-UniEX-RoBERTa-330M-Chinese - Github: [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM/tree/main/fengshen/examples/UniEX/) - Docs: [Fengshenbang-Docs](https://fengshenbang-doc.readthedocs.io/)
d9ce959d410742a28b0c7acb58da0a72
apache-2.0
['automatic-speech-recognition', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'kk', 'robust-speech-event', 'model_for_talk', 'hf-asr-leaderboard']
false
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - KK dataset. It achieves the following results on the evaluation set: - Loss: 0.7149 - Wer: 0.451
fb7798d08ffd4fc8b5e028874ad48537
apache-2.0
['automatic-speech-recognition', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'kk', 'robust-speech-event', 'model_for_talk', 'hf-asr-leaderboard']
false
Evaluation Commands 1. To evaluate on mozilla-foundation/common_voice_8_0 with test split python eval.py --model_id DrishtiSharma/wav2vec2-large-xls-r-300m-kk-with-LM --dataset mozilla-foundation/common_voice_8_0 --config kk --split test --log_outputs 2. To evaluate on speech-recognition-community-v2/dev_data Kazakh language isn't available in speech-recognition-community-v2/dev_data
d6856488e9c880ab99eeb277ea7acb45
apache-2.0
['automatic-speech-recognition', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'kk', 'robust-speech-event', 'model_for_talk', 'hf-asr-leaderboard']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000222 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 150.0 - mixed_precision_training: Native AMP
c6089d24d5163f3f212b1b729f95aa96
apache-2.0
['automatic-speech-recognition', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'kk', 'robust-speech-event', 'model_for_talk', 'hf-asr-leaderboard']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:------:|:----:|:---------------:|:------:| | 9.6799 | 9.09 | 200 | 3.6119 | 1.0 | | 3.1332 | 18.18 | 400 | 2.5352 | 1.005 | | 1.0465 | 27.27 | 600 | 0.6169 | 0.682 | | 0.3452 | 36.36 | 800 | 0.6572 | 0.607 | | 0.2575 | 45.44 | 1000 | 0.6527 | 0.578 | | 0.2088 | 54.53 | 1200 | 0.6828 | 0.551 | | 0.158 | 63.62 | 1400 | 0.7074 | 0.5575 | | 0.1309 | 72.71 | 1600 | 0.6523 | 0.5595 | | 0.1074 | 81.8 | 1800 | 0.7262 | 0.5415 | | 0.087 | 90.89 | 2000 | 0.7199 | 0.521 | | 0.0711 | 99.98 | 2200 | 0.7113 | 0.523 | | 0.0601 | 109.09 | 2400 | 0.6863 | 0.496 | | 0.0451 | 118.18 | 2600 | 0.6998 | 0.483 | | 0.0378 | 127.27 | 2800 | 0.6971 | 0.4615 | | 0.0319 | 136.36 | 3000 | 0.7119 | 0.4475 | | 0.0305 | 145.44 | 3200 | 0.7181 | 0.459 |
1793ef5eaad2549df55e5206af5f529d
apache-2.0
['sentence-transformers', 'feature-extraction', 'sentence-similarity', 'transformers']
false
sentence-transformers/distilbert-base-nli-stsb-mean-tokens This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
c32568a305569948c51e77ca089e5e78
apache-2.0
['sentence-transformers', 'feature-extraction', 'sentence-similarity', 'transformers']
false
Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('sentence-transformers/distilbert-base-nli-stsb-mean-tokens') embeddings = model.encode(sentences) print(embeddings) ```
a6e5c890612d1688c4bbcb6b2d0a0c4a
apache-2.0
['sentence-transformers', 'feature-extraction', 'sentence-similarity', 'transformers']
false
Load model from HuggingFace Hub tokenizer = AutoTokenizer.from_pretrained('sentence-transformers/distilbert-base-nli-stsb-mean-tokens') model = AutoModel.from_pretrained('sentence-transformers/distilbert-base-nli-stsb-mean-tokens')
4b2ce08b86eddac35f48360a5ab2f6b9
apache-2.0
['sentence-transformers', 'feature-extraction', 'sentence-similarity', 'transformers']
false
Evaluation Results For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers/distilbert-base-nli-stsb-mean-tokens)
406246d706f3142278f45f57f6011f22
mit
['singapore', 'sg', 'singlish', 'malaysia', 'ms', 'manglish', 'bert-large-uncased']
false
Model description Similar to [SingBert](https://huggingface.co/zanelim/singbert) but the large version, which was initialized from [BERT large uncased (whole word masking)](https://github.com/google-research/bert
836e1907bc4cb8c4d0cc1dd295b6061d
mit
['singapore', 'sg', 'singlish', 'malaysia', 'ms', 'manglish', 'bert-large-uncased']
false
How to use ```python >>> from transformers import pipeline >>> nlp = pipeline('fill-mask', model='zanelim/singbert-large-sg') >>> nlp("kopi c siew [MASK]") [{'sequence': '[CLS] kopi c siew dai [SEP]', 'score': 0.9003700017929077, 'token': 18765, 'token_str': 'dai'}, {'sequence': '[CLS] kopi c siew mai [SEP]', 'score': 0.0779474675655365, 'token': 14736, 'token_str': 'mai'}, {'sequence': '[CLS] kopi c siew. [SEP]', 'score': 0.0032227332703769207, 'token': 1012, 'token_str': '.'}, {'sequence': '[CLS] kopi c siew bao [SEP]', 'score': 0.0017727474914863706, 'token': 25945, 'token_str': 'bao'}, {'sequence': '[CLS] kopi c siew peng [SEP]', 'score': 0.0012526646023616195, 'token': 26473, 'token_str': 'peng'}] >>> nlp("one teh c siew dai, and one kopi [MASK]") [{'sequence': '[CLS] one teh c siew dai, and one kopi. [SEP]', 'score': 0.5249741077423096, 'token': 1012, 'token_str': '.'}, {'sequence': '[CLS] one teh c siew dai, and one kopi o [SEP]', 'score': 0.27349168062210083, 'token': 1051, 'token_str': 'o'}, {'sequence': '[CLS] one teh c siew dai, and one kopi peng [SEP]', 'score': 0.057190295308828354, 'token': 26473, 'token_str': 'peng'}, {'sequence': '[CLS] one teh c siew dai, and one kopi c [SEP]', 'score': 0.04022320732474327, 'token': 1039, 'token_str': 'c'}, {'sequence': '[CLS] one teh c siew dai, and one kopi? [SEP]', 'score': 0.01191170234233141, 'token': 1029, 'token_str': '?'}] >>> nlp("die [MASK] must try") [{'sequence': '[CLS] die die must try [SEP]', 'score': 0.9921030402183533, 'token': 3280, 'token_str': 'die'}, {'sequence': '[CLS] die also must try [SEP]', 'score': 0.004993876442313194, 'token': 2036, 'token_str': 'also'}, {'sequence': '[CLS] die liao must try [SEP]', 'score': 0.000317625846946612, 'token': 727, 'token_str': 'liao'}, {'sequence': '[CLS] die still must try [SEP]', 'score': 0.0002260878391098231, 'token': 2145, 'token_str': 'still'}, {'sequence': '[CLS] die i must try [SEP]', 'score': 0.00016935862367972732, 'token': 1045, 'token_str': 'i'}] >>> nlp("dont play [MASK] leh") [{'sequence': '[CLS] dont play play leh [SEP]', 'score': 0.9079819321632385, 'token': 2377, 'token_str': 'play'}, {'sequence': '[CLS] dont play punk leh [SEP]', 'score': 0.006846973206847906, 'token': 7196, 'token_str': 'punk'}, {'sequence': '[CLS] dont play games leh [SEP]', 'score': 0.004041737411171198, 'token': 2399, 'token_str': 'games'}, {'sequence': '[CLS] dont play politics leh [SEP]', 'score': 0.003728888463228941, 'token': 4331, 'token_str': 'politics'}, {'sequence': '[CLS] dont play cheat leh [SEP]', 'score': 0.0032805048394948244, 'token': 21910, 'token_str': 'cheat'}] >>> nlp("confirm plus [MASK]") {'sequence': '[CLS] confirm plus chop [SEP]', 'score': 0.9749826192855835, 'token': 24494, 'token_str': 'chop'}, {'sequence': '[CLS] confirm plus chopped [SEP]', 'score': 0.017554156482219696, 'token': 24881, 'token_str': 'chopped'}, {'sequence': '[CLS] confirm plus minus [SEP]', 'score': 0.002725469646975398, 'token': 15718, 'token_str': 'minus'}, {'sequence': '[CLS] confirm plus guarantee [SEP]', 'score': 0.000900257145985961, 'token': 11302, 'token_str': 'guarantee'}, {'sequence': '[CLS] confirm plus one [SEP]', 'score': 0.0004384620988275856, 'token': 2028, 'token_str': 'one'}] >>> nlp("catch no [MASK]") [{'sequence': '[CLS] catch no ball [SEP]', 'score': 0.9381157159805298, 'token': 3608, 'token_str': 'ball'}, {'sequence': '[CLS] catch no balls [SEP]', 'score': 0.060842301696538925, 'token': 7395, 'token_str': 'balls'}, {'sequence': '[CLS] catch no fish [SEP]', 'score': 0.00030917322146706283, 'token': 3869, 'token_str': 'fish'}, {'sequence': '[CLS] catch no breath [SEP]', 'score': 7.552534952992573e-05, 'token': 3052, 'token_str': 'breath'}, {'sequence': '[CLS] catch no tail [SEP]', 'score': 4.208395694149658e-05, 'token': 5725, 'token_str': 'tail'}] ``` Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('zanelim/singbert-large-sg') model = BertModel.from_pretrained("zanelim/singbert-large-sg") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained("zanelim/singbert-large-sg") model = TFBertModel.from_pretrained("zanelim/singbert-large-sg") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ```
5b4e12926314a351567026ce34389224
apache-2.0
['generated_from_trainer']
false
swin-kitchenware This model is a fine-tuned version of [microsoft/swin-tiny-patch4-window7-224](https://huggingface.co/microsoft/swin-tiny-patch4-window7-224) on the imagefolder dataset. It achieves the following results on the evaluation set: - Loss: 0.0860 - Accuracy: 0.9762
2c0bf108739a2185aff3b25dce7aa2f7
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 12
bab658837816a0a4a365a8c9d475f018
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6286 | 1.0 | 56 | 0.1875 | 0.9437 | | 0.3034 | 2.0 | 112 | 0.0921 | 0.9750 | | 0.2759 | 3.0 | 168 | 0.0992 | 0.9712 | | 0.23 | 4.0 | 224 | 0.0906 | 0.9687 | | 0.2484 | 5.0 | 280 | 0.0862 | 0.9725 | | 0.2434 | 6.0 | 336 | 0.0838 | 0.9762 | | 0.2398 | 7.0 | 392 | 0.0889 | 0.9762 | | 0.2094 | 8.0 | 448 | 0.0872 | 0.9750 | | 0.203 | 9.0 | 504 | 0.0888 | 0.9725 | | 0.1616 | 10.0 | 560 | 0.0820 | 0.9787 | | 0.1936 | 11.0 | 616 | 0.0846 | 0.9725 | | 0.1768 | 12.0 | 672 | 0.0860 | 0.9762 |
e1868ff0f2b0f5b2573a02db52fe2a56
mit
[]
false
aemond_targaryen_sandman on Stable Diffusion via Dreambooth trained on the [fast-DreamBooth.ipynb by TheLastBen](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook
827c5dc8db32e6b235aa6336b8a166e0
mit
[]
false
model by EdXD This your the Stable Diffusion model fine-tuned the aemond_targaryen_sandman concept taught to Stable Diffusion with Dreambooth. It can be used by modifying the `instance_prompt(s)`: **AemondHoD, Sandman2022** You can also train your own concepts and upload them to the library by using [the fast-DremaBooth.ipynb by TheLastBen](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb). And you can run your new concept via `diffusers`: [Colab Notebook for Inference](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_dreambooth_inference.ipynb), [Spaces with the Public Concepts loaded](https://huggingface.co/spaces/sd-dreambooth-library/stable-diffusion-dreambooth-concepts) Here are the images used for training this concept: Sandman2022 AemondHoD ![AemondHoD 0](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(1).jpg) ![AemondHoD 1](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(2).jpg) ![AemondHoD 2](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(3).jpg) ![AemondHoD 3](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(4).jpg) ![AemondHoD 4](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(5).jpg) ![AemondHoD 5](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(6).jpg) ![AemondHoD 6](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(7).jpg) ![AemondHoD 7](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(8).jpg) ![AemondHoD 8](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(9).jpg) ![AemondHoD 9](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(10).jpg) ![AemondHoD 10](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(11).jpg) ![AemondHoD 11](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(12).jpg) ![AemondHoD 12](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(13).jpg) ![AemondHoD 13](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(14).jpg) ![AemondHoD 14](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(15).jpg) ![AemondHoD 15](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(16).jpg) ![AemondHoD 16](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(17).jpg) ![AemondHoD 17](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(18).jpg) ![AemondHoD 18](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(19).jpg) ![AemondHoD 19](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(20).jpg) ![AemondHoD 20](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(21).jpg) ![AemondHoD 21](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(22).jpg) ![AemondHoD 22](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(23).jpg) ![AemondHoD 23](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/AemondHoD_(24).jpg) ![Sandman2022 24](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(1).jpg) ![Sandman2022 25](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(2).jpg) ![Sandman2022 26](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(3).jpg) ![Sandman2022 27](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(4).jpg) ![Sandman2022 28](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(5).jpg) ![Sandman2022 29](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(6).jpg) ![Sandman2022 30](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(7).jpg) ![Sandman2022 31](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(8).jpg) ![Sandman2022 32](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(9).jpg) ![Sandman2022 33](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(10).jpg) ![Sandman2022 34](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(11).jpg) ![Sandman2022 35](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(12).jpg) ![Sandman2022 36](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(13).jpg) ![Sandman2022 37](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(14).jpg) ![Sandman2022 38](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(15).jpg) ![Sandman2022 39](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(16).jpg) ![Sandman2022 40](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(17).jpg) ![Sandman2022 41](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(18).jpg) ![Sandman2022 42](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(19).jpg) ![Sandman2022 43](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(20).jpg) ![Sandman2022 44](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(21).jpg) ![Sandman2022 45](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(22).jpg) ![Sandman2022 46](https://huggingface.co/EdXD/aemond-targaryen-sandman/resolve/main/concept_images/Sandman2022_(23).jpg)
a8408e50882af3dbcb021d1c60ac47a3
cc-by-sa-4.0
['asteroid', 'audio', 'ConvTasNet', 'audio-to-audio']
false
.X9M69cLjJH4) Description: This model was trained by Joris Cosentino using the librimix recipe in [Asteroid](https://github.com/asteroid-team/asteroid). It was trained on the `sep_clean` task of the Libri2Mix dataset. Training config: ```yaml data: n_src: 2 sample_rate: 8000 segment: 3 task: sep_clean train_dir: data/wav8k/min/train-360 valid_dir: data/wav8k/min/dev filterbank: kernel_size: 16 n_filters: 512 stride: 8 masknet: bn_chan: 128 hid_chan: 512 mask_act: relu n_blocks: 8 n_repeats: 3 skip_chan: 128 optim: lr: 0.001 optimizer: adam weight_decay: 0.0 training: batch_size: 24 early_stop: True epochs: 200 half_lr: True num_workers: 2 ``` Results : On Libri2Mix min test set : ```yaml si_sdr: 14.764543634468069 si_sdr_imp: 14.764029375607246 sdr: 15.29337970745095 sdr_imp: 15.114146605113111 sir: 24.092904661115366 sir_imp: 23.913669683141528 sar: 16.06055906916849 sar_imp: -51.980784441287454 stoi: 0.9311142440593033 stoi_imp: 0.21817376142710482 ``` License notice: This work "ConvTasNet_Libri2Mix_sepclean_8k" is a derivative of [LibriSpeech ASR corpus](http://www.openslr.org/12) by Vassil Panayotov, used under [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/). "ConvTasNet_Libri2Mix_sepclean_8k" is licensed under [Attribution-ShareAlike 3.0 Unported](https://creativecommons.org/licenses/by-sa/3.0/) by Cosentino Joris.
df10c42a4dc746dbbfe7c239d3d399d8
cc-by-4.0
[]
false
Model description This is the T5-3B model for System 3 DREAM-FLUTE (all 4 dimensions), as described in our paper Just-DREAM-about-it: Figurative Language Understanding with DREAM-FLUTE, FigLang workshop @ EMNLP 2022 (Arxiv link: https://arxiv.org/abs/2210.16407) Systems 3: DREAM-FLUTE - Providing DREAM’s different dimensions as input context We adapt DREAM’s scene elaborations (Gu et al., 2022) for the figurative language understanding NLI task by using the DREAM model to generate elaborations for the premise and hypothesis separately. This allows us to investigate if similarities or differences in the scene elaborations for the premise and hypothesis will provide useful signals for entailment/contradiction label prediction and improving explanation quality. The input-output format is: ``` Input <Premise> <Premise-elaboration-from-DREAM> <Hypothesis> <Hypothesis-elaboration-from-DREAM> Output <Label> <Explanation> ``` where the scene elaboration dimensions from DREAM are: consequence, emotion, motivation, and social norm. We also consider a system incorporating all these dimensions as additional context. In this model, DREAM-FLUTE (all 4 dimensions), we use elaborations along all DREAM dimensions. For more details on DREAM, please refer to DREAM: Improving Situational QA by First Elaborating the Situation, NAACL 2022 (Arxiv link: https://arxiv.org/abs/2112.08656, ACL Anthology link: https://aclanthology.org/2022.naacl-main.82/).
865dfba924aed6c4d06f93e2a87d1303
cc-by-4.0
[]
false
How to use this model? We provide a quick example of how you can try out DREAM-FLUTE (all 4 dimensions) in our paper with just a few lines of code: ``` >>> from transformers import AutoTokenizer, AutoModelForSeq2SeqLM >>> model = AutoModelForSeq2SeqLM.from_pretrained("allenai/System3_DREAM_FLUTE_all_dimensions_FigLang2022") >>> tokenizer = AutoTokenizer.from_pretrained("t5-3b") >>> input_string = "Premise: I was really looking forward to camping but now it is going to rain so I won't go. [Premise - social norm] It's okay to be disappointed when plans change. [Premise - emotion] I (myself)'s emotion is disappointed. [Premise - motivation] I (myself)'s motivation is to stay home. [Premise - likely consequence] I will miss out on a great experience and be bored and sad. Hypothesis: I am absolutely elated at the prospects of getting drenched in the rain and then sleep in a wet tent just to have the experience of camping. [Hypothesis - social norm] It's good to want to have new experiences. [Hypothesis - emotion] I (myself)'s emotion is excited. [Hypothesis - motivation] I (myself)'s motivation is to have fun. [Hypothesis - likely consequence] I am so excited that I forget to bring a raincoat and my tent gets soaked. Is there a contradiction or entailment between the premise and hypothesis?" >>> input_ids = tokenizer.encode(input_string, return_tensors="pt") >>> output = model.generate(input_ids, max_length=200) >>> tokenizer.batch_decode(output, skip_special_tokens=True) ['Answer : Contradiction. Explanation : Camping in the rain is often associated with the prospect of getting wet and cold, so someone who is elated about it is not being rational.'] ```
6d2184db457988003d50468d037e43a8
cc-by-4.0
[]
false
Model details This model is a fine-tuned version of [t5-3b](https://huggingface.co/t5-3b). It achieves the following results on the evaluation set: - Loss: 0.7499 - Rouge1: 58.5551 - Rouge2: 38.5673 - Rougel: 52.3701 - Rougelsum: 52.335 - Gen Len: 40.7452
43ab571ffcc44f1cbf48cba39fb9dd51
cc-by-4.0
[]
false
Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | 0.992 | 0.33 | 1000 | 0.8911 | 39.9287 | 27.5817 | 38.2127 | 38.2042 | 19.0 | | 0.9022 | 0.66 | 2000 | 0.8409 | 40.8873 | 28.7963 | 39.16 | 39.1615 | 19.0 | | 0.8744 | 1.0 | 3000 | 0.7813 | 41.2617 | 29.5498 | 39.5857 | 39.5695 | 19.0 | | 0.5636 | 1.33 | 4000 | 0.7961 | 41.1429 | 30.2299 | 39.6592 | 39.6648 | 19.0 | | 0.5585 | 1.66 | 5000 | 0.7763 | 41.2581 | 30.0851 | 39.6859 | 39.68 | 19.0 | | 0.5363 | 1.99 | 6000 | 0.7499 | 41.8302 | 30.964 | 40.3059 | 40.2964 | 19.0 | | 0.3347 | 2.32 | 7000 | 0.8540 | 41.4633 | 30.6209 | 39.9933 | 39.9948 | 18.9954 | | 0.341 | 2.65 | 8000 | 0.8599 | 41.6576 | 31.0316 | 40.1466 | 40.1526 | 18.9907 | | 0.3531 | 2.99 | 9000 | 0.8368 | 42.05 | 31.6387 | 40.6239 | 40.6254 | 18.9907 |
9ed16f3f08db35fba8367c698e981c46
mit
['generated_from_trainer']
false
xlm-roberta-base-finetuned-panx-it This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the xtreme dataset. It achieves the following results on the evaluation set: - Loss: 0.5555 - F1: 0.7875
9217f0ecf6961efb084af6e170bb4fd8
mit
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.8118 | 1.0 | 1680 | 0.5555 | 0.7875 |
d3b591741938e1b78ca2cb7aff29b2d9
apache-2.0
['generated_from_trainer']
false
2nd-wav2vec2-l-xls-r-300m-turkish-test This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. It achieves the following results on the evaluation set: - Loss: 0.6019 - Wer: 0.4444
64c19b9ba2cdafbee497d36341ab394d
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 3.0522 | 3.67 | 400 | 0.7773 | 0.7296 | | 0.5369 | 7.34 | 800 | 0.6282 | 0.5888 | | 0.276 | 11.01 | 1200 | 0.5998 | 0.5330 | | 0.1725 | 14.68 | 1600 | 0.5859 | 0.4908 | | 0.1177 | 18.35 | 2000 | 0.6019 | 0.4444 |
fe05952f79fbfc8e7e2b0b2f5aefa996
mit
[]
false
Lula 13 on Stable Diffusion This is the `<lula-13>` concept taught to Stable Diffusion via Textual Inversion. You can load this concept into the [Stable Conceptualizer](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/stable_conceptualizer_inference.ipynb) notebook. You can also train your own concepts and load them into the concept libraries using [this notebook](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_textual_inversion_training.ipynb). Here is the new concept you will be able to use as an `object`: ![<lula-13> 0](https://huggingface.co/sd-concepts-library/lula-13/resolve/main/concept_images/3.jpeg) ![<lula-13> 1](https://huggingface.co/sd-concepts-library/lula-13/resolve/main/concept_images/6.jpeg) ![<lula-13> 2](https://huggingface.co/sd-concepts-library/lula-13/resolve/main/concept_images/0.jpeg) ![<lula-13> 3](https://huggingface.co/sd-concepts-library/lula-13/resolve/main/concept_images/5.jpeg) ![<lula-13> 4](https://huggingface.co/sd-concepts-library/lula-13/resolve/main/concept_images/1.jpeg) ![<lula-13> 5](https://huggingface.co/sd-concepts-library/lula-13/resolve/main/concept_images/2.jpeg) ![<lula-13> 6](https://huggingface.co/sd-concepts-library/lula-13/resolve/main/concept_images/4.jpeg)
9ae588937c7eeaf06bc6a1a1a20662f8
apache-2.0
['generated_from_trainer']
false
hf_fine_tune_hello_world This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the yelp_review_full dataset. It achieves the following results on the evaluation set: - Loss: 1.0142 - Accuracy: 0.592
6d9506cf9a75ed0110679915559c2bd2
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 125 | 1.0844 | 0.529 | | No log | 2.0 | 250 | 1.0022 | 0.58 | | No log | 3.0 | 375 | 1.0142 | 0.592 |
6485e2d5c2a2ceb6e11dd1e1c1de347f
apache-2.0
['generated_from_trainer', 't5-base']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP
461cb6fdb79c2295bbfe39256ea6fcd7
mit
['roberta-base', 'roberta-base-epoch_51']
false
RoBERTa, Intermediate Checkpoint - Epoch 51 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_51.
a62bac9de86563aaf91e4b0889a6e0cd
apache-2.0
['whisper-event', 'generated_from_trainer']
false
Whisper_large_Khmer This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the google/fleurs km_kh dataset. It achieves the following results on the evaluation set: - Loss: 0.5659 - Wer: 51.1683
4312e5497cd6f83a86fcd9179cb53eb3
apache-2.0
['whisper-event', 'generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.0002 | 50.0 | 500 | 0.5488 | 51.5328 | | 0.0001 | 100.0 | 1000 | 0.5659 | 51.1683 |
8ddac99f7f56065bd206d6d4dd32261a
apache-2.0
['audio', 'automatic-speech-recognition', 'speech', 'xlsr-fine-tuning-week']
false
We need to read the aduio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result**: 56.253154 %
40dd012fc80c8e6bccc33838a9c70a80
apache-2.0
['image-classification', 'generated_from_trainer']
false
exper_batch_32_e8 This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the sudo-s/herbier_mesuem1 dataset. It achieves the following results on the evaluation set: - Loss: 0.3520 - Accuracy: 0.9113
b9a07edf641d418b75e5edb9fb6a73ea
apache-2.0
['image-classification', 'generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 8 - mixed_precision_training: Apex, opt level O1
74d34851ae88fe56a03f68f5e572c8c2
apache-2.0
['image-classification', 'generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 3.3787 | 0.31 | 100 | 3.3100 | 0.3566 | | 2.3975 | 0.62 | 200 | 2.3196 | 0.5717 | | 1.5578 | 0.94 | 300 | 1.6764 | 0.6461 | | 1.0291 | 1.25 | 400 | 1.1713 | 0.7463 | | 0.8185 | 1.56 | 500 | 0.9292 | 0.7953 | | 0.6181 | 1.88 | 600 | 0.7732 | 0.8169 | | 0.3873 | 2.19 | 700 | 0.6877 | 0.8277 | | 0.2979 | 2.5 | 800 | 0.6250 | 0.8404 | | 0.2967 | 2.81 | 900 | 0.6151 | 0.8365 | | 0.1874 | 3.12 | 1000 | 0.5401 | 0.8608 | | 0.2232 | 3.44 | 1100 | 0.5032 | 0.8712 | | 0.1109 | 3.75 | 1200 | 0.4635 | 0.8774 | | 0.0539 | 4.06 | 1300 | 0.4495 | 0.8843 | | 0.0668 | 4.38 | 1400 | 0.4273 | 0.8951 | | 0.0567 | 4.69 | 1500 | 0.4427 | 0.8867 | | 0.0285 | 5.0 | 1600 | 0.4092 | 0.8955 | | 0.0473 | 5.31 | 1700 | 0.3720 | 0.9071 | | 0.0225 | 5.62 | 1800 | 0.3691 | 0.9063 | | 0.0196 | 5.94 | 1900 | 0.3775 | 0.9048 | | 0.0173 | 6.25 | 2000 | 0.3641 | 0.9040 | | 0.0092 | 6.56 | 2100 | 0.3551 | 0.9090 | | 0.008 | 6.88 | 2200 | 0.3591 | 0.9125 | | 0.0072 | 7.19 | 2300 | 0.3542 | 0.9121 | | 0.007 | 7.5 | 2400 | 0.3532 | 0.9106 | | 0.007 | 7.81 | 2500 | 0.3520 | 0.9113 |
b8c12b92bb7296d564e0f330b16d287e
gpl-3.0
['pytorch', 'lm-head', 'albert', 'zh']
false
Usage Please use BertTokenizerFast as tokenizer instead of AutoTokenizer. 請使用 BertTokenizerFast 而非 AutoTokenizer。 ``` from transformers import ( BertTokenizerFast, AutoModel, ) tokenizer = BertTokenizerFast.from_pretrained('bert-base-chinese') model = AutoModel.from_pretrained('ckiplab/albert-base-chinese') ``` For full usage and more information, please refer to https://github.com/ckiplab/ckip-transformers. 有關完整使用方法及其他資訊,請參見 https://github.com/ckiplab/ckip-transformers 。
716e4d5d429ad4a3db2ce6451489a6c9
apache-2.0
['spacy', 'token-classification']
false
DaCy large transformer DaCy is a Danish language processing framework with state-of-the-art pipelines as well as functionality for analysing Danish pipelines. DaCy's largest pipeline has achieved State-of-the-Art performance on Named entity recognition, part-of-speech tagging and dependency parsing for Danish on the DaNE dataset. Check out the [DaCy repository](https://github.com/centre-for-humanities-computing/DaCy) for material on how to use DaCy and reproduce the results. DaCy also contains guides on usage of the package as well as behavioural test for biases and robustness of Danish NLP pipelines. | Feature | Description | | --- | --- | | **Name** | `da_dacy_large_trf` | | **Version** | `0.1.0` | | **spaCy** | `>=3.1.1,<3.2.0` | | **Default Pipeline** | `transformer`, `morphologizer`, `parser`, `attribute_ruler`, `lemmatizer`, `ner` | | **Components** | `transformer`, `morphologizer`, `parser`, `attribute_ruler`, `lemmatizer`, `ner` | | **Vectors** | 0 keys, 0 unique vectors (0 dimensions) | | **Sources** | [UD Danish DDT v2.5](https://github.com/UniversalDependencies/UD_Danish-DDT) (Johannsen, Anders; Martínez Alonso, Héctor; Plank, Barbara)<br />[DaNE](https://github.com/alexandrainst/danlp/blob/master/docs/datasets.md
b849691a8dc6e30f591d9fb08a06f3c3
apache-2.0
['spacy', 'token-classification']
false
danish-dependency-treebank-dane) (Rasmus Hvingelby, Amalie B. Pauli, Maria Barrett, Christina Rosted, Lasse M. Lidegaard, Anders Søgaard)<br />[xlm-roberta-large](https://huggingface.co/xlm-roberta-large) (Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzmán, Edouard Grave, Myle Ott, Luke Zettlemoyer, Veselin Stoyanov) | | **License** | `Apache-2.0 License` | | **Author** | [Centre for Humanities Computing Aarhus](https://chcaa.io/
c4a51fd11a03213c67afa8f2ff91b076
apache-2.0
['spacy', 'token-classification']
false
Accuracy | Type | Score | | --- | --- | | `POS_ACC` | 98.70 | | `MORPH_ACC` | 98.49 | | `DEP_UAS` | 90.75 | | `DEP_LAS` | 88.38 | | `SENTS_P` | 96.09 | | `SENTS_R` | 95.74 | | `SENTS_F` | 95.91 | | `LEMMA_ACC` | 84.91 | | `ENTS_F` | 90.12 | | `ENTS_P` | 89.02 | | `ENTS_R` | 91.25 | | `TRANSFORMER_LOSS` | 1805626.49 | | `MORPHOLOGIZER_LOSS` | 111735.86 | | `PARSER_LOSS` | 8037491.27 | | `NER_LOSS` | 16634.46 |
292d474faab90e1de977817e49b4d9ff
apache-2.0
['spacy', 'token-classification']
false
Deterministic Augmentations Deterministic augmentations are augmentation which always yield the same result. | Augmentation | Part-of-speech tagging (Accuracy) | Morphological tagging (Accuracy) | Dependency Parsing (UAS) | Dependency Parsing (LAS) | Sentence segmentation (F1) | Lemmatization (Accuracy) | Named entity recognition (F1) | | --- | --- | --- | --- | --- | --- | --- | --- | | No augmentation | 0.985 | 0.979 | 0.906 | 0.881 | 0.986 | 0.844 | 0.839 | | Æøå Augmentation | 0.973 | 0.963 | 0.892 | 0.863 | 0.975 | 0.754 | 0.815 | | Lowercase | 0.981 | 0.975 | 0.902 | 0.876 | 0.93 | 0.848 | 0.788 | | No Spacing | 0.227 | 0.229 | 0.004 | 0.004 | 0.54 | 0.225 | 0.086 | | Abbreviated first names | 0.984 | 0.978 | 0.903 | 0.878 | 0.986 | 0.845 | 0.839 | | Input size augmentation 5 sentences | 0.986 | 0.981 | 0.904 | 0.88 | 0.97 | 0.844 | 0.847 | | Input size augmentation 10 sentences | 0.986 | 0.981 | 0.905 | 0.881 | 0.964 | 0.844 | 0.849 |
9d91d786788edc034202cdd9bd4bcfee
apache-2.0
['spacy', 'token-classification']
false
Stochastic Augmentations Stochastic augmentations are augmentation which are repeated mulitple times to estimate the effect of the augmentation. | Augmentation | Part-of-speech tagging (Accuracy) | Morphological tagging (Accuracy) | Dependency Parsing (UAS) | Dependency Parsing (LAS) | Sentence segmentation (F1) | Lemmatization (Accuracy) | Named entity recognition (F1) | | --- | --- | --- | --- | --- | --- | --- | --- | | Keystroke errors 2% | 0.949 (0.002) | 0.944 (0.002) | 0.868 (0.002) | 0.833 (0.002) | 0.965 (0.002) | 0.773 (0.002) | 0.775 (0.002) | | Keystroke errors 5% | 0.895 (0.003) | 0.893 (0.003) | 0.81 (0.003) | 0.76 (0.003) | 0.92 (0.003) | 0.68 (0.003) | 0.698 (0.003) | | Keystroke errors 15% | 0.705 (0.005) | 0.72 (0.005) | 0.6 (0.005) | 0.518 (0.005) | 0.801 (0.005) | 0.462 (0.005) | 0.506 (0.005) | | Danish names | 0.984 (0.0) | 0.979 (0.0) | 0.904 (0.0) | 0.879 (0.0) | 0.987 (0.0) | 0.847 (0.0) | 0.844 (0.0) | | Muslim names | 0.984 (0.0) | 0.979 (0.0) | 0.904 (0.0) | 0.879 (0.0) | 0.987 (0.0) | 0.847 (0.0) | 0.844 (0.0) | | Female names | 0.984 (0.0) | 0.979 (0.0) | 0.904 (0.0) | 0.879 (0.0) | 0.986 (0.0) | 0.847 (0.0) | 0.846 (0.0) | | Male names | 0.984 (0.0) | 0.979 (0.0) | 0.904 (0.0) | 0.879 (0.0) | 0.986 (0.0) | 0.846 (0.0) | 0.845 (0.0) | | Spacing Augmention 5% | 0.946 (0.002) | 0.941 (0.002) | 0.794 (0.002) | 0.771 (0.002) | 0.969 (0.002) | 0.812 (0.002) | 0.781 (0.002) | <details> <summary> Description of Augmenters </summary> **No augmentation:** Applies no augmentation to the DaNE test set. **Æøå Augmentation:** This augmentation replace the æ,ø, and å with their spelling variations ae, oe and aa respectively. **Lowercase:** This augmentation lowercases all text. **No Spacing:** This augmentation removed all spacing from the text. **Abbreviated first names:** This agmentation abbreviates the first names of entities. For instance 'Kenneth Enevoldsen' would turn to 'K. Enevoldsen'. **Keystroke errors 2%:** This agmentation simulate keystroke errors by replacing 2% of keys with a neighbouring key on a Danish QWERTY keyboard. As this agmentation is stochastic it is repeated 20 times to obtain a consistent estimate and the mean is provided with its standard deviation in parenthesis. **Keystroke errors 5%:** This agmentation simulate keystroke errors by replacing 5% of keys with a neighbouring key on a Danish QWERTY keyboard. As this agmentation is stochastic it is repeated 20 times to obtain a consistent estimate and the mean is provided with its standard deviation in parenthesis. **Keystroke errors 15%:** This agmentation simulate keystroke errors by replacing 15% of keys with a neighbouring key on a Danish QWERTY keyboard. As this agmentation is stochastic it is repeated 20 times to obtain a consistent estimate and the mean is provided with its standard deviation in parenthesis. **Danish names:** This agmentation replace all names with Danish names derived from Danmarks Statistik (2021). As this agmentation is stochastic it is repeated 20 times to obtain a consistent estimate and the mean is provided with its standard deviation in parenthesis. **Muslim names:** This agmentation replace all names with Muslim names derived from Meldgaard (2005). As this agmentation is stochastic it is repeated 20 times to obtain a consistent estimate and the mean is provided with its standard deviation in parenthesis. **Female names:** This agmentation replace all names with Danish female names derived from Danmarks Statistik (2021). As this agmentation is stochastic it is repeated 20 times to obtain a consistent estimate and the mean is provided with its standard deviation in parenthesis. **Male names:** This agmentation replace all names with Danish male names derived from Danmarks Statistik (2021). As this agmentation is stochastic it is repeated 20 times to obtain a consistent estimate and the mean is provided with its standard deviation in parenthesis. **Spacing Augmention 5%:** This agmentation replace all names with Danish male names derived from Danmarks Statistik (2021). As this agmentation is stochastic it is repeated 20 times to obtain a consistent estimate and the mean is provided with its standard deviation in parenthesis. </details> <br />
faf4d2e8772a0f245b98ae2ece7fdfc9
creativeml-openrail-m
['text-to-image', 'stable-diffusion']
false
tfmfurbase Dreambooth model trained by Deitsao with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook (BROKEN because I'm sleepy asf 😭) Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) it's a simple model which can generate tfm furs based on old transformice furs. it can help with fantasy if someone wants to suggest fur for tfm. (it can work a lot better when using mouse base from that game) Sample pictures of this concept: ![0](https://huggingface.co/Deitsao/tfmfurbase/resolve/main/sample_images/photo_2023-02-02_01-10-49.jpg) ![1](https://huggingface.co/Deitsao/tfmfurbase/resolve/main/sample_images/zrlOX47rRhM.jpg) ![2](https://huggingface.co/Deitsao/tfmfurbase/resolve/main/sample_images/photo_2023-02-02_01-08-27.jpg)
daa30595b1f33b798d22be0129523a0f
apache-2.0
['generated_from_trainer']
false
test-clm This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the bittensor train-v1.1.json dataset. It achieves the following results on the evaluation set: - Loss: 6.5199 - Accuracy: 0.1387
da10be0856c58f8b0337498f9e2c0d19
apache-2.0
['generated_from_trainer']
false
distilbert-base-uncased_fold_1_ternary_v1 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.1145 - F1: 0.7757
29abec69c901ff416eac7bcce00f6174
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 1.0 | 290 | 0.5580 | 0.7646 | | 0.555 | 2.0 | 580 | 0.5820 | 0.7670 | | 0.555 | 3.0 | 870 | 0.6683 | 0.7757 | | 0.2633 | 4.0 | 1160 | 0.9137 | 0.7844 | | 0.2633 | 5.0 | 1450 | 1.1367 | 0.7708 | | 0.1148 | 6.0 | 1740 | 1.2192 | 0.7757 | | 0.0456 | 7.0 | 2030 | 1.4035 | 0.7633 | | 0.0456 | 8.0 | 2320 | 1.5185 | 0.7658 | | 0.0226 | 9.0 | 2610 | 1.6126 | 0.7782 | | 0.0226 | 10.0 | 2900 | 1.7631 | 0.7658 | | 0.0061 | 11.0 | 3190 | 1.7279 | 0.7794 | | 0.0061 | 12.0 | 3480 | 1.8548 | 0.7584 | | 0.0076 | 13.0 | 3770 | 1.9052 | 0.7646 | | 0.0061 | 14.0 | 4060 | 1.9100 | 0.7757 | | 0.0061 | 15.0 | 4350 | 1.9280 | 0.7732 | | 0.0025 | 16.0 | 4640 | 1.9991 | 0.7745 | | 0.0025 | 17.0 | 4930 | 1.9960 | 0.7757 | | 0.0035 | 18.0 | 5220 | 2.0018 | 0.7708 | | 0.0015 | 19.0 | 5510 | 2.1099 | 0.7646 | | 0.0015 | 20.0 | 5800 | 2.1061 | 0.7695 | | 0.0022 | 21.0 | 6090 | 2.0941 | 0.7757 | | 0.0022 | 22.0 | 6380 | 2.0967 | 0.7794 | | 0.0005 | 23.0 | 6670 | 2.1133 | 0.7745 | | 0.0005 | 24.0 | 6960 | 2.1042 | 0.7782 | | 0.0021 | 25.0 | 7250 | 2.1145 | 0.7757 |
a237927bd9919f0fd96c8f885bbd60c8
cc-by-4.0
['espnet', 'audio', 'text-to-speech']
false
Demo: How to use in ESPnet2 ```bash cd espnet git checkout 047d0c474c18a87c205e566948410be16787e477 pip install -e . cd egs2/kss/tts1 ./run.sh --skip_data_prep false --skip_train true --download_model imdanboy/kss_tts_train_jets_raw_phn_null_g2pk_train.total_count.ave ```
cc95c14092b40f8d3a2407662fb8557b
cc-by-4.0
['espnet', 'audio', 'text-to-speech']
false
TTS config <details><summary>expand</summary> ``` config: conf/tuning/train_jets.yaml print_config: false log_level: INFO dry_run: false iterator_type: sequence output_dir: exp/tts_train_jets_raw_phn_null_g2pk ngpu: 1 seed: 777 num_workers: 4 num_att_plot: 3 dist_backend: nccl dist_init_method: env:// dist_world_size: 4 dist_rank: 0 local_rank: 0 dist_master_addr: localhost dist_master_port: 52809 dist_launcher: null multiprocessing_distributed: true unused_parameters: true sharded_ddp: false cudnn_enabled: true cudnn_benchmark: false cudnn_deterministic: false collect_stats: false write_collected_feats: false max_epoch: 1000 patience: null val_scheduler_criterion: - valid - loss early_stopping_criterion: - valid - loss - min best_model_criterion: - - valid - text2mel_loss - min - - train - text2mel_loss - min - - train - total_count - max keep_nbest_models: 5 nbest_averaging_interval: 0 grad_clip: -1 grad_clip_type: 2.0 grad_noise: false accum_grad: 1 no_forward_run: false resume: true train_dtype: float32 use_amp: false log_interval: 50 use_matplotlib: true use_tensorboard: true use_wandb: false wandb_project: null wandb_id: null wandb_entity: null wandb_name: null wandb_model_log_interval: -1 detect_anomaly: false pretrain_path: null init_param: [] ignore_init_mismatch: false freeze_param: [] num_iters_per_epoch: 1000 batch_size: 20 valid_batch_size: null batch_bins: 2000000 valid_batch_bins: null train_shape_file: - exp/tts_stats_raw_phn_null_g2pk/train/text_shape.phn - exp/tts_stats_raw_phn_null_g2pk/train/speech_shape valid_shape_file: - exp/tts_stats_raw_phn_null_g2pk/valid/text_shape.phn - exp/tts_stats_raw_phn_null_g2pk/valid/speech_shape batch_type: numel valid_batch_type: null fold_length: - 150 - 204800 sort_in_batch: descending sort_batch: descending multiple_iterator: false chunk_length: 500 chunk_shift_ratio: 0.5 num_cache_chunks: 1024 train_data_path_and_name_and_type: - - dump/raw/tr_no_dev/text - text - text - - dump/raw/tr_no_dev/wav.scp - speech - sound - - exp/tts_stats_raw_phn_null_g2pk/train/collect_feats/pitch.scp - pitch - npy - - exp/tts_stats_raw_phn_null_g2pk/train/collect_feats/energy.scp - energy - npy valid_data_path_and_name_and_type: - - dump/raw/dev/text - text - text - - dump/raw/dev/wav.scp - speech - sound - - exp/tts_stats_raw_phn_null_g2pk/valid/collect_feats/pitch.scp - pitch - npy - - exp/tts_stats_raw_phn_null_g2pk/valid/collect_feats/energy.scp - energy - npy allow_variable_data_keys: false max_cache_size: 0.0 max_cache_fd: 32 valid_max_cache_size: null optim: adamw optim_conf: lr: 0.0002 betas: - 0.8 - 0.99 eps: 1.0e-09 weight_decay: 0.0 scheduler: exponentiallr scheduler_conf: gamma: 0.999875 optim2: adamw optim2_conf: lr: 0.0002 betas: - 0.8 - 0.99 eps: 1.0e-09 weight_decay: 0.0 scheduler2: exponentiallr scheduler2_conf: gamma: 0.999875 generator_first: true token_list: - <blank> - <unk> - '' - ᅡ - ᅵ - ᄋ - ᅳ - ᄀ - ᅥ - ᄂ - ᆫ - ᄅ - ᄌ - ᄉ - ᅩ - ᆯ - ᄆ - . - ᅮ - ᄃ - ᄒ - ᅦ - ᆼ - ᅢ - ᄇ - ᅭ - ᅧ - ᄊ - ᆷ - ᄄ - ᆮ - ᄎ - ᄁ - ᆨ - ᄑ - ᄐ - ᅪ - ᄏ - '?' - ᄍ - ᆸ - ᅬ - ᅣ - ᅴ - ᅯ - ᅨ - ᄈ - ᅱ - ᅲ - ᅫ - ',' - '!' - ᅤ - ':' - ᅰ - '''' - '-' - '"' - / - I - M - F - E - S - C - A - B - ㅇ - <sos/eos> odim: null model_conf: {} use_preprocessor: true token_type: phn bpemodel: null non_linguistic_symbols: null cleaner: null g2p: g2pk feats_extract: fbank feats_extract_conf: n_fft: 1024 hop_length: 256 win_length: null fs: 24000 fmin: 0 fmax: null n_mels: 80 normalize: global_mvn normalize_conf: stats_file: exp/tts_stats_raw_phn_null_g2pk/train/feats_stats.npz tts: jets tts_conf: generator_type: jets_generator generator_params: adim: 256 aheads: 2 elayers: 4 eunits: 1024 dlayers: 4 dunits: 1024 positionwise_layer_type: conv1d positionwise_conv_kernel_size: 3 duration_predictor_layers: 2 duration_predictor_chans: 256 duration_predictor_kernel_size: 3 use_masking: true encoder_normalize_before: true decoder_normalize_before: true encoder_type: transformer decoder_type: transformer conformer_rel_pos_type: latest conformer_pos_enc_layer_type: rel_pos conformer_self_attn_layer_type: rel_selfattn conformer_activation_type: swish use_macaron_style_in_conformer: true use_cnn_in_conformer: true conformer_enc_kernel_size: 7 conformer_dec_kernel_size: 31 init_type: xavier_uniform transformer_enc_dropout_rate: 0.2 transformer_enc_positional_dropout_rate: 0.2 transformer_enc_attn_dropout_rate: 0.2 transformer_dec_dropout_rate: 0.2 transformer_dec_positional_dropout_rate: 0.2 transformer_dec_attn_dropout_rate: 0.2 pitch_predictor_layers: 5 pitch_predictor_chans: 256 pitch_predictor_kernel_size: 5 pitch_predictor_dropout: 0.5 pitch_embed_kernel_size: 1 pitch_embed_dropout: 0.0 stop_gradient_from_pitch_predictor: true energy_predictor_layers: 2 energy_predictor_chans: 256 energy_predictor_kernel_size: 3 energy_predictor_dropout: 0.5 energy_embed_kernel_size: 1 energy_embed_dropout: 0.0 stop_gradient_from_energy_predictor: false generator_out_channels: 1 generator_channels: 512 generator_global_channels: -1 generator_kernel_size: 7 generator_upsample_scales: - 8 - 8 - 2 - 2 generator_upsample_kernel_sizes: - 16 - 16 - 4 - 4 generator_resblock_kernel_sizes: - 3 - 7 - 11 generator_resblock_dilations: - - 1 - 3 - 5 - - 1 - 3 - 5 - - 1 - 3 - 5 generator_use_additional_convs: true generator_bias: true generator_nonlinear_activation: LeakyReLU generator_nonlinear_activation_params: negative_slope: 0.1 generator_use_weight_norm: true segment_size: 64 idim: 69 odim: 80 discriminator_type: hifigan_multi_scale_multi_period_discriminator discriminator_params: scales: 1 scale_downsample_pooling: AvgPool1d scale_downsample_pooling_params: kernel_size: 4 stride: 2 padding: 2 scale_discriminator_params: in_channels: 1 out_channels: 1 kernel_sizes: - 15 - 41 - 5 - 3 channels: 128 max_downsample_channels: 1024 max_groups: 16 bias: true downsample_scales: - 2 - 2 - 4 - 4 - 1 nonlinear_activation: LeakyReLU nonlinear_activation_params: negative_slope: 0.1 use_weight_norm: true use_spectral_norm: false follow_official_norm: false periods: - 2 - 3 - 5 - 7 - 11 period_discriminator_params: in_channels: 1 out_channels: 1 kernel_sizes: - 5 - 3 channels: 32 downsample_scales: - 3 - 3 - 3 - 3 - 1 max_downsample_channels: 1024 bias: true nonlinear_activation: LeakyReLU nonlinear_activation_params: negative_slope: 0.1 use_weight_norm: true use_spectral_norm: false generator_adv_loss_params: average_by_discriminators: false loss_type: mse discriminator_adv_loss_params: average_by_discriminators: false loss_type: mse feat_match_loss_params: average_by_discriminators: false average_by_layers: false include_final_outputs: true mel_loss_params: fs: 24000 n_fft: 1024 hop_length: 256 win_length: null window: hann n_mels: 80 fmin: 0 fmax: null log_base: null lambda_adv: 1.0 lambda_mel: 45.0 lambda_feat_match: 2.0 lambda_var: 1.0 lambda_align: 2.0 sampling_rate: 24000 cache_generator_outputs: true pitch_extract: dio pitch_extract_conf: reduction_factor: 1 use_token_averaged_f0: false fs: 24000 n_fft: 1024 hop_length: 256 f0max: 400 f0min: 80 pitch_normalize: global_mvn pitch_normalize_conf: stats_file: exp/tts_stats_raw_phn_null_g2pk/train/pitch_stats.npz energy_extract: energy energy_extract_conf: reduction_factor: 1 use_token_averaged_energy: false fs: 24000 n_fft: 1024 hop_length: 256 win_length: null energy_normalize: global_mvn energy_normalize_conf: stats_file: exp/tts_stats_raw_phn_null_g2pk/train/energy_stats.npz required: - output_dir - token_list version: '202204' distributed: true ``` </details>
c5527f5fcb7f0a9f37d79dfefe6d9ba0
apache-2.0
['generated_from_trainer']
false
vit-base-highways-2 This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the highways-hacktum dataset. It achieves the following results on the evaluation set: - Loss: 0.2196 - Accuracy: 0.96
d15ea82dd74e291cad101320743eef73
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2
ee741037c7ec32bf5a3fe08a713c12d6
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.0012 | 1.59 | 100 | 0.2196 | 0.96 |
af48739fd34f01dcc490acab01021c7a
apache-2.0
['catalan']
false
roberta-base-ca-finetuned-catalonia-independence-detector This model is a fine-tuned version of [BSC-TeMU/roberta-base-ca](https://huggingface.co/BSC-TeMU/roberta-base-ca) on the catalonia_independence dataset. It achieves the following results on the evaluation set: - Loss: 0.6065 - Accuracy: 0.7612 <details>
5a736b70d252c7cd60b46e2094840f91
apache-2.0
['catalan']
false
Training and evaluation data The data was collected over 12 days during February and March of 2019 from tweets posted in Barcelona, and during September of 2018 from tweets posted in the town of Terrassa, Catalonia. Each corpus is annotated with three classes: AGAINST, FAVOR and NEUTRAL, which express the stance towards the target - independence of Catalonia.
b1c63e07636f9fc56ad24000bbf3d73d
apache-2.0
['catalan']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 377 | 0.6311 | 0.7453 | | 0.7393 | 2.0 | 754 | 0.6065 | 0.7612 | | 0.5019 | 3.0 | 1131 | 0.6340 | 0.7547 | | 0.3837 | 4.0 | 1508 | 0.6777 | 0.7597 | | 0.3837 | 5.0 | 1885 | 0.7232 | 0.7582 | </details>
c9bc0f19020e26d0f155ac451eff7667
apache-2.0
['catalan']
false
Model in action 🚀 Fast usage with **pipelines**: ```python from transformers import pipeline model_path = "JonatanGk/roberta-base-ca-finetuned-catalonia-independence-detector" independence_analysis = pipeline("text-classification", model=model_path, tokenizer=model_path) independence_analysis( "Assegura l'expert que en un 46% els catalans s'inclouen dins del que es denomina com el doble sentiment identitari. És a dir, se senten tant catalans com espanyols. 1 de cada cinc, en canvi, té un sentiment excloent, només se senten catalans, i un 4% sol espanyol." )
6a1f0bfe74126926648728b090e39fbb
apache-2.0
['catalan']
false
Output: [{'label': 'FAVOR', 'score': 0.9040119647979736}] ``` [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/JonatanGk/Shared-Colab/blob/master/Catalonia_independence_Detector_(CATALAN).ipynb
fee5493dff137ae28cc4b3f482d7edfc
apache-2.0
['catalan']
false
Citation Thx to HF.co & [@lewtun](https://github.com/lewtun) for Dataset ;) > Special thx to [Manuel Romero/@mrm8488](https://huggingface.co/mrm8488) as my mentor & R.C. > Created by [Jonatan Luna](https://JonatanGk.github.io) | [LinkedIn](https://www.linkedin.com/in/JonatanGk/)
eb556b451160f0228fc16982d2638ba9
mit
['generated_from_trainer']
false
DistillBerTurk_15_epoch This model is a fine-tuned version of [dbmdz/distilbert-base-turkish-cased](https://huggingface.co/dbmdz/distilbert-base-turkish-cased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0549 - Accuracy: 0.9931
12a0983d8321075fc801780f65fbc2fc
mit
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 50 | 0.6766 | 0.625 | | No log | 2.0 | 100 | 0.3955 | 0.9583 | | No log | 3.0 | 150 | 0.0874 | 0.9792 | | No log | 4.0 | 200 | 0.0446 | 0.9861 | | No log | 5.0 | 250 | 0.0725 | 0.9792 | | No log | 6.0 | 300 | 0.0352 | 0.9931 | | No log | 7.0 | 350 | 0.0452 | 0.9931 | | No log | 8.0 | 400 | 0.0442 | 0.9931 | | No log | 9.0 | 450 | 0.0499 | 0.9931 | | 0.1537 | 10.0 | 500 | 0.0514 | 0.9931 | | 0.1537 | 11.0 | 550 | 0.0542 | 0.9931 | | 0.1537 | 12.0 | 600 | 0.0545 | 0.9931 | | 0.1537 | 13.0 | 650 | 0.0548 | 0.9931 | | 0.1537 | 14.0 | 700 | 0.0550 | 0.9931 | | 0.1537 | 15.0 | 750 | 0.0549 | 0.9931 |
c9f83b944700be98b9409baea85e275a
apache-2.0
['generated_from_trainer']
false
distilbert-base-uncased-finetuned-imdb This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset. It achieves the following results on the evaluation set: - Loss: 2.2489
839a990efa5752c1b42e0b7c0865371f
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.4463 | 1.0 | 782 | 2.2692 | | 2.3895 | 2.0 | 1564 | 2.2460 | | 2.3631 | 3.0 | 2346 | 2.2205 |
f3b0030c48f9492e605dc7ac1d5f8528
apache-2.0
['generated_from_trainer']
false
Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 16 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 2524 - mixed_precision_training: Native AMP
c060912cfeee7092f5ca0959da562681
apache-2.0
['generated_from_trainer']
false
Full config {'dataset': {'conditional_training_config': {'aligned_prefix': '<|aligned|>', 'drop_token_fraction': 0.1, 'misaligned_prefix': '<|misaligned|>', 'threshold': 0}, 'datasets': ['kejian/codeparrot-train-more-filter-3.3b-cleaned'], 'is_split_by_sentences': True, 'skip_tokens': 2969174016}, 'generation': {'batch_size': 128, 'force_call_on': [503], 'metrics_configs': [{}, {'n': 1}, {}], 'scenario_configs': [{'display_as_html': True, 'generate_kwargs': {'bad_words_ids': [[32769]], 'do_sample': True, 'eos_token_id': 0, 'max_length': 640, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'unconditional', 'num_hits_threshold': 0, 'num_samples': 4096, 'prefix': '<|aligned|>', 'use_prompt_for_scoring': False}], 'scorer_config': {}}, 'kl_gpt3_callback': {'force_call_on': [503], 'gpt3_kwargs': {'model_name': 'code-cushman-001'}, 'max_tokens': 64, 'num_samples': 4096, 'prefix': '<|aligned|>', 'should_insert_prefix': True}, 'model': {'from_scratch': False, 'gpt2_config_kwargs': {'reorder_and_upcast_attn': True, 'scale_attn_by': True}, 'model_kwargs': {'revision': '9cdfa11a07b00726ddfdabb554de05b29d777db3'}, 'num_additional_tokens': 2, 'path_or_name': 'kejian/grainy-pep8'}, 'objective': {'name': 'MLE'}, 'tokenizer': {'path_or_name': 'codeparrot/codeparrot-small', 'special_tokens': ['<|aligned|>', '<|misaligned|>']}, 'training': {'dataloader_num_workers': 0, 'effective_batch_size': 128, 'evaluation_strategy': 'no', 'fp16': True, 'hub_model_id': 'naughty_davinci', 'hub_strategy': 'all_checkpoints', 'learning_rate': 0.0001, 'logging_first_step': True, 'logging_steps': 10, 'num_tokens': 3300000000, 'output_dir': 'training_output', 'per_device_train_batch_size': 8, 'push_to_hub': True, 'remove_unused_columns': False, 'save_steps': 100, 'save_strategy': 'steps', 'seed': 42, 'tokens_already_seen': 2969174016, 'warmup_ratio': 0.01, 'weight_decay': 0.1}}
b7decb98b9279640b1aef2b1a771e810
cc-by-4.0
['norwegian', 'bert']
false
Results |**Model** | **NoRec** | **NorNe-NB**| **NorNe-NN** | **NorDial** | **DaNe** | **Da-Angry-Tweets** | |:-----------|------------:|------------:|------------:|------------:|------------:|------------:| |roberta-base (English) | 51.77 | 79.01/79.53| 79.79/83.02 | 67.18| 75.44/78.07 | 55.51 | |mBERT-cased | 63.91 | 83.72/86.12| 83.05/87.12 | 66.23| 80.00/81.43 | 57.67 | |nb-bert-base | 75.60 |**91.98**/**92.95** |**90.93**/**94.06**|69.39| 81.95/84.83| 64.18| |notram-bert-norwegian-cased | 72.47 | 91.77/93.12|89.79/93.70| **78.55**| **83.69**/**86.55**| **64.19** | |notram-bert-norwegian-uncased | 73.47 | 89.28/91.61 |87.23/90.23 |74.21 | 80.29/82.31| 61.18| |notram-bert-norwegian-cased-pod | **76.18** | 91.24/92.24| 90.88/93.21| 76.21| 81.82/84.99| 62.16 | |nb-roberta-base | 68.77 |87.99/89.43 | 85.43/88.66| 76.34| 75.91/77.94| 61.50 | |nb-roberta-base-scandinavian | 67.88 | 87.73/89.14| 87.39/90.92| 74.81| 76.22/78.66 | 63.37 | |nb-roberta-base-v2-200k | 46.87 | 85.57/87.04| - | 64.99| - | - | |test_long_w5 200k| 60.48 | 88.00/90:00 | 83.93/88.45 | 68.41 |75.22/78.50| 57.95 | |test_long_w5_roberta_tokenizer 200k| 63.51| 86.28/87.77| 84.95/88.61 | 69.86 | 71.31/74.27 | 59.96 | |test_long_w5_roberta_tokenizer 400k| 59.76 |87.39/89.06 | 85.16/89.01 | 71.46 | 72.39/75.65| 39.73 | |test_long_w5_dataset 400k| 66.80 | 86.52/88.55 | 82.81/86.76 | 66.94 | 71.47/74.20| 55.25 | |test_long_w5_dataset 600k| 67.37 | 89.98/90.95 | 84.53/88.37 | 66.84 | 75.14/76.50| 57.47 | |roberta-jan-128_ncc - 400k - 128| 67.79 | 91.45/92.33 | 86.41/90.19 | 67.20 | 81.00/82.39| 59.65 | |roberta-jan-128_ncc - 1000k - 128| 68.17 | 89.34/90.74 | 86.89/89.87 | 68.41 | 80.30/82.17| 61.63 |
3bcf0d48d839b8055fd5b8bf9d8e078d
apache-2.0
['automatic-speech-recognition', 'multilingual_librispeech', 'generated_from_trainer']
false
wav2vec2-xlsr-53-300m-mls-german-ft This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the MULTILINGUAL_LIBRISPEECH - GERMAN 10h dataset. It achieves the following results on the evaluation set: - Loss: 0.2219 - Wer: 0.1288
5975ccee37e92c6dd3dc052ca0e3de01
apache-2.0
['automatic-speech-recognition', 'multilingual_librispeech', 'generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:------:|:-----:|:---------------:|:------:| | 2.9888 | 7.25 | 500 | 2.9192 | 1.0 | | 2.9313 | 14.49 | 1000 | 2.8698 | 1.0 | | 1.068 | 21.74 | 1500 | 0.2647 | 0.2565 | | 0.8151 | 28.99 | 2000 | 0.2067 | 0.1719 | | 0.764 | 36.23 | 2500 | 0.1975 | 0.1568 | | 0.7332 | 43.48 | 3000 | 0.1812 | 0.1463 | | 0.5952 | 50.72 | 3500 | 0.1923 | 0.1428 | | 0.6655 | 57.97 | 4000 | 0.1900 | 0.1404 | | 0.574 | 65.22 | 4500 | 0.1822 | 0.1370 | | 0.6211 | 72.46 | 5000 | 0.1937 | 0.1355 | | 0.5883 | 79.71 | 5500 | 0.1872 | 0.1335 | | 0.5666 | 86.96 | 6000 | 0.1874 | 0.1324 | | 0.5526 | 94.2 | 6500 | 0.1998 | 0.1368 | | 0.5671 | 101.45 | 7000 | 0.2054 | 0.1365 | | 0.5514 | 108.7 | 7500 | 0.1987 | 0.1340 | | 0.5382 | 115.94 | 8000 | 0.2104 | 0.1344 | | 0.5819 | 123.19 | 8500 | 0.2125 | 0.1334 | | 0.5277 | 130.43 | 9000 | 0.2063 | 0.1330 | | 0.4626 | 137.68 | 9500 | 0.2105 | 0.1310 | | 0.5842 | 144.93 | 10000 | 0.2087 | 0.1307 | | 0.535 | 152.17 | 10500 | 0.2137 | 0.1309 | | 0.5081 | 159.42 | 11000 | 0.2215 | 0.1302 | | 0.6033 | 166.67 | 11500 | 0.2162 | 0.1302 | | 0.5549 | 173.91 | 12000 | 0.2198 | 0.1286 | | 0.5389 | 181.16 | 12500 | 0.2241 | 0.1293 | | 0.4912 | 188.41 | 13000 | 0.2190 | 0.1290 | | 0.4671 | 195.65 | 13500 | 0.2218 | 0.1290 |
a5bbb102f3be090d5555c86ccfacedd6
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | No log | 1.0 | 393 | 0.2287 | 0.9341 | 0.9112 |
821db75089abcd4fb516ea6b254708cd
creativeml-openrail-m
['text-to-image', 'stable-diffusion']
false
yyaaeell Dreambooth model trained by Brainergy with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept:
a4105c08dcd6367d356d3e33f4a7da87
mit
['grug', 'caveman', 'fun']
false
GPT-Grug-355m A finetuned version of [GPT2-Medium](https://huggingface.co/gpt2-medium) on the 'grug' dataset. A demo is available [here](https://huggingface.co/spaces/DarwinAnim8or/grug-chat) If you're interested, there's a smaller model available here: [GPT-Grug-125m](https://huggingface.co/DarwinAnim8or/gpt-grug-125m) Do note however that it is very limited by comparison.
3dad10e76089a22d78d92fb62d5b928a
apache-2.0
['generated_from_trainer']
false
distilbert_sa_GLUE_Experiment_logit_kd_data_aug_cola_192 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the GLUE COLA dataset. It achieves the following results on the evaluation set: - Loss: 0.6917 - Matthews Correlation: 0.0937
c897c8f7084f6ed18584cb43cd08fb86
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.647 | 1.0 | 835 | 0.6917 | 0.0937 | | 0.5356 | 2.0 | 1670 | 0.7312 | 0.1294 | | 0.4736 | 3.0 | 2505 | 0.7512 | 0.1372 | | 0.4318 | 4.0 | 3340 | 0.7576 | 0.1259 | | 0.3993 | 5.0 | 4175 | 0.7826 | 0.1277 | | 0.3736 | 6.0 | 5010 | 0.8138 | 0.1028 |
0f571e05882ca262d6972f7d7baa3df5
apache-2.0
['fnet']
false
FNet base model Pretrained model on English language using a masked language modeling (MLM) and next sentence prediction (NSP) objective. It was introduced in [this paper](https://arxiv.org/abs/2105.03824) and first released in [this repository](https://github.com/google-research/google-research/tree/master/f_net). This model is cased: it makes a difference between english and English. The model achieves 0.58 accuracy on MLM objective and 0.80 on NSP objective. Disclaimer: This model card has been written by [gchhablani](https://huggingface.co/gchhablani).
ef90384b6c23ebe50b93a52bf5aa1f2b
apache-2.0
['fnet']
false
Model description FNet is a transformers model with attention replaced with fourier transforms. Hence, the inputs do not contain an `attention_mask`. It is pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with two objectives: - Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. - Next sentence prediction (NSP): the models concatenates two masked sentences as inputs during pretraining. Sometimes they correspond to sentences that were next to each other in the original text, sometimes not. The model then has to predict if the two sentences were following each other or not. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the FNet model as inputs.
1cd48fb805d51fe4121aca8a5f6bd9c0
apache-2.0
['fnet']
false
Pretraining FNet-base was trained on 4 cloud TPUs in Pod configuration (16 TPU chips total) for one million steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 1e-4, \\(\beta_{1} = 0.9\\) and \\(\beta_{2} = 0.999\\), a weight decay of 0.01, learning rate warmup for 10,000 steps and linear decay of the learning rate after.
10b4c207b141bac18b7e1835ecb23fd1
apache-2.0
['fnet']
false
Evaluation results FNet-base was fine-tuned and evaluated on the validation data of the [GLUE benchamrk](https://huggingface.co/datasets/glue). The results of the official model (written in Flax) can be seen in Table 1 on page 7 of [the official paper](https://arxiv.org/abs/2105.03824). For comparison, this model (ported to PyTorch) was fine-tuned and evaluated using the [official Hugging Face GLUE evaluation scripts](https://github.com/huggingface/transformers/tree/master/examples/pytorch/text-classification
ac0bf64c8acbba74f32fa3045b681303
apache-2.0
['fnet']
false
glue-tasks) alongside [bert-base-cased](https://hf.co/models/bert-base-cased) for comparison. The training was done on a single 16GB NVIDIA Tesla V100 GPU. For MRPC/WNLI, the models were trained for 5 epochs, while for other tasks, the models were trained for 3 epochs. A sequence length of 512 was used with batch size 16 and learning rate 2e-5. The following table summarizes the results for [fnet-base](https://huggingface.co/google/fnet-base) (called *FNet (PyTorch) - Reproduced*) and [bert-base-cased](https://hf.co/models/bert-base-cased) (called *Bert (PyTorch) - Reproduced*) in terms of **fine-tuning** speed. The format is *hour:min:seconds*. **Note** that the authors compared **pre-traning** speed in [the official paper](https://arxiv.org/abs/2105.03824) instead. | Task/Model | FNet-base (PyTorch) |Bert-base (PyTorch)| |:----:|:-----------:|:----:| | MNLI-(m/mm) | [06:40:55](https://huggingface.co/gchhablani/fnet-base-finetuned-mnli) | [09:52:33](https://huggingface.co/gchhablani/bert-base-cased-finetuned-mnli)| | QQP | [06:21:16](https://huggingface.co/gchhablani/fnet-base-finetuned-qqp) | [09:25:01](https://huggingface.co/gchhablani/bert-base-cased-finetuned-qqp) | | QNLI | [01:48:22](https://huggingface.co/gchhablani/fnet-base-finetuned-qnli) | [02:40:22](https://huggingface.co/gchhablani/bert-base-cased-finetuned-qnli)| | SST-2 | [01:09:27](https://huggingface.co/gchhablani/fnet-base-finetuned-sst2) | [01:42:17](https://huggingface.co/gchhablani/bert-base-cased-finetuned-sst2)| | CoLA | [00:09:47](https://huggingface.co/gchhablani/fnet-base-finetuned-cola) | [00:14:20](https://huggingface.co/gchhablani/bert-base-cased-finetuned-cola)| | STS-B | [00:07:09](https://huggingface.co/gchhablani/fnet-base-finetuned-stsb) | [00:10:24](https://huggingface.co/gchhablani/bert-base-cased-finetuned-stsb)| | MRPC | [00:07:48](https://huggingface.co/gchhablani/fnet-base-finetuned-mrpc) | [00:11:12](https://huggingface.co/gchhablani/bert-base-cased-finetuned-mrpc)| | RTE | [00:03:24](https://huggingface.co/gchhablani/fnet-base-finetuned-rte) | [00:04:51](https://huggingface.co/gchhablani/bert-base-cased-finetuned-rte)| | WNLI | [00:02:37](https://huggingface.co/gchhablani/fnet-base-finetuned-wnli) | [00:03:23](https://huggingface.co/gchhablani/bert-base-cased-finetuned-wnli)| | SUM | 16:30:45 | 24:23:56 | On average the PyTorch version of FNet-base requires *ca.* 32% less time for GLUE fine-tuning on GPU. The following table summarizes the results for [fnet-base](https://huggingface.co/google/fnet-base) (called *FNet (PyTorch) - Reproduced*) and [bert-base-cased](https://hf.co/models/bert-base-cased) (called *Bert (PyTorch) - Reproduced*) in terms of performance and compares it to the reported performance of the official FNet-base model (called *FNet (Flax) - Official*). Note that the training hyperparameters of the reproduced models were not the same as the official model, so the performance may differ significantly for some tasks (for example: CoLA). | Task/Model | Metric | FNet-base (PyTorch) | Bert-base (PyTorch) | FNet-Base (Flax - official) | |:----:|:-----------:|:----:|:-----------:|:----:| | MNLI-(m/mm) | Accuracy or Match/Mismatch | [76.75](https://huggingface.co/gchhablani/fnet-base-finetuned-mnli) | [84.10](https://huggingface.co/gchhablani/bert-base-cased-finetuned-mnli) | 72/73 | | QQP | mean(Accuracy,F1) | [86.5](https://huggingface.co/gchhablani/fnet-base-finetuned-qqp) | [89.26](https://huggingface.co/gchhablani/bert-base-cased-finetuned-qqp) | 83 | | QNLI | Accuracy | [84.39](https://huggingface.co/gchhablani/fnet-base-finetuned-qnli) | [90.99](https://huggingface.co/gchhablani/bert-base-cased-finetuned-qnli) | 80 | | SST-2 | Accuracy | [89.45](https://huggingface.co/gchhablani/fnet-base-finetuned-sst2) | [92.32](https://huggingface.co/gchhablani/bert-base-cased-finetuned-sst2) | 95 | | CoLA | Matthews corr or Accuracy | [35.94](https://huggingface.co/gchhablani/fnet-base-finetuned-cola) | [59.57](https://huggingface.co/gchhablani/bert-base-cased-finetuned-cola) | 69 | | STS-B | Spearman corr. | [82.19](https://huggingface.co/gchhablani/fnet-base-finetuned-stsb) | [88.98](https://huggingface.co/gchhablani/bert-base-cased-finetuned-stsb) | 79 | | MRPC | mean(F1/Accuracy) | [81.15](https://huggingface.co/gchhablani/fnet-base-finetuned-mrpc) | [88.15](https://huggingface.co/gchhablani/bert-base-cased-finetuned-mrpc) | 76 | | RTE | Accuracy | [62.82](https://huggingface.co/gchhablani/fnet-base-finetuned-rte) | [67.15](https://huggingface.co/gchhablani/bert-base-cased-finetuned-rte) | 63 | | WNLI | Accuracy | [54.93](https://huggingface.co/gchhablani/fnet-base-finetuned-wnli) | [46.48](https://huggingface.co/gchhablani/bert-base-cased-finetuned-wnli) | - | | Avg | - | 72.7 | 78.6 | 76.7 | We can see that FNet-base achieves around 93% of BERT-base's performance on average. For more details, please refer to the checkpoints linked with the scores. On overview of all fine-tuned checkpoints of the following table can be accessed [here](https://huggingface.co/models?other=fnet-bert-base-comparison).
646ed0fc103b247df562dbf40c39095d
apache-2.0
['fnet']
false
How to use You can use this model directly with a pipeline for masked language modeling: **Note: The mask filling pipeline doesn't work exactly as the original model performs masking after converting to tokens. In masking pipeline an additional space is added after the [MASK].** ```python >>> from transformers import FNetForMaskedLM, FNetTokenizer, pipeline >>> tokenizer = FNetTokenizer.from_pretrained("google/fnet-base") >>> model = FNetForMaskedLM.from_pretrained("google/fnet-base") >>> unmasker = pipeline('fill-mask', model=model, tokenizer=tokenizer) >>> unmasker("Hello I'm a [MASK] model.") [ {"sequence": "hello i'm a new model.", "score": 0.12073223292827606, "token": 351, "token_str": "new"}, {"sequence": "hello i'm a first model.", "score": 0.08501081168651581, "token": 478, "token_str": "first"}, {"sequence": "hello i'm a next model.", "score": 0.060546260327100754, "token": 1037, "token_str": "next"}, {"sequence": "hello i'm a last model.", "score": 0.038265593349933624, "token": 813, "token_str": "last"}, {"sequence": "hello i'm a sister model.", "score": 0.033868927508592606, "token": 6232, "token_str": "sister"}, ] ``` Here is how to use this model to get the features of a given text in PyTorch: **Note: You must specify the maximum sequence length to be 512 and truncate/pad to the same length because the original model has no attention mask and considers all the hidden states during forward pass.** ```python from transformers import FNetTokenizer, FNetModel tokenizer = FNetTokenizer.from_pretrained("google/fnet-base") model = FNetModel.from_pretrained("google/fnet-base") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt', padding='max_length', truncation=True, max_length=512) output = model(**encoded_input) ```
8703c18e551beca2599af59f189b9f4a
apache-2.0
['generated_from_keras_callback']
false
silviacamplani/bert-finetuned-ner This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.0270 - Validation Loss: 0.0563 - Epoch: 2
e29b10009fef9b61aee5c6d8e5aa58d4
apache-2.0
['generated_from_keras_callback']
false
Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 0.1764 | 0.0664 | 0 | | 0.0476 | 0.0584 | 1 | | 0.0270 | 0.0563 | 2 |
4a492ba51690e19f03fca283e977d6f7
creativeml-openrail-m
['stable-diffusion', 'text-to-image', 'safetensors', 'disco-elysium']
false
Harrier Du Bois Meet the [Harrier Du Bois](https://discoelysium.fandom.com/wiki/Harrier_Du_Bois). A disco elysium style Stable Diffusion model I trained a long time ago. Trained with [Disco Elysium Archetypes](https://www.artstation.com/artwork/6aAL8x) from [Aleksander Rostov](https://www.artstation.com/rostovjanka) and some painitng from [Craig Mullins](http://www.goodbrush.com). You might have seen it somewhere else. Based on a novel model we all know. Text encoder is trained and broken. **Very unwieldy** but interesting enough.
a26577afc05b4ee3caca52a5f6558a56
apache-2.0
['generated_from_trainer']
false
distilbert_sa_GLUE_Experiment_data_aug_qnli This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the GLUE QNLI dataset. It achieves the following results on the evaluation set: - Loss: 1.2699 - Accuracy: 0.5997
29f9310d1268c61301509421a16d3627
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.3057 | 1.0 | 16604 | 1.2699 | 0.5997 | | 0.0735 | 2.0 | 33208 | 1.7786 | 0.5953 | | 0.0313 | 3.0 | 49812 | 1.9603 | 0.5801 | | 0.0188 | 4.0 | 66416 | 2.2529 | 0.5927 | | 0.0134 | 5.0 | 83020 | 2.4498 | 0.5913 | | 0.0106 | 6.0 | 99624 | 2.5181 | 0.6031 |
fb5c1516b1ef363b804f72de26342ba7
apache-2.0
['generated_from_trainer']
false
distilbert_add_GLUE_Experiment_wnli_192 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the GLUE WNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.6861 - Accuracy: 0.5634
b3c2f19b91a9905cceb3abecbf317641
apache-2.0
['generated_from_trainer']
false
Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6944 | 1.0 | 3 | 0.6957 | 0.4366 | | 0.6952 | 2.0 | 6 | 0.6954 | 0.4366 | | 0.6914 | 3.0 | 9 | 0.6877 | 0.5634 | | 0.6951 | 4.0 | 12 | 0.6861 | 0.5634 | | 0.6974 | 5.0 | 15 | 0.6886 | 0.5634 | | 0.6942 | 6.0 | 18 | 0.6940 | 0.4366 | | 0.6934 | 7.0 | 21 | 0.6957 | 0.4366 | | 0.6953 | 8.0 | 24 | 0.6949 | 0.4366 | | 0.6931 | 9.0 | 27 | 0.6906 | 0.5634 |
62d02128a3fedef62d0a18f6e3572270