modelId
stringlengths
4
81
tags
list
pipeline_tag
stringclasses
17 values
config
dict
downloads
int64
0
59.7M
first_commit
timestamp[ns, tz=UTC]
card
stringlengths
51
438k
DeepPavlov/distilrubert-tiny-cased-conversational-v1
[ "pytorch", "distilbert", "ru", "arxiv:2205.02340", "transformers" ]
null
{ "architectures": null, "model_type": "distilbert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "n...
9,141
2022-11-16T10:20:11Z
--- license: openrail library_name: diffusers tags: - TPU - JAX - Flax - stable-diffusion - text-to-image language: - en ---
DeltaHub/adapter_t5-3b_qnli
[ "pytorch", "transformers" ]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
3
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - emotion metrics: - accuracy - f1 model-index: - name: bert_uncased_L-2_H-128_A-2-finetuned-emotion results: - task: name: Text Classification type: text-classification dataset: name: emotion type: emotion args: defa...
DemangeJeremy/4-sentiments-with-flaubert
[ "pytorch", "flaubert", "text-classification", "fr", "transformers", "sentiments", "french", "flaubert-large" ]
text-classification
{ "architectures": [ "FlaubertForSequenceClassification" ], "model_type": "flaubert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, ...
226
null
Access to model AlexKozachuk/Kitchen is restricted and you are not in the authorized list. Visit https://huggingface.co/AlexKozachuk/Kitchen to ask for access.
Deniskin/emailer_medium_300
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
14
null
# This is min-stable-diffusion weights file ## I hope you enjoyed. I hope you can discovery light!!! #### weight file notes 1) wd-1-3-penultimate-ucg-cont.pt is waifu-diffusion-v1-4 weight 2) mdjrny-v4.pt is midjourney-v4-diffusion weight 3) stable_diffusion_v1_4.pt is CompVis/stable-diffusion-v1-4 4) stable_diffusi...
Deniskin/essays_small_2000
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imagefolder metrics: - accuracy model-index: - name: vit-large-patch32-384-finetuned-melanoma results: - task: name: Image Classification type: image-classification dataset: name: imagefolder type: imagefolder confi...
DeskDown/MarianMixFT_en-ms
[ "pytorch", "marian", "text2text-generation", "transformers", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "MarianMTModel" ], "model_type": "marian", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
5
null
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity - transformers --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 512 dimensional dense vector space and can be used for tasks like cluste...
DeskDown/MarianMixFT_en-my
[ "pytorch", "marian", "text2text-generation", "transformers", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "MarianMTModel" ], "model_type": "marian", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
7
null
--- license: creativeml-openrail-m tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image --- Example result: =============== # Using whitemanedb_step_3500.ckpt ![whitemanedb_step_3500 1](https://huggingface.co/sd-dreambooth-library/sally-whitemanev/resolve/main/concept_images/tmpez9ubybz.png) # Using d...
DeskDown/MarianMix_en-zh-10
[ "pytorch", "tensorboard", "marian", "text2text-generation", "transformers", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "MarianMTModel" ], "model_type": "marian", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
3
null
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: distilbert-base-uncased-finetuned-emotion results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, t...
Dhritam/Zova-bot
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- tags: - generated_from_trainer model-index: - name: unifiedqa-v2-t5-base-1363200-finetuned-causalqa-squad results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # u...
Dibyaranjan/nl_image_search
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: cc-by-nc-4.0 tags: - galactica widget: - text: "The Transformer architecture [START_REF]" - text: "The Schwarzschild radius is defined as: \\[" - text: "A force of 0.6N is applied to an object, which accelerates at 3m/s. What is its mass? <work>" - text: "Lecture 1: The Ising Model\n\n" - text: "[START_I_...
Digakive/Hsgshs
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: apache-2.0 language: - uk - sv tags: - generated_from_trainer - translation model-index: - name: mt-uk-sv-finetuned results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove thi...
Dilmk2/DialoGPT-small-harrypotter
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
13
null
--- library_name: stable-baselines3 tags: - seals/MountainCar-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - metrics: - type: mean_reward value: -100.60 +/- 5.75 name: mean_reward task: type: reinforcement-learning na...
DingleyMaillotUrgell/homer-bot
[ "pytorch", "gpt2", "text-generation", "en", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
12
null
--- language: id license: mit datasets: - oscar - wikipedia - id_newspapers_2018 widget: - text: Saya [MASK] makan nasi goreng. - text: Kucing itu sedang bermain dengan [MASK]. pipeline_tag: fill-mask --- # Indonesian small BigBird model ## Source Code Source code to create this model is available at [https://github...
DivyanshuSheth/T5-Seq2Seq-Final
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: apache-2.0 --- ## Prompt Trigger Keywords Use `darkprincess638 person` to trigger the character, this works best at the start of the prompt ## Examples Here are some random generations to show the flexibility of the model with different prompts ![Generated Images with this model](https://i.imgur.com/V...
Dizoid/Lll
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: cc-by-nc-4.0 tags: - galactica widget: - text: "The Transformer architecture [START_REF]" - text: "The Schwarzschild radius is defined as: \\[" - text: "A force of 0.6N is applied to an object, which accelerates at 3m/s. What is its mass? <work>" - text: "Lecture 1: The Ising Model\n\n" - text: "[START_I_...
Dkwkk/Da
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- language: en tags: - newspapers - library - historic - glam - mdma license: mit metrics: - pseudo-perplexity widget: - text: "1820 [DATE] We received a letter from [MASK] Majesty." - text: "1850 [DATE] We received a letter from [MASK] Majesty." - text: "[MASK] [DATE] The Franco-Prussian war is a matter of great con...
Dmitry12/sber
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: other tags: - generated_from_trainer model-index: - name: nvidia-segformer-b0-finetuned-ade-512-512-finetuned-ISIC17 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove th...
Doiman/DialoGPT-medium-harrypotter
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
13
null
--- license: mit tags: - generated_from_trainer model-index: - name: aw-gpt results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # aw-gpt This model is a fine-tuned ...
DongHai/DialoGPT-small-rick
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
9
null
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semanti...
DongHyoungLee/distilbert-base-uncased-finetuned-cola
[ "pytorch", "tensorboard", "distilbert", "text-classification", "dataset:glue", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index" ]
text-classification
{ "architectures": [ "DistilBertForSequenceClassification" ], "model_type": "distilbert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, ...
27
null
--- license: agpl-3.0 pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity --- # sentence-t5-base-nlpl-code_search_net This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can b...
Donghyun/L2_BERT
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: cc-by-nc-4.0 tags: - galactica widget: - text: "The Transformer architecture [START_REF]" - text: "The Schwarzschild radius is defined as: \\[" - text: "A force of 0.6N is applied to an object, which accelerates at 3m/s. What is its mass? <work>" - text: "Lecture 1: The Ising Model\n\n" - text: "[START_I_S...
Dongjae/mrc2reader
[ "pytorch", "xlm-roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "XLMRobertaForQuestionAnswering" ], "model_type": "xlm-roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, ...
3
null
--- language: en tags: - newspapers - library - historic - glam - mdma license: mit metrics: - pseudo-perplexity widget: - text: "[1820] [SEP] We received a letter from [MASK] Majesty." - text: "[1850] [SEP] We received a letter from [MASK] Majesty." - text: "[MASK] [SEP] The Franco-Prussian war is a matter of great co...
Dongmin/testmodel
[ "pytorch", "t5", "text2text-generation", "transformers", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "T5ForConditionalGeneration" ], "model_type": "t5", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": true, "length_penalty": 2, "max_length": 200, "min_length": 30, "no_repeat_ngram_s...
11
null
--- language: en tags: - newspapers - library - historic - glam - mdma license: mit metrics: - pseudo-perplexity widget: - text: "1820 [DATE] We received a letter from [MASK] Majesty." - text: "1850 [DATE] We received a letter from [MASK] Majesty." - text: "[MASK] [DATE] The Franco-Prussian war is a matter of great con...
Waynehillsdev/Waynehills_summary_tensorflow
[ "tf", "t5", "text2text-generation", "transformers", "generated_from_keras_callback", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "T5ForConditionalGeneration" ], "model_type": "t5", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_n...
5
null
--- license: mit tags: - generated_from_trainer model-index: - name: bert-base-german-cased-issues-128-finetuned results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> ...
Doohae/q_encoder
[ "pytorch" ]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
3
null
--- tags: - generated_from_trainer datasets: - rvl_cdip metrics: - accuracy model-index: - name: invoicevsadvertisement results: - task: name: Image Classification type: image-classification dataset: name: rvl_cdip type: rvl_cdip config: default split: train args: defau...
Doohae/roberta
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_re...
3
null
--- language: - en license: creativeml-openrail-m tags: - stable-diffusion - text-to-image --- # finetuned on dark, moody, "victorian" imagery (ノ◕ヮ◕)ノ*:・゚✧ [<img src="https://colab.research.google.com/assets/colab-badge.svg">](https://colab.research.google.com/drive/13E3i6_Z1BWd3e6f71-TNd5bk8eGqaeZf?usp=sharing) !...
albert-base-v2
[ "pytorch", "tf", "jax", "rust", "safetensors", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
4,785,283
2022-11-16T16:27:33Z
--- license: mit --- ### abstract_patterns_in_nature on Stable Diffusion via Dreambooth trained on the [fast-DreamBooth.ipynb by TheLastBen](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook #### model by apurik-parv This your the Stable Diffusion model ...
albert-large-v1
[ "pytorch", "tf", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
687
2022-11-16T16:28:58Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imdb metrics: - accuracy - f1 model-index: - name: bert_uncased_L-2_H-128_A-2-finetuned-emotion-finetuned-tweet results: - task: name: Text Classification type: text-classification dataset: name: imdb type: imdb con...
albert-xlarge-v1
[ "pytorch", "tf", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
341
2022-11-16T16:41:41Z
data: https://github.com/BigSalmon2/InformalToFormalDataset Text Generation Informal Formal ``` from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("BigSalmon/InformalToFormalLincoln90Paraphrase") model = AutoModelForCausalLM.from_pretrained("BigSalmon/InformalToForm...
albert-xlarge-v2
[ "pytorch", "tf", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
2,973
2022-11-16T16:43:22Z
--- license: mit --- ##MODEL BY ShadoWxShinigamI Use Token - mdjrny-shttr at the beginning of your prompt [ Prompt Engineering not required ]; If some object doesn't work, increase the prompt weight of the object to 1.6 Training - 2500 steps, Batch size 2, 512x512, v1-5 Base, 26 images (52 Flipped) Examples:- Lion ...
bert-base-cased-finetuned-mrpc
[ "pytorch", "tf", "jax", "bert", "fill-mask", "transformers", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "BertForMaskedLM" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
11,644
2022-11-16T16:55:38Z
--- license: cc-by-4.0 metrics: - bleu4 - meteor - rouge-l - bertscore - moverscore language: en datasets: - lmqg/qag_tweetqa pipeline_tag: text2text-generation tags: - questions and answers generation widget: - text: "Beyonce further expanded her acting career, starring as blues singer Etta James in the 2008 musical ...
bert-base-chinese
[ "pytorch", "tf", "jax", "safetensors", "bert", "fill-mask", "zh", "arxiv:1810.04805", "transformers", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "BertForMaskedLM" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
3,377,486
2022-11-16T16:57:47Z
--- license: apache-2.0 language: - bs - sv tags: - translation model-index: - name: mt-bs-sv-finetuned results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mt-bs...
bert-base-german-cased
[ "pytorch", "tf", "jax", "safetensors", "bert", "fill-mask", "de", "transformers", "exbert", "license:mit", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "BertForMaskedLM" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
175,983
2022-11-16T17:02:45Z
--- license: mit tags: - generated_from_trainer model-index: - name: BERiT_2000_custom_architecture_40_epochs_ls_.1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --...
bert-base-german-dbmdz-uncased
[ "pytorch", "jax", "safetensors", "bert", "fill-mask", "de", "transformers", "license:mit", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "BertForMaskedLM" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
68,305
2022-11-16T17:05:17Z
--- license: mit tags: - generated_from_trainer metrics: - f1 model-index: - name: frases-bertimbau-v0.4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # frases...
bert-base-multilingual-cased
[ "pytorch", "tf", "jax", "safetensors", "bert", "fill-mask", "multilingual", "af", "sq", "ar", "an", "hy", "ast", "az", "ba", "eu", "bar", "be", "bn", "inc", "bs", "br", "bg", "my", "ca", "ceb", "ce", "zh", "cv", "hr", "cs", "da", "nl", "en", "et", ...
fill-mask
{ "architectures": [ "BertForMaskedLM" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
4,749,504
2022-11-16T17:07:36Z
--- license: creativeml-openrail-m tags: - text-to-image --- ### Vulvine_Look_v02 on Stable Diffusion via Dreambooth trained on the [fast-DreamBooth.ipynb by TheLastBen](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook #### Model by LaCambre This your t...
bert-base-multilingual-uncased
[ "pytorch", "tf", "jax", "safetensors", "bert", "fill-mask", "multilingual", "af", "sq", "ar", "an", "hy", "ast", "az", "ba", "eu", "bar", "be", "bn", "inc", "bs", "br", "bg", "my", "ca", "ceb", "ce", "zh", "cv", "hr", "cs", "da", "nl", "en", "et", ...
fill-mask
{ "architectures": [ "BertForMaskedLM" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
328,585
2022-11-16T17:15:07Z
--- language: - hi license: apache-2.0 tags: - hf-asr-leaderboard - generated_from_trainer datasets: - mozilla-foundation/common_voice_11_0 metrics: - wer model-index: - name: Whisper Small Hi - Sanchit Gandhi results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dat...
bert-large-cased-whole-word-masking
[ "pytorch", "tf", "jax", "bert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1810.04805", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "BertForMaskedLM" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
2,316
2022-11-16T17:21:00Z
--- language: - en tags: - stable-diffusion - text-to-image license: creativeml-openrail-m --- #### Realistic Lucy Edgerunners models Trained on SD1.5 Our beloved Lucy who tends to realism ! There are two models, the main is more Lucy accurate but less style transferrable, the second '-creative' tend to do real...
bert-large-cased
[ "pytorch", "tf", "jax", "safetensors", "bert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1810.04805", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "BertForMaskedLM" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
388,769
null
Access to model lucieackley/setfit-welcome-msg is restricted and you are not in the authorized list. Visit https://huggingface.co/lucieackley/setfit-welcome-msg to ask for access.
bert-large-uncased-whole-word-masking-finetuned-squad
[ "pytorch", "tf", "jax", "safetensors", "bert", "question-answering", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1810.04805", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_n...
480,510
2022-11-16T17:27:48Z
--- tags: - stanza - token-classification library_name: stanza language: myv license: apache-2.0 --- # Stanza model for Erzya (myv) Stanza is a collection of accurate and efficient tools for the linguistic analysis of many human languages. Starting from raw text to syntactic analysis and entity recognition, Stanza brin...
camembert-base
[ "pytorch", "tf", "safetensors", "camembert", "fill-mask", "fr", "dataset:oscar", "arxiv:1911.03894", "transformers", "license:mit", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "CamembertForMaskedLM" ], "model_type": "camembert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_...
1,440,898
2022-11-16T17:38:00Z
--- license: mit datasets: Den4ikAI/mailruQA-big widget: - text: "Q: Что такое любовь?\n A:" example_title: test --- Здесь будут выкладываться чекпоинты модели rugpt3-medium обученной на данных с otvet.mail.ru Датасет для обучения [тык](https://huggingface.co/datasets/Den4ikAI/mailruQA-big) Мини-версия (в два раза ...
distilbert-base-cased-distilled-squad
[ "pytorch", "tf", "rust", "safetensors", "openvino", "distilbert", "question-answering", "en", "dataset:squad", "arxiv:1910.01108", "arxiv:1910.09700", "transformers", "license:apache-2.0", "model-index", "autotrain_compatible", "has_space" ]
question-answering
{ "architectures": [ "DistilBertForQuestionAnswering" ], "model_type": "distilbert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, ...
257,745
2022-11-16T17:41:21Z
--- license: mit --- This is still a WiP. But if anyone still wants to try it out, the prompt token is mdjrny-splttr Updated Style - Midjourney-v4-PaintArt [Link](https://huggingface.co/ShadoWxShinigamI/Midjourney-v4-PaintArt "Paint Art")
AHussain0418/distillbert-truth-detector
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
2022-11-17T02:14:14Z
# L0 Regularizer PyTorch adaptation of Louizos (2017) to wrap PyTorch modules. This code provides a method for wrapping an existing PyTorch module and regularizing it according to the L0 norm. It is an adaptation and extension of code from the repository [here](https://github.com/AMLab-Amsterdam/L0_regularization). Th...
AZTEC/Arcane
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
2022-11-17T07:43:43Z
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity - transformers --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like cluste...
Pinwheel/wav2vec2-large-xlsr-53-hi
[ "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "transformers" ]
automatic-speech-recognition
{ "architectures": [ "Wav2Vec2ForCTC" ], "model_type": "wav2vec2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_s...
9
2022-11-17T08:31:25Z
--- tags: - endpoints-template license: apache-2.0 --- # Multi-Model GPU Inference with Hugging Face Inference Endpoints Multi-model Inference Endpoints provide a way to deploy multiple models onto the same infrastructure for a scalable and cost-effective inference. On multi-model Inference Endpoints, we load a list ...
Abdullaziz/model1
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
2022-11-17T09:06:47Z
--- license: cc --- This model is a Pytorch version of [uklfr/gottbert-base](https://huggingface.co/uklfr/gottbert-base). All credits to their developers.
AhmedHassan19/model
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: apache-2.0 --- ## Prompt Trigger Use the keywords `darkprincess638 person` to trigger the character, best if used at start of prompt ## Examples This is some of the stuff that can be generated with this model ![Generation examples](https://i.imgur.com/ZI3vHVs.png)
Akbarariza/Anjar
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- tags: - LunarLander-v2 - ppo - deep-reinforcement-learning - reinforcement-learning - custom-implementation - deep-rl-course model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metr...
Akira-Yana/distilbert-base-uncased-finetuned-cola
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: apache-2.0 --- ## Model description An answer classification model for boolean questions based on XLM-RoBERTa. The answer classifier takes as input a boolean question and a passage, and returns a label (yes, no-answer, no). The model was initialized with [xlm-roberta-large](https://huggingface.co/xl...
Aleksandra/herbert-base-cased-finetuned-squad
[ "pytorch", "tensorboard", "bert", "question-answering", "transformers", "generated_from_trainer", "license:cc-by-4.0", "autotrain_compatible" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_n...
8
2022-11-18T00:07:30Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: bert-finetuned-expression_epoch5 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and co...
Alessandro/model_name
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - squad_v2 model-index: - name: distilbert-base-uncased-finetuned-squad-seed-42 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it,...
AlexDemon/Alex
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
2022-11-18T01:17:04Z
--- license: apache-2.0 --- ## Trigger Prompt Use the keywords `darkprincess638 person` to trigger the character, best used at start of prompt ## Examples Some examples of what you can generate with this model ![Sample generations](https://i.imgur.com/b3tpprk.png)
AlexN/xls-r-300m-fr
[ "pytorch", "wav2vec2", "automatic-speech-recognition", "fr", "dataset:mozilla-foundation/common_voice_8_0", "transformers", "generated_from_trainer", "hf-asr-leaderboard", "mozilla-foundation/common_voice_8_0", "robust-speech-event", "model-index" ]
automatic-speech-recognition
{ "architectures": [ "Wav2Vec2ForCTC" ], "model_type": "wav2vec2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_s...
17
null
--- license: creativeml-openrail-m tags: - text-to-image --- ### Sonic06-Diffusion on Stable Diffusion via Dreambooth trained on the [fast-DreamBooth.ipynb by TheLastBen](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook #### Model by Laughify This is a...
AlexN/xls-r-300m-pt
[ "pytorch", "wav2vec2", "automatic-speech-recognition", "pt", "dataset:mozilla-foundation/common_voice_8_0", "transformers", "robust-speech-event", "mozilla-foundation/common_voice_8_0", "generated_from_trainer", "hf-asr-leaderboard", "license:apache-2.0", "model-index" ]
automatic-speech-recognition
{ "architectures": [ "Wav2Vec2ForCTC" ], "model_type": "wav2vec2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_s...
15
null
--- inference: true language: - en tags: - stable-diffusion - text-to-image license: creativeml-openrail-m --- # -Stable Diffusion fine tuned on Fantastic Mr Fox screencaps- Use prompt: 'fantasticmrfox' # This model is very versatile but without negative prompts it will mostly produce images of foxes or other ...
AlexaRyck/KEITH
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- library_name: stable-baselines3 tags: - AntBulletEnv-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: AntBulletEnv-v0 type: AntBulletEnv-v0 ...
Alexander-Learn/bert-finetuned-ner-accelerate
[ "pytorch", "bert", "token-classification", "transformers", "autotrain_compatible" ]
token-classification
{ "architectures": [ "BertForTokenClassification" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat...
4
null
--- language: en license: apache-2.0 library_name: diffusers tags: [] datasets: huggan/smithsonian_butterflies_subset metrics: [] --- <!-- This model card has been generated automatically according to the information the training script had access to. You should probably proofread and complete it, then remove this com...
Alexander-Learn/bert-finetuned-squad
[ "pytorch", "tensorboard", "bert", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_n...
7
null
--- language: en thumbnail: http://www.huggingtweets.com/rundizzy-s4m31p4n-tyler02020202/1668738792600/predictions.png tags: - huggingtweets widget: - text: "My dream is" --- <div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; ma...
Alexandru/creative_copilot
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imagefolder metrics: - accuracy model-index: - name: swin-tiny-patch4-window7-224-finetuned-eurosat results: - task: name: Image Classification type: image-classification dataset: name: imagefolder type: imagefolder ...
AliPotter24/a
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
2022-11-18T03:36:19Z
--- language: en thumbnail: http://www.huggingtweets.com/dril-s4m31p4n-wnbagirlfriend/1668742659829/predictions.png tags: - huggingtweets widget: - text: "My dream is" --- <div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margi...
Aliraza47/BERT
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: agpl-3.0 --- stanford_v0.0.1 segments muscle, bone, VAT and SAT on L3 axial CT slices. abCT_v0.0.1 segments muscle, IMAT, VAT, and SAT on L3 axial CT slices.
Alireza-rw/testbot
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- library_name: stable-baselines3 tags: - HalfCheetahBulletEnv-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: HalfCheetahBulletEnv-v0 type: ...
Alireza1044/albert-base-v2-rte
[ "pytorch", "tensorboard", "albert", "text-classification", "en", "dataset:glue", "transformers", "generated_from_trainer", "license:apache-2.0" ]
text-classification
{ "architectures": [ "AlbertForSequenceClassification" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no...
30
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - squad_v2 model-index: - name: albert-base-v2-finetuned-squad-seed-1024 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then r...
Alireza1044/albert-base-v2-sst2
[ "pytorch", "tensorboard", "albert", "text-classification", "en", "dataset:glue", "transformers", "generated_from_trainer", "license:apache-2.0" ]
text-classification
{ "architectures": [ "AlbertForSequenceClassification" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no...
52
null
--- license: odc-by --- Basically, generate the images by saying "dnd[RACE] person" I know some arent people, but it's what I've got to work with. ;) Make sure there are no spaces, or punctuation in the "dnd[RACE HERE]" section, so "a portrait of dndYuanTi person, intricate, elegant, highly detailed, digital painting...
Amalq/distilroberta-base-finetuned-MentalHealth
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: dpkmnit/bert-finetuned-squad results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # dpkmn...
AndrewChar/model-QA-5-epoch-RU
[ "tf", "distilbert", "question-answering", "ru", "dataset:sberquad", "transformers", "generated_from_keras_callback", "autotrain_compatible" ]
question-answering
{ "architectures": [ "DistilBertForQuestionAnswering" ], "model_type": "distilbert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, ...
109
null
--- tags: - text-to-image library_name: generic --- # Text To Image repository template This is a template repository for text to image to support generic inference with Hugging Face Hub generic Inference API. There are two required steps 1. Specify the requirements by defining a `requirements.txt` file. 2. Implement...
Andrey1989/mt5-small-finetuned-mlsum-es
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: mit tags: - nowcasting - forecasting - timeseries - remote-sensing --- # Nowcasting CNN ## Model description 3d conv model, that takes in different data streams architecture is roughly 1. satellite image time series goes into many 3d convolution layers. 2. nwp time series goes i...
Andrey1989/mt5-small-finetuned-mlsum-fr
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
2022-11-18T09:10:34Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: distilbert-base-uncased-finetuned-ner results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread a...
Andrija/RobertaFastBPE
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
2022-11-18T09:21:10Z
--- thumbnail: https://imgur.com/DkGWTA2.png language: - en tags: - stable-diffusion - text-to-image license: creativeml-openrail-m inference: false --- # Diffusion model This model is trained with detailed semi realistic images via my anime model. # Sample generations This model is made to get semi realistic, reali...
AnjanBiswas/distilbert-base-uncased-finetuned-emotion
[ "pytorch", "distilbert", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "DistilBertForSequenceClassification" ], "model_type": "distilbert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, ...
37
null
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: vikram15/bert-finetuned-squad results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # vikr...
Anonymous/ReasonBERT-TAPAS
[ "pytorch", "tapas", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "TapasModel" ], "model_type": "tapas", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": n...
7
null
Access to model dwmit/ja_classification is restricted and you are not in the authorized list. Visit https://huggingface.co/dwmit/ja_classification to ask for access.
Anonymous0230/model_name
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams...
0
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imdb metrics: - accuracy - f1 model-index: - name: NLP-sentiment-project-2000-samples results: - task: name: Text Classification type: text-classification dataset: name: imdb type: imdb config: plain_text spli...
AnonymousNLP/pretrained-model-1
[ "pytorch", "gpt2", "transformers" ]
null
{ "architectures": [ "GPT2DoubleHeadsModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram...
4
null
--- license: apache-2.0 language: - sv - no tags: - generated_from_trainer metrics: - bleu model-index: - name: mt-no-sv-finetuned results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove t...
AnonymousSub/AR_rule_based_roberta_twostagetriplet_epochs_1_shard_1
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
6
null
--- license: cc0-1.0 --- I trained the model on 65 images from the film. I used the template prompt of `painting` while training the model, so use that word in the prompt as well. It is not required. `painting of a spaceship by ghost-in-the-shell-style` or `a spaceship by ghost-in-the-shell-style`
AnonymousSub/AR_rule_based_roberta_twostagetriplet_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
2
null
--- license: creativeml-openrail-m --- **Milo Manara Style** This is the Alpha release of a Stable Diffusion model trained to achieve the style of the Italian illustration master Milo Manara. Use the token **in the style of ->Manara** in your prompts for the style. **Sample result** ![SD-tmpnmoxk0x9.jpg](https://s3....
AnonymousSub/SR_rule_based_bert_quadruplet_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": nul...
1
null
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: distilbert-base-uncased-finetuned-bbc-news results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, ...
AnonymousSub/SR_rule_based_bert_triplet_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": nul...
6
null
--- license: apache-2.0 tags: - distigpt2 - hearthstone metrics: - bleu - dvitel/codebleu - exact_match - chrf datasets: - dvitel/hearthstone model-index: - name: h1 results: - task: type: text-generation name: Python Code Synthesis dataset: type: dvitel/hearthstone name: HearthStone ...
AnonymousSub/SR_rule_based_roberta_bert_triplet_epochs_1_shard_1
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
2
null
--- license: apache-2.0 tags: - distigpt2 - hearthstone metrics: - bleu - dvitel/codebleu - exact_match - chrf datasets: - dvitel/hearthstone model-index: - name: h0 results: - task: type: text-generation name: Python Code Synthesis dataset: type: dvitel/hearthstone name: HearthStone ...
AnonymousSub/SR_rule_based_roberta_hier_quadruplet_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
5
null
--- license: unknown --- # Soda Stream finetuned style Model Produced from publicly available pictures in landscape, portrait and square format. ## Model info The models included was trained on "multi-resolution" images. ## Using the model * common subject prompt tokens: `<wathever> by soda stream` ## Example pro...
AnonymousSub/SR_rule_based_roberta_hier_triplet_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
6
2022-11-18T15:50:51Z
--- language: - en license: creativeml-openrail-m thumbnail: "https://huggingface.co/nitrosocke/Ghibli-Diffusion/resolve/main/images/ghibli-diffusion-thumbnail.jpg" tags: - stable-diffusion - text-to-image - image-to-image - diffusers --- ### Ghibli Diffusion This is the fine-tuned Stable Diffusion model trained on i...
AnonymousSub/SR_rule_based_roberta_only_classfn_epochs_1_shard_1
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
2
2022-11-18T15:53:37Z
--- tags: - generated_from_trainer metrics: - f1 model-index: - name: bert-base-chinese-finetuned-food results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-bas...
AnonymousSub/SR_rule_based_roberta_only_classfn_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
3
2022-11-18T15:58:21Z
--- tags: - image-classification - pytorch - huggingpics metrics: - accuracy model-index: - name: Clasificador-Ojos results: - task: name: Image Classification type: image-classification metrics: - name: Accuracy type: accuracy value: 0.7727272510528564 --- # Clasificador-Ojo...
AnonymousSub/SR_rule_based_roberta_only_classfn_twostage_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
4
null
--- license: apache-2.0 language: ta tags: - audio - automatic-speech-recognition - speech - xlsr-fine-tuning-week - hf-asr-leaderboard - tamil language model-index: - name: XLSR Wav2Vec2 Tamil by Manan Dey results: - task: name: Speech Recognition type: automatic-speech-recognition dataset: ...
AnonymousSub/SR_rule_based_roberta_twostagetriplet_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
4
null
--- language: "en" tags: - text-to-speech - TTS - speech-synthesis - fastspeech2 - speechbrain license: "apache-2.0" datasets: - LJSpeech metrics: - mos --- <iframe src="https://ghbtns.com/github-btn.html?user=speechbrain&repo=speechbrain&type=star&count=true&size=large&v=2" frameborder="0" scrolling="0" width="170" h...
AnonymousSub/SR_rule_based_twostage_quadruplet_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": nul...
3
null
--- language: - en - hi - multilingual tags: - generated_from_trainer licence: cc-by-sa-4.0 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # muril-en-hi-codemixed muril-en...
AnonymousSub/bert_mean_diff_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": nul...
4
null
--- tags: - image-classification - pytorch - huggingpics metrics: - accuracy model-index: - name: Clasificador-Ojos-XD results: - task: name: Image Classification type: image-classification metrics: - name: Accuracy type: accuracy value: 0.9696969985961914 --- # Clasificador-...
AnonymousSub/bert_snips
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": nul...
5
2022-11-18T17:52:52Z
--- license: creativeml-openrail-m language: - en tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image --- # Miyo-Waifu-Diffusion This model is a fine-tuned Waifu-Diffusion v1.3 by dreambooth. that can generate illustrations of Miyo Harada from THE IDOLM@STER CINDERELLA GIRLS. To use at a minimum,Pl...
AnonymousSub/bert_triplet_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": nul...
1
null
--- license: apache-2.0 tags: - image-classification - generated_from_trainer datasets: - imagefolder metrics: - accuracy model-index: - name: beit-base-ches-demo-v0 results: - task: name: Image Classification type: image-classification dataset: name: imagefolder type: imagefolder ...
AnonymousSub/cline-emanuals-s10-AR
[ "pytorch", "roberta", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "RobertaForSequenceClassification" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "...
27
null
--- license: unknown --- # Ans Huh finetuned style Model Produced from publicly available pictures in landscape, portrait and square format. ## Model info The models included was trained on "multi-resolution" images. ## Using the model * common subject prompt tokens: `<wathever> watercolor by ans huh` ## Example ...
AnonymousSub/cline-papers-roberta-0.585
[ "pytorch", "roberta", "transformers" ]
null
{ "architectures": [ "LecbertForPreTraining" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_n...
1
null
--- license: unknown --- # Gauzy Storm finetuned style Model Produced from publicly available pictures in landscape, portrait and square format. This model is focussed on creating animal hybrid artwork. ## Model info The models included was trained on "multi-resolution" images. ## Using the model * common subject ...
AnonymousSub/cline-techqa
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_re...
6
null
Access to model ERRORCOMPANY/Lino is restricted and you are not in the authorized list. Visit https://huggingface.co/ERRORCOMPANY/Lino to ask for access.
AnonymousSub/dummy_1
[ "pytorch", "bert", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "BertForSequenceClassification" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_rep...
33
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imagefolder metrics: - accuracy model-index: - name: vit-base-patch16-224-in21k-finetuned-eurosat results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofr...
AnonymousSub/dummy_2_parent
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": nul...
3
null
--- license: mit tags: - generated_from_trainer datasets: - imagefolder model-index: - name: dof-Rai2-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # dof-Rai...
AnonymousSub/rule_based_bert_quadruplet_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": nul...
8
2022-11-18T22:05:31Z
--- license: mit tags: - vision - image-classification datasets: - imagenet-1k widget: - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg example_title: Tiger - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg example_title: Teapot - src: https://hugg...
AnonymousSub/rule_based_bert_quadruplet_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": nul...
8
null
--- license: mit tags: - vision - image-classification datasets: - imagenet-21k - imagenet-1k widget: - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg example_title: Tiger - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg example_title: Teapot - sr...
AnonymousSub/rule_based_bert_quadruplet_epochs_1_shard_1_squad2.0
[ "pytorch", "bert", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_n...
3
null
--- license: mit tags: - vision - image-classification datasets: - imagenet-21k - imagenet-1k widget: - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg example_title: Tiger - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg example_title: Teapot - sr...
AnonymousSub/rule_based_hier_triplet_0.1_epochs_1_shard_1_squad2.0
[ "pytorch", "bert", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_n...
2
null
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semanti...
AnonymousSub/rule_based_roberta_bert_quadruplet_epochs_1_shard_1_squad2.0
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_re...
2
null
--- license: creativeml-openrail-m --- `m_ross artstyle,`class token License This embedding is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage. The CreativeML OpenRAIL License specifies: You can't use the embedding to deliberately produce nor share illegal or...
AnonymousSub/rule_based_roberta_bert_quadruplet_epochs_1_shard_1_wikiqa
[ "pytorch", "roberta", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "RobertaForSequenceClassification" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "...
23
null
--- license: cc-by-4.0 metrics: - bleu4 - meteor - rouge-l - bertscore - moverscore language: en datasets: - lmqg/qag_tweetqa pipeline_tag: text2text-generation tags: - questions and answers generation widget: - text: "Beyonce further expanded her acting career, starring as blues singer Etta James in the 2008 musical ...
AnonymousSub/rule_based_roberta_twostagetriplet_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size...
1
null
Access to model aazurita/selfv2 is restricted and you are not in the authorized list. Visit https://huggingface.co/aazurita/selfv2 to ask for access.
AnonymousSub/rule_based_roberta_twostagetriplet_epochs_1_shard_1_squad2.0
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_re...
4
null
--- license: apache-2.0 tags: - generated_from_trainer metrics: - bleu model-index: - name: nmt-mpst-id-en-lr_1e-05-ep_20-seq_128_bs-32 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then rem...
AnonymousSub/rule_based_roberta_twostagetriplet_epochs_1_shard_1_wikiqa
[ "pytorch", "roberta", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "RobertaForSequenceClassification" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "...
24
null
--- license: apache-2.0 tags: - generated_from_trainer metrics: - bleu model-index: - name: nmt-mpst-id-en-lr_0.0001-ep_20-seq_128_bs-32 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then re...