modelId
stringlengths
4
81
tags
list
pipeline_tag
stringclasses
17 values
config
dict
downloads
int64
0
59.7M
first_commit
timestamp[ns, tz=UTC]
card
stringlengths
51
438k
AlekseyKulnevich/Pegasus-Summarization
[ "pytorch", "pegasus", "text2text-generation", "transformers", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "PegasusForConditionalGeneration" ], "model_type": "pegasus", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
7
2023-04-16T02:28:37Z
--- license: other --- Llama 7B model to be used in huggingface/transformers. Scripts Used: https://github.com/huggingface/transformers/pull/21955 Attention: "transformers_version": "4.27.0.dev0" commit time: 2023.4.16
AlexDemon/Alex
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
2023-04-16T02:31:23Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - aeslc metrics: - rouge model-index: - name: bart-large-finetuned-aeslc-test results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: aeslc type: aeslc config: default split: test args: default metrics: - name: Rouge1 type: rouge value: 34.1259 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-large-finetuned-aeslc-test This model is a fine-tuned version of [facebook/bart-large](https://huggingface.co/facebook/bart-large) on the aeslc dataset. It achieves the following results on the evaluation set: - Loss: 2.4993 - Rouge1: 34.1259 - Rouge2: 18.262 - Rougel: 33.121 - Rougelsum: 33.1402 - Gen Len: 10.1516 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:------:|:---------:|:-------:| | 3.1135 | 1.0 | 980 | 2.4993 | 34.1259 | 18.262 | 33.121 | 33.1402 | 10.1516 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
AlexMaclean/sentence-compression-roberta
[ "pytorch", "roberta", "token-classification", "transformers", "generated_from_trainer", "license:mit", "autotrain_compatible" ]
token-classification
{ "architectures": [ "RobertaForTokenClassification" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
13
2023-04-16T02:35:54Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - audiofolder metrics: - accuracy model-index: - name: wav2vec2-base-finetuned-stop-classification-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-base-finetuned-stop-classification-1 This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the audiofolder dataset. It achieves the following results on the evaluation set: - Loss: 0.2121 - Accuracy: 0.9285 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 256 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6921 | 0.99 | 18 | 0.6781 | 0.6185 | | 0.612 | 1.97 | 36 | 0.5797 | 0.7207 | | 0.5046 | 2.96 | 54 | 0.3820 | 0.8569 | | 0.3956 | 4.0 | 73 | 0.2827 | 0.9012 | | 0.3428 | 4.99 | 91 | 0.2915 | 0.8903 | | 0.3222 | 5.97 | 109 | 0.2333 | 0.9162 | | 0.3033 | 6.96 | 127 | 0.2403 | 0.9162 | | 0.2743 | 8.0 | 146 | 0.2129 | 0.9237 | | 0.2494 | 8.99 | 164 | 0.2121 | 0.9285 | | 0.2543 | 9.86 | 180 | 0.2199 | 0.9251 | ### Framework versions - Transformers 4.27.4 - Pytorch 1.13.0 - Datasets 2.7.1 - Tokenizers 0.13.2
AlexN/xls-r-300m-pt
[ "pytorch", "wav2vec2", "automatic-speech-recognition", "pt", "dataset:mozilla-foundation/common_voice_8_0", "transformers", "robust-speech-event", "mozilla-foundation/common_voice_8_0", "generated_from_trainer", "hf-asr-leaderboard", "license:apache-2.0", "model-index" ]
automatic-speech-recognition
{ "architectures": [ "Wav2Vec2ForCTC" ], "model_type": "wav2vec2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
15
2023-04-16T02:46:23Z
--- license: artistic-2.0 tags: - chemistry - biology - medical - gpt2 --- # DrugGPT A generative drug design model based on GPT2. <img src="https://img.shields.io/badge/license-Artistic%20License%202.0-green"><img src="https://img.shields.io/badge/python-3.7-blue"><img src="https://img.shields.io/github/stars/LIYUESEN/druggpt?style=social"> ## Deployment 1. Clone ```shell git clone https://github.com/LIYUESEN/druggpt.git cd druggpt ``` Or you can visit our [GitHub repo](https://github.com/LIYUESEN/druggpt) and click *Code>Download ZIP* to download this repo. 2. Create virtual environment ```shell conda create -n druggpt python=3.7 conda activate druggpt ``` 3. Download python dependencies ```shell pip install datasets transformers scipy scikit-learn pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu117 conda install -c openbabel openbabel ``` ## Example usage Run the script with the desired arguments, such as the protein sequence, ligand prompt, number of molecules to generate, and output directory. - If you want to input a protein FASTA file ```shell python drug_generator.py -f bcl2.fasta -n 50 ``` - If you want to input the amino acid sequence of the protein ```shell python drug_generator.py -p MAKQPSDVSSECDREGRQLQPAERPPQLRPGAPTSLQTEPQGNPEGNHGGEGDSCPHGSPQGPLAPPASPGPFATRSPLFIFMRRSSLLSRSSSGYFSFDTDRSPAPMSCDKSTQTPSPPCQAFNHYLSAMASMRQAEPADMRPEIWIAQELRRIGDEFNAYYARRVFLNNYQAAEDHPRMVILRLLRYIVRLVWRMH -n 50 ``` - If you want to provide a prompt for the ligand ```shell python drug_generator.py -f bcl2.fasta -l COc1ccc(cc1)C(=O) -n 50 ``` - Note: If you are running in a Linux environment, you need to enclose the ligand's prompt with single quotes (''). ```shell python drug_generator.py -f bcl2.fasta -l 'COc1ccc(cc1)C(=O)' -n 50 ``` ## License [Artistic License 2.0](https://opensource.org/license/artistic-license-2-0-php/)
AlexeyIgnatov/albert-xlarge-v2-squad-v2
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
2023-04-16T03:09:40Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: flan-t5-large-da-multiwoz2.1_80-ep25-nonstop results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # flan-t5-large-da-multiwoz2.1_80-ep25-nonstop This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4562 - Accuracy: 34.0522 - Num: 7365 - Gen Len: 15.9048 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 64 - seed: 1799 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 25 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Num | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:--------:|:----:|:-------:| | 0.8766 | 10.81 | 400 | 0.4640 | 33.3924 | 7365 | 16.2136 | | 0.417 | 21.62 | 800 | 0.4552 | 34.0814 | 7365 | 15.8768 | ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.0+cu111 - Datasets 2.5.1 - Tokenizers 0.12.1
AlexeyYazev/my-awesome-model
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
2023-04-16T03:09:41Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: flan-t5-large-da-multiwoz2.0_80-ep25-nonstop results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # flan-t5-large-da-multiwoz2.0_80-ep25-nonstop This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4549 - Accuracy: 34.6057 - Num: 7358 - Gen Len: 16.1846 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 64 - seed: 1799 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 25 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Num | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:--------:|:----:|:-------:| | 0.8792 | 10.81 | 400 | 0.4653 | 33.0413 | 7358 | 16.1449 | | 0.4126 | 21.62 | 800 | 0.4535 | 34.5898 | 7358 | 16.078 | ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.0+cu111 - Datasets 2.5.1 - Tokenizers 0.12.1
AlgoveraAI/dcgan
[ "pytorch", "transformers" ]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
12
null
--- license: unknown --- nostalgiaClearベースにAngYuzuAnimeOdとTooOldPcGirlをまぜたもの
AmirBialer/amirbialer-Classifier
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 238.80 +/- 59.36 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
Amrrs/south-indian-foods
[ "pytorch", "tensorboard", "vit", "image-classification", "transformers", "huggingpics", "model-index", "autotrain_compatible" ]
image-classification
{ "architectures": [ "ViTForImageClassification" ], "model_type": "vit", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
21
null
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: bert-large-uncased-Hate_Offensive_or_Normal_Speech results: [] language: - en --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-large-uncased-Hate_Offensive_or_Normal_Speech This model is a fine-tuned version of [bert-large-uncased](https://huggingface.co/bert-large-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0443 - Accuracy: 0.9869 - Weighted f1: 0.9869 - Micro f1: 0.9869 - Macro f1: 0.9863 - Weighted recall: 0.9869 - Micro recall: 0.9869 - Macro recall: 0.9857 - Weighted precision: 0.9869 - Micro precision: 0.9869 - Macro precision: 0.9870 ## Model description For more information on how it was created, check out the following link: https://github.com/DunnBC22/NLP_Projects/blob/main/Multiclass%20Classification/Transformer%20Comparison/Hate%20%26%20Offensive%20Speech%20-%20BERT-Large.ipynb ### Associated Models This project is part of a comparison that included the following models: - https://huggingface.co/DunnBC22/bert-base-uncased-Hate_Offensive_or_Normal_Speech - https://huggingface.co/DunnBC22/distilbert-base-uncased-Hate_Offensive_or_Normal_Speech - https://huggingface.co/DunnBC22/fBERT-Hate_Offensive_or_Normal_Speech - https://huggingface.co/DunnBC22/hateBERT-Hate_Offensive_or_Normal_Speech ## Intended uses & limitations This model is intended to demonstrate my ability to solve a complex problem using technology. The main limitation is the quality of the data source. ## Training and evaluation data Dataset Source: https://www.kaggle.com/datasets/subhajournal/normal-hate-and-offensive-speeches ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Weighted f1 | Micro f1 | Macro f1 | Weighted recall | Micro recall | Macro recall | Weighted precision | Micro precision | Macro precision | |:-------------:|:-----:|:----:|:---------------:|:--------:|:-----------:|:--------:|:--------:|:---------------:|:------------:|:------------:|:------------------:|:---------------:|:---------------:| | 0.7991 | 1.0 | 39 | 0.4235 | 0.7430 | 0.7100 | 0.7430 | 0.6902 | 0.7430 | 0.7430 | 0.7049 | 0.7782 | 0.7430 | 0.7886 | | 0.2156 | 2.0 | 78 | 0.1072 | 0.9607 | 0.9605 | 0.9607 | 0.9585 | 0.9607 | 0.9607 | 0.9569 | 0.9607 | 0.9607 | 0.9605 | | 0.0518 | 3.0 | 117 | 0.0518 | 0.9869 | 0.9869 | 0.9869 | 0.9863 | 0.9869 | 0.9869 | 0.9857 | 0.9869 | 0.9869 | 0.9870 | | 0.0242 | 4.0 | 156 | 0.0500 | 0.9853 | 0.9852 | 0.9853 | 0.9845 | 0.9853 | 0.9853 | 0.9841 | 0.9853 | 0.9853 | 0.9850 | | 0.0163 | 5.0 | 195 | 0.0443 | 0.9869 | 0.9869 | 0.9869 | 0.9863 | 0.9869 | 0.9869 | 0.9857 | 0.9869 | 0.9869 | 0.9870 | ### Framework versions - Transformers 4.26.0 - Pytorch 1.12.1 - Datasets 2.8.0 - Tokenizers 0.12.1
AnonymousSub/SR_consert
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
2
null
--- library_name: stable-baselines3 tags: - AntBulletEnv-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: AntBulletEnv-v0 type: AntBulletEnv-v0 metrics: - type: mean_reward value: 1416.13 +/- 523.94 name: mean_reward verified: false --- # **A2C** Agent playing **AntBulletEnv-v0** This is a trained model of a **A2C** agent playing **AntBulletEnv-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
AnonymousSub/SR_declutr
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
6
null
### more info: https://github.com/D1026/Chinese-Tiger-LoRA ### test examples: https://github.com/D1026/Chinese-Tiger-LoRA/blob/main/generate_by_hf_ckpt.py #### Template: "prompt_input": "Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n\n### Instruction:\n{instruction}\n\n### Input:\n{input}\n\n### Response:\n", "prompt_no_input": "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### Instruction:\n{instruction}\n\n### Response:\n", "response_split": "### Response:"
AnonymousSub/SR_rule_based_bert_triplet_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
6
null
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: dataScienceChallenge results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # dataScienceChallenge This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
AnonymousSub/SR_rule_based_roberta_hier_quadruplet_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
5
2023-04-16T08:54:02Z
--- license: afl-3.0 language: - kk - en metrics: - bleu --- Description: finetuned mt5-base Dataset: https://github.com/Helsinki-NLP/Tatoeba-Challenge/blob/master/data/README.md Evaluation: BLEU EN-KK 11.5; KK-EN 22.68
AnonymousSub/SR_rule_based_roberta_only_classfn_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
3
2023-04-16T09:02:46Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 262.93 +/- 16.07 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
AnonymousSub/SR_rule_based_roberta_twostagetriplet_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
2023-04-15T03:18:41Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - wer model-index: - name: whisper-small-hi results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # whisper-small-hi This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4276 - Wer: 33.1118 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 4000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.0887 | 2.44 | 1000 | 0.2922 | 34.9361 | | 0.0202 | 4.89 | 2000 | 0.3473 | 34.1869 | | 0.0013 | 7.33 | 3000 | 0.4052 | 33.2007 | | 0.0005 | 9.78 | 4000 | 0.4276 | 33.1118 | ### Framework versions - Transformers 4.29.0.dev0 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
AnonymousSub/bert_mean_diff_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- library_name: stable-baselines3 tags: - AntBulletEnv-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: AntBulletEnv-v0 type: AntBulletEnv-v0 metrics: - type: mean_reward value: 2027.56 +/- 44.00 name: mean_reward verified: false --- # **A2C** Agent playing **AntBulletEnv-v0** This is a trained model of a **A2C** agent playing **AntBulletEnv-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
AnonymousSub/bert_triplet_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
2
null
--- language: - en tags: - summarization license: mit datasets: - multi_news model-index: - name: ppiiesle3y/fined-tuned-bart results: - task: type: summarization name: Summarization dataset: name: multi_news type: multi_news split: train metrics: - name: ROUGE-1 type: rouge value: 43.7065 verified: true - name: ROUGE-2 type: rouge value: 16.5533 verified: true - name: ROUGE-L type: rouge value: 24.7588 verified: true - name: ROUGE-LSUM type: rouge value: 37.7586 verified: true - name: loss type: loss value: 2.00663 verified: true - name: gen_len type: gen_len value: 129.1379 verified: true --- # TL;DR AT2 Applied Natural Language Processing Assignment ## PROJECT OBJECTIVES This project aims to use NLP technology to summarise longer passages of text into succinct and accurate summations. ## PROJECT OUTCOMES AND INSIGHTS The expected outcomes from the project is a model that is able to intake a larger body of text and provide a shortened summary that is both succinct and accurate. This will benefit most human readers by making it more efficient gain understanding from written text. Applications for this technology include as a study aide, for people in roles where they are required to quickly assess documents such as book publishers reading through manuscripts to assess if they are fit for publishing or script readers etc. The most significant impact this project has is to increase information assimilation in a compressed timeframe, thus saving time.
AnonymousSub/cline-emanuals-techqa
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - wnut_17 metrics: - precision - recall - f1 - accuracy model-index: - name: my_awesome_wnut_model results: - task: name: Token Classification type: token-classification dataset: name: wnut_17 type: wnut_17 config: wnut_17 split: test args: wnut_17 metrics: - name: Precision type: precision value: 0.5218120805369127 - name: Recall type: recall value: 0.2882298424467099 - name: F1 type: f1 value: 0.37134328358208957 - name: Accuracy type: accuracy value: 0.9401051686546108 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_awesome_wnut_model This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the wnut_17 dataset. It achieves the following results on the evaluation set: - Loss: 0.2889 - Precision: 0.5218 - Recall: 0.2882 - F1: 0.3713 - Accuracy: 0.9401 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 213 | 0.2993 | 0.3997 | 0.2122 | 0.2772 | 0.9363 | | No log | 2.0 | 426 | 0.2889 | 0.5218 | 0.2882 | 0.3713 | 0.9401 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
AnonymousSub/cline-s10-AR
[ "pytorch", "roberta", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "RobertaForSequenceClassification" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
31
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - billsum metrics: - rouge model-index: - name: my_awesome_billsum_model results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: billsum type: billsum config: default split: ca_test args: default metrics: - name: Rouge1 type: rouge value: 0.1426 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_awesome_billsum_model This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the billsum dataset. It achieves the following results on the evaluation set: - Loss: 2.4784 - Rouge1: 0.1426 - Rouge2: 0.052 - Rougel: 0.1192 - Rougelsum: 0.1191 - Gen Len: 19.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 62 | 2.7668 | 0.1263 | 0.0355 | 0.1055 | 0.1054 | 19.0 | | No log | 2.0 | 124 | 2.5587 | 0.1347 | 0.0466 | 0.1134 | 0.1134 | 19.0 | | No log | 3.0 | 186 | 2.4951 | 0.1396 | 0.0492 | 0.1161 | 0.116 | 19.0 | | No log | 4.0 | 248 | 2.4784 | 0.1426 | 0.052 | 0.1192 | 0.1191 | 19.0 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
AnonymousSub/cline-s10-SR
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: bert-large-cased-sigir-support-refute-no-label-40-2nd-test-LR10-8-5 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-large-cased-sigir-support-refute-no-label-40-2nd-test-LR10-8-5 This model is a fine-tuned version of [jojoUla/bert-large-cased-sigir-support-refute-no-label-40](https://huggingface.co/jojoUla/bert-large-cased-sigir-support-refute-no-label-40) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.7921 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 4e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 8.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 6.5771 | 1.0 | 1 | 4.3782 | | 6.0377 | 2.0 | 2 | 8.7456 | | 6.4309 | 3.0 | 3 | 5.6265 | | 6.1044 | 4.0 | 4 | 1.0357 | | 3.2773 | 5.0 | 5 | 3.2681 | | 2.7925 | 6.0 | 6 | 0.0914 | | 2.5397 | 7.0 | 7 | 1.5701 | | 1.8551 | 8.0 | 8 | 0.9440 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
AnonymousSub/cline_emanuals
[ "pytorch", "roberta", "transformers" ]
null
{ "architectures": [ "LecbertForPreTraining" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
3
null
Access to model writeup/dont-know is restricted and you are not in the authorized list. Visit https://huggingface.co/writeup/dont-know to ask for access.
AnonymousSub/cline_wikiqa
[ "pytorch", "roberta", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "RobertaForSequenceClassification" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
27
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - cnn_dailymail metrics: - rouge model-index: - name: flan-t5-xl-deepspeed-zero3-summary results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: cnn_dailymail type: cnn_dailymail config: 3.0.0 split: test args: 3.0.0 metrics: - name: Rouge1 type: rouge value: 42.6105 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # flan-t5-xl-deepspeed-zero3-summary This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the cnn_dailymail dataset. It achieves the following results on the evaluation set: - Loss: 1.4054 - Rouge1: 42.6105 - Rouge2: 20.2181 - Rougel: 29.7866 - Rougelsum: 39.4431 - Gen Len: 98.0013 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 10 - total_train_batch_size: 80 - total_eval_batch_size: 80 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | 1.6011 | 1.0 | 3589 | 1.4054 | 42.6105 | 20.2181 | 29.7866 | 39.4431 | 98.0013 | ### Framework versions - Transformers 4.26.0 - Pytorch 1.12.0 - Datasets 2.9.0 - Tokenizers 0.13.3
AnonymousSub/declutr-emanuals-s10-AR
[ "pytorch", "roberta", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "RobertaForSequenceClassification" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
29
null
--- license: bigscience-bloom-rail-1.0 tags: - generated_from_trainer model-index: - name: bloomz-1b1-eli5-pretrained results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bloomz-1b1-eli5-pretrained This model is a fine-tuned version of [bigscience/bloomz-1b1](https://huggingface.co/bigscience/bloomz-1b1) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 3.1743 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - training_steps: 2500 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.2783 | 0.4 | 1000 | 3.1924 | | 3.1211 | 0.8 | 2000 | 3.1743 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu117 - Datasets 2.11.0 - Tokenizers 0.13.3
AnonymousSub/declutr-emanuals-techqa
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: bert-large-cased-sigir-support-refute-no-label-40-2nd-test-LR10-8-6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-large-cased-sigir-support-refute-no-label-40-2nd-test-LR10-8-6 This model is a fine-tuned version of [jojoUla/bert-large-cased-sigir-support-refute-no-label-40](https://huggingface.co/jojoUla/bert-large-cased-sigir-support-refute-no-label-40) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.7754 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 4e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 8.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.8817 | 1.0 | 1 | 6.4648 | | 7.1116 | 2.0 | 2 | 8.1799 | | 5.2011 | 3.0 | 3 | 2.5263 | | 3.0932 | 4.0 | 4 | 1.0840 | | 2.5374 | 5.0 | 5 | 5.7930 | | 2.035 | 6.0 | 6 | 0.0070 | | 1.5178 | 7.0 | 7 | 0.9903 | | 1.6442 | 8.0 | 8 | 0.4513 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
AnonymousSub/declutr-s10-SR
[ "pytorch", "roberta", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "RobertaForSequenceClassification" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
36
null
--- license: creativeml-openrail-m --- # AC_H-3 **AC_H-3是和Alice-Yozakura共同制作,有非常好的背景和光影** **若生成出来的图片和预览图不同,请下载EasyNegative模型** **(** https://huggingface.co/datasets/gsdf/EasyNegative **)** **将他放入你的Stable Diffusion WebUI里面的embeddings文件夹里** ----------------------------------------------------- # AC_H-3预览图 ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE9.png) ``` 1girl,dag ears,fox tail,(loli:1.5),white dress,socks,(ribbon tied ankle|pink),Sit on the bed, Negative prompt: EasyNegative,extra fingers,fewer fingers,watermark,Invisible watermark,username,Signature,jpeg artifacts,Bad feet,Bad hands,Extra legs,Three Legs,Three feet, Steps: 28, Sampler: DPM++ 2M Karras, CFG scale: 7, Seed: 2478275294, Size: 768x384, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires resize: 1280x640, Hires steps: 30, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE8.png) ``` 1girl,dag_ears,fox_tail,(loli:1.5),white dress,White stockings,(white background: 1.8),[(glass jar:1.2),(girl in jar:1.4):(girl)], Negative prompt: Abnormal human structure,Impossible human body structure,amputation,Extra limbs,Extra fingers,Extra legs,Extra ears,3leg,Fewer fingers,Watermark,Lnvisible watermark,Username,Signature,Jpeg artifacts,Bad feet,Bad hands,high-heeled, Navel,Pencil skirt,sofa,(Katyusha maid headdress),cap,Leg lift, EasyNegative, Steps: 28, Sampler: DPM++ 2S a Karras, CFG scale: 4.5, Seed: 3182655935, Size: 480x720, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires upscale: 1.6666666666666667, Hires steps: 20, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE7.png) ``` 1girl,fox_ears,fox_tail,white hair,long hair,white dress,(loli:1.5),White stockings,no shoes,[(Transparent background:1.5)::5],illustration,Blue ambient light,Magic circle,Accumulator,Dynamic surround,Liquid aperture,Bright and colorful,Ionizer Negative prompt: EasyNegative,extra fingers,fewer fingers,watermark,Invisible watermark,username,Signature,jpeg artifacts,Bad feet,Bad hands,Extra legs,Three Legs,Three feet,(NSFW),Multiple people,More than one person,2girl,More than two legs, Steps: 28, Sampler: DPM++ 2S a Karras, CFG scale: 7, Seed: 2429212780, Size: 480x720, Model hash: 74a62f9313, Model: AC_H-3-RD-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires upscale: 1.6, Hires steps: 30, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE11.png) ``` evening, beautiful sunset, summer evening Negative prompt: EasyNegative, Steps: 35, Sampler: UniPC, CFG scale: 7, Seed: 2837482015, Size: 960x540, Model hash: eb4ffeebdf, Model: AC_H-3-RC-half, Denoising strength: 0.5, Clip skip: 2, ENSD: 31337, Hires upscale: 2, Hires steps: 8, Hires upscaler: Latent ``` --- ![sample1](https://s3.amazonaws.com/moonup/production/uploads/63fd94ea3c880680af48b776/SkD5BtPEZueJkeyD07YiZ.png) ``` great grandfather,ripe,(Older:1.25),muscle, Negative prompt: EasyNegative,Impossible human body structure,Extra limbs,Extra fingers,Extra legs,Extra ears,Fewer fingers,Watermark,Lnvisible watermark,Username,Signature,Jpeg artifacts,Bad feet,Bad hands,high-heeled shoes, Steps: 28, Sampler: Euler a, CFG scale: 6, Seed: 4065660254, Size: 480x720, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires upscale: 1.6, Hires steps: 30, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE14.png) ``` room,computer,coffee,flower vase,book,coming,1girl,long hair,(white dress|gorgeous|luxuriant),White pantyhose,(Crural wrist|ribbon:1.2|black),(ankle|ribbon|black)no shoes,blue eyes,(loli:1.5),expressionless, Negative prompt: Abnormal human structure,Impossible human body structure,amputation,Extra limbs,Extra fingers,Extra legs,Extra ears,3leg,Fewer fingers,Watermark,Lnvisible watermark,Username,Signature,Jpeg artifacts,Bad feet,Bad hands,high-heeled, Navel,Pencil skirt,sofa,(Katyusha maid headdress),cap,Leg lift, EasyNegative, Steps: 20, Sampler: Euler a, CFG scale: 4.5, Seed: 124256572, Size: 768x512, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires upscale: 1.6, Hires steps: 30, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE12.png) ``` evening, beautiful sunset, summer evening Negative prompt: EasyNegative, Steps: 35, Sampler: UniPC, CFG scale: 7, Seed: 3138791891, Size: 960x540, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.5, Clip skip: 2, ENSD: 31337, Hires upscale: 2, Hires steps: 8, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE3.png) --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE4.png) --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE6.png) --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE2.png) --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE1.png)
AnonymousSub/dummy_2
[ "pytorch", "bert", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "BertForSequenceClassification" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
39
null
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: Taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.56 +/- 2.71 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="Shot846/Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
AnonymousSub/hier_triplet_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
--- pipeline_tag: question-answering license: apache-2.0 datasets: - kejian/codesearchnet-python-linelen40-full metrics: - accuracy library_name: transformers ---
AnonymousSub/hier_triplet_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: my_awesome_eli5_clm-model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_awesome_eli5_clm-model This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 3.7357 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.8787 | 1.0 | 1128 | 3.7557 | | 3.7814 | 2.0 | 2256 | 3.7393 | | 3.7346 | 3.0 | 3384 | 3.7357 | ### Framework versions - Transformers 4.27.1 - Pytorch 2.0.0+cu117 - Datasets 2.10.1 - Tokenizers 0.13.2
AnonymousSub/rule_based_bert_hier_diff_equal_wts_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
6
null
--- license: apache-2.0 tags: - generated_from_trainer datasets: - cnn_dailymail metrics: - rouge model-index: - name: cnn_news_summary_model_trained_on_reduced_data results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: cnn_dailymail type: cnn_dailymail config: 3.0.0 split: train[:3%] args: 3.0.0 metrics: - name: Rouge1 type: rouge value: 0.2176 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # cnn_news_summary_model_trained_on_reduced_data This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the cnn_dailymail dataset. It achieves the following results on the evaluation set: - Loss: 1.5899 - Rouge1: 0.2176 - Rouge2: 0.0942 - Rougel: 0.1832 - Rougelsum: 0.1828 - Gen Len: 19.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 431 | 1.5995 | 0.2168 | 0.0933 | 0.1824 | 0.1823 | 19.0 | | 1.7979 | 2.0 | 862 | 1.5925 | 0.2177 | 0.0942 | 0.1835 | 0.1832 | 19.0 | | 1.7936 | 3.0 | 1293 | 1.5899 | 0.2176 | 0.0942 | 0.1832 | 0.1828 | 19.0 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
AnonymousSub/rule_based_bert_mean_diff_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- tags: - conversational --- # Deadpool DialoGPT Model
AnonymousSub/rule_based_bert_quadruplet_epochs_1_shard_1_wikiqa
[ "pytorch", "bert", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "BertForSequenceClassification" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
33
null
--- license: mit library_name: sklearn tags: - sklearn - skops - tabular-classification model_format: pickle model_file: model.pkl widget: structuredData: BsmtFinSF1: - 1280 - 1464 - 0 BsmtUnfSF: - 402 - 536 - 795 Condition2: - Norm - Norm - Norm ExterQual: - Ex - Gd - Gd Foundation: - PConc - PConc - PConc GarageCars: - 3 - 3 - 1 GarageType: - BuiltIn - Attchd - Detchd Heating: - GasA - GasA - GasA HeatingQC: - Ex - Ex - TA HouseStyle: - 2Story - 1Story - 2.5Fin MSSubClass: - 60 - 20 - 75 MasVnrArea: - 272.0 - 246.0 - 0.0 MasVnrType: - Stone - Stone - .nan MiscFeature: - .nan - .nan - .nan MoSold: - 8 - 7 - 3 OverallQual: - 10 - 8 - 4 Street: - Pave - Pave - Pave TotalBsmtSF: - 1682 - 2000 - 795 YearRemodAdd: - 2008 - 2005 - 1950 YrSold: - 2008 - 2007 - 2006 --- # Model description This is a Lasso regression model trained on ames housing dataset from OpenML ## Intended uses & limitations This model is not ready to be used in production. ## Training Procedure [More Information Needed] ### Hyperparameters <details> <summary> Click to expand </summary> | Hyperparameter | Value | |-----------------------------------------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | memory | | | steps | [('columntransformer', ColumnTransformer(transformers=[('pipeline',<br /> Pipeline(steps=[('standardscaler',<br /> StandardScaler()),<br /> ('simpleimputer',<br /> SimpleImputer(add_indicator=True))]),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000001CF5D97B7C0>),<br /> ('onehotencoder',<br /> OneHotEncoder(handle_unknown='ignore'),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000001CF128511E0>)])), ('lassocv', LassoCV())] | | verbose | False | | columntransformer | ColumnTransformer(transformers=[('pipeline',<br /> Pipeline(steps=[('standardscaler',<br /> StandardScaler()),<br /> ('simpleimputer',<br /> SimpleImputer(add_indicator=True))]),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000001CF5D97B7C0>),<br /> ('onehotencoder',<br /> OneHotEncoder(handle_unknown='ignore'),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000001CF128511E0>)]) | | lassocv | LassoCV() | | columntransformer__n_jobs | | | columntransformer__remainder | drop | | columntransformer__sparse_threshold | 0.3 | | columntransformer__transformer_weights | | | columntransformer__transformers | [('pipeline', Pipeline(steps=[('standardscaler', StandardScaler()),<br /> ('simpleimputer', SimpleImputer(add_indicator=True))]), <sklearn.compose._column_transformer.make_column_selector object at 0x000001CF5D97B7C0>), ('onehotencoder', OneHotEncoder(handle_unknown='ignore'), <sklearn.compose._column_transformer.make_column_selector object at 0x000001CF128511E0>)] | | columntransformer__verbose | False | | columntransformer__verbose_feature_names_out | True | | columntransformer__pipeline | Pipeline(steps=[('standardscaler', StandardScaler()),<br /> ('simpleimputer', SimpleImputer(add_indicator=True))]) | | columntransformer__onehotencoder | OneHotEncoder(handle_unknown='ignore') | | columntransformer__pipeline__memory | | | columntransformer__pipeline__steps | [('standardscaler', StandardScaler()), ('simpleimputer', SimpleImputer(add_indicator=True))] | | columntransformer__pipeline__verbose | False | | columntransformer__pipeline__standardscaler | StandardScaler() | | columntransformer__pipeline__simpleimputer | SimpleImputer(add_indicator=True) | | columntransformer__pipeline__standardscaler__copy | True | | columntransformer__pipeline__standardscaler__with_mean | True | | columntransformer__pipeline__standardscaler__with_std | True | | columntransformer__pipeline__simpleimputer__add_indicator | True | | columntransformer__pipeline__simpleimputer__copy | True | | columntransformer__pipeline__simpleimputer__fill_value | | | columntransformer__pipeline__simpleimputer__keep_empty_features | False | | columntransformer__pipeline__simpleimputer__missing_values | nan | | columntransformer__pipeline__simpleimputer__strategy | mean | | columntransformer__pipeline__simpleimputer__verbose | deprecated | | columntransformer__onehotencoder__categories | auto | | columntransformer__onehotencoder__drop | | | columntransformer__onehotencoder__dtype | <class 'numpy.float64'> | | columntransformer__onehotencoder__handle_unknown | ignore | | columntransformer__onehotencoder__max_categories | | | columntransformer__onehotencoder__min_frequency | | | columntransformer__onehotencoder__sparse | deprecated | | columntransformer__onehotencoder__sparse_output | True | | lassocv__alphas | | | lassocv__copy_X | True | | lassocv__cv | | | lassocv__eps | 0.001 | | lassocv__fit_intercept | True | | lassocv__max_iter | 1000 | | lassocv__n_alphas | 100 | | lassocv__n_jobs | | | lassocv__positive | False | | lassocv__precompute | auto | | lassocv__random_state | | | lassocv__selection | cyclic | | lassocv__tol | 0.0001 | | lassocv__verbose | False | </details> ### Model Plot <style>#sk-container-id-1 {color: black;background-color: white;}#sk-container-id-1 pre{padding: 0;}#sk-container-id-1 div.sk-toggleable {background-color: white;}#sk-container-id-1 label.sk-toggleable__label {cursor: pointer;display: block;width: 100%;margin-bottom: 0;padding: 0.3em;box-sizing: border-box;text-align: center;}#sk-container-id-1 label.sk-toggleable__label-arrow:before {content: "▸";float: left;margin-right: 0.25em;color: #696969;}#sk-container-id-1 label.sk-toggleable__label-arrow:hover:before {color: black;}#sk-container-id-1 div.sk-estimator:hover label.sk-toggleable__label-arrow:before {color: black;}#sk-container-id-1 div.sk-toggleable__content {max-height: 0;max-width: 0;overflow: hidden;text-align: left;background-color: #f0f8ff;}#sk-container-id-1 div.sk-toggleable__content pre {margin: 0.2em;color: black;border-radius: 0.25em;background-color: #f0f8ff;}#sk-container-id-1 input.sk-toggleable__control:checked~div.sk-toggleable__content {max-height: 200px;max-width: 100%;overflow: auto;}#sk-container-id-1 input.sk-toggleable__control:checked~label.sk-toggleable__label-arrow:before {content: "▾";}#sk-container-id-1 div.sk-estimator input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-label input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 input.sk-hidden--visually {border: 0;clip: rect(1px 1px 1px 1px);clip: rect(1px, 1px, 1px, 1px);height: 1px;margin: -1px;overflow: hidden;padding: 0;position: absolute;width: 1px;}#sk-container-id-1 div.sk-estimator {font-family: monospace;background-color: #f0f8ff;border: 1px dotted black;border-radius: 0.25em;box-sizing: border-box;margin-bottom: 0.5em;}#sk-container-id-1 div.sk-estimator:hover {background-color: #d4ebff;}#sk-container-id-1 div.sk-parallel-item::after {content: "";width: 100%;border-bottom: 1px solid gray;flex-grow: 1;}#sk-container-id-1 div.sk-label:hover label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-serial::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: 0;}#sk-container-id-1 div.sk-serial {display: flex;flex-direction: column;align-items: center;background-color: white;padding-right: 0.2em;padding-left: 0.2em;position: relative;}#sk-container-id-1 div.sk-item {position: relative;z-index: 1;}#sk-container-id-1 div.sk-parallel {display: flex;align-items: stretch;justify-content: center;background-color: white;position: relative;}#sk-container-id-1 div.sk-item::before, #sk-container-id-1 div.sk-parallel-item::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: -1;}#sk-container-id-1 div.sk-parallel-item {display: flex;flex-direction: column;z-index: 1;position: relative;background-color: white;}#sk-container-id-1 div.sk-parallel-item:first-child::after {align-self: flex-end;width: 50%;}#sk-container-id-1 div.sk-parallel-item:last-child::after {align-self: flex-start;width: 50%;}#sk-container-id-1 div.sk-parallel-item:only-child::after {width: 0;}#sk-container-id-1 div.sk-dashed-wrapped {border: 1px dashed gray;margin: 0 0.4em 0.5em 0.4em;box-sizing: border-box;padding-bottom: 0.4em;background-color: white;}#sk-container-id-1 div.sk-label label {font-family: monospace;font-weight: bold;display: inline-block;line-height: 1.2em;}#sk-container-id-1 div.sk-label-container {text-align: center;}#sk-container-id-1 div.sk-container {/* jupyter's `normalize.less` sets `[hidden] { display: none; }` but bootstrap.min.css set `[hidden] { display: none !important; }` so we also need the `!important` here to be able to override the default hidden behavior on the sphinx rendered scikit-learn.org. See: https://github.com/scikit-learn/scikit-learn/issues/21755 */display: inline-block !important;position: relative;}#sk-container-id-1 div.sk-text-repr-fallback {display: none;}</style><div id="sk-container-id-1" class="sk-top-container" style="overflow: auto;"><div class="sk-text-repr-fallback"><pre>Pipeline(steps=[(&#x27;columntransformer&#x27;,ColumnTransformer(transformers=[(&#x27;pipeline&#x27;,Pipeline(steps=[(&#x27;standardscaler&#x27;,StandardScaler()),(&#x27;simpleimputer&#x27;,SimpleImputer(add_indicator=True))]),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001CF5D97B7C0&gt;),(&#x27;onehotencoder&#x27;,OneHotEncoder(handle_unknown=&#x27;ignore&#x27;),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001CF128511E0&gt;)])),(&#x27;lassocv&#x27;, LassoCV())])</pre><b>In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook. <br />On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.</b></div><div class="sk-container" hidden><div class="sk-item sk-dashed-wrapped"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-1" type="checkbox" ><label for="sk-estimator-id-1" class="sk-toggleable__label sk-toggleable__label-arrow">Pipeline</label><div class="sk-toggleable__content"><pre>Pipeline(steps=[(&#x27;columntransformer&#x27;,ColumnTransformer(transformers=[(&#x27;pipeline&#x27;,Pipeline(steps=[(&#x27;standardscaler&#x27;,StandardScaler()),(&#x27;simpleimputer&#x27;,SimpleImputer(add_indicator=True))]),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001CF5D97B7C0&gt;),(&#x27;onehotencoder&#x27;,OneHotEncoder(handle_unknown=&#x27;ignore&#x27;),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001CF128511E0&gt;)])),(&#x27;lassocv&#x27;, LassoCV())])</pre></div></div></div><div class="sk-serial"><div class="sk-item sk-dashed-wrapped"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-2" type="checkbox" ><label for="sk-estimator-id-2" class="sk-toggleable__label sk-toggleable__label-arrow">columntransformer: ColumnTransformer</label><div class="sk-toggleable__content"><pre>ColumnTransformer(transformers=[(&#x27;pipeline&#x27;,Pipeline(steps=[(&#x27;standardscaler&#x27;,StandardScaler()),(&#x27;simpleimputer&#x27;,SimpleImputer(add_indicator=True))]),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001CF5D97B7C0&gt;),(&#x27;onehotencoder&#x27;,OneHotEncoder(handle_unknown=&#x27;ignore&#x27;),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001CF128511E0&gt;)])</pre></div></div></div><div class="sk-parallel"><div class="sk-parallel-item"><div class="sk-item"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-3" type="checkbox" ><label for="sk-estimator-id-3" class="sk-toggleable__label sk-toggleable__label-arrow">pipeline</label><div class="sk-toggleable__content"><pre>&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001CF5D97B7C0&gt;</pre></div></div></div><div class="sk-serial"><div class="sk-item"><div class="sk-serial"><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-4" type="checkbox" ><label for="sk-estimator-id-4" class="sk-toggleable__label sk-toggleable__label-arrow">StandardScaler</label><div class="sk-toggleable__content"><pre>StandardScaler()</pre></div></div></div><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-5" type="checkbox" ><label for="sk-estimator-id-5" class="sk-toggleable__label sk-toggleable__label-arrow">SimpleImputer</label><div class="sk-toggleable__content"><pre>SimpleImputer(add_indicator=True)</pre></div></div></div></div></div></div></div></div><div class="sk-parallel-item"><div class="sk-item"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-6" type="checkbox" ><label for="sk-estimator-id-6" class="sk-toggleable__label sk-toggleable__label-arrow">onehotencoder</label><div class="sk-toggleable__content"><pre>&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001CF128511E0&gt;</pre></div></div></div><div class="sk-serial"><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-7" type="checkbox" ><label for="sk-estimator-id-7" class="sk-toggleable__label sk-toggleable__label-arrow">OneHotEncoder</label><div class="sk-toggleable__content"><pre>OneHotEncoder(handle_unknown=&#x27;ignore&#x27;)</pre></div></div></div></div></div></div></div></div><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-8" type="checkbox" ><label for="sk-estimator-id-8" class="sk-toggleable__label sk-toggleable__label-arrow">LassoCV</label><div class="sk-toggleable__content"><pre>LassoCV()</pre></div></div></div></div></div></div></div> ## Evaluation Results | Metric | Value | |----------|----------| | R2 score | 0.753308 | | MAE | 0.112742 | # How to Get Started with the Model Use the following code to get started: ```python import joblib from skops.hub_utils import download import json import pandas as pd download(repo_id="haizad/ames-housing-lasso-predictor", dst='ames-housing-lasso-predictor') pipeline = joblib.load( "ames-housing-lasso-predictor/model.pkl") with open("ames-housing-lasso-predictor/config.json") as f: config = json.load(f) pipeline.predict(pd.DataFrame.from_dict(config["sklearn"]["example_input"])) ``` # Model Card Authors This model card is written by following authors: [More Information Needed] # Model Card Contact You can contact the model card authors through following channels: [More Information Needed] # Citation Below you can find information related to citation. **BibTeX:** ``` [More Information Needed] ``` # Evaluation ![evaluation](prediction_error.png)
AnonymousSub/rule_based_bert_triplet_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
//Stable_Diffusion.ipynb// #Le texte commenté (#) en vert ne correspondent pas à du code exécutable. Il s’agit d’indications que nous avons rajouté afin de vous aider à comprendre notre modèle. #On vérifie le type de GPU et de VRAM disponibles. #GPU = puce informatique qui effectue des calculs mathématiques rapides, principalement pour le rendu d'images. #VRAM = mémoire vidéo. !nvidia-smi --query-gpu=name,memory.total,memory.free --format=csv,noheader #On importe le modèle pré-entrainé. !wget -q https://github.com/ShivamShrirao/diffusers/raw/main/examples/dreambooth/train_dreambooth.py !wget -q https://github.com/ShivamShrirao/diffusers/raw/main/scripts/convert_diffusers_to_original_stable_diffusion.py %pip install -qq git+https://github.com/ShivamShrirao/diffusers %pip install -q -U --pre triton %pip install -q accelerate transformers ftfy bitsandbytes==0.35.0 gradio natsort safetensors #On se connecte à HuggingFace, bibliothèque open source contenant des modèles pré-formés. !mkdir -p ~/.huggingface HUGGINGFACE_TOKEN = "" #@param {type:"string"} !echo -n "{HUGGINGFACE_TOKEN}" > ~/.huggingface/token #On installe les xformers à partir de wheels précompilés. # xformers = bibliothèque de modèles pré-entraînés. # wheels = composants de l'écosystème Python qui aident à faire fonctionner les installations de packs. %pip install --no-deps -q https://github.com/brian6091/xformers-wheels/releases/download/0.0.15.dev0%2B4c06c79/xformers-0.0.15.dev0+4c06c79.d20221205-cp38-cp38-linux_x86_64.whl #On sauvegarde les paramètres du modèle dans notre Google drive. save_to_gdrive = False if save_to_gdrive: from google.colab import drive drive.mount('/content/drive') MODEL_NAME = "runwayml/stable-diffusion-v1-5" OUTPUT_DIR = "stable_diffusion_weights/zwx" if save_to_gdrive: OUTPUT_DIR = "/content/drive/MyDrive/" + OUTPUT_DIR else: OUTPUT_DIR = "/content/" + OUTPUT_DIR print(f"[*] Weights will be saved at {OUTPUT_DIR}") !mkdir -p $OUTPUT_DIR #On crée une liste à partir de notre corpus d’images. concepts_list = [ { "instance_prompt": "photo of sacristy room", "class_prompt": "photo of a room", "instance_data_dir": "/content/data/sacristy", "class_data_dir": "/content/data/room" }, { "instance_prompt": "photo of screens furniture", "class_prompt": "photo of a furniture", "instance_data_dir": "/content/data/screens", "class_data_dir": "/content/data/furniture" } ] import json import os for c in concepts_list: os.makedirs(c["instance_data_dir"], exist_ok=True) with open("concepts_list.json", "w") as f: json.dump(concepts_list, f, indent=4) #On télécharge notre corpus d’images. import os from google.colab import files import shutil for c in concepts_list: print(f"Uploading instance images for `{c['instance_prompt']}`") uploaded = files.upload() for filename in uploaded.keys(): dst_path = os.path.join(c['instance_data_dir'], filename) shutil.move(filename, dst_path) #On entraine notre corpus d’images. !accelerate launch train_dreambooth.py \ --pretrained_model_name_or_path=$MODEL_NAME \ --pretrained_vae_name_or_path="stabilityai/sd-vae-ft-mse" \ --output_dir=$OUTPUT_DIR \ --revision="fp16" \ --with_prior_preservation --prior_loss_weight=1.0 \ --seed=1337 \ --resolution=512 \ --train_batch_size=1 \ --train_text_encoder \ --mixed_precision="fp16" \ --use_8bit_adam \ --gradient_accumulation_steps=1 \ --learning_rate=1e-6 \ --lr_scheduler="constant" \ --lr_warmup_steps=0 \ --num_class_images=50 \ --sample_batch_size=4 \ --max_train_steps=800 \ --save_interval=10000 \ --concepts_list="concepts_list.json" #On renseigne le chemin d’accès vers les paramètres du modèle pré-entrainé. WEIGHTS_DIR = "" #@param {type:"string"} if WEIGHTS_DIR == "": from natsort import natsorted from glob import glob import os WEIGHTS_DIR = natsorted(glob(OUTPUT_DIR + os.sep + "*"))[-1] print(f"[*] WEIGHTS_DIR={WEIGHTS_DIR}") #On renseigne le chemin d’accès vers les paramètres du modèle pré-entrainé. WEIGHTS_DIR = "" #@param {type:"string"} if WEIGHTS_DIR == "": from natsort import natsorted from glob import glob import os WEIGHTS_DIR = natsorted(glob(OUTPUT_DIR + os.sep + "*"))[-1] print(f"[*] WEIGHTS_DIR={WEIGHTS_DIR}") #On demande à utiliser le modèle pré-entrainé. import torch from torch import autocast from diffusers import StableDiffusionPipeline, DDIMScheduler from IPython.display import display model_path = WEIGHTS_DIR # If you want to use previously trained model saved in gdrive, replace this with the full path of model in gdrive pipe = StableDiffusionPipeline.from_pretrained(model_path, safety_checker=None, torch_dtype=torch.float16).to("cuda") pipe.scheduler = DDIMScheduler.from_config(pipe.scheduler.config) pipe.enable_xformers_memory_efficient_attention() g_cuda = None #On vient générer nos images à partir d’une entrée textuelle, le prompt. prompt = "photo of a sacristy formed by three red satin screens" negative_prompt = "" num_samples = 4 guidance_scale = 7.5 num_inference_steps = 24 height = 512 width = 512 with autocast("cuda"), torch.inference_mode(): images = pipe( prompt, height=height, width=width, negative_prompt=negative_prompt, num_images_per_prompt=num_samples, num_inference_steps=num_inference_steps, guidance_scale=guidance_scale, generator=g_cuda ).images for img in images: display(img)
AnonymousSub/rule_based_bert_triplet_epochs_1_shard_1_squad2.0
[ "pytorch", "bert", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
3
null
--- library_name: sample-factory tags: - deep-reinforcement-learning - reinforcement-learning - sample-factory model-index: - name: APPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: doom_health_gathering_supreme type: doom_health_gathering_supreme metrics: - type: mean_reward value: 11.03 +/- 5.23 name: mean_reward verified: false --- A(n) **APPO** model trained on the **doom_health_gathering_supreme** environment. This model was trained using Sample-Factory 2.0: https://github.com/alex-petrenko/sample-factory. Documentation for how to use Sample-Factory can be found at https://www.samplefactory.dev/ ## Downloading the model After installing Sample-Factory, download the model with: ``` python -m sample_factory.huggingface.load_from_hub -r Maulik-P/rl_course_vizdoom_health_gathering_supreme ``` ## Using the model To run the model after download, use the `enjoy` script corresponding to this environment: ``` python -m .usr.local.lib.python3.9.dist-packages.ipykernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme ``` You can also upload models to the Hugging Face Hub using the same script with the `--push_to_hub` flag. See https://www.samplefactory.dev/10-huggingface/huggingface/ for more details ## Training with this model To continue training with this model, use the `train` script corresponding to this environment: ``` python -m .usr.local.lib.python3.9.dist-packages.ipykernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme --restart_behavior=resume --train_for_env_steps=10000000000 ``` Note, you may have to adjust `--train_for_env_steps` to a suitably high number as the experiment will resume at the number of steps it concluded at.
AnonymousSub/rule_based_hier_quadruplet_0.1_epochs_1_shard_1_squad2.0
[ "pytorch", "bert", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- license: "cc-by-nc-4.0" tags: - vision - video-classification --- # VideoMAE (huge-sized model, fine-tuned on Kinetics-400) VideoMAE model pre-trained for 1600 epochs in a self-supervised way and fine-tuned in a supervised way on Kinetics-400. It was introduced in the paper [VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training](https://arxiv.org/abs/2203.12602) by Tong et al. and first released in [this repository](https://github.com/MCG-NJU/VideoMAE). Disclaimer: The team releasing VideoMAE did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description VideoMAE is an extension of [Masked Autoencoders (MAE)](https://arxiv.org/abs/2111.06377) to video. The architecture of the model is very similar to that of a standard Vision Transformer (ViT), with a decoder on top for predicting pixel values for masked patches. Videos are presented to the model as a sequence of fixed-size patches (resolution 16x16), which are linearly embedded. One also adds a [CLS] token to the beginning of a sequence to use it for classification tasks. One also adds fixed sinus/cosinus position embeddings before feeding the sequence to the layers of the Transformer encoder. By pre-training the model, it learns an inner representation of videos that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled videos for instance, you can train a standard classifier by placing a linear layer on top of the pre-trained encoder. One typically places a linear layer on top of the [CLS] token, as the last hidden state of this token can be seen as a representation of an entire video. ## Intended uses & limitations You can use the raw model for video classification into one of the 400 possible Kinetics-400 labels. ### How to use Here is how to use this model to classify a video: ```python from transformers import VideoMAEImageProcessor, VideoMAEForVideoClassification import numpy as np import torch video = list(np.random.randn(16, 3, 224, 224)) processor = VideoMAEImageProcessor.from_pretrained("MCG-NJU/videomae-huge-finetuned-kinetics") model = VideoMAEForVideoClassification.from_pretrained("MCG-NJU/videomae-huge-finetuned-kinetics") inputs = processor(video, return_tensors="pt") with torch.no_grad(): outputs = model(**inputs) logits = outputs.logits predicted_class_idx = logits.argmax(-1).item() print("Predicted class:", model.config.id2label[predicted_class_idx]) ``` For more code examples, we refer to the [documentation](https://huggingface.co/transformers/main/model_doc/videomae.html#). ## Training data (to do, feel free to open a PR) ## Training procedure ### Preprocessing (to do, feel free to open a PR) ### Pretraining (to do, feel free to open a PR) ## Evaluation results This model obtains a top-1 accuracy of 86.6 and a top-5 accuracy of 97.1 on the test set of Kinetics-400. ### BibTeX entry and citation info ```bibtex misc{https://doi.org/10.48550/arxiv.2203.12602, doi = {10.48550/ARXIV.2203.12602}, url = {https://arxiv.org/abs/2203.12602}, author = {Tong, Zhan and Song, Yibing and Wang, Jue and Wang, Limin}, keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ```
AnonymousSub/rule_based_hier_quadruplet_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
--- license: mit library_name: sklearn tags: - sklearn - skops - tabular-classification model_format: pickle model_file: model.pkl widget: structuredData: BsmtFinSF1: - 1280 - 1464 - 0 BsmtUnfSF: - 402 - 536 - 795 Condition2: - Norm - Norm - Norm ExterQual: - Ex - Gd - Gd Foundation: - PConc - PConc - PConc GarageCars: - 3 - 3 - 1 GarageType: - BuiltIn - Attchd - Detchd Heating: - GasA - GasA - GasA HeatingQC: - Ex - Ex - TA HouseStyle: - 2Story - 1Story - 2.5Fin MSSubClass: - 60 - 20 - 75 MasVnrArea: - 272.0 - 246.0 - 0.0 MasVnrType: - Stone - Stone - .nan MiscFeature: - .nan - .nan - .nan MoSold: - 8 - 7 - 3 OverallQual: - 10 - 8 - 4 Street: - Pave - Pave - Pave TotalBsmtSF: - 1682 - 2000 - 795 YearRemodAdd: - 2008 - 2005 - 1950 YrSold: - 2008 - 2007 - 2006 --- # Model description This is a gradient boosted regression model trained on ames housing dataset from OpenML. ## Intended uses & limitations This model is not ready to be used in production. ## Training Procedure [More Information Needed] ### Hyperparameters <details> <summary> Click to expand </summary> | Hyperparameter | Value | |----------------------------------------------------------|-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | memory | | | steps | [('columntransformer', ColumnTransformer(transformers=[('simpleimputer',<br /> SimpleImputer(add_indicator=True),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000002A2B7A2B730>),<br /> ('ordinalencoder',<br /> OrdinalEncoder(encoded_missing_value=-2,<br /> handle_unknown='use_encoded_value',<br /> unknown_value=-1),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000002A2EC9B9180>)])), ('histgradientboostingregressor', HistGradientBoostingRegressor(random_state=0))] | | verbose | False | | columntransformer | ColumnTransformer(transformers=[('simpleimputer',<br /> SimpleImputer(add_indicator=True),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000002A2B7A2B730>),<br /> ('ordinalencoder',<br /> OrdinalEncoder(encoded_missing_value=-2,<br /> handle_unknown='use_encoded_value',<br /> unknown_value=-1),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000002A2EC9B9180>)]) | | histgradientboostingregressor | HistGradientBoostingRegressor(random_state=0) | | columntransformer__n_jobs | | | columntransformer__remainder | drop | | columntransformer__sparse_threshold | 0.3 | | columntransformer__transformer_weights | | | columntransformer__transformers | [('simpleimputer', SimpleImputer(add_indicator=True), <sklearn.compose._column_transformer.make_column_selector object at 0x000002A2B7A2B730>), ('ordinalencoder', OrdinalEncoder(encoded_missing_value=-2, handle_unknown='use_encoded_value',<br /> unknown_value=-1), <sklearn.compose._column_transformer.make_column_selector object at 0x000002A2EC9B9180>)] | | columntransformer__verbose | False | | columntransformer__verbose_feature_names_out | True | | columntransformer__simpleimputer | SimpleImputer(add_indicator=True) | | columntransformer__ordinalencoder | OrdinalEncoder(encoded_missing_value=-2, handle_unknown='use_encoded_value',<br /> unknown_value=-1) | | columntransformer__simpleimputer__add_indicator | True | | columntransformer__simpleimputer__copy | True | | columntransformer__simpleimputer__fill_value | | | columntransformer__simpleimputer__keep_empty_features | False | | columntransformer__simpleimputer__missing_values | nan | | columntransformer__simpleimputer__strategy | mean | | columntransformer__simpleimputer__verbose | deprecated | | columntransformer__ordinalencoder__categories | auto | | columntransformer__ordinalencoder__dtype | <class 'numpy.float64'> | | columntransformer__ordinalencoder__encoded_missing_value | -2 | | columntransformer__ordinalencoder__handle_unknown | use_encoded_value | | columntransformer__ordinalencoder__unknown_value | -1 | | histgradientboostingregressor__categorical_features | | | histgradientboostingregressor__early_stopping | auto | | histgradientboostingregressor__interaction_cst | | | histgradientboostingregressor__l2_regularization | 0.0 | | histgradientboostingregressor__learning_rate | 0.1 | | histgradientboostingregressor__loss | squared_error | | histgradientboostingregressor__max_bins | 255 | | histgradientboostingregressor__max_depth | | | histgradientboostingregressor__max_iter | 100 | | histgradientboostingregressor__max_leaf_nodes | 31 | | histgradientboostingregressor__min_samples_leaf | 20 | | histgradientboostingregressor__monotonic_cst | | | histgradientboostingregressor__n_iter_no_change | 10 | | histgradientboostingregressor__quantile | | | histgradientboostingregressor__random_state | 0 | | histgradientboostingregressor__scoring | loss | | histgradientboostingregressor__tol | 1e-07 | | histgradientboostingregressor__validation_fraction | 0.1 | | histgradientboostingregressor__verbose | 0 | | histgradientboostingregressor__warm_start | False | </details> ### Model Plot <style>#sk-container-id-1 {color: black;background-color: white;}#sk-container-id-1 pre{padding: 0;}#sk-container-id-1 div.sk-toggleable {background-color: white;}#sk-container-id-1 label.sk-toggleable__label {cursor: pointer;display: block;width: 100%;margin-bottom: 0;padding: 0.3em;box-sizing: border-box;text-align: center;}#sk-container-id-1 label.sk-toggleable__label-arrow:before {content: "▸";float: left;margin-right: 0.25em;color: #696969;}#sk-container-id-1 label.sk-toggleable__label-arrow:hover:before {color: black;}#sk-container-id-1 div.sk-estimator:hover label.sk-toggleable__label-arrow:before {color: black;}#sk-container-id-1 div.sk-toggleable__content {max-height: 0;max-width: 0;overflow: hidden;text-align: left;background-color: #f0f8ff;}#sk-container-id-1 div.sk-toggleable__content pre {margin: 0.2em;color: black;border-radius: 0.25em;background-color: #f0f8ff;}#sk-container-id-1 input.sk-toggleable__control:checked~div.sk-toggleable__content {max-height: 200px;max-width: 100%;overflow: auto;}#sk-container-id-1 input.sk-toggleable__control:checked~label.sk-toggleable__label-arrow:before {content: "▾";}#sk-container-id-1 div.sk-estimator input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-label input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 input.sk-hidden--visually {border: 0;clip: rect(1px 1px 1px 1px);clip: rect(1px, 1px, 1px, 1px);height: 1px;margin: -1px;overflow: hidden;padding: 0;position: absolute;width: 1px;}#sk-container-id-1 div.sk-estimator {font-family: monospace;background-color: #f0f8ff;border: 1px dotted black;border-radius: 0.25em;box-sizing: border-box;margin-bottom: 0.5em;}#sk-container-id-1 div.sk-estimator:hover {background-color: #d4ebff;}#sk-container-id-1 div.sk-parallel-item::after {content: "";width: 100%;border-bottom: 1px solid gray;flex-grow: 1;}#sk-container-id-1 div.sk-label:hover label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-serial::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: 0;}#sk-container-id-1 div.sk-serial {display: flex;flex-direction: column;align-items: center;background-color: white;padding-right: 0.2em;padding-left: 0.2em;position: relative;}#sk-container-id-1 div.sk-item {position: relative;z-index: 1;}#sk-container-id-1 div.sk-parallel {display: flex;align-items: stretch;justify-content: center;background-color: white;position: relative;}#sk-container-id-1 div.sk-item::before, #sk-container-id-1 div.sk-parallel-item::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: -1;}#sk-container-id-1 div.sk-parallel-item {display: flex;flex-direction: column;z-index: 1;position: relative;background-color: white;}#sk-container-id-1 div.sk-parallel-item:first-child::after {align-self: flex-end;width: 50%;}#sk-container-id-1 div.sk-parallel-item:last-child::after {align-self: flex-start;width: 50%;}#sk-container-id-1 div.sk-parallel-item:only-child::after {width: 0;}#sk-container-id-1 div.sk-dashed-wrapped {border: 1px dashed gray;margin: 0 0.4em 0.5em 0.4em;box-sizing: border-box;padding-bottom: 0.4em;background-color: white;}#sk-container-id-1 div.sk-label label {font-family: monospace;font-weight: bold;display: inline-block;line-height: 1.2em;}#sk-container-id-1 div.sk-label-container {text-align: center;}#sk-container-id-1 div.sk-container {/* jupyter's `normalize.less` sets `[hidden] { display: none; }` but bootstrap.min.css set `[hidden] { display: none !important; }` so we also need the `!important` here to be able to override the default hidden behavior on the sphinx rendered scikit-learn.org. See: https://github.com/scikit-learn/scikit-learn/issues/21755 */display: inline-block !important;position: relative;}#sk-container-id-1 div.sk-text-repr-fallback {display: none;}</style><div id="sk-container-id-1" class="sk-top-container" style="overflow: auto;"><div class="sk-text-repr-fallback"><pre>Pipeline(steps=[(&#x27;columntransformer&#x27;,ColumnTransformer(transformers=[(&#x27;simpleimputer&#x27;,SimpleImputer(add_indicator=True),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000002A2B7A2B730&gt;),(&#x27;ordinalencoder&#x27;,OrdinalEncoder(encoded_missing_value=-2,handle_unknown=&#x27;use_encoded_value&#x27;,unknown_value=-1),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000002A2EC9B9180&gt;)])),(&#x27;histgradientboostingregressor&#x27;,HistGradientBoostingRegressor(random_state=0))])</pre><b>In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook. <br />On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.</b></div><div class="sk-container" hidden><div class="sk-item sk-dashed-wrapped"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-1" type="checkbox" ><label for="sk-estimator-id-1" class="sk-toggleable__label sk-toggleable__label-arrow">Pipeline</label><div class="sk-toggleable__content"><pre>Pipeline(steps=[(&#x27;columntransformer&#x27;,ColumnTransformer(transformers=[(&#x27;simpleimputer&#x27;,SimpleImputer(add_indicator=True),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000002A2B7A2B730&gt;),(&#x27;ordinalencoder&#x27;,OrdinalEncoder(encoded_missing_value=-2,handle_unknown=&#x27;use_encoded_value&#x27;,unknown_value=-1),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000002A2EC9B9180&gt;)])),(&#x27;histgradientboostingregressor&#x27;,HistGradientBoostingRegressor(random_state=0))])</pre></div></div></div><div class="sk-serial"><div class="sk-item sk-dashed-wrapped"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-2" type="checkbox" ><label for="sk-estimator-id-2" class="sk-toggleable__label sk-toggleable__label-arrow">columntransformer: ColumnTransformer</label><div class="sk-toggleable__content"><pre>ColumnTransformer(transformers=[(&#x27;simpleimputer&#x27;,SimpleImputer(add_indicator=True),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000002A2B7A2B730&gt;),(&#x27;ordinalencoder&#x27;,OrdinalEncoder(encoded_missing_value=-2,handle_unknown=&#x27;use_encoded_value&#x27;,unknown_value=-1),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000002A2EC9B9180&gt;)])</pre></div></div></div><div class="sk-parallel"><div class="sk-parallel-item"><div class="sk-item"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-3" type="checkbox" ><label for="sk-estimator-id-3" class="sk-toggleable__label sk-toggleable__label-arrow">simpleimputer</label><div class="sk-toggleable__content"><pre>&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000002A2B7A2B730&gt;</pre></div></div></div><div class="sk-serial"><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-4" type="checkbox" ><label for="sk-estimator-id-4" class="sk-toggleable__label sk-toggleable__label-arrow">SimpleImputer</label><div class="sk-toggleable__content"><pre>SimpleImputer(add_indicator=True)</pre></div></div></div></div></div></div><div class="sk-parallel-item"><div class="sk-item"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-5" type="checkbox" ><label for="sk-estimator-id-5" class="sk-toggleable__label sk-toggleable__label-arrow">ordinalencoder</label><div class="sk-toggleable__content"><pre>&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000002A2EC9B9180&gt;</pre></div></div></div><div class="sk-serial"><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-6" type="checkbox" ><label for="sk-estimator-id-6" class="sk-toggleable__label sk-toggleable__label-arrow">OrdinalEncoder</label><div class="sk-toggleable__content"><pre>OrdinalEncoder(encoded_missing_value=-2, handle_unknown=&#x27;use_encoded_value&#x27;,unknown_value=-1)</pre></div></div></div></div></div></div></div></div><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-7" type="checkbox" ><label for="sk-estimator-id-7" class="sk-toggleable__label sk-toggleable__label-arrow">HistGradientBoostingRegressor</label><div class="sk-toggleable__content"><pre>HistGradientBoostingRegressor(random_state=0)</pre></div></div></div></div></div></div></div> ## Evaluation Results | Metric | Value | |----------|----------| | R2 score | 0.838471 | | MAE | 0.111495 | # How to Get Started with the Model Use the following code to get started: ```python import joblib from skops.hub_utils import download import json import pandas as pd download(repo_id="haizad/ames-housing-gbdt-predictor", dst='ames-housing-gbdt-predictor') pipeline = joblib.load( "ames-housing-gbdt-predictor/model.pkl") with open("ames-housing-gbdt-predictor/config.json") as f: config = json.load(f) pipeline.predict(pd.DataFrame.from_dict(config["sklearn"]["example_input"])) ``` # Model Card Authors This model card is written by following authors: [More Information Needed] # Model Card Contact You can contact the model card authors through following channels: [More Information Needed] # Citation Below you can find information related to citation. **BibTeX:** ``` [More Information Needed] ``` # Intended uses & limitations This model is not ready to be used in production. # Evaluation ![Evaluation](evaluation.png)
AnonymousSub/rule_based_hier_quadruplet_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- license: "cc-by-nc-4.0" tags: - vision - video-classification --- # VideoMAE (small-sized model, fine-tuned on Kinetics-400) VideoMAE model pre-trained for 1600 epochs in a self-supervised way and fine-tuned in a supervised way on Kinetics-400. It was introduced in the paper [VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training](https://arxiv.org/abs/2203.12602) by Tong et al. and first released in [this repository](https://github.com/MCG-NJU/VideoMAE). Disclaimer: The team releasing VideoMAE did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description VideoMAE is an extension of [Masked Autoencoders (MAE)](https://arxiv.org/abs/2111.06377) to video. The architecture of the model is very similar to that of a standard Vision Transformer (ViT), with a decoder on top for predicting pixel values for masked patches. Videos are presented to the model as a sequence of fixed-size patches (resolution 16x16), which are linearly embedded. One also adds a [CLS] token to the beginning of a sequence to use it for classification tasks. One also adds fixed sinus/cosinus position embeddings before feeding the sequence to the layers of the Transformer encoder. By pre-training the model, it learns an inner representation of videos that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled videos for instance, you can train a standard classifier by placing a linear layer on top of the pre-trained encoder. One typically places a linear layer on top of the [CLS] token, as the last hidden state of this token can be seen as a representation of an entire video. ## Intended uses & limitations You can use the raw model for video classification into one of the 400 possible Kinetics-400 labels. ### How to use Here is how to use this model to classify a video: ```python from transformers import VideoMAEImageProcessor, VideoMAEForVideoClassification import numpy as np import torch video = list(np.random.randn(16, 3, 224, 224)) processor = VideoMAEImageProcessor.from_pretrained("MCG-NJU/videomae-small-finetuned-kinetics") model = VideoMAEForVideoClassification.from_pretrained("MCG-NJU/videomae-small-finetuned-kinetics") inputs = processor(video, return_tensors="pt") with torch.no_grad(): outputs = model(**inputs) logits = outputs.logits predicted_class_idx = logits.argmax(-1).item() print("Predicted class:", model.config.id2label[predicted_class_idx]) ``` For more code examples, we refer to the [documentation](https://huggingface.co/transformers/main/model_doc/videomae.html#). ## Training data (to do, feel free to open a PR) ## Training procedure ### Preprocessing (to do, feel free to open a PR) ### Pretraining (to do, feel free to open a PR) ## Evaluation results This model obtains a top-1 accuracy of 79.0 and a top-5 accuracy of 93.8 on the test set of Kinetics-400. ### BibTeX entry and citation info ```bibtex misc{https://doi.org/10.48550/arxiv.2203.12602, doi = {10.48550/ARXIV.2203.12602}, url = {https://arxiv.org/abs/2203.12602}, author = {Tong, Zhan and Song, Yibing and Wang, Jue and Wang, Limin}, keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ```
AnonymousSub/rule_based_hier_triplet_0.1_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- license: mit library_name: sklearn tags: - sklearn - skops - tabular-classification model_format: pickle model_file: model.pkl widget: structuredData: BsmtFinSF1: - 1280 - 1464 - 0 BsmtUnfSF: - 402 - 536 - 795 Condition2: - Norm - Norm - Norm ExterQual: - Ex - Gd - Gd Foundation: - PConc - PConc - PConc GarageCars: - 3 - 3 - 1 GarageType: - BuiltIn - Attchd - Detchd Heating: - GasA - GasA - GasA HeatingQC: - Ex - Ex - TA HouseStyle: - 2Story - 1Story - 2.5Fin MSSubClass: - 60 - 20 - 75 MasVnrArea: - 272.0 - 246.0 - 0.0 MasVnrType: - Stone - Stone - .nan MiscFeature: - .nan - .nan - .nan MoSold: - 8 - 7 - 3 OverallQual: - 10 - 8 - 4 Street: - Pave - Pave - Pave TotalBsmtSF: - 1682 - 2000 - 795 YearRemodAdd: - 2008 - 2005 - 1950 YrSold: - 2008 - 2007 - 2006 --- # Model description This is a random forest regression model trained on ames housing dataset from OpenML. ## Intended uses & limitations This model is not ready to be used in production. ## Training Procedure [More Information Needed] ### Hyperparameters <details> <summary> Click to expand </summary> | Hyperparameter | Value | |----------------------------------------------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | memory | | | steps | [('columntransformer', ColumnTransformer(transformers=[('simpleimputer',<br /> SimpleImputer(add_indicator=True),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000001EF7028B6D0>),<br /> ('ordinalencoder',<br /> OrdinalEncoder(encoded_missing_value=-2,<br /> handle_unknown='use_encoded_value',<br /> unknown_value=-1),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000001EF252211B0>)])), ('randomforestregressor', RandomForestRegressor(random_state=42))] | | verbose | False | | columntransformer | ColumnTransformer(transformers=[('simpleimputer',<br /> SimpleImputer(add_indicator=True),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000001EF7028B6D0>),<br /> ('ordinalencoder',<br /> OrdinalEncoder(encoded_missing_value=-2,<br /> handle_unknown='use_encoded_value',<br /> unknown_value=-1),<br /> <sklearn.compose._column_transformer.make_column_selector object at 0x000001EF252211B0>)]) | | randomforestregressor | RandomForestRegressor(random_state=42) | | columntransformer__n_jobs | | | columntransformer__remainder | drop | | columntransformer__sparse_threshold | 0.3 | | columntransformer__transformer_weights | | | columntransformer__transformers | [('simpleimputer', SimpleImputer(add_indicator=True), <sklearn.compose._column_transformer.make_column_selector object at 0x000001EF7028B6D0>), ('ordinalencoder', OrdinalEncoder(encoded_missing_value=-2, handle_unknown='use_encoded_value',<br /> unknown_value=-1), <sklearn.compose._column_transformer.make_column_selector object at 0x000001EF252211B0>)] | | columntransformer__verbose | False | | columntransformer__verbose_feature_names_out | True | | columntransformer__simpleimputer | SimpleImputer(add_indicator=True) | | columntransformer__ordinalencoder | OrdinalEncoder(encoded_missing_value=-2, handle_unknown='use_encoded_value',<br /> unknown_value=-1) | | columntransformer__simpleimputer__add_indicator | True | | columntransformer__simpleimputer__copy | True | | columntransformer__simpleimputer__fill_value | | | columntransformer__simpleimputer__keep_empty_features | False | | columntransformer__simpleimputer__missing_values | nan | | columntransformer__simpleimputer__strategy | mean | | columntransformer__simpleimputer__verbose | deprecated | | columntransformer__ordinalencoder__categories | auto | | columntransformer__ordinalencoder__dtype | <class 'numpy.float64'> | | columntransformer__ordinalencoder__encoded_missing_value | -2 | | columntransformer__ordinalencoder__handle_unknown | use_encoded_value | | columntransformer__ordinalencoder__unknown_value | -1 | | randomforestregressor__bootstrap | True | | randomforestregressor__ccp_alpha | 0.0 | | randomforestregressor__criterion | squared_error | | randomforestregressor__max_depth | | | randomforestregressor__max_features | 1.0 | | randomforestregressor__max_leaf_nodes | | | randomforestregressor__max_samples | | | randomforestregressor__min_impurity_decrease | 0.0 | | randomforestregressor__min_samples_leaf | 1 | | randomforestregressor__min_samples_split | 2 | | randomforestregressor__min_weight_fraction_leaf | 0.0 | | randomforestregressor__n_estimators | 100 | | randomforestregressor__n_jobs | | | randomforestregressor__oob_score | False | | randomforestregressor__random_state | 42 | | randomforestregressor__verbose | 0 | | randomforestregressor__warm_start | False | </details> ### Model Plot <style>#sk-container-id-1 {color: black;background-color: white;}#sk-container-id-1 pre{padding: 0;}#sk-container-id-1 div.sk-toggleable {background-color: white;}#sk-container-id-1 label.sk-toggleable__label {cursor: pointer;display: block;width: 100%;margin-bottom: 0;padding: 0.3em;box-sizing: border-box;text-align: center;}#sk-container-id-1 label.sk-toggleable__label-arrow:before {content: "▸";float: left;margin-right: 0.25em;color: #696969;}#sk-container-id-1 label.sk-toggleable__label-arrow:hover:before {color: black;}#sk-container-id-1 div.sk-estimator:hover label.sk-toggleable__label-arrow:before {color: black;}#sk-container-id-1 div.sk-toggleable__content {max-height: 0;max-width: 0;overflow: hidden;text-align: left;background-color: #f0f8ff;}#sk-container-id-1 div.sk-toggleable__content pre {margin: 0.2em;color: black;border-radius: 0.25em;background-color: #f0f8ff;}#sk-container-id-1 input.sk-toggleable__control:checked~div.sk-toggleable__content {max-height: 200px;max-width: 100%;overflow: auto;}#sk-container-id-1 input.sk-toggleable__control:checked~label.sk-toggleable__label-arrow:before {content: "▾";}#sk-container-id-1 div.sk-estimator input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-label input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 input.sk-hidden--visually {border: 0;clip: rect(1px 1px 1px 1px);clip: rect(1px, 1px, 1px, 1px);height: 1px;margin: -1px;overflow: hidden;padding: 0;position: absolute;width: 1px;}#sk-container-id-1 div.sk-estimator {font-family: monospace;background-color: #f0f8ff;border: 1px dotted black;border-radius: 0.25em;box-sizing: border-box;margin-bottom: 0.5em;}#sk-container-id-1 div.sk-estimator:hover {background-color: #d4ebff;}#sk-container-id-1 div.sk-parallel-item::after {content: "";width: 100%;border-bottom: 1px solid gray;flex-grow: 1;}#sk-container-id-1 div.sk-label:hover label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-serial::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: 0;}#sk-container-id-1 div.sk-serial {display: flex;flex-direction: column;align-items: center;background-color: white;padding-right: 0.2em;padding-left: 0.2em;position: relative;}#sk-container-id-1 div.sk-item {position: relative;z-index: 1;}#sk-container-id-1 div.sk-parallel {display: flex;align-items: stretch;justify-content: center;background-color: white;position: relative;}#sk-container-id-1 div.sk-item::before, #sk-container-id-1 div.sk-parallel-item::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: -1;}#sk-container-id-1 div.sk-parallel-item {display: flex;flex-direction: column;z-index: 1;position: relative;background-color: white;}#sk-container-id-1 div.sk-parallel-item:first-child::after {align-self: flex-end;width: 50%;}#sk-container-id-1 div.sk-parallel-item:last-child::after {align-self: flex-start;width: 50%;}#sk-container-id-1 div.sk-parallel-item:only-child::after {width: 0;}#sk-container-id-1 div.sk-dashed-wrapped {border: 1px dashed gray;margin: 0 0.4em 0.5em 0.4em;box-sizing: border-box;padding-bottom: 0.4em;background-color: white;}#sk-container-id-1 div.sk-label label {font-family: monospace;font-weight: bold;display: inline-block;line-height: 1.2em;}#sk-container-id-1 div.sk-label-container {text-align: center;}#sk-container-id-1 div.sk-container {/* jupyter's `normalize.less` sets `[hidden] { display: none; }` but bootstrap.min.css set `[hidden] { display: none !important; }` so we also need the `!important` here to be able to override the default hidden behavior on the sphinx rendered scikit-learn.org. See: https://github.com/scikit-learn/scikit-learn/issues/21755 */display: inline-block !important;position: relative;}#sk-container-id-1 div.sk-text-repr-fallback {display: none;}</style><div id="sk-container-id-1" class="sk-top-container" style="overflow: auto;"><div class="sk-text-repr-fallback"><pre>Pipeline(steps=[(&#x27;columntransformer&#x27;,ColumnTransformer(transformers=[(&#x27;simpleimputer&#x27;,SimpleImputer(add_indicator=True),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001EF7028B6D0&gt;),(&#x27;ordinalencoder&#x27;,OrdinalEncoder(encoded_missing_value=-2,handle_unknown=&#x27;use_encoded_value&#x27;,unknown_value=-1),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001EF252211B0&gt;)])),(&#x27;randomforestregressor&#x27;,RandomForestRegressor(random_state=42))])</pre><b>In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook. <br />On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.</b></div><div class="sk-container" hidden><div class="sk-item sk-dashed-wrapped"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-1" type="checkbox" ><label for="sk-estimator-id-1" class="sk-toggleable__label sk-toggleable__label-arrow">Pipeline</label><div class="sk-toggleable__content"><pre>Pipeline(steps=[(&#x27;columntransformer&#x27;,ColumnTransformer(transformers=[(&#x27;simpleimputer&#x27;,SimpleImputer(add_indicator=True),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001EF7028B6D0&gt;),(&#x27;ordinalencoder&#x27;,OrdinalEncoder(encoded_missing_value=-2,handle_unknown=&#x27;use_encoded_value&#x27;,unknown_value=-1),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001EF252211B0&gt;)])),(&#x27;randomforestregressor&#x27;,RandomForestRegressor(random_state=42))])</pre></div></div></div><div class="sk-serial"><div class="sk-item sk-dashed-wrapped"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-2" type="checkbox" ><label for="sk-estimator-id-2" class="sk-toggleable__label sk-toggleable__label-arrow">columntransformer: ColumnTransformer</label><div class="sk-toggleable__content"><pre>ColumnTransformer(transformers=[(&#x27;simpleimputer&#x27;,SimpleImputer(add_indicator=True),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001EF7028B6D0&gt;),(&#x27;ordinalencoder&#x27;,OrdinalEncoder(encoded_missing_value=-2,handle_unknown=&#x27;use_encoded_value&#x27;,unknown_value=-1),&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001EF252211B0&gt;)])</pre></div></div></div><div class="sk-parallel"><div class="sk-parallel-item"><div class="sk-item"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-3" type="checkbox" ><label for="sk-estimator-id-3" class="sk-toggleable__label sk-toggleable__label-arrow">simpleimputer</label><div class="sk-toggleable__content"><pre>&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001EF7028B6D0&gt;</pre></div></div></div><div class="sk-serial"><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-4" type="checkbox" ><label for="sk-estimator-id-4" class="sk-toggleable__label sk-toggleable__label-arrow">SimpleImputer</label><div class="sk-toggleable__content"><pre>SimpleImputer(add_indicator=True)</pre></div></div></div></div></div></div><div class="sk-parallel-item"><div class="sk-item"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-5" type="checkbox" ><label for="sk-estimator-id-5" class="sk-toggleable__label sk-toggleable__label-arrow">ordinalencoder</label><div class="sk-toggleable__content"><pre>&lt;sklearn.compose._column_transformer.make_column_selector object at 0x000001EF252211B0&gt;</pre></div></div></div><div class="sk-serial"><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-6" type="checkbox" ><label for="sk-estimator-id-6" class="sk-toggleable__label sk-toggleable__label-arrow">OrdinalEncoder</label><div class="sk-toggleable__content"><pre>OrdinalEncoder(encoded_missing_value=-2, handle_unknown=&#x27;use_encoded_value&#x27;,unknown_value=-1)</pre></div></div></div></div></div></div></div></div><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-7" type="checkbox" ><label for="sk-estimator-id-7" class="sk-toggleable__label sk-toggleable__label-arrow">RandomForestRegressor</label><div class="sk-toggleable__content"><pre>RandomForestRegressor(random_state=42)</pre></div></div></div></div></div></div></div> ## Evaluation Results | Metric | Value | |----------|----------| | R2 score | 0.831021 | | MAE | 0.111169 | # How to Get Started with the Model Use the following code to get started: ```python import joblib from skops.hub_utils import download import json import pandas as pd download(repo_id="haizad/ames-housing-random-forest-predictor", dst='ames-housing-random-forest-predictor') pipeline = joblib.load( "ames-housing-random-forest-predictor/model.pkl") with open("ames-housing-random-forest-predictor/config.json") as f: config = json.load(f) pipeline.predict(pd.DataFrame.from_dict(config["sklearn"]["example_input"])) ``` # Model Card Authors This model card is written by following authors: [More Information Needed] # Model Card Contact You can contact the model card authors through following channels: [More Information Needed] # Citation Below you can find information related to citation. **BibTeX:** ``` [More Information Needed] ``` # Intended uses & limitations This model is not ready to be used in production. # Evaluation ![Evaluation](evaluation.png)
AnonymousSub/rule_based_hier_triplet_0.1_epochs_1_shard_1_squad2.0
[ "pytorch", "bert", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
2
null
--- license: "cc-by-nc-4.0" tags: - vision - video-classification --- # VideoMAE (small-sized model, fine-tuned on SSV2) VideoMAE model pre-trained for 2400 epochs in a self-supervised way and fine-tuned in a supervised way on Something-Something V2. It was introduced in the paper [VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training](https://arxiv.org/abs/2203.12602) by Tong et al. and first released in [this repository](https://github.com/MCG-NJU/VideoMAE). Disclaimer: The team releasing VideoMAE did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description VideoMAE is an extension of [Masked Autoencoders (MAE)](https://arxiv.org/abs/2111.06377) to video. The architecture of the model is very similar to that of a standard Vision Transformer (ViT), with a decoder on top for predicting pixel values for masked patches. Videos are presented to the model as a sequence of fixed-size patches (resolution 16x16), which are linearly embedded. One also adds a [CLS] token to the beginning of a sequence to use it for classification tasks. One also adds fixed sinus/cosinus position embeddings before feeding the sequence to the layers of the Transformer encoder. By pre-training the model, it learns an inner representation of videos that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled videos for instance, you can train a standard classifier by placing a linear layer on top of the pre-trained encoder. One typically places a linear layer on top of the [CLS] token, as the last hidden state of this token can be seen as a representation of an entire video. ## Intended uses & limitations You can use the raw model for video classification into one of the 174 possible Something-Something V2 labels. ### How to use Here is how to use this model to classify a video: ```python from transformers import VideoMAEFeatureExtractor, VideoMAEForVideoClassification import numpy as np import torch video = list(np.random.randn(16, 3, 224, 224)) feature_extractor = VideoMAEFeatureExtractor.from_pretrained("MCG-NJU/videomae-small-finetuned-ssv2") model = VideoMAEForVideoClassification.from_pretrained("MCG-NJU/videomae-small-finetuned-ssv2") inputs = feature_extractor(video, return_tensors="pt") with torch.no_grad(): outputs = model(**inputs) logits = outputs.logits predicted_class_idx = logits.argmax(-1).item() print("Predicted class:", model.config.id2label[predicted_class_idx]) ``` For more code examples, we refer to the [documentation](https://huggingface.co/transformers/main/model_doc/videomae.html#). ## Training data (to do, feel free to open a PR) ## Training procedure ### Preprocessing (to do, feel free to open a PR) ### Pretraining (to do, feel free to open a PR) ## Evaluation results This model obtains a top-1 accuracy of 66.8 and a top-5 accuracy of 90.3 on the test set of Something-Something V2. ### BibTeX entry and citation info ```bibtex misc{https://doi.org/10.48550/arxiv.2203.12602, doi = {10.48550/ARXIV.2203.12602}, url = {https://arxiv.org/abs/2203.12602}, author = {Tong, Zhan and Song, Yibing and Wang, Jue and Wang, Limin}, keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ```
AnonymousSub/rule_based_hier_triplet_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
--- license: bigscience-openrail-m language: - zh pipeline_tag: text2text-generation thumbnail: Chinese Lyrics Generation with Masked Sequence-to-Sequence Pretraining. --- # Chinese Generation with Masked Sequence-to-Sequence Pretraining This repository demostrates a format-controllable Chinese lyric generator, fine-tuned on [Chinese-Lyric-Corpus](https://github.com/gaussic/Chinese-Lyric-Corpus) using a [MASS](https://arxiv.org/abs/1905.02450)-like strategy. # Usage ## Initialization ```python from transformers import MT5ForConditionalGeneration, MT5Tokenizer, Text2TextGenerationPipeline model_path = "zake7749/chinese-lyrics-generation-mass" model = MT5ForConditionalGeneration.from_pretrained(model_path) tokenizer = MT5Tokenizer.from_pretrained(model_path) pipe = Text2TextGenerationPipeline(model=model, tokenizer=tokenizer) ``` ## Generate lyrics with a template ```python template = "風花雪月。像XXXXXXXXXX。日升月落。仿若XXXXXXXXXX。" lyric = pipe(template, max_length=128, top_p=0.8, do_sample=True, repetition_penalty=1.2)[0]['generated_text'] print(lyric) # 風花雪月。像你在我的夢裡慢慢散落。日升月落。仿若我宿命無法陪隨你走過。 template = "XXXXXXX留戀。XXXXXXX。XXX燈火XXXX。XXX手牽手XXXX。" lyric = pipe(template, max_length=128, top_p=0.8, do_sample=True, repetition_penalty=1.2)[0]['generated_text'] print(lyric) # 我們說好一生不留戀。我們相約在夏天。我們的燈火相偎相牽。我們說好手牽手到永遠。 ``` ## Acrostic ```python template = "分XXXXXX。手XXXXXXXXX。之XXXXXXX。後XXXXXXXXX。" lyric = pipe(template, max_length=128, top_p=0.8, do_sample=True, repetition_penalty=1.2)[0]['generated_text'] print(lyric) # 分開後激情浮現。手牽著手走過的那一天。之間有太多的危險。後悔一點點,傷心一片。 ``` ## Completion ```python template = "餘生的光陰牽你手前行。我們共赴一場光年的旅行。XXXXXXXXXX。XXXXXXXXXXXX。" lyric = pipe(template, max_length=128, top_p=0.8, do_sample=True, repetition_penalty=1.2)[0]['generated_text'] print(lyric) # 餘生的光陰牽你手前行。我們共赴一場光年的旅行。走過的經歷新舊的記憶。都是帶著珍珠淚水無法代替。 ``` ## Random Generation ```python import random num_example = 5 min_sentence_num, max_sentence_num = 2, 5 min_characher_num, max_character_num = 4, 10 for example_id in range(num_example): num_sentences = random.randint(min_sentence_num, max_sentence_num) num_words = ["X" * random.randint(min_characher_num, max_character_num) for _ in range(num_sentences)] template = "。".join(num_words) + "。" lyric = pipe(template, max_length=128, top_p=0.8, do_sample=True, repetition_penalty=1.2)[0]['generated_text'] print(f"{example_id + 1}. {lyric}") # 1. 愛不愛我。讓自己難過。你的擁抱是那麼多。 # 2. 那一天我們重相見。你已站在那個熟悉的街邊。讓我魂牽夢繞在肩。有你的明天。不再留戀。飛過天邊。 # 3. 誰知我們入骨的相思。深深地被俘虜。苦澀滋味含在茶中傾訴。餘情未了落幕。愛到痛處奢望幸福。 # 4. 為什麼你一直讓我傷心。總覺得對你太著迷。 # 5. 一點可憐。還在期待你會出現。哪怕只是匆匆一眼。 ``` # Note 1. The model is still under training, so sometimes it might not follow the template explicitly, especially for long sequences generation. 2. The model would output `,` as a pause in the lyric, for example `我的愛,像潮水。`. If you don't need the pause, you can add the id of `,` to `bad_words_ids`. 3. The model was only fine-tuned on traditional Chinese corpus which leads to a bit unstable performance in simplified Chinese. 4. When there are no/few keywords in the given input, the model may **combine snippets from real world songs** to fit the template. # Disclaimer This lyric generator is for academic purposes only. Users of this model should exercise caution and carefully evaluate the results before using them for any commercial or non-academic purpose. We are not liable for any damages or losses resulting from the use or misuse of the model.
AnonymousSub/rule_based_only_classfn_epochs_1_shard_1
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### CTensen Dreambooth model trained by Strangematter with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept:
AnonymousSub/rule_based_only_classfn_epochs_1_shard_10
[ "pytorch", "bert", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "BertModel" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
7
null
--- tags: - LunarLander-v2 - ppo - deep-reinforcement-learning - reinforcement-learning - custom-implementation - deep-rl-course model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: -109.89 +/- 63.45 name: mean_reward verified: false --- # PPO Agent Playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2. # Hyperparameters ```python {'exp_name': 'ppo' 'seed': 1 'torch_deterministic': True 'cuda': True 'track': False 'wandb_project_name': 'cleanRL' 'wandb_entity': None 'capture_video': False 'env_id': 'LunarLander-v2' 'total_timesteps': 100000 'learning_rate': 0.00025 'num_envs': 4 'num_steps': 128 'anneal_lr': True 'gae': True 'gamma': 0.99 'gae_lambda': 0.95 'num_minibatches': 4 'update_epochs': 4 'norm_adv': True 'clip_coef': 0.2 'clip_vloss': True 'ent_coef': 0.01 'vf_coef': 0.5 'max_grad_norm': 0.5 'target_kl': None 'repo_id': 'Isaacgv/Lunar-v2' 'batch_size': 512 'minibatch_size': 128} ```
AnonymousSub/rule_based_only_classfn_epochs_1_shard_1_squad2.0
[ "pytorch", "bert", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
2
null
--- license: cc-by-nc-4.0 language: - en tags: - art - texture - game-development - asset-creation - pbr - stable_diffusion --- # Introducing Texture Hell This is my first publicly released sd model. It is meant specifically to create albedo/diffuse textures for use in video games and animations. It has been fully finetuned on the sd2.1 768 base model to provide you with high resolution results. Let's dive into the details. Please read my description first before using this model, as it is rather diffrent to use than the usual sd models. ## Training and prompt tips The dataset consists of 350 high-quality images sourced from Poly Heaven, ensuring a diverse and representative range of albedo textures. Clip 1 was used throughout training. Prompts where comma seperated tag lists (no clip/bert full text descriptions). The most import one is "texture", followed by any material you are interested. Though because of the limited dataset, there are lacking domains (e.g. flesh and cloth). I plan to improve on diversity on subsequent versions. A full list of tags used on images with a frequency usage statistic is available -here- to help you identify potentially interesting tags. But there is probably a lot of room to experiment with other words, i was only able to do rudimentary testing. Very Important: there are some aerial textures inside the dataset (ie. beach taken by drone from very far up) for landscape textures. If you don't want this type of texture, put "aerial" into negative prompt. All example images have full prompt information inside them for more information about samplers, steps etc. ## Tiling and WebUI Please note that the textures generated do not perfectly tile. While there is a tiling option available in the WebUI, i do not recommend using it. Instead, i have found a "reasonable" manual workflow to turn them into seamless textures. A tutorial can be found on my Huggingface page: https://huggingface.co/Nazzaroth2/texture-hell I hope to improve the inherent tiling ability of the model in further versions, especially a bigger dataset and good tagging should help here. ## Other Textures My first try involded generating the full set of textures needed for the normal pbr-based render pipeline(normal, height, roughness, etc). The initial results are promising, though the reduction in resolution and limited dataset made me change my plans. This might change in the future. In the meantime you can use other software to create these textures from the albedo image. I personally used the free option Materialize https://boundingboxsoftware.com/materialize . But there are of course also paid versions like Substance Designer. I am curious to see what you all are able to make with this model and hope it can help you in your projects. If you have any ideas for improvements or a specific material type you think is lacking inside the model, i am happy to hear from you. Happy texturing! P.S. Why Texture Hell? Because hell is way more interesting than boring old heaven :3 # How to Create Seamless Textures ## Additional Software used - auto111 webui - image editing software (photoshop or gimp) - lama cleaner (https://github.com/Sanster/lama-cleaner) ## Step 1: Create the Texture Create your desired texture with the usual text2image workflow. Do NOT use tiling option yet. ## Step 2: Inpaint the Texture 1. Send the image to the inpainting tab. 2. Handpaint a mask around the edges or use one of my pre-made mask images 3. Now activate the tiling option 4. Adjust the denoising strength. A value of 0.75 is a good starting point, but feel free to experiment as needed. 5. Save the inpainted image. ## Step 3: Offset the Edges in an Image Editor 1. Open the inpainted image in the image editor of your choice. 2. Use the offset modifier to push the edges into the center of the image (check online tutorials for specific instructions on using the offset modifier in your chosen image editor). 3. Save the modified image. ## Step 4: Clean Up the Texture with LAMA-cleaner 1. Open the offset image in LAMA-cleaner. 2. Paint over parts that do not line up perfectly or remove any larger objects that cannot be fixed. Lama or Zits are my preffered models, but you can also experiment here. 3. Save the cleaned image. ## Step 5: Additional Cleanup (Optional) Depending on your quality requirements, you may want to open the cleaned image in your image editor again and further clean up any imperfections. Continue this process until you are satisfied with the results. ## Step 6: Restore the Original Texture 1. Open the cleaned image in your image editor. 2. Use the offset modifier again to return the image to its original state. The texture should now have seamless edges. 3. Save the final seamless texture and use other software like materialize https://boundingboxsoftware.com/materialize/ to create missing textures like normal-map and roughness
AnonymousSub/rule_based_only_classfn_epochs_1_shard_1_wikiqa
[ "pytorch", "bert", "text-classification", "transformers" ]
text-classification
{ "architectures": [ "BertForSequenceClassification" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
32
null
--- license: other tags: - generated_from_trainer model-index: - name: vicuna_pashtu results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vicuna_pashtu This model is a fine-tuned version of [AlekseyKorshuk/vicuna-7b](https://huggingface.co/AlekseyKorshuk/vicuna-7b) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - training_steps: 300 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.28.0.dev0 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
AnonymousSub/rule_based_roberta_hier_quadruplet_0.1_epochs_1_shard_1_squad2.0
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- library_name: sample-factory tags: - deep-reinforcement-learning - reinforcement-learning - sample-factory model-index: - name: APPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: doom_health_gathering_supreme type: doom_health_gathering_supreme metrics: - type: mean_reward value: 10.80 +/- 4.99 name: mean_reward verified: false --- A(n) **APPO** model trained on the **doom_health_gathering_supreme** environment. This model was trained using Sample-Factory 2.0: https://github.com/alex-petrenko/sample-factory. Documentation for how to use Sample-Factory can be found at https://www.samplefactory.dev/ ## Downloading the model After installing Sample-Factory, download the model with: ``` python -m sample_factory.huggingface.load_from_hub -r Isaacgv/rl_course_vizdoom_health_gathering_supreme ``` ## Using the model To run the model after download, use the `enjoy` script corresponding to this environment: ``` python -m .usr.local.lib.python3.9.dist-packages.ipykernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme ``` You can also upload models to the Hugging Face Hub using the same script with the `--push_to_hub` flag. See https://www.samplefactory.dev/10-huggingface/huggingface/ for more details ## Training with this model To continue training with this model, use the `train` script corresponding to this environment: ``` python -m .usr.local.lib.python3.9.dist-packages.ipykernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme --restart_behavior=resume --train_for_env_steps=10000000000 ``` Note, you may have to adjust `--train_for_env_steps` to a suitably high number as the experiment will resume at the number of steps it concluded at.
AnonymousSub/rule_based_roberta_hier_quadruplet_epochs_1_shard_1
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
1
null
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: flan-t5-large-extraction-all-cnn_8000-ep25-nonstop results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # flan-t5-large-extraction-all-cnn_8000-ep25-nonstop This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.8671 - Hint Hit Num: 2.008 - Hint Precision: 0.3399 - Num: 5.895 - Gen Len: 18.991 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 64 - seed: 1799 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 25 ### Training results | Training Loss | Epoch | Step | Validation Loss | Hint Hit Num | Hint Precision | Num | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:------------:|:--------------:|:-----:|:-------:| | 2.119 | 0.4 | 200 | 1.7746 | 1.966 | 0.3387 | 5.814 | 18.99 | | 1.9135 | 0.8 | 400 | 1.7157 | 1.76 | 0.3129 | 5.617 | 18.987 | | 1.85 | 1.2 | 600 | 1.7140 | 1.913 | 0.3327 | 5.732 | 18.995 | | 1.7963 | 1.6 | 800 | 1.7022 | 1.887 | 0.3291 | 5.702 | 18.994 | | 1.7784 | 2.0 | 1000 | 1.6911 | 1.875 | 0.3268 | 5.711 | 18.989 | | 1.711 | 2.4 | 1200 | 1.6935 | 1.932 | 0.3354 | 5.749 | 18.994 | | 1.7186 | 2.8 | 1400 | 1.6721 | 1.979 | 0.3427 | 5.791 | 18.997 | | 1.6704 | 3.2 | 1600 | 1.7007 | 1.945 | 0.334 | 5.792 | 18.994 | | 1.6484 | 3.6 | 1800 | 1.6900 | 1.896 | 0.3282 | 5.751 | 18.994 | | 1.6334 | 4.0 | 2000 | 1.6732 | 1.879 | 0.3283 | 5.698 | 18.994 | | 1.5761 | 4.4 | 2200 | 1.6869 | 1.97 | 0.3357 | 5.861 | 18.992 | | 1.5882 | 4.8 | 2400 | 1.6784 | 1.952 | 0.3354 | 5.792 | 18.992 | | 1.558 | 5.2 | 2600 | 1.7012 | 1.984 | 0.3394 | 5.83 | 19.0 | | 1.5339 | 5.6 | 2800 | 1.7013 | 1.898 | 0.3245 | 5.82 | 18.991 | | 1.5419 | 6.0 | 3000 | 1.6850 | 1.952 | 0.3377 | 5.766 | 18.992 | | 1.4884 | 6.4 | 3200 | 1.7009 | 1.967 | 0.3375 | 5.812 | 18.991 | | 1.4857 | 6.8 | 3400 | 1.7038 | 1.913 | 0.3289 | 5.805 | 18.992 | | 1.4655 | 7.2 | 3600 | 1.7103 | 1.956 | 0.3347 | 5.82 | 18.992 | | 1.4578 | 7.6 | 3800 | 1.7235 | 1.946 | 0.3318 | 5.837 | 18.999 | | 1.443 | 8.0 | 4000 | 1.7176 | 1.963 | 0.3347 | 5.828 | 18.991 | | 1.42 | 8.4 | 4200 | 1.7305 | 1.977 | 0.3404 | 5.809 | 18.996 | | 1.4155 | 8.8 | 4400 | 1.7267 | 1.988 | 0.3408 | 5.816 | 18.997 | | 1.3753 | 9.2 | 4600 | 1.7418 | 1.992 | 0.3427 | 5.804 | 19.0 | | 1.3853 | 9.6 | 4800 | 1.7360 | 2.013 | 0.3461 | 5.818 | 18.992 | | 1.3768 | 10.0 | 5000 | 1.7280 | 1.994 | 0.3397 | 5.874 | 18.992 | | 1.3465 | 10.4 | 5200 | 1.7530 | 2.01 | 0.3424 | 5.855 | 18.992 | | 1.3445 | 10.8 | 5400 | 1.7416 | 1.996 | 0.3438 | 5.814 | 18.992 | | 1.3321 | 11.2 | 5600 | 1.7653 | 2.014 | 0.3434 | 5.861 | 18.992 | | 1.3092 | 11.6 | 5800 | 1.7705 | 2.007 | 0.3423 | 5.861 | 18.983 | | 1.3263 | 12.0 | 6000 | 1.7617 | 1.988 | 0.3412 | 5.815 | 18.986 | | 1.2847 | 12.4 | 6200 | 1.7816 | 1.988 | 0.3407 | 5.815 | 18.992 | | 1.2942 | 12.8 | 6400 | 1.7905 | 1.987 | 0.3395 | 5.83 | 18.991 | | 1.2784 | 13.2 | 6600 | 1.7795 | 2.028 | 0.3436 | 5.899 | 18.992 | | 1.2562 | 13.6 | 6800 | 1.7861 | 1.97 | 0.3371 | 5.825 | 18.989 | | 1.2776 | 14.0 | 7000 | 1.7899 | 2.02 | 0.3431 | 5.871 | 18.992 | | 1.2524 | 14.4 | 7200 | 1.8054 | 2.038 | 0.3435 | 5.916 | 18.992 | | 1.2402 | 14.8 | 7400 | 1.8072 | 2.034 | 0.3459 | 5.872 | 18.995 | | 1.2352 | 15.2 | 7600 | 1.8123 | 2.014 | 0.3431 | 5.861 | 18.987 | | 1.2195 | 15.6 | 7800 | 1.8196 | 2.034 | 0.3444 | 5.869 | 18.987 | | 1.23 | 16.0 | 8000 | 1.8115 | 1.979 | 0.338 | 5.85 | 18.989 | | 1.2047 | 16.4 | 8200 | 1.8129 | 2.02 | 0.3428 | 5.888 | 18.99 | | 1.2155 | 16.8 | 8400 | 1.8178 | 1.978 | 0.335 | 5.883 | 18.991 | | 1.2028 | 17.2 | 8600 | 1.8293 | 2.017 | 0.3418 | 5.88 | 18.992 | | 1.189 | 17.6 | 8800 | 1.8303 | 1.983 | 0.3374 | 5.858 | 18.992 | | 1.195 | 18.0 | 9000 | 1.8367 | 2.021 | 0.3423 | 5.883 | 18.992 | | 1.1837 | 18.4 | 9200 | 1.8388 | 2.015 | 0.3403 | 5.893 | 18.999 | | 1.1668 | 18.8 | 9400 | 1.8388 | 2.023 | 0.342 | 5.903 | 18.991 | | 1.1568 | 19.2 | 9600 | 1.8514 | 2.036 | 0.3458 | 5.876 | 18.99 | | 1.1783 | 19.6 | 9800 | 1.8419 | 2.042 | 0.3458 | 5.902 | 18.985 | | 1.1674 | 20.0 | 10000 | 1.8433 | 1.992 | 0.3394 | 5.868 | 18.991 | | 1.1515 | 20.4 | 10200 | 1.8601 | 2.004 | 0.3404 | 5.881 | 18.985 | | 1.1478 | 20.8 | 10400 | 1.8520 | 2.032 | 0.3437 | 5.897 | 18.991 | | 1.1634 | 21.2 | 10600 | 1.8582 | 2.013 | 0.3398 | 5.926 | 18.985 | | 1.138 | 21.6 | 10800 | 1.8571 | 2.006 | 0.3399 | 5.902 | 18.985 | | 1.1609 | 22.0 | 11000 | 1.8557 | 2.006 | 0.3402 | 5.899 | 18.991 | | 1.1306 | 22.4 | 11200 | 1.8622 | 2.02 | 0.3431 | 5.894 | 18.99 | | 1.1485 | 22.8 | 11400 | 1.8619 | 2.003 | 0.3402 | 5.872 | 18.992 | | 1.1239 | 23.2 | 11600 | 1.8648 | 2.004 | 0.3405 | 5.879 | 18.992 | | 1.1427 | 23.6 | 11800 | 1.8651 | 2.003 | 0.3397 | 5.897 | 18.991 | | 1.1451 | 24.0 | 12000 | 1.8631 | 2.008 | 0.3404 | 5.89 | 18.991 | | 1.1342 | 24.4 | 12200 | 1.8654 | 2.004 | 0.3397 | 5.884 | 18.99 | | 1.1289 | 24.8 | 12400 | 1.8672 | 2.005 | 0.3399 | 5.888 | 18.991 | ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.0+cu111 - Datasets 2.5.1 - Tokenizers 0.12.1
AnonymousSub/rule_based_roberta_hier_quadruplet_epochs_1_shard_10
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
6
null
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: flan-t5-large-extraction-all-dm_8000-ep25-nonstop results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # flan-t5-large-extraction-all-dm_8000-ep25-nonstop This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.5739 - Hint Hit Num: 2.184 - Hint Precision: 0.4154 - Num: 5.218 - Gen Len: 18.738 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 24 - eval_batch_size: 96 - seed: 1799 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 25 ### Training results | Training Loss | Epoch | Step | Validation Loss | Hint Hit Num | Hint Precision | Num | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------------:|:--------------:|:-----:|:-------:| | 2.1513 | 0.6 | 200 | 1.5120 | 2.679 | 0.4838 | 5.551 | 18.94 | | 1.9342 | 1.2 | 400 | 1.4818 | 2.563 | 0.4677 | 5.469 | 18.925 | | 1.8591 | 1.8 | 600 | 1.4593 | 2.494 | 0.4607 | 5.396 | 18.899 | | 1.7973 | 2.4 | 800 | 1.4597 | 2.389 | 0.4515 | 5.265 | 18.836 | | 1.7824 | 2.99 | 1000 | 1.4494 | 2.387 | 0.4419 | 5.399 | 18.891 | | 1.7263 | 3.59 | 1200 | 1.4597 | 2.278 | 0.4301 | 5.261 | 18.875 | | 1.711 | 4.19 | 1400 | 1.4673 | 2.292 | 0.4314 | 5.272 | 18.826 | | 1.6631 | 4.79 | 1600 | 1.4638 | 2.185 | 0.4163 | 5.177 | 18.832 | | 1.6494 | 5.39 | 1800 | 1.4625 | 2.287 | 0.431 | 5.278 | 18.841 | | 1.6328 | 5.99 | 2000 | 1.4584 | 2.209 | 0.4211 | 5.185 | 18.842 | | 1.6008 | 6.59 | 2200 | 1.4677 | 2.299 | 0.4374 | 5.233 | 18.777 | | 1.5646 | 7.19 | 2400 | 1.4902 | 2.182 | 0.4224 | 5.137 | 18.71 | | 1.574 | 7.78 | 2600 | 1.4777 | 2.211 | 0.4235 | 5.19 | 18.781 | | 1.5348 | 8.38 | 2800 | 1.4796 | 2.314 | 0.4311 | 5.317 | 18.792 | | 1.5224 | 8.98 | 3000 | 1.4799 | 2.197 | 0.4212 | 5.17 | 18.805 | | 1.4857 | 9.58 | 3200 | 1.4897 | 2.256 | 0.4296 | 5.221 | 18.755 | | 1.4948 | 10.18 | 3400 | 1.5030 | 2.206 | 0.4203 | 5.201 | 18.76 | | 1.4667 | 10.78 | 3600 | 1.4956 | 2.269 | 0.4319 | 5.203 | 18.772 | | 1.4492 | 11.38 | 3800 | 1.5098 | 2.208 | 0.4191 | 5.235 | 18.801 | | 1.4454 | 11.98 | 4000 | 1.5064 | 2.187 | 0.4153 | 5.22 | 18.799 | | 1.4125 | 12.57 | 4200 | 1.5173 | 2.175 | 0.4164 | 5.182 | 18.766 | | 1.426 | 13.17 | 4400 | 1.5299 | 2.162 | 0.414 | 5.189 | 18.772 | | 1.3944 | 13.77 | 4600 | 1.5297 | 2.199 | 0.4182 | 5.224 | 18.797 | | 1.382 | 14.37 | 4800 | 1.5301 | 2.204 | 0.4217 | 5.197 | 18.799 | | 1.3836 | 14.97 | 5000 | 1.5303 | 2.188 | 0.4185 | 5.209 | 18.764 | | 1.358 | 15.57 | 5200 | 1.5293 | 2.264 | 0.4283 | 5.261 | 18.812 | | 1.3645 | 16.17 | 5400 | 1.5411 | 2.195 | 0.42 | 5.19 | 18.753 | | 1.3455 | 16.77 | 5600 | 1.5417 | 2.267 | 0.4286 | 5.251 | 18.76 | | 1.3395 | 17.37 | 5800 | 1.5436 | 2.207 | 0.4217 | 5.19 | 18.738 | | 1.3302 | 17.96 | 6000 | 1.5468 | 2.268 | 0.4256 | 5.288 | 18.765 | | 1.3329 | 18.56 | 6200 | 1.5488 | 2.265 | 0.4251 | 5.299 | 18.788 | | 1.299 | 19.16 | 6400 | 1.5582 | 2.245 | 0.4253 | 5.25 | 18.717 | | 1.3141 | 19.76 | 6600 | 1.5562 | 2.211 | 0.421 | 5.195 | 18.742 | | 1.318 | 20.36 | 6800 | 1.5597 | 2.22 | 0.4204 | 5.24 | 18.776 | | 1.2905 | 20.96 | 7000 | 1.5605 | 2.228 | 0.4224 | 5.24 | 18.745 | | 1.2967 | 21.56 | 7200 | 1.5679 | 2.199 | 0.4149 | 5.255 | 18.798 | | 1.2896 | 22.16 | 7400 | 1.5667 | 2.218 | 0.4212 | 5.229 | 18.736 | | 1.2886 | 22.75 | 7600 | 1.5663 | 2.212 | 0.4175 | 5.262 | 18.8 | | 1.2818 | 23.35 | 7800 | 1.5718 | 2.211 | 0.4193 | 5.228 | 18.757 | | 1.2893 | 23.95 | 8000 | 1.5730 | 2.185 | 0.4155 | 5.215 | 18.737 | | 1.2772 | 24.55 | 8200 | 1.5736 | 2.186 | 0.4153 | 5.224 | 18.753 | ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.0+cu111 - Datasets 2.5.1 - Tokenizers 0.12.1
AnonymousSub/rule_based_roberta_twostage_quadruplet_epochs_1_shard_1
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
5
null
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-CartPole-v1 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 500.00 +/- 0.00 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
AnonymousSub/rule_based_roberta_twostagetriplet_epochs_1_shard_1
[ "pytorch", "roberta", "feature-extraction", "transformers" ]
feature-extraction
{ "architectures": [ "RobertaModel" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
2
null
--- language: en tags: - seq2seq - t5 - text-generation - recipe-generation pipeline_tag: text2text-generation widget: - text: "provolone cheese, bacon, bread, ginger" - text: "sugar, crunchy jif peanut butter, cornflakes" - text: "sweet butter, confectioners sugar, flaked coconut, condensed milk, nuts, vanilla, dipping chocolate" - text: "macaroni, butter, salt, bacon, milk, flour, pepper, cream corn" - text: "hamburger, sausage, onion, regular, american cheese, colby cheese" - text: "chicken breasts, onion, garlic, great northern beans, black beans, green chilies, broccoli, garlic oil, butter, cajun seasoning, salt, oregano, thyme, black pepper, basil, worcestershire sauce, chicken broth, sour cream, chardonnay wine" - text: "serrano peppers, garlic, celery, oregano, canola oil, vinegar, water, kosher salt, salt, black pepper" --- ![avatar](chef-transformer.png) # Chef Transformer (T5) > This is part of the [Flax/Jax Community Week](https://discuss.huggingface.co/t/recipe-generation-model/7475), organized by [HuggingFace](https://huggingface.co/) and TPU usage sponsored by Google. Want to give it a try? Then what's the wait, head over to Hugging Face Spaces [here](https://huggingface.co/spaces/flax-community/chef-transformer). ## Team Members - Mehrdad Farahani ([m3hrdadfi](https://huggingface.co/m3hrdadfi)) - Kartik Godawat ([dk-crazydiv](https://huggingface.co/dk-crazydiv)) - Haswanth Aekula ([hassiahk](https://huggingface.co/hassiahk)) - Deepak Pandian ([rays2pix](https://huggingface.co/rays2pix)) - Nicholas Broad ([nbroad](https://huggingface.co/nbroad)) ## Dataset [RecipeNLG: A Cooking Recipes Dataset for Semi-Structured Text Generation](https://recipenlg.cs.put.poznan.pl/). This dataset contains **2,231,142** cooking recipes (>2 millions) with size of **2.14 GB**. It's processed in more careful way. ### Example ```json { "NER": [ "oyster crackers", "salad dressing", "lemon pepper", "dill weed", "garlic powder", "salad oil" ], "directions": [ "Combine salad dressing mix and oil.", "Add dill weed, garlic powder and lemon pepper.", "Pour over crackers; stir to coat.", "Place in warm oven.", "Use very low temperature for 15 to 20 minutes." ], "ingredients": [ "12 to 16 oz. plain oyster crackers", "1 pkg. Hidden Valley Ranch salad dressing mix", "1/4 tsp. lemon pepper", "1/2 to 1 tsp. dill weed", "1/4 tsp. garlic powder", "3/4 to 1 c. salad oil" ], "link": "www.cookbooks.com/Recipe-Details.aspx?id=648947", "source": "Gathered", "title": "Hidden Valley Ranch Oyster Crackers" } ``` ## How To Use ```bash # Installing requirements pip install transformers ``` ```python from transformers import FlaxAutoModelForSeq2SeqLM from transformers import AutoTokenizer MODEL_NAME_OR_PATH = "flax-community/t5-recipe-generation" tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME_OR_PATH, use_fast=True) model = FlaxAutoModelForSeq2SeqLM.from_pretrained(MODEL_NAME_OR_PATH) prefix = "items: " # generation_kwargs = { # "max_length": 512, # "min_length": 64, # "no_repeat_ngram_size": 3, # "early_stopping": True, # "num_beams": 5, # "length_penalty": 1.5, # } generation_kwargs = { "max_length": 512, "min_length": 64, "no_repeat_ngram_size": 3, "do_sample": True, "top_k": 60, "top_p": 0.95 } special_tokens = tokenizer.all_special_tokens tokens_map = { "<sep>": "--", "<section>": "\n" } def skip_special_tokens(text, special_tokens): for token in special_tokens: text = text.replace(token, "") return text def target_postprocessing(texts, special_tokens): if not isinstance(texts, list): texts = [texts] new_texts = [] for text in texts: text = skip_special_tokens(text, special_tokens) for k, v in tokens_map.items(): text = text.replace(k, v) new_texts.append(text) return new_texts def generation_function(texts): _inputs = texts if isinstance(texts, list) else [texts] inputs = [prefix + inp for inp in _inputs] inputs = tokenizer( inputs, max_length=256, padding="max_length", truncation=True, return_tensors="jax" ) input_ids = inputs.input_ids attention_mask = inputs.attention_mask output_ids = model.generate( input_ids=input_ids, attention_mask=attention_mask, **generation_kwargs ) generated = output_ids.sequences generated_recipe = target_postprocessing( tokenizer.batch_decode(generated, skip_special_tokens=False), special_tokens ) return generated_recipe ``` ```python items = [ "macaroni, butter, salt, bacon, milk, flour, pepper, cream corn", "provolone cheese, bacon, bread, ginger" ] generated = generation_function(items) for text in generated: sections = text.split("\n") for section in sections: section = section.strip() if section.startswith("title:"): section = section.replace("title:", "") headline = "TITLE" elif section.startswith("ingredients:"): section = section.replace("ingredients:", "") headline = "INGREDIENTS" elif section.startswith("directions:"): section = section.replace("directions:", "") headline = "DIRECTIONS" if headline == "TITLE": print(f"[{headline}]: {section.strip().capitalize()}") else: section_info = [f" - {i+1}: {info.strip().capitalize()}" for i, info in enumerate(section.split("--"))] print(f"[{headline}]:") print("\n".join(section_info)) print("-" * 130) ``` Output: ```text [TITLE]: Macaroni and corn [INGREDIENTS]: - 1: 2 c. macaroni - 2: 2 tbsp. butter - 3: 1 tsp. salt - 4: 4 slices bacon - 5: 2 c. milk - 6: 2 tbsp. flour - 7: 1/4 tsp. pepper - 8: 1 can cream corn [DIRECTIONS]: - 1: Cook macaroni in boiling salted water until tender. - 2: Drain. - 3: Melt butter in saucepan. - 4: Blend in flour, salt and pepper. - 5: Add milk all at once. - 6: Cook and stir until thickened and bubbly. - 7: Stir in corn and bacon. - 8: Pour over macaroni and mix well. ---------------------------------------------------------------------------------------------------------------------------------- [TITLE]: Grilled provolone and bacon sandwich [INGREDIENTS]: - 1: 2 slices provolone cheese - 2: 2 slices bacon - 3: 2 slices sourdough bread - 4: 2 slices pickled ginger [DIRECTIONS]: - 1: Place a slice of provolone cheese on one slice of bread. - 2: Top with a slice of bacon. - 3: Top with a slice of pickled ginger. - 4: Top with the other slice of bread. - 5: Heat a skillet over medium heat. - 6: Place the sandwich in the skillet and cook until the cheese is melted and the bread is golden brown. ---------------------------------------------------------------------------------------------------------------------------------- ``` ## Evaluation Since the test set is not available, we will evaluate the model based on a shared test set. This test set consists of 5% of the whole test (*= 5,000 records*), and we will generate five recipes for each input(*= 25,000 records*). The following table summarizes the scores obtained by the **Chef Transformer** and **RecipeNLG** as our baseline. | Model | COSIM | WER | ROUGE-2 | BLEU | GLEU | METEOR | |:------------------------------------------------------------------------:|:----------:|:----------:|:----------:|:----------:|:----------:|:----------:| | [RecipeNLG](https://huggingface.co/mbien/recipenlg) | 0.5723 | 1.2125 | 0.1354 | 0.1164 | 0.1503 | 0.2309 | | [Chef Transformer](huggingface.co/flax-community/t5-recipe-generation) * | **0.7282** | **0.7613** | **0.2470** | **0.3245** | **0.2624** | **0.4150** | *From the 5 generated recipes corresponding to each NER (food items), only the highest score was taken into account in the WER, COSIM, and ROUGE metrics. At the same time, BLEU, GLEU, Meteor were designed to have many possible references.* ## Copyright Special thanks to those who provided these fantastic materials. - [Anatomy](https://www.flaticon.com/free-icon) - [Chef Hat](https://www.vecteezy.com/members/jellyfishwater) - [Moira Nazzari](https://pixabay.com/photos/food-dessert-cake-eggs-butter-3048440/) - [Instagram Post](https://www.freepik.com/free-psd/recipes-ad-social-media-post-template_11520617.htm)
AnonymousSub/rule_based_roberta_twostagetriplet_epochs_1_shard_1_squad2.0
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### HaydenP Dreambooth model trained by Strangematter with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept:
Arghyad/Loki_small
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- tags: - spacy - token-classification language: - grc model-index: - name: grc_perseus_trf results: - task: name: TAG type: token-classification metrics: - name: TAG (XPOS) Accuracy type: accuracy value: 0.9684210526 - task: name: POS type: token-classification metrics: - name: POS (UPOS) Accuracy type: accuracy value: 0.9671560876 - task: name: MORPH type: token-classification metrics: - name: Morph (UFeats) Accuracy type: accuracy value: 0.9500790603 - task: name: LEMMA type: token-classification metrics: - name: Lemma Accuracy type: accuracy value: 0.9524110672 - task: name: UNLABELED_DEPENDENCIES type: token-classification metrics: - name: Unlabeled Attachment Score (UAS) type: f_score value: 0.8171657794 - task: name: LABELED_DEPENDENCIES type: token-classification metrics: - name: Labeled Attachment Score (LAS) type: f_score value: 0.7701700389 - task: name: SENTS type: token-classification metrics: - name: Sentences F-Score type: f_score value: 0.9881422925 --- | Feature | Description | | --- | --- | | **Name** | `grc_perseus_trf` | | **Version** | `3.5.2` | | **spaCy** | `>=3.5.2,<3.6.0` | | **Default Pipeline** | `transformer`, `morphologizer`, `tagger`, `parser`, `senter`, `lemmatizer`, `attribute_ruler` | | **Components** | `transformer`, `morphologizer`, `tagger`, `parser`, `senter`, `lemmatizer`, `attribute_ruler` | | **Vectors** | -1 keys, 200000 unique vectors (300 dimensions) | | **Sources** | n/a | | **License** | n/a | | **Author** | [n/a]() | ### Label Scheme <details> <summary>View label scheme (912 labels for 3 components)</summary> | Component | Labels | | --- | --- | | **`morphologizer`** | `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `POS=ADV`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=NOUN`, `POS=VERB\|Tense=Past\|VerbForm=Inf\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=NOUN`, `POS=PUNCT`, `POS=CCONJ`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=NOUN`, `POS=ADP`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=DET`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=NOUN`, `Case=Gen\|Gender=Neut\|Number=Plur\|POS=NOUN`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=NOUN`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=NOUN`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=NOUN`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=PROPN`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=PROPN`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=PROPN`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=NOUN`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=NOUN`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=DET`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=NOUN`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=PRON\|Person=1`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=NOUN`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=DET`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=DET`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `POS=SCONJ`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=PROPN`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=DET`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=DET`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=PRON`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=PRON`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=PROPN`, `Case=Gen\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=DET`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=PRON`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=DET`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=PROPN`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=NOUN`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=PRON`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=PRON`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=NOUN`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Gender=Neut\|Number=Sing\|POS=NOUN`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Gender=Fem\|Number=Plur\|POS=NOUN`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=NOUN`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=PRON\|Person=1`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=PROPN`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=PRON`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=PROPN`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Gender=Neut\|Number=Plur\|POS=DET`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=PRON`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=NOUN`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=PRON`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=DET`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=DET`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=NOUN`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=PROPN`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Gender=Masc\|Number=Sing\|POS=NOUN`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=DET`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=PROPN`, `POS=VERB\|Tense=Pres\|VerbForm=Inf\|Voice=Act`, `POS=VERB\|Tense=Pres\|VerbForm=Inf\|Voice=Mid`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=AUX\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=PRON`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=PRON\|Person=1`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=NOUN`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=NOUN`, `Case=Dat\|Gender=Fem\|Number=Plur\|POS=DET`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=PROPN`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=PROPN`, `Aspect=Perf\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=PRON`, `POS=VERB\|Tense=Past\|VerbForm=Inf\|Voice=Pass`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=PROPN`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=NOUN`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Aspect=Perf\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Gender=Neut\|Number=Sing\|POS=DET`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=DET`, `POS=VERB\|Tense=Past\|VerbForm=Inf\|Voice=Mid`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=DET`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=PRON`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=DET`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=DET`, `Case=Dat\|Gender=Neut\|Number=Plur\|POS=PRON`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=DET`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Case=Dat\|Gender=Neut\|Number=Plur\|POS=DET`, `Case=Dat\|Gender=Neut\|Number=Plur\|POS=NOUN`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=NOUN`, `Case=Dat\|Gender=Neut\|Number=Sing\|POS=ADJ`, `POS=ADJ`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|POS=VERB\|Tense=Past\|VerbForm=Inf\|Voice=Mid`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=PRON\|Person=2`, `POS=INTJ`, `Case=Voc\|Gender=Masc\|Number=Sing\|POS=PROPN`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=DET`, `Mood=Sub\|Number=Plur\|POS=AUX\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `POS=AUX\|Tense=Pres\|VerbForm=Inf\|Voice=Act`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=PROPN`, `Case=Dat\|Gender=Neut\|Number=Sing\|POS=PROPN`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Case=Dat\|Gender=Neut\|Number=Plur\|POS=PROPN`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=DET`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=PRON\|Person=1`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=PRON\|Person=2`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Degree=Cmp\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=PRON\|Person=3`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Degree=Cmp\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Voc\|Gender=Fem\|Number=Sing\|POS=NOUN`, `Case=Voc\|Gender=Masc\|Number=Plur\|POS=NOUN`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=PRON`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=PROPN`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=PRON`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=PRON`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|POS=VERB\|Tense=Past\|VerbForm=Inf\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Imp\|Number=Plur\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=AUX\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Imp\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=PRON`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=PRON`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Degree=Cmp\|POS=ADV`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Gender=Neut\|Number=Sing\|POS=PRON`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Mood=Opt\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Gen\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=PRON\|Person=1`, `Case=Gen\|Gender=Neut\|Number=Plur\|POS=PRON`, `Case=Acc\|Degree=Cmp\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=DET`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pqp\|VerbForm=Fin\|Voice=Act`, `POS=NUM`, `Gender=Masc\|Number=Plur\|POS=NUM`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=PRON`, `Aspect=Perf\|Case=Gen\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Nom\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Degree=Sup\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=DET`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Degree=Sup\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pqp\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Voc\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pqp\|VerbForm=Fin\|Voice=Mid`, `POS=X`, `Case=Acc\|Degree=Sup\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Degree=Cmp\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=PRON\|Person=2`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=AUX\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=PROPN`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Number=Plur\|POS=PRON\|Person=2`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=PRON\|Person=2`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=AUX\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Acc\|Degree=Cmp\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Acc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=PRON`, `Degree=Sup\|POS=ADV`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=PRON`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Gender=Fem\|Number=Sing\|POS=PROPN`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Voc\|Gender=Neut\|Number=Sing\|POS=NOUN`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Pqp\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Case=Acc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Acc\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Degree=Sup\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Aspect=Perf\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Degree=Sup\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `POS=VERB\|Tense=Fut\|VerbForm=Inf\|Voice=Act`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=PRON`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Degree=Cmp\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Dat\|Gender=Fem\|Number=Plur\|POS=PRON`, `POS=DET`, `Case=Voc\|Gender=Fem\|Number=Plur\|POS=NOUN`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Case=Gen\|Number=Sing\|POS=PRON`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=AUX\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part`, `Case=Nom\|Number=Sing\|POS=NOUN`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Degree=Cmp\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=AUX\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Number=Sing\|POS=PRON`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Gender=Masc\|Number=Sing\|POS=NOUN`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Pqp\|VerbForm=Fin\|Voice=Mid`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Fut\|VerbForm=Fin\|Voice=Pass`, `Aspect=Perf\|Mood=Sub\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Degree=Cmp\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Pass`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Aspect=Perf\|Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Gender=Masc\|Number=Sing\|POS=ADJ`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `POS=PRON`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Degree=Sup\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Case=Voc\|Degree=Sup\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Imp\|Number=Plur\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin`, `Case=Gen\|Degree=Sup\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Aspect=Perf\|Mood=Imp\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Sub\|Number=Sing\|POS=AUX\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Case=Gen\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `POS=VERB\|Tense=Fut\|VerbForm=Inf\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Voc\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Dual\|POS=NOUN`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Gen\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `Case=Dat\|Gender=Fem\|Number=Dual\|POS=PRON`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=NOUN`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=DET`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=ADJ`, `Case=Dat\|Gender=Fem\|Number=Dual\|POS=ADJ`, `Case=Dat\|Gender=Fem\|Number=Dual\|POS=NOUN`, `Case=Nom\|Gender=Fem\|Number=Dual\|POS=ADJ`, `Case=Nom\|Gender=Fem\|Number=Dual\|POS=PRON`, `Aspect=Perf\|Case=Nom\|Gender=Fem\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Fem\|Number=Dual\|POS=NOUN`, `Case=Nom\|Gender=Fem\|Number=Dual\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Dat\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Mood=Imp\|Number=Sing\|POS=AUX\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=1\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Gender=Masc\|Number=Dual\|POS=DET`, `Case=Gen\|Gender=Masc\|Number=Dual\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=PRON`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Gender=Masc\|Number=Dual\|POS=ADJ`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Gen\|Degree=Sup\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Case=Dat\|Degree=Cmp\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=AUX\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Case=Dat\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Nom\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|Voice=Act`, `Aspect=Perf\|Case=Dat\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=X`, `Case=Nom\|Degree=Cmp\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pqp\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Aspect=Perf\|Mood=Imp\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Acc\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Dat\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `Case=Nom\|Degree=Cmp\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Degree=Cmp\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=AUX\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Dual\|POS=DET`, `Case=Acc\|Gender=Fem\|Number=Dual\|POS=NOUN`, `Case=Acc\|Gender=Fem\|Number=Dual\|POS=PRON`, `Case=Acc\|Gender=Fem\|Number=Dual\|POS=ADJ`, `Aspect=Perf\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=PROPN`, `Case=Acc\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=PROPN`, `Aspect=Perf\|Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Degree=Sup\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Mood=Sub\|Number=Sing\|POS=AUX\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Voc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Number=Sing\|POS=PRON`, `Aspect=Perf\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Mood=Imp\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Voc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Number=Sing\|POS=NOUN`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Pass`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Acc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Dat\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Mood=Imp\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Degree=Cmp\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `POS=VERB\|Tense=Past\|VerbForm=Inf`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Imp\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Gen\|Degree=Sup\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Case=Nom\|Degree=Sup\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Case=Gen\|Gender=Fem\|Number=Dual\|POS=NOUN`, `Case=Nom\|Number=Sing\|POS=PRON`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Imp\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Mood=Sub\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Degree=Cmp\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Mood=Sub\|Number=Sing\|POS=AUX\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Number=Dual\|POS=PRON`, `Aspect=Perf\|Case=Nom\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Mood=Imp\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Gender=Neut\|Number=Plur\|POS=NOUN`, `Mood=Opt\|Number=Sing\|POS=AUX\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Case=Nom\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Aspect=Perf\|Mood=Sub\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Case=Gen\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Degree=Cmp\|Number=Sing\|POS=ADJ`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=X`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Case=Gen\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Gender=Neut\|Number=Dual\|POS=ADJ`, `Aspect=Imp\|Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Mood=Opt\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Plur\|POS=PROPN`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Degree=Sup\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Dat\|Number=Plur\|POS=PRON`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=AUX\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Mood=Opt\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=AUX\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=1\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Gender=Fem\|Number=Dual\|POS=PRON`, `Case=Gen\|Gender=Fem\|Number=Dual\|POS=DET`, `Case=Gen\|Gender=Fem\|Number=Dual\|POS=ADJ`, `Case=Gen\|Gender=Fem\|Number=Dual\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=3\|Tense=Fut\|VerbForm=Fin`, `Case=Gen\|Degree=Cmp\|Number=Sing\|POS=ADJ`, `Mood=Imp\|POS=VERB\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Imp\|Number=Dual\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Voc\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Imp\|Number=Sing\|POS=AUX\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Degree=Cmp\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Case=Dat\|Degree=Sup\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Case=Dat\|Gender=Masc\|Number=Dual\|POS=NOUN`, `Aspect=Perf\|Case=Acc\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Fem\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Degree=Cmp\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Mood=Ind\|Number=Dual\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Degree=Cmp\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Aspect=Perf\|Case=Gen\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Voc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Voc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Acc\|Gender=Masc\|Number=Dual\|POS=PRON`, `Case=Nom\|Gender=Neut\|Number=Dual\|POS=NOUN`, `Case=Dat\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Pqp\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Degree=Sup\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Aspect=Perf\|Case=Voc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Gen\|Number=Plur\|POS=PROPN`, `Case=Nom\|Gender=Neut\|Number=Dual\|POS=ADJ`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pqp\|VerbForm=Fin\|Voice=Mid`, `Case=Nom\|Degree=Cmp\|Gender=Masc\|Number=Dual\|POS=ADJ`, `Mood=Ind\|Number=Dual\|POS=AUX\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Degree=Cmp\|Number=Plur\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Imp\|Number=Dual\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Dual\|POS=PRON`, `Mood=Imp\|Number=Plur\|POS=AUX\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Mood=Sub\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Number=Sing\|POS=ADJ`, `Aspect=Perf\|Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Number=Dual\|POS=PRON`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Case=Dat\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Fem\|POS=NOUN`, `Case=Acc\|Gender=Neut\|POS=NOUN`, `Aspect=Perf\|POS=VERB\|Tense=Past\|VerbForm=Inf\|Voice=Pass`, `Case=Dat\|Degree=Cmp\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pqp\|VerbForm=Fin\|Voice=Pass`, `Case=Dat\|Number=Plur\|POS=NOUN`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Voc\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Case=Nom\|Gender=Masc\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Voc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Gender=Masc\|Number=Plur\|POS=ADJ`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=1\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `POS=VERB\|Tense=Fut\|VerbForm=Inf\|Voice=Pass`, `Case=Acc\|Gender=Masc\|Number=Dual\|POS=ADJ`, `Case=Acc\|Gender=Masc\|Number=Dual\|POS=NOUN`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Aspect=Imp\|Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Gender=Neut\|Number=Dual\|POS=NOUN`, `Case=Gen\|Gender=Masc\|Number=Dual\|POS=PRON`, `Case=Voc\|Gender=Fem\|Number=Plur\|POS=PROPN`, `Case=Voc\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Number=Sing\|POS=ADJ`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Gen\|Number=Plur\|POS=NOUN`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Nom\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=AUX\|Person=3\|Tense=Past\|VerbForm=Fin`, `Aspect=Perf\|Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=PROPN`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Number=Sing\|POS=NOUN`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Imp\|Mood=Ind\|Number=Dual\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Pqp\|VerbForm=Fin\|Voice=Mid`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pqp\|VerbForm=Fin\|Voice=Pass`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `POS=AUX\|Tense=Fut\|VerbForm=Inf\|Voice=Mid`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Voc\|Gender=Masc\|Number=Plur\|POS=PROPN`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Mood=Sub\|Number=Dual\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Dual\|POS=ADJ`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pqp\|VerbForm=Fin`, `POS=VERB\|VerbForm=Inf\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part`, `Case=Acc\|Gender=Masc\|Number=Dual\|POS=PROPN`, `Aspect=Perf\|Case=Acc\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Voc\|Gender=Masc\|Number=Dual\|POS=PROPN`, `Case=Nom\|Number=Plur\|POS=PRON`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|POS=VERB\|Tense=Past\|VerbForm=Inf`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Neut\|Number=Dual\|POS=DET`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=AUX\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Degree=Cmp\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Degree=Sup\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Case=Acc\|Degree=Sup\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Pqp\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Degree=Sup\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Mood=Sub\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Gen\|Degree=Cmp\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=1\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=X\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Degree=Sup\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Pqp\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Gender=Neut\|Number=Dual\|POS=DET`, `Case=Nom\|Gender=Neut\|POS=NOUN`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|POS=VERB\|Tense=Fut\|VerbForm=Inf\|Voice=Act`, `Case=Voc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Pass`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Mood=Imp\|Number=Plur\|POS=AUX\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Degree=Sup\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Case=Voc\|Gender=Masc\|Number=Sing\|POS=PRON`, `Aspect=Perf\|Case=Acc\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Mood=Imp\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Case=Voc\|Gender=Masc\|Number=Dual\|POS=ADJ`, `Mood=Sub\|Number=Dual\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Aspect=Perf\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Dat\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Aspect=Perf\|Case=Nom\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Aspect=Perf\|Mood=Sub\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Pass`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Aspect=Perf\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Mood=Imp\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Dat\|Degree=Sup\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Dual\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `POS=VERB\|Tense=Pres\|VerbForm=Inf\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=X\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Voc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Opt\|Number=Plur\|POS=AUX\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=AUX\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Mood=Opt\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Mood=Imp\|Number=Dual\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Gender=Masc\|Number=Dual\|POS=PRON`, `Mood=Sub\|Number=Dual\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Mood=Imp\|Number=Dual\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Mood=Imp\|Number=Dual\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Mid`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Sub\|Number=Plur\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=AUX\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Aspect=Perf\|Case=Acc\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Degree=Sup\|Gender=Fem\|Number=Sing\|POS=ADJ`, `Mood=Opt\|Number=Dual\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `Mood=Imp\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=X\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Mood=Opt\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Gen\|Gender=Neut\|Number=Plur\|POS=PROPN`, `Case=Voc\|Number=Sing\|POS=NOUN`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part`, `Aspect=Perf\|Mood=Opt\|Number=Sing\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=1\|Tense=Past\|VerbForm=Fin`, `Aspect=Imp\|Mood=Ind\|Number=Dual\|POS=VERB\|Person=2\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Number=Plur\|POS=PRON`, `Case=Voc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Mood=Opt\|Number=Plur\|POS=VERB\|Person=1\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Aspect=Perf\|Case=Nom\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Gen\|Degree=Cmp\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Case=Acc\|Gender=Neut\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=1\|Tense=Fut\|VerbForm=Fin`, `Mood=Sub\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin`, `Aspect=Imp\|Mood=Ind\|Number=Sing\|POS=AUX\|Person=1\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Voc\|Gender=Masc\|Number=Dual\|POS=NOUN`, `Aspect=Imp\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Mood=Imp\|Number=Dual\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part`, `Aspect=Perf\|Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Mood=Ind\|Number=Dual\|POS=VERB\|Person=2\|Tense=Fut\|VerbForm=Fin\|Voice=Mid`, `Case=Dat\|Degree=Sup\|Gender=Neut\|Number=Sing\|POS=ADJ`, `Mood=Opt\|Number=Sing\|POS=AUX\|Person=2\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Mood=Imp\|Number=Plur\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `Case=Dat\|Number=Sing\|POS=PRON\|Person=2`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=PRON\|Person=1`, `Case=Nom\|Number=Sing\|POS=PRON\|Person=2`, `Case=Gen\|Number=Sing\|POS=PRON\|Person=2`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=PRON\|Person=1`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=PRON\|Person=1`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `Case=Acc\|Number=Sing\|POS=PRON\|Person=2`, `Case=Nom\|Gender=Fem\|Number=Sing\|POS=PRON\|Person=1`, `Case=Dat\|Number=Plur\|POS=PRON\|Person=2`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Case=Voc\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=PROPN`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=PRON\|Person=1`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=PRON\|Person=3`, `Aspect=Perf\|Mood=Ind\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Pass`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=PRON\|Person=1`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=PRON\|Person=1`, `Degree=Sup\|POS=ADJ`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=PRON\|Person=2`, `Case=Nom\|Gender=Masc\|Mood=Ind\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Number=Plur\|POS=X`, `Case=Dat\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Gen\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Pass`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Pass`, `Aspect=Perf\|Case=Gen\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=3\|Tense=Pres\|VerbForm=Fin\|Voice=Pass`, `Mood=Ind\|Number=Plur\|POS=AUX\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Neut\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=NUM`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Dat\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Mood=Ind\|Number=Plur\|POS=VERB\|Person=1\|Tense=Pqp\|VerbForm=Fin\|Voice=Mid`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Masc\|Mood=Ind\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Sing\|POS=AUX\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Gender=Neut\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Gen\|Gender=Masc\|Number=Sing\|POS=PRON\|Person=1`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=PRON\|Person=2`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=3\|Tense=Fut\|VerbForm=Fin\|Voice=Act`, `Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Gen\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Gender=Masc\|Number=Plur\|POS=NUM`, `Case=Voc\|Degree=Cmp\|Gender=Masc\|Number=Sing\|POS=ADJ`, `Case=Gen\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Mid`, `POS=VERB\|Tense=Pres\|VerbForm=Inf`, `Gender=Masc\|Number=Sing\|POS=PROPN`, `Aspect=Perf\|Case=Voc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=PROPN`, `Case=Nom\|Gender=Fem\|Number=Dual\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Acc\|Number=Dual\|POS=NOUN`, `Aspect=Imp\|Mood=Ind\|Number=Dual\|POS=AUX\|Person=3\|Tense=Past\|VerbForm=Fin\|Voice=Act`, `Case=Acc\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Gen\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Nom\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Pres\|VerbForm=Part`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Degree=Sup\|Gender=Masc\|Number=Plur\|POS=ADJ`, `Case=Nom\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Pass`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Dat\|Degree=Sup\|Gender=Neut\|Number=Plur\|POS=ADJ`, `Case=Nom\|Degree=Cmp\|Gender=Fem\|Number=Plur\|POS=ADJ`, `Aspect=Perf\|Case=Dat\|Gender=Neut\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Act`, `Aspect=Perf\|Case=Dat\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Case=Dat\|Gender=Fem\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Inf\|Voice=Act`, `Case=Dat\|Gender=Fem\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Dat\|Gender=Neut\|Number=Sing\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Gender=Neut\|Number=Plur\|POS=X`, `Case=Dat\|Gender=Neut\|Number=Plur\|POS=AUX\|Tense=Pres\|VerbForm=Part\|Voice=Act`, `Case=Nom\|Number=Sing\|POS=PROPN`, `Case=Acc\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Pass`, `Case=Gen\|Gender=Masc\|Number=Plur\|POS=VERB\|Tense=Fut\|VerbForm=Part\|Voice=Act`, `Case=Acc\|Gender=Masc\|Number=Plur\|POS=VERB\|VerbForm=Part\|Voice=Mid`, `Case=Gen\|Gender=Fem\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Pass`, `Case=Gen\|Gender=Neut\|Number=Plur\|POS=VERB\|Tense=Past\|VerbForm=Part\|Voice=Mid`, `Aspect=Perf\|Case=Nom\|Gender=Masc\|Number=Sing\|POS=VERB\|Tense=Past\|VerbForm=Part`, `Case=Acc\|Gender=Fem\|POS=VERB\|Tense=Pres\|VerbForm=Part\|Voice=Mid`, `Mood=Opt\|Number=Sing\|POS=VERB\|Person=3\|Tense=Past\|VerbForm=Fin`, `Case=Acc\|Number=Plur\|POS=PRON`, `Mood=Ind\|Number=Sing\|POS=AUX\|Person=2\|Tense=Fut\|VerbForm=Fin` | | **`tagger`** | `---------`, `--p---fa-`, `--s---ma-`, `-3paia---`, `-3paim---`, `-3siia---`, `A-`, `C-`, `D-`, `F-`, `G-`, `I-`, `M-`, `Nb`, `Ne`, `P-`, `R-`, `V-`, `Z` | | **`parser`** | `ROOT`, `acl`, `advcl`, `advmod`, `amod`, `appos`, `case`, `cc`, `ccomp`, `conj`, `cop`, `csubj`, `dep`, `det`, `discourse`, `iobj`, `mark`, `nmod`, `nsubj`, `nummod`, `obj`, `obl`, `parataxis`, `punct`, `vocative`, `xcomp` | </details> ### Accuracy | Type | Score | | --- | --- | | `POS_ACC` | 96.72 | | `MORPH_ACC` | 95.01 | | `TAG_ACC` | 96.84 | | `DEP_UAS` | 81.72 | | `DEP_LAS` | 77.02 | | `SENTS_P` | 98.68 | | `SENTS_R` | 98.94 | | `SENTS_F` | 98.81 | | `LEMMA_ACC` | 95.24 | | `TRANSFORMER_LOSS` | 132835.18 | | `MORPHOLOGIZER_LOSS` | 6393.54 | | `TAGGER_LOSS` | 3844.03 | | `PARSER_LOSS` | 1953813.72 |
Atampy26/GPT-Glacier
[ "pytorch", "gpt_neo", "text-generation", "transformers" ]
text-generation
{ "architectures": [ "GPTNeoForCausalLM" ], "model_type": "gpt_neo", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
5
null
--- tags: - FrozenLake-v1-4x4-no_slippery - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-FrozenLake-v1-4x4-noSlippery results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: FrozenLake-v1-4x4-no_slippery type: FrozenLake-v1-4x4-no_slippery metrics: - type: mean_reward value: 1.00 +/- 0.00 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="adrienJeg/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
Augustvember/WOKKAWOKKA
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
12
null
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-Pyramids 2. Step 1: Find your model_id: alvarobb/ppo-PyramidsRNG 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
Augustvember/wokka2
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
12
2023-04-16T17:22:15Z
--- tags: - unity-ml-agents - ml-agents - deep-reinforcement-learning - reinforcement-learning - ML-Agents-SoccerTwos library_name: ml-agents --- # **poca** Agent playing **SoccerTwos** This is a trained model of a **poca** agent playing **SoccerTwos** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-SoccerTwos 2. Step 1: Write your model_id: Mithul/poca-SoccerTwos 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
Augustvember/wokkabottest2
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
13
2023-04-16T17:27:51Z
--- license: mit tags: - pytorch - diffusers - unconditional-image-generation - diffusion-models-class --- # Model Card for Unit 1 of the [Diffusion Models Class 🧨](https://github.com/huggingface/diffusion-models-class) This model is a diffusion model for unconditional image generation of cute 🦋. ## Usage ```python from diffusers import DDPMPipeline pipeline = DDPMPipeline.from_pretrained('utyug1/sd-class-butterflies-32') image = pipeline().images[0] image ```
Augustvember/your-model-name
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
2023-04-16T17:29:45Z
--- tags: - unity-ml-agents - ml-agents - deep-reinforcement-learning - reinforcement-learning - ML-Agents-SoccerTwos library_name: ml-agents --- # **poca** Agent playing **SoccerTwos** This is a trained model of a **poca** agent playing **SoccerTwos** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-SoccerTwos 2. Step 1: Write your model_id: mdapri/poca-SoccerTwos 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
Aybars/ModelOnTquad
[ "pytorch", "bert", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "BertForQuestionAnswering" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
2023-04-16T17:55:37Z
--- license: openrail language: - en metrics: - accuracy pipeline_tag: text-classification library_name: adapter-transformers ---
Ayham/ernie_gpt2_summarization_cnn_dailymail
[ "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "dataset:cnn_dailymail", "transformers", "generated_from_trainer", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "EncoderDecoderModel" ], "model_type": "encoder-decoder", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
13
null
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 269.21 +/- 20.44 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
Ayham/roberta_distilgpt2_summarization_cnn_dailymail
[ "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "dataset:cnn_dailymail", "transformers", "generated_from_trainer", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "EncoderDecoderModel" ], "model_type": "encoder-decoder", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
2023-04-16T18:41:37Z
--- license: creativeml-openrail-m language: - ja --- Nakagin Capsule Tower Building Lora File We are learning three parts: the building exterior, the room, and the bathroom. Please use the value of Lora at about 0.8 to 1.2 If adding elements to the image weakens Lora's expression, increase Lora's number to ~1.2 or add tags for the elements included in the data to reinforce it Here are the calling and additional tags for each scene: room Invocation tag: nkr room Additional tags: indoors, realistic photo background window, console, bed, CRT TV exterior call tag: nkg building additional tag: outdoors, realistic, photo background road, city, Bathroom Invocation tag: nkb, bathroom Additional tags: indoors, realistic photo background toilet, sink, mirror, bathtub, etc. 中銀カプセルタワービルのLoraファイルです ビル外観、部屋、バスルームの三か所を学習しています Loraの数値は0.8~1.2程度で使用して下さい 画像に要素を追加することでLoraの表現が弱くなる場合、Loraの数値を~1.2に上げるかデータに含まれる要素のタグを加えて補強してください 各シーンの呼び出しタグと追加タグは次のとおりです 部屋 呼び出しタグ:nkr room 追加タグ: indoors,realistic, photo background window, console, bed, CRT TV 外観 呼び出しタグ:nkg, building, 追加タグ: outdoors, realistic, photo background road, city バスルーム 呼び出しタグ: nkb, bathroom 追加タグ: indoors,realistic, photo background toilet, sink, mirror,bathtubなど 学習元はNAIとchilloutです
Ayham/roberta_gpt2_new_max64_summarization_cnndm
[ "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "dataset:cnn_dailymail", "transformers", "generated_from_trainer", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "EncoderDecoderModel" ], "model_type": "encoder-decoder", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
2023-04-16T18:42:11Z
--- language: en thumbnail: https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true tags: - huggingtweets widget: - text: "My dream is" --- <div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1554121036349702147/KJlo7HqZ_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">時期赤 🇺🇸</div> <div style="text-align: center; font-size: 14px;">@jekred2</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from 時期赤 🇺🇸. | Data | 時期赤 🇺🇸 | | --- | --- | | Tweets downloaded | 3245 | | Retweets | 885 | | Short tweets | 261 | | Tweets kept | 2099 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/k6ifq2s8/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @jekred2's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/qh3s0b3s) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/qh3s0b3s/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/jekred2') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
Ayham/roberta_roberta_summarization_cnn_dailymail
[ "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "dataset:cnn_dailymail", "transformers", "generated_from_trainer", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "EncoderDecoderModel" ], "model_type": "encoder-decoder", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
3
2023-04-16T18:51:44Z
*.7z filter=lfs diff=lfs merge=lfs -text *.arrow filter=lfs diff=lfs merge=lfs -text *.bin filter=lfs diff=lfs merge=lfs -text *.bz2 filter=lfs diff=lfs merge=lfs -text *.ckpt filter=lfs diff=lfs merge=lfs -text *.ftz filter=lfs diff=lfs merge=lfs -text *.gz filter=lfs diff=lfs merge=lfs -text *.h5 filter=lfs diff=lfs merge=lfs -text *.joblib filter=lfs diff=lfs merge=lfs -text *.lfs.* filter=lfs diff=lfs merge=lfs -text *.mlmodel filter=lfs diff=lfs merge=lfs -text *.model filter=lfs diff=lfs merge=lfs -text *.msgpack filter=lfs diff=lfs merge=lfs -text *.npy filter=lfs diff=lfs merge=lfs -text *.npz filter=lfs diff=lfs merge=lfs -text *.onnx filter=lfs diff=lfs merge=lfs -text *.ot filter=lfs diff=lfs merge=lfs -text *.parquet filter=lfs diff=lfs merge=lfs -text *.pb filter=lfs diff=lfs merge=lfs -text *.pickle filter=lfs diff=lfs merge=lfs -text *.pkl filter=lfs diff=lfs merge=lfs -text *.pt filter=lfs diff=lfs merge=lfs -text *.pth filter=lfs diff=lfs merge=lfs -text *.rar filter=lfs diff=lfs merge=lfs -text *.safetensors filter=lfs diff=lfs merge=lfs -text saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.tar.* filter=lfs diff=lfs merge=lfs -text *.tflite filter=lfs diff=lfs merge=lfs -text *.tgz filter=lfs diff=lfs merge=lfs -text *.wasm filter=lfs diff=lfs merge=lfs -text *.xz filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text
Ayham/robertagpt2_xsum4
[ "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "transformers", "generated_from_trainer", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "EncoderDecoderModel" ], "model_type": "encoder-decoder", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
2023-04-16T20:02:14Z
--- license: other inference: false --- [llama.cpp PR required](https://github.com/ggerganov/llama.cpp/pull/1004)
Ayham/xlnet_bert_summarization_cnn_dailymail
[ "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "dataset:cnn_dailymail", "transformers", "generated_from_trainer", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "EncoderDecoderModel" ], "model_type": "encoder-decoder", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
7
2023-04-16T19:03:58Z
--- library_name: stable-baselines3 tags: - AntBulletEnv-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: AntBulletEnv-v0 type: AntBulletEnv-v0 metrics: - type: mean_reward value: 2836.29 +/- 12.24 name: mean_reward verified: false --- # **A2C** Agent playing **AntBulletEnv-v0** This is a trained model of a **A2C** agent playing **AntBulletEnv-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
Ayham/xlnet_roberta_new_summarization_cnn_dailymail
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
2023-04-16T19:09:34Z
--- tags: - autotrain - summarization language: - unk widget: - text: "I love AutoTrain 🤗" datasets: - Udit191/autotrain-data-summarizationpegasus co2_eq_emissions: emissions: 4.220134504425799 --- # Model Trained Using AutoTrain - Problem type: Summarization - Model ID: 50007120073 - CO2 Emissions (in grams): 4.2201 ## Validation Metrics - Loss: 2.619 - Rouge1: 36.984 - Rouge2: 16.807 - RougeL: 24.810 - RougeLsum: 32.737 - Gen Len: 57.700 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_HUGGINGFACE_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoTrain"}' https://api-inference.huggingface.co/Udit191/autotrain-summarizationpegasus-50007120073 ```
Ayham/xlnetgpt2_xsum7
[ "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "transformers", "generated_from_trainer", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "EncoderDecoderModel" ], "model_type": "encoder-decoder", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
--- license: apache-2.0 tags: - setfit - sentence-transformers - text-classification pipeline_tag: text-classification --- # nihiluis/legal-components-mpnet This is a [SetFit model](https://github.com/huggingface/setfit) that can be used for text classification. The model has been trained using an efficient few-shot learning technique that involves: 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning. 2. Training a classification head with features from the fine-tuned Sentence Transformer. ## Usage To use this model for inference, first install the SetFit library: ```bash python -m pip install setfit ``` You can then run inference as follows: ```python from setfit import SetFitModel # Download from Hub and run inference model = SetFitModel.from_pretrained("nihiluis/legal-components-mpnet") # Run inference preds = model(["i loved the spiderman movie!", "pineapple on pizza is the worst 🤮"]) ``` ## BibTeX entry and citation info ```bibtex @article{https://doi.org/10.48550/arxiv.2209.11055, doi = {10.48550/ARXIV.2209.11055}, url = {https://arxiv.org/abs/2209.11055}, author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren}, keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Efficient Few-Shot Learning Without Prompts}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ```
Aymene/opus-mt-en-ro-finetuned-en-to-ro
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
2023-04-16T19:15:34Z
--- library_name: stable-baselines3 tags: - AntBulletEnv-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: AntBulletEnv-v0 type: AntBulletEnv-v0 metrics: - type: mean_reward value: 1921.04 +/- 321.41 name: mean_reward verified: false --- # **A2C** Agent playing **AntBulletEnv-v0** This is a trained model of a **A2C** agent playing **AntBulletEnv-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
Ayoola/pytorch_model
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- license: apache-2.0 tags: - generated_from_trainer metrics: - bleu model-index: - name: my_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_model This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.2877 - Bleu: 0.0001 - Gen Len: 17.7878 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:------:|:-------:| | 0.3656 | 1.0 | 1000 | 0.3225 | 0.0 | 18.175 | | 0.3395 | 2.0 | 2000 | 0.3070 | 0.0 | 18.193 | | 0.3266 | 3.0 | 3000 | 0.2998 | 0.0 | 18.0902 | | 0.3142 | 4.0 | 4000 | 0.2957 | 0.0 | 17.8615 | | 0.3155 | 5.0 | 5000 | 0.2925 | 0.0 | 17.8932 | | 0.3045 | 6.0 | 6000 | 0.2908 | 0.0 | 17.8382 | | 0.3093 | 7.0 | 7000 | 0.2894 | 0.0 | 17.8515 | | 0.3012 | 8.0 | 8000 | 0.2885 | 0.0001 | 17.827 | | 0.3007 | 9.0 | 9000 | 0.2878 | 0.0 | 17.7822 | | 0.302 | 10.0 | 10000 | 0.2877 | 0.0001 | 17.7878 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
Ayran/DialoGPT-medium-harry-potter-1-through-3
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
12
null
--- tags: - autotrain - summarization language: - unk widget: - text: "I love AutoTrain 🤗" datasets: - Udit191/autotrain-data-summarization co2_eq_emissions: emissions: 1.0944595053594732 --- # Model Trained Using AutoTrain - Problem type: Summarization - Model ID: 50010120083 - CO2 Emissions (in grams): 1.0945 ## Validation Metrics - Loss: 2.930 - Rouge1: 17.853 - Rouge2: 6.331 - RougeL: 14.314 - RougeLsum: 16.139 - Gen Len: 20.000 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_HUGGINGFACE_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoTrain"}' https://api-inference.huggingface.co/Udit191/autotrain-summarization-50010120083 ```
Ayran/DialoGPT-medium-harry-potter-1-through-4-plus-6-e18
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
12
2023-04-16T19:28:29Z
--- license: creativeml-openrail-m base_model: runwayml/stable-diffusion-v1-5 tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - controlnet - jax-diffusers-event inference: true --- # controlnet- birgermoell/controlnet-fill-circle These are controlnet weights trained on runwayml/stable-diffusion-v1-5 with new type of conditioning. You can find some example images in the following. prompt: red circle with blue background ![images_0)](./images_0.png) prompt: cyan circle with brown floral background ![images_1)](./images_1.png)
Ayran/DialoGPT-medium-harry-potter-1-through-4-plus-6
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
12
2023-04-16T19:29:41Z
--- license: mit datasets: - botisan-ai/cantonese-mandarin-translations language: - yue - zh metrics: - bleu - chrf library_name: transformers pipeline_tag: translation tags: - BART --- # Guide Please read the `finetune.ipynb` for the main training code.
Ayran/DialoGPT-small-gandalf
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
11
2023-04-16T19:29:44Z
--- tags: - Pixelcopter-PLE-v0 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-PixelCopter results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Pixelcopter-PLE-v0 type: Pixelcopter-PLE-v0 metrics: - type: mean_reward value: 72.10 +/- 38.56 name: mean_reward verified: false --- # **Reinforce** Agent playing **Pixelcopter-PLE-v0** This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
AyushPJ/ai-club-inductions-21-nlp-roBERTa-base-squad-v2
[ "pytorch", "roberta", "question-answering", "transformers", "generated_from_trainer", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
--- tags: - autotrain - summarization language: - unk widget: - text: "I love AutoTrain 🤗" datasets: - Udit191/autotrain-data-summarization-bart co2_eq_emissions: emissions: 0.005523336238976097 --- # Model Trained Using AutoTrain - Problem type: Summarization - Model ID: 50017120109 - CO2 Emissions (in grams): 0.0055 ## Validation Metrics - Loss: 2.538 - Rouge1: 49.903 - Rouge2: 20.621 - RougeL: 27.927 - RougeLsum: 44.133 - Gen Len: 142.000 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_HUGGINGFACE_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoTrain"}' https://api-inference.huggingface.co/Udit191/autotrain-summarization-bart-50017120109 ```
AyushPJ/test-squad-trained-finetuned-squad
[ "pytorch", "tensorboard", "distilbert", "question-answering", "dataset:squad", "transformers", "generated_from_trainer", "autotrain_compatible" ]
question-answering
{ "architectures": [ "DistilBertForQuestionAnswering" ], "model_type": "distilbert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
2023-04-16T19:46:59Z
--- library_name: ml-agents tags: - SnowballTarget - deep-reinforcement-learning - reinforcement-learning - ML-Agents-SnowballTarget --- # **ppo** Agent playing **SnowballTarget** This is a trained model of a **ppo** agent playing **SnowballTarget** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-SnowballTarget 2. Step 1: Find your model_id: mgarciav/ppo-SnowballTarget 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
Azaghast/GPT2-SCP-Miscellaneous
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": true, "max_length": 50 }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
5
null
--- # For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1 # Doc / guide: https://huggingface.co/docs/hub/model-cards datasets: - MU-NLPC/Calc-gsm8k - gsm8k metrics: - exact_match - rouge model-index: - name: Calc-FLAN-3B-GSM8K results: - task: type: question-answering name: Question Answering dataset: type: gsm8k name: GSM8K split: validation metrics: - type: exact_match value: 0.495 - type: rouge value: 0.655 license: apache-2.0 language: - en --- # Model Card for Calculator-FLAN-T5-3B_GSM8K <!-- Provide a quick summary of what the model is/does. --> This model generates reasoning chains over mathematical questions while **using an external tool: Sympy calculator**. ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> With the idea to offload a symbolic reasoning from the stochastic language model, we train this model to utilize a calculator **for all applicable numeric operations**. This is achieved by training the model to construct calls to the tool's API in this format: ```html <gadget id="calculator">100/2</gadget> <output>50</output> ``` where `<gadget>` segment triggers a call of the tool, which is subsequently served by extending model's decoder input context by adding the output of the tool within the `<output>` segment. - **Developed by:** Marek Kadlcik & Michal Stefanik, Masaryk University - **Model type:** Autoregressive Encoder-Decoder - **Language(s):** en - **Finetuned from:** google/flan-t5-xl ### Model Sources <!-- Provide the basic links for the model. --> - **Repository:** https://github.com/prompteus/gadgets - **Paper:** Stay tuned! ## Usage Additionally to conventional generation, using Tool-augmented generation requires (1) implementation of the tool(s) and (2) a customization of generate() method augmenting input context on-demand with the outputs of the tools. You can find these two components implemented in the attached **gadget_assisted_model.py** and **gadget.py** in this model's repo and the project's [home repo](https://github.com/prompteus/gadgets). After adding these two scripts to your directory, you can use the model as follows: ```python from gadget_assisted_model import GadgetAssistedModel from gadget import Calculator from transformers import T5ForConditionalGeneration, T5Tokenizer class GadgetAssistedT5(GadgetAssistedModel, T5ForConditionalGeneration): # GadgetAssistedModel overrides the standard generate() from transformers pass model = GadgetAssistedT5.from_pretrained("MU-NLPC/Calc-FLAN-3B-GSM8K") tokenizer = T5Tokenizer.from_pretrained("MU-NLPC/Calc-FLAN-3B-GSM8K") model.prepare_for_generate(tokenizer, enabled_gadgets=[Calculator()], default_max_tokens=512) query = """ The profit from a business transaction is shared among 2 business partners, Mike and Johnson in the ratio 2:5 respectively. If Johnson got $2500, how much will Mike have after spending some of his share on a shirt that costs $200? """ inputs = tokenizer(query, return_tensors="pt") output_ids = model.generate(**inputs) tokenizer.decode(output_ids[0], spaces_between_special_tokens=False) ``` This returns: ```html According to the ratio, Mike got 2/5*$2500 = $<gadget id="calculator">2/5*2500</gadget><output>1_000</output> 1000 Mike will have $1000-$200 = $<gadget id="calculator">1000-200</gadget><output>800</output> 800 after buying a shirt. Final result is<result>800</result></s> ``` ### Out-of-Scope Usage <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> Note that given the limited scope of the exercises' complexity in the training, this model will not work well for tasks requiring more complex algebraic operations, including equations, variables and operations outside the scope of (+-*/). ## Training Details ### Training Data <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> This model was trained on oru [Calculator-augmented set of GSM8K](https://huggingface.co/datasets/gsm8k) in a standard auto-regressive setup i.e. for a conditional next-token prediction with teacher-forced prefix. ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> The model was fine-tuned from [google/flan-t5-xl](https://huggingface.co/google/flan-t5-xl) for 10,400 steps aiming to maximise exact-match ration on a validation split of the questions from [gsm8k dataset](https://huggingface.co/datasets/gsm8k). We fine-tune only 10% of the parameters finding that this circumvents overfitting to relatively small training dataset. The full training configuration can be identified from the [training script](https://github.com/prompteus/gadgets/blob/9185d1fc4b4812321179f8e5cad3e2f2a764f1df/examples/train_gsm8k_flan-t5-slice.py). ### Contact If you'd like to chat about this, please get in touch via email at kadlcik`<at>`mail.muni.cz or stefanik.m`<at>`mail.muni.cz
Azuris/DialoGPT-medium-senorita
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
14
null
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: bert-paper-classifier results: [] --- # bert-paper-classifier This model is a fine-tuned version of [microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract](https://huggingface.co/microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract) on the dataset from [González-Márquez et al., 2023](https://www.biorxiv.org/content/10.1101/2023.04.10.536208v1). ## Intended uses & limitations This model is intended to predict the category given the paper title (and optionally its abstract) — for the biomedical papers. For example, it is likely to predict `virology` as a category for the paper with a title containing `COVID-19`. So far only a subset of the PubMed dataset has been used for training. Future improvements to this model can come with using the full dataset with a combination of titles and abstracts for the fine-tuning as well as extending the training set to the preprints from bioRxiv and/or arXiv. ## Training procedure The code for the model fine-tuning can be found [in the respective notebook](https://huggingface.co/oracat/bert-paper-classifier/blob/main/finetuning-pubmed.ipynb). ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 128 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu117 - Datasets 2.11.0 - Tokenizers 0.13.3
Azuris/DialoGPT-small-envy
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
14
null
--- tags: - generated_from_trainer metrics: - bleu model-index: - name: nllb-ecolindo results: [] datasets: - yonathanstwn/ecolindo language: - id - en --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # nllb-ecolindo This model was trained from scratch on the EColIndo dataset. It achieves the following results on the evaluation set: - Loss: 0.8704 - Bleu: 37.2396 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 4000 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | |:-------------:|:-----:|:------:|:---------------:|:-------:| | 1.3017 | 1.0 | 21019 | 0.9831 | 35.0676 | | 1.0088 | 2.0 | 42038 | 0.9318 | 36.3191 | | 0.9472 | 3.0 | 63057 | 0.9090 | 36.5221 | | 0.9078 | 4.0 | 84076 | 0.8919 | 36.3949 | | 0.8789 | 5.0 | 105095 | 0.8833 | 37.3689 | | 0.8576 | 6.0 | 126114 | 0.8785 | 37.2262 | | 0.8403 | 7.0 | 147133 | 0.8748 | 37.2933 | | 0.8281 | 8.0 | 168152 | 0.8717 | 37.2952 | | 0.8186 | 9.0 | 189171 | 0.8709 | 37.2801 | | 0.813 | 10.0 | 210190 | 0.8704 | 37.2396 | ### Framework versions - Transformers 4.26.1 - Pytorch 2.0.0 - Datasets 2.10.1 - Tokenizers 0.11.0
Backedman/DialoGPT-small-Anika
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
6
null
--- # For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1 # Doc / guide: https://huggingface.co/docs/hub/model-cards {} --- # Model Card for Model ID Meta Segment-anything This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Data Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Badr/model1
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-Pyramids 2. Step 1: Find your model_id: mgarciav/Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
Bagus/ser-japanese
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- library_name: ml-agents tags: - Huggy - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Huggy --- # **ppo** Agent playing **Huggy** This is a trained model of a **ppo** agent playing **Huggy** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-Huggy 2. Step 1: Find your model_id: RagnaChris/ppo-Huggy 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
BalajiSathesh/DialoGPT-small-harrypotter
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
--- license: mit tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: sexism-identification-coroseof results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # sexism-identification-coroseof This model is a fine-tuned version of [dumitrescustefan/bert-base-romanian-uncased-v1](https://huggingface.co/dumitrescustefan/bert-base-romanian-uncased-v1) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.6960 - Accuracy: 0.8499 - F1: 0.8537 - Balanced Accuracy: 0.6139 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Balanced Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:-----------------:| | No log | 1.0 | 488 | 0.9896 | 0.8125 | 0.8263 | 0.6059 | | 0.9572 | 2.0 | 976 | 0.8694 | 0.7992 | 0.8202 | 0.7183 | | 0.5835 | 3.0 | 1464 | 1.1954 | 0.8388 | 0.8477 | 0.6485 | | 0.2833 | 4.0 | 1952 | 1.6960 | 0.8499 | 0.8537 | 0.6139 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu117 - Datasets 2.11.0 - Tokenizers 0.13.3
Banshee/LukeSkywalker
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: Finetuned_FLAN-T5_VALUE_adapterfusion_lr5e-5_bs64 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Finetuned_FLAN-T5_VALUE_adapterfusion_lr5e-5_bs64 This model is a fine-tuned version of [liuyanchen1015/FLAN-T5_GLUE_finetuning_lr3e-4](https://huggingface.co/liuyanchen1015/FLAN-T5_GLUE_finetuning_lr3e-4) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0879 - Accuracy: 0.8776 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.0597 | 0.07 | 1000 | 0.0890 | 0.8752 | | 0.055 | 0.14 | 2000 | 0.0886 | 0.8768 | | 0.0545 | 0.2 | 3000 | 0.0870 | 0.8746 | | 0.0526 | 0.27 | 4000 | 0.0881 | 0.8756 | | 0.0515 | 0.34 | 5000 | 0.0876 | 0.8756 | | 0.0501 | 0.41 | 6000 | 0.0885 | 0.8788 | | 0.0497 | 0.47 | 7000 | 0.0908 | 0.8756 | | 0.0498 | 0.54 | 8000 | 0.0899 | 0.8776 | | 0.0509 | 0.61 | 9000 | 0.0902 | 0.8754 | | 0.0504 | 0.68 | 10000 | 0.0898 | 0.8736 | | 0.049 | 0.74 | 11000 | 0.0879 | 0.8734 | | 0.0494 | 0.81 | 12000 | 0.0878 | 0.8746 | | 0.0477 | 0.88 | 13000 | 0.0883 | 0.8766 | | 0.0498 | 0.95 | 14000 | 0.0879 | 0.8776 | ### Framework versions - Transformers 4.26.1 - Pytorch 1.13.0+cu117 - Datasets 2.10.1 - Tokenizers 0.12.1
Banshee/dialoGPT-luke-small
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
2023-04-16T20:42:07Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - emotion metrics: - accuracy model-index: - name: bert-base-uncased-clasificator-emotions results: - task: name: Text Classification type: text-classification dataset: name: emotion type: emotion config: split split: validation args: split metrics: - name: Accuracy type: accuracy value: 0.9335 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-uncased-clasificator-emotions This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.1825 - Accuracy: 0.9335 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.1692 | 1.0 | 250 | 0.1858 | 0.931 | | 0.1201 | 2.0 | 500 | 0.1818 | 0.9315 | | 0.0829 | 3.0 | 750 | 0.1800 | 0.933 | | 0.0568 | 4.0 | 1000 | 0.1825 | 0.9335 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
Barleysack/AERoberta
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
7
null
--- language: - en license: creativeml-openrail-m thumbnail: "https://huggingface.co/Guizmus/SDArt_ChaosAndOrder768/resolve/main/showcase.jpg" tags: - stable-diffusion - text-to-image - image-to-image --- # SDArt : Chaos And Order (version based on 2.1 768px) ![Showcase](https://huggingface.co/Guizmus/SDArt_ChaosAndOrder768/resolve/main/showcase.jpg) ## Theme > The storm was a maelstrom of darkness and red, a tempest of writhing fury that threatened to swallow everything in its path. Against this chaos stood the ordered landscape, an unwavering bastion of precision and clarity, holding back the furious assault. The battle raged on, each side locked in an unforgiving struggle for dominance. > Amidst Chaos & Order, a figure emerged. A young girl with eyes that glowed like the sun stood tall and unafraid as the battle unfolded. With a calm determination, she held out her hands and began to guide the energies around her. The forces of Chaos & Order clashed and collided, lightning bolts and sharp lines slashing through the air with reckless abandon. But the girl remained unflinching, her movements fluid and precise as she weaved her way through the tumultuous battlefield. > The opposing forces began to merge as swirling clouds of Chaos began to dissipate, softening into curves and patterns that blended seamlessly with the structured landscape of Order. The sky bloomed into a tapestry of colors, a reflection emerging forth from two opposites. > In the end, it was neither Chaos or Order that prevailed, but the harmony that emerged from their interplay. And the young girl, with her eyes shining like stars, knew that this was the true power of Balance. 💢 Two forces, one outcome. Chaos and Order clash in a visually stunning battle for dominance! 🔳 *Encapsulate the tension between chaos and order in a dynamic, visually striking composition. The human experience of the push-pull struggle between two powerful forces.* **Challenges:** * Use contrasting colors and shapes to create a sense of conflict between two or more opposing forces. * At least one aspect that represents Chaos and one aspect that represents Order. * Create a sense of movement and energy that captures the dynamic interplay between Chaos & Order. ## Model description This is a model related to the "Challenge of the WeekEnd" contest on [Stable Diffusion discord](https://discord.gg/stablediffusion). I try to make a model out of all the submission for people to continue enjoy the theme after the even, and see a little of their designs in other people's creations. The token stays "SDArt" and I balance the learning on the low side, so that it doesn't just replicate creations. The total dataset is made of 22 pictures. It was trained on [Stable diffusion 2.1 768px](https://huggingface.co/stabilityai/stable-diffusion-2-1). I used [EveryDream](https://github.com/victorchall/EveryDream2trainer) to do the training, 100 total repeat per picture. The pictures were tagged using the token "SDArt", and an arbitrary token I choose. The dataset is provided below, as well as a list of usernames and their corresponding token. The recommended sampling is k_Euler_a or DPM++ 2M Karras on 20 steps, CFGS 7.5 . [The model is also available here](https://huggingface.co/Guizmus/SDArt_ChaosAndOrder) in a version trained on 1.5 as a base. ## Trained tokens * SDArt * rcn * in * aten * opi * omd * kuro * nlwx * asot * psst * buka * muc * kts * utm * avel * mss * guin * pgs * crit * mlas * phol * dds * pte * rean ## Download links [SafeTensors](https://huggingface.co/Guizmus/SDArt_ChaosAndOrder768/resolve/main/SDArt_ChaosAndOrder768.safetensors) [CKPT](https://huggingface.co/Guizmus/SDArt_ChaosAndOrder768/resolve/main/SDArt_ChaosAndOrder768.ckpt) [Config (yaml)](https://huggingface.co/Guizmus/SDArt_ChaosAndOrder768/resolve/main/SDArt_ChaosAndOrder768.yaml) [Dataset](https://huggingface.co/Guizmus/SDArt_ChaosAndOrder768/resolve/main/dataset.zip) ## 🧨 Diffusers This model can be used just like any other Stable Diffusion model. For more information, please have a look at the [Stable Diffusion](https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion). You can also export the model to [ONNX](https://huggingface.co/docs/diffusers/optimization/onnx), [MPS](https://huggingface.co/docs/diffusers/optimization/mps) and/or [FLAX/JAX](). ```python from diffusers import StableDiffusionPipeline import torch model_id = "Guizmus/SDArt_ChaosAndOrder768" pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16) pipe = pipe.to("cuda") prompt = "SDArt mlas" image = pipe(prompt).images[0] image.save("./SDArt.png") ```
Barleysack/AERoberta2
[ "pytorch", "roberta", "question-answering", "transformers", "autotrain_compatible" ]
question-answering
{ "architectures": [ "RobertaForQuestionAnswering" ], "model_type": "roberta", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
2
null
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: yt_videos_comments results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # yt_videos_comments This model is a fine-tuned version of [gpt2-large](https://huggingface.co/gpt2-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.0918 - Accuracy: 0.6277 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 4 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - total_eval_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - training_steps: 2000 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 2.1201 | 1.53 | 500 | 2.1152 | 0.6220 | | 2.016 | 3.07 | 1000 | 2.0957 | 0.6254 | | 1.9383 | 4.6 | 1500 | 2.0898 | 0.6271 | | 1.8823 | 6.14 | 2000 | 2.0918 | 0.6277 | ### Framework versions - Transformers 4.29.0.dev0 - Pytorch 2.0.0-rc1 - Datasets 2.11.0 - Tokenizers 0.13.3
Bella4322/Sarah
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- license: cc-by-4.0 tags: - generated_from_keras_callback model-index: - name: Aas324532534/qa_squadshifts results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # Aas324532534/qa_squadshifts This model is a fine-tuned version of [deepset/bert-base-cased-squad2](https://huggingface.co/deepset/bert-base-cased-squad2) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 1.1692 - Epoch: 1 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': True, 'is_legacy_optimizer': False, 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 1224, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False} - training_precision: float32 ### Training results | Train Loss | Epoch | |:----------:|:-----:| | 1.6919 | 0 | | 1.1692 | 1 | ### Framework versions - Transformers 4.28.1 - TensorFlow 2.12.0 - Datasets 2.11.0 - Tokenizers 0.13.3
BenQLange/HF_bot
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
from PIL import Image, ImageDraw, ImageFont # Crea una nueva imagen de fondo blanco de 500x500 píxeles img = Image.new('RGB', (500, 500), color='white') # Abre la imagen del dispositivo de realidad aumentada device_img = Image.open('device.png') # Escala la imagen del dispositivo a 300x300 píxeles device_img = device_img.resize((300, 300)) # Pega la imagen del dispositivo en el centro de la imagen de fondo blanco img.paste(device_img, (100, 100)) # Crea un objeto ImageDraw para escribir texto en la imagen draw = ImageDraw.Draw(img) # Carga una fuente de texto font = ImageFont.truetype('arial.ttf', size=20) # Escribe el texto de información del restaurante y calificaciones en la imagen draw.text((100, 400), "Restaurante: Nombre del restaurante", fill='black', font=font) draw.text((100, 430), "Calificación: 4.5 estrellas", fill='black', font=font) # Guarda la imagen en un archivo llamado "restaurant_AR.png" img.save('restaurant_AR.png')
Berzemu/Coco
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- license: apache-2.0 tags: - setfit - sentence-transformers - text-classification pipeline_tag: text-classification --- # /var/folders/8x/qp375g154zg3h3ktpt_8tyqw0000gn/T/tmpto2q3f4b/JoshELambert/fishconflict_tanzania This is a [SetFit model](https://github.com/huggingface/setfit) that can be used for text classification. The model has been trained using an efficient few-shot learning technique that involves: 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning. 2. Training a classification head with features from the fine-tuned Sentence Transformer. ## Usage To use this model for inference, first install the SetFit library: ```bash python -m pip install setfit ``` You can then run inference as follows: ```python from setfit import SetFitModel # Download from Hub and run inference model = SetFitModel.from_pretrained("/var/folders/8x/qp375g154zg3h3ktpt_8tyqw0000gn/T/tmpto2q3f4b/JoshELambert/fishconflict_tanzania") # Run inference preds = model(["i loved the spiderman movie!", "pineapple on pizza is the worst 🤮"]) ``` ## BibTeX entry and citation info ```bibtex @article{https://doi.org/10.48550/arxiv.2209.11055, doi = {10.48550/ARXIV.2209.11055}, url = {https://arxiv.org/abs/2209.11055}, author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren}, keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Efficient Few-Shot Learning Without Prompts}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ```
Bhuvana/t5-base-spellchecker
[ "pytorch", "t5", "text2text-generation", "transformers", "autotrain_compatible" ]
text2text-generation
{ "architectures": [ "T5ForConditionalGeneration" ], "model_type": "t5", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": true, "length_penalty": 2, "max_length": 200, "min_length": 30, "no_repeat_ngram_size": 3, "num_beams": 4, "prefix": "summarize: " }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": true, "max_length": 300, "num_beams": 4, "prefix": "translate English to German: " }, "translation_en_to_fr": { "early_stopping": true, "max_length": 300, "num_beams": 4, "prefix": "translate English to French: " }, "translation_en_to_ro": { "early_stopping": true, "max_length": 300, "num_beams": 4, "prefix": "translate English to Romanian: " } } }
93
null
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: akshaymathur777/text_summarization results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # akshaymathur777/text_summarization This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 9.7223 - Validation Loss: 4.3413 - Epoch: 0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5.6e-05, 'decay_steps': 9672, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01} - training_precision: mixed_float16 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 9.7223 | 4.3413 | 0 | ### Framework versions - Transformers 4.28.1 - TensorFlow 2.12.0 - Datasets 2.11.0 - Tokenizers 0.13.3
Biasface/DDDC2
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": 1000 }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
10
null
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 255.08 +/- 19.74 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
BigBoy/model
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- license: apache-2.0 tags: - summarization - generated_from_trainer metrics: - rouge model-index: - name: bart-base-finetuned-multi-news results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-base-finetuned-multi-news This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.6353 - Rouge1: 15.1146 - Rouge2: 5.3873 - Rougel: 11.4132 - Rougelsum: 13.2739 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5.6e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 7 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:-------:|:---------:| | 2.9189 | 1.0 | 625 | 2.4645 | 15.2063 | 5.2852 | 11.5864 | 13.4208 | | 2.4697 | 2.0 | 1250 | 2.4706 | 15.3737 | 5.4725 | 11.7465 | 13.5681 | | 2.1831 | 3.0 | 1875 | 2.4789 | 14.8306 | 5.0857 | 11.2416 | 13.1072 | | 1.9598 | 4.0 | 2500 | 2.5299 | 15.1744 | 5.5465 | 11.6445 | 13.4053 | | 1.7777 | 5.0 | 3125 | 2.5799 | 14.9417 | 5.2124 | 11.3553 | 13.1401 | | 1.6454 | 6.0 | 3750 | 2.6028 | 14.9804 | 5.333 | 11.294 | 13.2385 | | 1.554 | 7.0 | 4375 | 2.6353 | 15.1146 | 5.3873 | 11.4132 | 13.2739 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
BigDaddyNe1L/Hhaa
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
2023-04-16T23:29:52Z
--- license: mit tags: - generated_from_trainer model-index: - name: gpt2-small-alpacaGPT4-epochs-5 results: [] widget: - text: "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n### Instruction:\nWhat does AI mean?\n### Response:\n" --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gpt2-small-alpacaGPT4-epochs-5 This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on [alpaca-gpt4](https://github.com/tloen/alpaca-lora.git) dataset. It achieves the following results on the evaluation set: - Loss: 1.6267 ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 16 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 20 - num_epochs: 5 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.8098 | 1.0 | 325 | 1.6921 | | 1.7541 | 2.0 | 650 | 1.6552 | | 1.7189 | 3.0 | 975 | 1.6379 | | 1.6939 | 4.0 | 1300 | 1.6296 | | 1.689 | 5.0 | 1625 | 1.6267 | ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3 ### Project name - gpt2_alpaca.ipynb ### Load Model and perform data tokenization code ```python # Fine Tuning import torch from datasets import load_dataset import transformers from transformers import GPT2LMHeadModel, GPT2Tokenizer, GPT2Config, Trainer, TrainingArguments # %% MODEL_NAME = "gpt2" CUTOFF_LEN = 256 # %% model = GPT2LMHeadModel.from_pretrained(MODEL_NAME) tokenizer = GPT2Tokenizer.from_pretrained(MODEL_NAME, add_special_tokens=True) tokenizer.pad_token_id = tokenizer.eos_token_id dataset = load_dataset("json", data_files="/content/alpaca-lora/alpaca_data_gpt4.json") def generate_prompt(data_point): instruction = data_point["instruction"] input_text = data_point["input"] output_text = data_point["output"] prompt = f"Below is an instruction that describes a task" if input_text: prompt += f", paired with an input that provides further context" prompt += ". Write a response that appropriately completes the request.\n" prompt += f"### Instruction:\n{instruction}\n" if input_text: prompt += f"### Input:\n{input_text}\n" prompt += f"### Response:\n{output_text}\n" return prompt def tokenize(item, add_eos_token=True): # print(generate_prompt(item)) result = tokenizer( generate_prompt(item), truncation=True, max_length=CUTOFF_LEN, padding="max_length", return_tensors=None, ) if ( result["input_ids"][-1] != tokenizer.eos_token_id and len(result["input_ids"]) < CUTOFF_LEN and add_eos_token ): result["input_ids"].append(tokenizer.eos_token_id) result["attention_mask"].append(1) result["labels"] = result["input_ids"].copy() return result train_val = dataset["train"].train_test_split(test_size=0.2, shuffle=True, seed=42) train_data = train_val["train"].shuffle().map(tokenize) val_data = train_val["test"].shuffle().map(tokenize) ``` ### Training code ```python MICRO_BATCH_SIZE = 8 BATCH_SIZE = 128 GRADIENT_ACCUMULATION_STEPS = BATCH_SIZE // MICRO_BATCH_SIZE EPOCHS = 5 LEARNING_RATE = 2e-5 EVAL_BATCH_SIZE = 4 LOGGING_STEPS = 64 # SAVE_STEPS = 10240 # Reduce it to a smaler value like 512 if you want to save checkpoints SAVE_STEPS = 512 # Reduce it to a smaler value like 512 if you want to save checkpoints SAVE_TOTAL_LIMIT = 2 hf_model_id='gpt2-small-alpacaGPT4-epochs-5' hf_user_id='fouadbakour' !python -c "from huggingface_hub.hf_api import HfFolder; HfFolder.save_token('****')" training_args = TrainingArguments( overwrite_output_dir=True, load_best_model_at_end=True, optim='adamw_torch', push_to_hub=True, save_strategy='epoch', evaluation_strategy='epoch', per_device_eval_batch_size=EVAL_BATCH_SIZE, per_device_train_batch_size=MICRO_BATCH_SIZE, gradient_accumulation_steps=GRADIENT_ACCUMULATION_STEPS, warmup_steps=20, weight_decay=0.1, logging_steps=LOGGING_STEPS, save_steps=SAVE_STEPS, save_total_limit=SAVE_TOTAL_LIMIT, num_train_epochs=EPOCHS, learning_rate=LEARNING_RATE, fp16=True, output_dir=f'./{hf_model_id}', ) training_args.set_push_to_hub(model_id=f'{hf_user_id}/{hf_model_id}', strategy='all_checkpoints') # New code trainer = Trainer( model=model, train_dataset=train_data, eval_dataset=val_data, args=training_args, data_collator=transformers.DataCollatorForLanguageModeling(tokenizer, mlm=False), ) model.config.use_cache = False # trainer.train(resume_from_checkpoint=False) trainer.train() trainer.save_model(f'./{hf_model_id}') tokenizer.push_to_hub(hf_model_id) ```
BigSalmon/GPTHeHe
[ "pytorch", "gpt2", "text-generation", "transformers", "has_space" ]
text-generation
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": true, "max_length": 50 }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
#@title <b>TavernAI</b> #@markdown <- Click For Start (≖ ‸ ≖ ✿) Model = "PPO_Pygway-V8p4_Dev-6b" #@param ["PPO_Pygway-V8p4_Dev-6b", "Dolly_Shygmalion-6b"] {allow-input: true} Version = "Official" KoboldAI_Provider = "Localtunnel" #@param ["Localtunnel", "Cloudflare"] use_google_drive = True #@param {type:"boolean"} Provider = KoboldAI_Provider !nvidia-smi import subprocess import time import sys import os import threading import shutil from google.colab import drive if use_google_drive: drive.mount('/content/drive/') if not os.path.exists("/content/drive/MyDrive/TavernAI/"): os.mkdir("/content/drive/MyDrive/TavernAI/") if not os.path.exists("/content/drive/MyDrive/TavernAI/characters/"): os.mkdir("/content/drive/MyDrive/TavernAI/characters/") if not os.path.exists("/content/drive/MyDrive/TavernAI/chats/"): os.mkdir("/content/drive/MyDrive/TavernAI/chats/") else: if not os.path.exists("/content/drive"): os.mkdir("/content/drive") if not os.path.exists("/content/drive/MyDrive/"): os.mkdir("/content/drive/MyDrive/") def copy_characters(use_google_drive=False): if not use_google_drive: return src_folder = "/TavernAIColab/public/characters" dst_folder = "/content/drive/MyDrive/TavernAI/characters" for filename in os.listdir(src_folder): src_file = os.path.join(src_folder, filename) dst_file = os.path.join(dst_folder, filename) if os.path.exists(dst_file): print(f"{dst_file} already exists. Skipping...") continue shutil.copy(src_file, dst_folder) print(f"{src_file} copied to {dst_folder}") Revision = "" if Model == "Dolly_Shygmalion-6b": Model = "TehVenom/Dolly_Shygmalion-6b" path = "" download = "" Version = "United" elif Model == "PPO_Pygway-V8p4_Dev-6b": Model = "TehVenom/PPO_Pygway-V8p4_Dev-6b" Revision = "--revision dev" path = "" Version = "United" download = "" if Provider == "Localtunnel": tunnel = "--localtunnel yes" else: tunnel = "" #Henk's KoboldAI script !wget https://koboldai.org/ckds && chmod +x ckds !./ckds --init only if Provider == "Localtunnel": p = subprocess.Popen(['/content/ckds', '--model', Model, '--localtunnel', 'yes'], stdout=subprocess.PIPE, stderr=subprocess.PIPE) else: p = subprocess.Popen(['/content/ckds', '--model', Model], stdout=subprocess.PIPE, stderr=subprocess.PIPE) #Do not repeat! Tricks performed by a professional! url = '' while True: line = p.stdout.readline().decode().strip() if "KoboldAI has finished loading and is available at the following link: " in line: print(line) url = line.split("KoboldAI has finished loading and is available at the following link: ")[1] print(url) break if "KoboldAI has finished loading and is available at the following link for UI 1: " in line: print(line) url = line.split("KoboldAI has finished loading and is available at the following link for UI 1: ")[1] print(url) break if not line: break print(line) if "INIT" in line and "Transformers" in line: print("Model loading... (It will take 2 - 5 minutes)") #TavernAI %cd / !curl -o- https://raw.githubusercontent.com/nvm-sh/nvm/v0.37.2/install.sh | bash !nvm install 19.1.0 !nvm use 19.1.0 !node -v !git clone https://github.com/TavernAI/TavernAIColab copy_characters(use_google_drive) %cd TavernAIColab !npm install time.sleep(1) %env colab=2 %env colaburl=$url if use_google_drive: %env googledrive=2 !nohup node server.js & time.sleep(3) print('KoboldAI LINK:') print(url) print('') print('###TavernAI LINK###') !lt --port 800
BigSalmon/GPTNeo350MInformalToFormalLincoln2
[ "pytorch", "gpt_neo", "text-generation", "transformers", "has_space" ]
text-generation
{ "architectures": [ "GPTNeoForCausalLM" ], "model_type": "gpt_neo", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
--- tags: - generated_from_trainer model-index: - name: trocr-base-handwritten-OCR-handwriting_recognition_v2 results: [] language: - en metrics: - cer pipeline_tag: image-to-text --- # trocr-base-handwritten-OCR-handwriting_recognition_v2 This model is a fine-tuned version of [microsoft/trocr-base-handwritten](https://huggingface.co/microsoft/trocr-base-handwritten). It achieves the following results on the evaluation set: - Loss: 0.2470 - Cer: 0.0360 ## Model description For more information on how it was created, check out the following link: https://github.com/DunnBC22/Vision_Audio_and_Multimodal_Projects/blob/main/Optical%20Character%20Recognition%20(OCR)/Handwriting%20Recognition/Handwriting%20Recognition_v2/Mini%20Handwriting%20OCR%20Project.ipynb ## Intended uses & limitations This model is intended to demonstrate my ability to solve a complex problem using technology. ## Training and evaluation data Dataset Source: https://www.kaggle.com/datasets/ssarkar445/handwriting-recognitionocr ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Cer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.4292 | 1.0 | 2500 | 0.4332 | 0.0679 | | 0.2521 | 2.0 | 5000 | 0.2767 | 0.0483 | | 0.1049 | 3.0 | 7500 | 0.2470 | 0.0360 | ### Framework versions - Transformers 4.26.0 - Pytorch 1.12.1 - Datasets 2.8.0 - Tokenizers 0.12.1
BigSalmon/GPTNeo350MInformalToFormalLincoln3
[ "pytorch", "gpt_neo", "text-generation", "transformers", "has_space" ]
text-generation
{ "architectures": [ "GPTNeoForCausalLM" ], "model_type": "gpt_neo", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
10
null
--- tags: - conversational --- Mostly just for testing. #House MD DialoGPT Model
BigSalmon/GPTT
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": true, "max_length": 50 }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
9
null
--- library_name: stable-baselines3 tags: - PandaReachDense-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: PandaReachDense-v2 type: PandaReachDense-v2 metrics: - type: mean_reward value: -1.80 +/- 0.65 name: mean_reward verified: false --- # **A2C** Agent playing **PandaReachDense-v2** This is a trained model of a **A2C** agent playing **PandaReachDense-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
BigSalmon/InfillFormalLincoln
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": true, "max_length": 50 }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
8
null
--- datasets: - SJTU-CL/ArguGPT language: - en metrics: - accuracy pipeline_tag: text-classification tags: - AIGC for education --- ## Citation Please cite our work [arXiv:2304.07666](https://arxiv.org/abs/2304.07666) as ``` @misc{liu2023argugpt, title={ArguGPT: evaluating, understanding and identifying argumentative essays generated by GPT models}, author={Yikang Liu and Ziyin Zhang and Wanyang Zhang and Shisen Yue and Xiaojing Zhao and Xinyuan Cheng and Yiwen Zhang and Hai Hu}, year={2023}, eprint={2304.07666}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
BigSalmon/InformalToFormalLincoln23
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": true, "max_length": 50 }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
5
null
--- library_name: ml-agents tags: - SnowballTarget - deep-reinforcement-learning - reinforcement-learning - ML-Agents-SnowballTarget --- # **ppo** Agent playing **SnowballTarget** This is a trained model of a **ppo** agent playing **SnowballTarget** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-SnowballTarget 2. Step 1: Find your model_id: arkadyark/ppo-SnowballTarget 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
BigSalmon/Lincoln4
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
{ "architectures": [ "GPT2LMHeadModel" ], "model_type": "gpt2", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": true, "max_length": 50 }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
11
null
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-Pyramids 2. Step 1: Find your model_id: arkadyark/ppo-Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
BigSalmon/MrLincoln7
[]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- license: creativeml-openrail-m duplicated_from: SakuraFoxKira/AC_H-3 --- # AC_H-3 **AC_H-3是和Alice-Yozakura共同制作,有非常好的背景和光影** ----------------------------------------------------- # AC_H-3预览图 ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE9.png) ``` 1girl,dag ears,fox tail,(loli:1.5),white dress,socks,(ribbon tied ankle|pink),Sit on the bed, Negative prompt: EasyNegative,extra fingers,fewer fingers,watermark,Invisible watermark,username,Signature,jpeg artifacts,Bad feet,Bad hands,Extra legs,Three Legs,Three feet, Steps: 28, Sampler: DPM++ 2M Karras, CFG scale: 7, Seed: 2478275294, Size: 768x384, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires resize: 1280x640, Hires steps: 30, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE8.png) ``` 1girl,dag_ears,fox_tail,(loli:1.5),white dress,White stockings,(white background: 1.8),[(glass jar:1.2),(girl in jar:1.4):(girl)], Negative prompt: Abnormal human structure,Impossible human body structure,amputation,Extra limbs,Extra fingers,Extra legs,Extra ears,3leg,Fewer fingers,Watermark,Lnvisible watermark,Username,Signature,Jpeg artifacts,Bad feet,Bad hands,high-heeled, Navel,Pencil skirt,sofa,(Katyusha maid headdress),cap,Leg lift, EasyNegative, Steps: 28, Sampler: DPM++ 2S a Karras, CFG scale: 4.5, Seed: 3182655935, Size: 480x720, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires upscale: 1.6666666666666667, Hires steps: 20, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE7.png) ``` 1girl,fox_ears,fox_tail,white hair,long hair,white dress,(loli:1.5),White stockings,no shoes,[(Transparent background:1.5)::5],illustration,Blue ambient light,Magic circle,Accumulator,Dynamic surround,Liquid aperture,Bright and colorful,Ionizer Negative prompt: EasyNegative,extra fingers,fewer fingers,watermark,Invisible watermark,username,Signature,jpeg artifacts,Bad feet,Bad hands,Extra legs,Three Legs,Three feet,(NSFW),Multiple people,More than one person,2girl,More than two legs, Steps: 28, Sampler: DPM++ 2S a Karras, CFG scale: 7, Seed: 2429212780, Size: 480x720, Model hash: 74a62f9313, Model: AC_H-3-RD-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires upscale: 1.6, Hires steps: 30, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE11.png) ``` evening, beautiful sunset, summer evening Negative prompt: EasyNegative, Steps: 35, Sampler: UniPC, CFG scale: 7, Seed: 2837482015, Size: 960x540, Model hash: eb4ffeebdf, Model: AC_H-3-RC-half, Denoising strength: 0.5, Clip skip: 2, ENSD: 31337, Hires upscale: 2, Hires steps: 8, Hires upscaler: Latent ``` --- ![sample1](https://s3.amazonaws.com/moonup/production/uploads/63fd94ea3c880680af48b776/SkD5BtPEZueJkeyD07YiZ.png) ``` great grandfather,ripe,(Older:1.25),muscle, Negative prompt: EasyNegative,Impossible human body structure,Extra limbs,Extra fingers,Extra legs,Extra ears,Fewer fingers,Watermark,Lnvisible watermark,Username,Signature,Jpeg artifacts,Bad feet,Bad hands,high-heeled shoes, Steps: 28, Sampler: Euler a, CFG scale: 6, Seed: 4065660254, Size: 480x720, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires upscale: 1.6, Hires steps: 30, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE14.png) ``` room,computer,coffee,flower vase,book,coming,1girl,long hair,(white dress|gorgeous|luxuriant),White pantyhose,(Crural wrist|ribbon:1.2|black),(ankle|ribbon|black)no shoes,blue eyes,(loli:1.5),expressionless, Negative prompt: Abnormal human structure,Impossible human body structure,amputation,Extra limbs,Extra fingers,Extra legs,Extra ears,3leg,Fewer fingers,Watermark,Lnvisible watermark,Username,Signature,Jpeg artifacts,Bad feet,Bad hands,high-heeled, Navel,Pencil skirt,sofa,(Katyusha maid headdress),cap,Leg lift, EasyNegative, Steps: 20, Sampler: Euler a, CFG scale: 4.5, Seed: 124256572, Size: 768x512, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.7, Clip skip: 2, ENSD: 31337, Hires upscale: 1.6, Hires steps: 30, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE12.png) ``` evening, beautiful sunset, summer evening Negative prompt: EasyNegative, Steps: 35, Sampler: UniPC, CFG scale: 7, Seed: 3138791891, Size: 960x540, Model hash: f2ee51f8bc, Model: AC_H-3-half, Denoising strength: 0.5, Clip skip: 2, ENSD: 31337, Hires upscale: 2, Hires steps: 8, Hires upscaler: Latent ``` --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE3.png) --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE4.png) --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE6.png) --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE2.png) --- ![sample1](https://huggingface.co/SakuraFoxKira/AC_H-3/resolve/main/%E9%A2%84%E8%A7%88%E5%9B%BE/%E9%A2%84%E8%A7%88%E5%9B%BE1.png)
BillelBenoudjit/jplu-wikiann
[ "fr", "dataset:wikiann", "model-index" ]
null
{ "architectures": null, "model_type": null, "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
0
null
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: unli results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # unli This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0752 - Accuracy: 0.9681 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.0808 | 1.0 | 1735 | 0.0737 | 0.9681 | | 0.0626 | 2.0 | 3470 | 0.0765 | 0.9681 | | 0.0453 | 3.0 | 5205 | 0.0752 | 0.9681 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.12.1 - Datasets 2.10.1 - Tokenizers 0.12.1
BitanBiswas/mbert-bengali-ner-finetuned-ner
[ "pytorch", "tensorboard", "bert", "token-classification", "transformers", "autotrain_compatible" ]
token-classification
{ "architectures": [ "BertForTokenClassification" ], "model_type": "bert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_size": null, "num_beams": null, "prefix": null }, "text-generation": { "do_sample": null, "max_length": null }, "translation_en_to_de": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_fr": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null }, "translation_en_to_ro": { "early_stopping": null, "max_length": null, "num_beams": null, "prefix": null } } }
4
null
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: copter-reinforce2 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 500.00 +/- 0.00 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction