modelId
stringlengths
4
111
lastModified
stringlengths
24
24
tags
list
pipeline_tag
stringlengths
5
30
author
stringlengths
2
34
config
null
securityStatus
null
id
stringlengths
4
111
likes
int64
0
9.53k
downloads
int64
2
73.6M
library_name
stringlengths
2
84
created
timestamp[us]
card
stringlengths
101
901k
card_len
int64
101
901k
embeddings
list
AliGhiasvand86/digit_recognition2
2023-07-30T11:05:29.000Z
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "huggingpics", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
AliGhiasvand86
null
null
AliGhiasvand86/digit_recognition2
0
2
transformers
2023-07-30T11:05:22
--- tags: - image-classification - pytorch - huggingpics metrics: - accuracy model-index: - name: digit_recognition2 results: - task: name: Image Classification type: image-classification metrics: - name: Accuracy type: accuracy value: 0.19801980257034302 --- # digit_recognition2 Autogenerated by HuggingPics🤗🖼️ Create your own image classifier for **anything** by running [the demo on Google Colab](https://colab.research.google.com/github/nateraw/huggingpics/blob/main/HuggingPics.ipynb). Report any issues with the demo at the [github repo](https://github.com/nateraw/huggingpics). ## Example Images #### number 1 ![number 1](images/number_1.jpg) #### number 2 ![number 2](images/number_2.jpg) #### number 3 ![number 3](images/number_3.jpg) #### number 4 ![number 4](images/number_4.jpg) #### number 5 ![number 5](images/number_5.jpg) #### number 6 ![number 6](images/number_6.jpg) #### number 7 ![number 7](images/number_7.jpg) #### number 8 ![number 8](images/number_8.jpg) #### number 9 ![number 9](images/number_9.jpg)
1,095
[ [ -0.043548583984375, -0.0286865234375, 0.022186279296875, 0.04534912109375, -0.050201416015625, 0.0005230903625488281, 0.0159454345703125, -0.03607177734375, 0.0269012451171875, 0.0064544677734375, -0.0210113525390625, -0.0377197265625, -0.04388427734375, 0.0...
mlabonne/llama-2-13b-guanaco
2023-07-30T15:57:40.000Z
[ "transformers", "pytorch", "llama", "text-generation", "dataset:timdettmers/openassistant-guanaco", "license:apache-2.0", "endpoints_compatible", "text-generation-inference", "region:us" ]
text-generation
mlabonne
null
null
mlabonne/llama-2-13b-guanaco
3
2
transformers
2023-07-30T14:13:37
--- license: apache-2.0 datasets: - timdettmers/openassistant-guanaco pipeline_tag: text-generation --- # Llama-2-13b-guanaco 📝 [Article](https://towardsdatascience.com/fine-tune-your-own-llama-2-model-in-a-colab-notebook-df9823a04a32) | 💻 [Colab](https://colab.research.google.com/drive/1PEQyJO1-f6j0S_XJ8DV50NkpzasXkrzd?usp=sharing) | 📄 [Script](https://gist.github.com/mlabonne/b5718e1b229ce6553564e3f56df72c5c) <center><img src="https://i.imgur.com/C2x7n2a.png" width="300"></center> This is a `llama-2-13b-chat-hf` model fine-tuned using QLoRA (4-bit precision) on the [`mlabonne/guanaco-llama2`](https://huggingface.co/datasets/mlabonne/guanaco-llama2) dataset. ## 🔧 Training It was trained on a Google Colab notebook with a T4 GPU and high RAM. ## 💻 Usage ``` python # pip install transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "mlabonne/llama-2-13b-miniguanaco" prompt = "What is a large language model?" tokenizer = AutoTokenizer.from_pretrained(model) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) sequences = pipeline( f'<s>[INST] {prompt} [/INST]', do_sample=True, top_k=10, num_return_sequences=1, eos_token_id=tokenizer.eos_token_id, max_length=200, ) for seq in sequences: print(f"Result: {seq['generated_text']}") ```
1,413
[ [ -0.0173492431640625, -0.0648193359375, 0.014984130859375, 0.035797119140625, -0.038360595703125, 0.018524169921875, -0.0123748779296875, -0.0274810791015625, 0.02581787109375, 0.007904052734375, -0.03607177734375, -0.0347900390625, -0.04986572265625, 0.00942...
csabad/ppo-LunarLander-v2
2023-07-30T14:42:50.000Z
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
csabad
null
null
csabad/ppo-LunarLander-v2
0
2
stable-baselines3
2023-07-30T14:42:27
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 223.00 +/- 20.20 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
784
[ [ -0.00023484230041503906, -0.02716064453125, 0.017059326171875, 0.023345947265625, -0.00606536865234375, 0.002735137939453125, 0.034454345703125, -0.012115478515625, 0.019866943359375, 0.06500244140625, -0.043212890625, -0.035247802734375, -0.0343017578125, -...
matnord/PPO-LunarLander
2023-07-30T15:44:04.000Z
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
matnord
null
null
matnord/PPO-LunarLander
0
2
stable-baselines3
2023-07-30T15:43:42
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 261.51 +/- 17.11 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
784
[ [ -0.00023484230041503906, -0.02716064453125, 0.017059326171875, 0.023345947265625, -0.00606536865234375, 0.002735137939453125, 0.034454345703125, -0.012115478515625, 0.019866943359375, 0.06500244140625, -0.043212890625, -0.035247802734375, -0.0343017578125, -...
terru3/ppo-LunarLander-v2
2023-07-30T18:40:46.000Z
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
terru3
null
null
terru3/ppo-LunarLander-v2
0
2
stable-baselines3
2023-07-30T18:40:22
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 241.98 +/- 28.84 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
784
[ [ -0.00023484230041503906, -0.02716064453125, 0.017059326171875, 0.023345947265625, -0.00606536865234375, 0.002735137939453125, 0.034454345703125, -0.012115478515625, 0.019866943359375, 0.06500244140625, -0.043212890625, -0.035247802734375, -0.0343017578125, -...
yiminghuang47/prot_bert-finetuned-mhc
2023-07-30T19:46:35.000Z
[ "transformers", "pytorch", "tensorboard", "bert", "fill-mask", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
yiminghuang47
null
null
yiminghuang47/prot_bert-finetuned-mhc
0
2
transformers
2023-07-30T19:33:20
--- base_model: Rostlab/prot_bert tags: - generated_from_trainer model-index: - name: prot_bert-finetuned-MHC results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # prot_bert-finetuned-MHC This model is a fine-tuned version of [Rostlab/prot_bert](https://huggingface.co/Rostlab/prot_bert) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.6702 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 101 | 2.7188 | | No log | 2.0 | 202 | 2.6904 | | No log | 3.0 | 303 | 2.6655 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.1 - Tokenizers 0.13.3
1,380
[ [ -0.033935546875, -0.04931640625, 0.01013946533203125, -0.01145172119140625, -0.032440185546875, -0.034698486328125, -0.002437591552734375, -0.0204010009765625, 0.01052093505859375, 0.0179290771484375, -0.0665283203125, -0.0301666259765625, -0.041900634765625, ...
emre/llama-2-13b-mini
2023-07-30T21:22:30.000Z
[ "transformers", "pytorch", "tensorboard", "llama", "text-generation", "dataset:mlabonne/guanaco-llama2-1k", "license:apache-2.0", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
text-generation
emre
null
null
emre/llama-2-13b-mini
3
2
transformers
2023-07-30T21:05:39
--- license: apache-2.0 datasets: - mlabonne/guanaco-llama2-1k pipeline_tag: text-generation --- # 🦙🧠 emre/llama-2-13b-mini This is a `Llama-2-13b-chat-hf` model fine-tuned using QLoRA (4-bit precision). ## 🔧 Training It was trained Colab Pro+. It is mainly designed for educational purposes, not for inference but can be used exclusively with BBVA Group, GarantiBBVA and its subsidiaries. Parameters: ``` max_seq_length = 2048 use_nested_quant = True bnb_4bit_compute_dtype=bfloat16 lora_r=8 lora_alpha=16 lora_dropout=0.05 per_device_train_batch_size=2 ``` ## 💻 Usage ``` python # pip install transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "emre/llama-2-13b-mini" prompt = "What is a large language model?" tokenizer = AutoTokenizer.from_pretrained(model) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) sequences = pipeline( f'<s>[INST] {prompt} [/INST]', do_sample=True, top_k=10, num_return_sequences=1, eos_token_id=tokenizer.eos_token_id, max_length=200, ) for seq in sequences: print(f"Result: {seq['generated_text']}") ```
1,211
[ [ -0.01513671875, -0.07281494140625, 0.005992889404296875, 0.029266357421875, -0.037353515625, 0.00978851318359375, -0.007282257080078125, -0.005588531494140625, 0.0237579345703125, 0.020721435546875, -0.049407958984375, -0.035736083984375, -0.049774169921875, ...
brunoboat/ppo-PyramidsRND
2023-07-30T22:06:55.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
brunoboat
null
null
brunoboat/ppo-PyramidsRND
0
2
ml-agents
2023-07-30T22:06:53
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: brunoboat/ppo-PyramidsRND 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,337
[ [ -0.041168212890625, -0.03424072265625, 0.0030574798583984375, 0.0140380859375, -0.011749267578125, 0.012847900390625, 0.017333984375, -0.0146942138671875, 0.033447265625, 0.0310516357421875, -0.040618896484375, -0.049774169921875, -0.0296478271484375, -0.015...
nealchandra/llama-2-7b-hf-lora-alpaca-json
2023-07-31T00:09:07.000Z
[ "peft", "dataset:tatsu-lab/alpaca", "region:us" ]
null
nealchandra
null
null
nealchandra/llama-2-7b-hf-lora-alpaca-json
2
2
peft
2023-07-30T22:29:10
--- library_name: peft datasets: - tatsu-lab/alpaca --- ## Overview This model is simple POC for JSON based text completion for instructions following tasks. It was trained on the 20,000 records from the Alpaca dataset with a simple prompt template to expect and return JSON inputs. The prompt template used is roughly like this: ```` ### INPUT: ```json {"instructions": "<INSTRUCTIONS>", "input": "<INPUT>"} ``` ### OUTPUT: ```json {"response": "<OUTPUT">} ``` ```` New-lines are escaped, which means you would want to prompt the model like this: ```` ### INPUT:\n```json\n{"instructions": "Explain what an alpaca is"}\n```\n### OUTPUT:\n ```` As you can see from this example, the `input` arg in the input JSON can be omitted if they are not needed. The training dataset include examples with and without additional inputs and the model was trained to handle both cases. Ultimately, you can expect the model to behave like an Alpaca finetune on top of llama-2-7b, the only difference is that it should reliably expect and respond in json format. ### Training procedure The adapter was trained for 5 epochs using QLoRA with an average training loss of 0.7535. The following hyperparameters were used: - Learning Rate: 2e-4 - Lora R: 16 - Lora Alpha: 16 - Lora Dropout: 0.05 - Target Modules: "q_proj", "k_proj", "v_proj", "o_proj" The following `bitsandbytes` quantization config was used during training: - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: True - bnb_4bit_compute_dtype: bfloat16 ### Framework versions - PEFT 0.4.0.dev0
1,730
[ [ -0.03546142578125, -0.059661865234375, 0.0491943359375, 0.009674072265625, -0.032806396484375, -0.02520751953125, 0.0023746490478515625, -0.0147705078125, 0.005767822265625, 0.044219970703125, -0.04913330078125, -0.060577392578125, -0.046600341796875, 0.0178...
narvind2003/llama-2-7b-miniguanaco
2023-07-31T06:06:46.000Z
[ "transformers", "pytorch", "llama", "text-generation", "dataset:mlabonne/guanaco-llama2", "license:mit", "endpoints_compatible", "text-generation-inference", "region:us" ]
text-generation
narvind2003
null
null
narvind2003/llama-2-7b-miniguanaco
1
2
transformers
2023-07-31T05:36:47
--- license: mit datasets: - mlabonne/guanaco-llama2 --- # Finetuning Llama2 7b on 1000 samples from the Guanaco dataset, using QLoRA - quantization + low rank adaptation. The Llama2 model from Meta has several variants. The smallest one containing 7 billion parameters has been finetuned here.
295
[ [ -0.02301025390625, -0.031646728515625, 0.036956787109375, 0.01348876953125, -0.04083251953125, -0.0202484130859375, 0.01137542724609375, -0.0255889892578125, 0.003978729248046875, 0.04315185546875, -0.0379638671875, -0.041778564453125, -0.0295867919921875, 0...
saurabh2086/ppo-LunarLander-v2
2023-07-31T07:14:52.000Z
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
saurabh2086
null
null
saurabh2086/ppo-LunarLander-v2
0
2
stable-baselines3
2023-07-31T07:14:30
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 242.65 +/- 16.98 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
784
[ [ -0.00023484230041503906, -0.02716064453125, 0.017059326171875, 0.023345947265625, -0.00606536865234375, 0.002735137939453125, 0.034454345703125, -0.012115478515625, 0.019866943359375, 0.06500244140625, -0.043212890625, -0.035247802734375, -0.0343017578125, -...
jkhan447/HateXplain-weighted-agreement-labeled
2023-07-31T09:38:17.000Z
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
text-classification
jkhan447
null
null
jkhan447/HateXplain-weighted-agreement-labeled
0
2
transformers
2023-07-31T08:42:29
--- license: apache-2.0 base_model: bert-base-uncased tags: - generated_from_trainer metrics: - accuracy model-index: - name: HateXplain-weighted-agreement-labeled results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # HateXplain-weighted-agreement-labeled This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.0701 - Accuracy: 0.6760 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.2 - Tokenizers 0.13.3
1,211
[ [ -0.040679931640625, -0.028594970703125, 0.0197296142578125, 0.0185546875, -0.029754638671875, -0.0364990234375, -0.01971435546875, -0.0272674560546875, 0.0240631103515625, 0.0194244384765625, -0.045928955078125, -0.043975830078125, -0.0662841796875, -0.01985...
mrmrob003/ppo-Pyramids
2023-07-31T11:40:32.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
mrmrob003
null
null
mrmrob003/ppo-Pyramids
0
2
ml-agents
2023-07-31T11:40:25
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: mrmrob003/ppo-Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,334
[ [ -0.040740966796875, -0.034027099609375, 0.0018978118896484375, 0.01398468017578125, -0.0106048583984375, 0.01271820068359375, 0.0172119140625, -0.01482391357421875, 0.03326416015625, 0.0302734375, -0.040557861328125, -0.049957275390625, -0.0301513671875, -0....
Kamelowy/Nous-Hermes-Llama2-13b-Kimiko-GPTQ
2023-07-31T11:54:37.000Z
[ "transformers", "llama", "text-generation", "llama-2", "en", "license:other", "text-generation-inference", "region:us" ]
text-generation
Kamelowy
null
null
Kamelowy/Nous-Hermes-Llama2-13b-Kimiko-GPTQ
2
2
transformers
2023-07-31T11:41:01
--- inference: false language: - en library_name: transformers pipeline_tag: text-generation tags: - llama - llama-2 license: other --- # Model Card: Nous-Hermes-Llama-2-13b-Kimiko-GPTQ This is a GPTQ 4-bit quantization of Nous-Hermes-Llama2-13b-Kimiko-Lora-Merged (https://huggingface.co/Doctor-Shotgun/Nous-Hermes-Llama2-13b-Kimiko-Lora-Merged) ## Usage: Intended to be prompted either with the Alpaca instruction format of the NH-L2 base model: ``` ### Instruction: <prompt> ### Response: <leave a newline blank for model to respond> ``` Or the Kimiko lora instruction format: ``` <<SYSTEM>> <character card and system prompt> <<USER>> <prompt> <<AIBOT>> <leave a newline blank for model to respond> ``` ## Bias, Risks, and Limitations The model will show biases similar to those observed in niche roleplaying forums on the Internet, besides those exhibited by the base model. It is not intended for supplying factual information or advice in any form. ## Training Details This model is a merge. Please refer to the link repositories of the base model and lora for details.
1,092
[ [ -0.0231170654296875, -0.050994873046875, 0.03594970703125, 0.009857177734375, -0.061981201171875, -0.0197601318359375, 0.027008056640625, -0.04144287109375, 0.029205322265625, 0.0653076171875, -0.062042236328125, -0.043182373046875, -0.0565185546875, -0.0039...
aroot/eng-deu-sent_delfy
2023-07-31T12:49:46.000Z
[ "transformers", "pytorch", "tensorboard", "mbart", "text2text-generation", "translation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
aroot
null
null
aroot/eng-deu-sent_delfy
0
2
transformers
2023-07-31T12:15:33
--- tags: - translation - generated_from_trainer metrics: - bleu model-index: - name: eng-deu-sent_delfy results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # eng-deu-sent_delfy This model is a fine-tuned version of [facebook/mbart-large-50-many-to-many-mmt](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.6917 - Bleu: 19.9632 - Chrf: 46.2921 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1 - Datasets 2.12.0 - Tokenizers 0.11.0
1,185
[ [ -0.0396728515625, -0.050384521484375, 0.0182342529296875, 0.018951416015625, -0.02618408203125, -0.03802490234375, -0.0160675048828125, -0.012908935546875, 0.015289306640625, 0.0247344970703125, -0.060150146484375, -0.035797119140625, -0.045135498046875, -0....
jariasn/ML-Agents-Pyramids
2023-07-31T17:04:01.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
jariasn
null
null
jariasn/ML-Agents-Pyramids
0
2
ml-agents
2023-07-31T17:01:23
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: jariasn/ML-Agents-Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,338
[ [ -0.041839599609375, -0.03509521484375, 0.001972198486328125, 0.01428985595703125, -0.00943756103515625, 0.0133056640625, 0.0162811279296875, -0.0157318115234375, 0.034423828125, 0.0302734375, -0.042022705078125, -0.05010986328125, -0.029510498046875, -0.0148...
speechbrain/tts-diffwave-ljspeech
2023-08-09T10:40:54.000Z
[ "speechbrain", "Vocoder", "DiffWave", "text-to-speech", "TTS", "speech-synthesis", "en", "dataset:LJSpeech", "arxiv:2009.09761", "arxiv:2106.04624", "license:apache-2.0", "region:us" ]
text-to-speech
speechbrain
null
null
speechbrain/tts-diffwave-ljspeech
1
2
speechbrain
2023-07-31T17:23:42
--- language: "en" inference: false tags: - Vocoder - DiffWave - text-to-speech - TTS - speech-synthesis - speechbrain license: "apache-2.0" datasets: - LJSpeech --- # Vocoder with DiffWave trained on LJSpeech This repository provides all the necessary tools for using a [DiffWave](https://arxiv.org/pdf/2009.09761.pdf) vocoder trained with [LJSpeech](https://keithito.com/LJ-Speech-Dataset/). The pre-trained model takes as input a spectrogram and generates a waveform as output. Typically, a vocoder is used after a TTS model that converts an input text into a spectrogram. The sampling frequency is 22050 Hz. ## Install SpeechBrain ```bash pip install speechbrain ``` Please notice that we encourage you to read our tutorials and learn more about [SpeechBrain](https://speechbrain.github.io). ### Using the Vocoder as reconstructor ```python import torch import torchaudio import speechbrain as sb from speechbrain.pretrained import DiffWaveVocoder from speechbrain.lobes.models.HifiGAN import mel_spectogram diffwave = DiffWaveVocoder.from_hparams(source="speechbrain/tts-diffwave-ljspeech", savedir="tmpdir") audio = sb.dataio.dataio.read_audio("speechbrain/tts-diffwave-ljspeech/LJ050-0075.wav") audio = torch.FloatTensor(audio) audio = audio.unsqueeze(0) mel = mel_spectogram( sample_rate=22050, hop_length=256, win_length=1024, n_fft=1024, n_mels=80, f_min=0, f_max=8000, power=1.0, normalized=False, norm="slaney", mel_scale="slaney", compression=True, audio=audio, ) # Running Vocoder (spectrogram-to-waveform), a fast sampling can be realized by passing user-defined variance schedules. According to the paper, high-quality audios can be generated with only 6 steps (instead of a total of 50). waveforms = diffwave.decode_batch( mel, hop_len=256, # upsample factor, should be the same as "hop_len" during the extraction of mel-spectrogram fast_sampling=True, # fast sampling is highly recommanded fast_sampling_noise_schedule=[0.0001, 0.001, 0.01, 0.05, 0.2, 0.5], # customized noise schedule ) torchaudio.save('reconstructed.wav', waveforms.squeeze(1), 22050) ``` ### Using the Vocoder with TTS ```python import torchaudio from speechbrain.pretrained import FastSpeech2 from speechbrain.pretrained import DiffWaveVocoder # Intialize TTS (FastSpeech2) and Vocoder (DiffWave) fastspeech2 = FastSpeech2.from_hparams(source="speechbrain/tts-fastspeech2-ljspeech", savedir="tmpdir_tts") diffwave = DiffWaveVocoder.from_hparams(source="speechbrain/tts-diffwave-ljspeech", savedir="tmpdir_vocoder") input_text = "This is a test run with FastSpeech and DiffWave." # Running the TTS mel_output, durations, pitch, energy = fastspeech2.encode_text( [input_text], pace=1.0, # scale up/down the speed pitch_rate=1.0, # scale up/down the pitch energy_rate=1.0, # scale up/down the energy ) # Running Vocoder (spectrogram-to-waveform), a fast sampling can be realized by passing user-defined variance schedules. According to the paper, high-quality audios can be generated with only 6 steps (instead of a total of 50). waveforms = diffwave.decode_batch( mel_output, hop_len=256, # upsample factor, should be the same as "hop_len" during the extraction of mel-spectrogram fast_sampling=True, # fast sampling is highly recommanded fast_sampling_noise_schedule=[0.0001, 0.001, 0.01, 0.05, 0.2, 0.5], # customized noise schedule ) # Save the waverform torchaudio.save('example_TTS.wav',waveforms.squeeze(1), 22050) ``` ### Inference on GPU To perform inference on the GPU, add `run_opts={"device":"cuda"}` when calling the `from_hparams` method. ### Training The model was trained with SpeechBrain. To train it from scratch follow these steps: 1. Clone SpeechBrain: ```bash git clone https://github.com/speechbrain/speechbrain/ ``` 2. Install it: ```bash cd speechbrain pip install -r requirements.txt pip install -e . ``` 3. Run Training: ```bash cd recipes/LJSpeech/TTS/vocoder/diffwave/ python train.py hparams/train.yaml --data_folder /path/to/LJspeech ``` You can find our training results (models, logs, etc) [here](https://www.dropbox.com/sh/tbhpn1xirtaix68/AACvYaVDiUGAKURf2o-fvgMoa?dl=0). ### Limitations The SpeechBrain team does not provide any warranty on the performance achieved by this model when used on other datasets. # **About SpeechBrain** - Website: https://speechbrain.github.io/ - Code: https://github.com/speechbrain/speechbrain/ - HuggingFace: https://huggingface.co/speechbrain/ # **Citing SpeechBrain** Please, cite SpeechBrain if you use it for your research or business. ```bibtex @misc{speechbrain, title={{SpeechBrain}: A General-Purpose Speech Toolkit}, author={Mirco Ravanelli and Titouan Parcollet and Peter Plantinga and Aku Rouhe and Samuele Cornell and Loren Lugosch and Cem Subakan and Nauman Dawalatabad and Abdelwahab Heba and Jianyuan Zhong and Ju-Chieh Chou and Sung-Lin Yeh and Szu-Wei Fu and Chien-Feng Liao and Elena Rastorgueva and François Grondin and William Aris and Hwidong Na and Yan Gao and Renato De Mori and Yoshua Bengio}, year={2021}, eprint={2106.04624}, archivePrefix={arXiv}, primaryClass={eess.AS}, note={arXiv:2106.04624} } ```
5,234
[ [ -0.040496826171875, -0.04132080078125, 0.00807952880859375, 0.0281982421875, -0.020355224609375, 0.00753021240234375, -0.01168060302734375, -0.035308837890625, 0.0283203125, 0.009765625, -0.0400390625, -0.03509521484375, -0.0443115234375, 0.01074981689453125...
ckandemir/speecht5_finetuned_voxpopuli_fr
2023-08-03T21:20:21.000Z
[ "transformers", "pytorch", "tensorboard", "speecht5", "text-to-audio", "generated_from_trainer", "text-to-speech", "dataset:voxpopuli", "license:mit", "endpoints_compatible", "has_space", "region:us" ]
text-to-speech
ckandemir
null
null
ckandemir/speecht5_finetuned_voxpopuli_fr
0
2
transformers
2023-07-31T21:24:05
--- license: mit base_model: microsoft/speecht5_tts tags: - generated_from_trainer datasets: - voxpopuli model-index: - name: speecht5_finetuned_voxpopuli_fr results: [] pipeline_tag: text-to-speech --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # speecht5_finetuned_voxpopuli_fr This model is a fine-tuned version of [microsoft/speecht5_tts](https://huggingface.co/microsoft/speecht5_tts) on the voxpopuli dataset. It achieves the following results on the evaluation set: - Loss: 0.4697 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 1000 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.765 | 0.23 | 50 | 0.6575 | | 0.687 | 0.47 | 100 | 0.6106 | | 0.6423 | 0.7 | 150 | 0.5548 | | 0.5792 | 0.94 | 200 | 0.5300 | | 0.5658 | 1.17 | 250 | 0.5186 | | 0.5558 | 1.41 | 300 | 0.5078 | | 0.5484 | 1.64 | 350 | 0.5029 | | 0.5427 | 1.87 | 400 | 0.4981 | | 0.5349 | 2.11 | 450 | 0.4921 | | 0.524 | 2.34 | 500 | 0.4906 | | 0.5243 | 2.58 | 550 | 0.4857 | | 0.5238 | 2.81 | 600 | 0.4835 | | 0.5104 | 3.05 | 650 | 0.4796 | | 0.516 | 3.28 | 700 | 0.4769 | | 0.5084 | 3.51 | 750 | 0.4763 | | 0.5029 | 3.75 | 800 | 0.4749 | | 0.5015 | 3.98 | 850 | 0.4725 | | 0.5045 | 4.22 | 900 | 0.4716 | | 0.503 | 4.45 | 950 | 0.4706 | | 0.5013 | 4.69 | 1000 | 0.4697 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.2 - Tokenizers 0.13.3
2,444
[ [ -0.038360595703125, -0.04083251953125, -0.0019311904907226562, 0.00592803955078125, -0.009002685546875, -0.01378631591796875, -0.003917694091796875, -0.007778167724609375, 0.009490966796875, 0.018829345703125, -0.050323486328125, -0.052154541015625, -0.043426513...
KM4STfulltext/HsscBERT_e5
2023-08-01T03:36:55.000Z
[ "transformers", "pytorch", "bert", "fill-mask", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
KM4STfulltext
null
null
KM4STfulltext/HsscBERT_e5
0
2
transformers
2023-08-01T03:22:49
--- tags: - generated_from_trainer metrics: - accuracy model-index: - name: HsscBERT_abs_and_full results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # HsscBERT_abs_and_full This model is a fine-tuned version of [/home/hscrc/pretrained_models/bert-base-chinese](https://huggingface.co//home/hscrc/pretrained_models/bert-base-chinese) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6037 - Accuracy: 0.8504 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 16 - seed: 42 - distributed_type: multi-GPU - num_devices: 9 - total_train_batch_size: 288 - total_eval_batch_size: 144 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5.0 ### Training results | Training Loss | Epoch | Step | Accuracy | Validation Loss | |:-------------:|:-----:|:-----:|:--------:|:---------------:| | 0.8163 | 0.19 | 5000 | 0.8326 | 0.6971 | | 0.7942 | 0.38 | 10000 | 0.8364 | 0.6761 | | 0.7817 | 0.57 | 15000 | 0.8384 | 0.6651 | | 0.7751 | 0.75 | 20000 | 0.8402 | 0.6563 | | 0.7654 | 0.94 | 25000 | 0.8415 | 0.6490 | | 0.7546 | 1.13 | 30000 | 0.8427 | 0.6441 | | 0.7527 | 1.32 | 35000 | 0.8434 | 0.6398 | | 0.7484 | 1.51 | 40000 | 0.8444 | 0.6345 | | 0.7443 | 1.7 | 45000 | 0.8450 | 0.6318 | | 0.74 | 1.88 | 50000 | 0.8456 | 0.6292 | | 0.738 | 2.07 | 55000 | 0.8460 | 0.6268 | | 0.734 | 2.26 | 60000 | 0.8464 | 0.6246 | | 0.7335 | 2.45 | 65000 | 0.8467 | 0.6229 | | 0.7299 | 2.64 | 70000 | 0.8470 | 0.6212 | | 0.7291 | 2.83 | 75000 | 0.8473 | 0.6201 | ### Framework versions - Transformers 4.26.1 - Pytorch 1.10.0+cu113 - Datasets 2.9.0 - Tokenizers 0.13.2
2,367
[ [ -0.03863525390625, -0.037506103515625, 0.0098419189453125, 0.0091552734375, -0.01284027099609375, -0.02001953125, -0.01175689697265625, -0.0129852294921875, 0.023834228515625, 0.0249176025390625, -0.053985595703125, -0.05218505859375, -0.044219970703125, -0....
GrazittiInteractive/llama-2-13b
2023-08-10T10:49:14.000Z
[ "transformers", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-2", "en", "dataset:meta-llama/Llama-2-13b", "license:other", "text-generation-inference", "region:us" ]
text-generation
GrazittiInteractive
null
null
GrazittiInteractive/llama-2-13b
1
2
transformers
2023-08-01T07:39:26
--- inference: false language: - en pipeline_tag: text-generation datasets: - meta-llama/Llama-2-13b tags: - facebook - meta - pytorch - llama - llama-2 model_type: llama license: other --- # Meta's Llama 2 13B GGML A 4 Bit GGML format quantized version of base mode Llama-2-13b taken from https://huggingface.co/meta-llama/Llama-2-13b, reduced from 24.2 GB to 7.37GB These files are GGML format model files for [Meta's Llama 2 13B](https://huggingface.co/meta-llama/Llama-2-13b). GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as: * [KoboldCpp](https://github.com/LostRuins/koboldcpp), a powerful GGML web UI with full GPU acceleration out of the box. Especially good for story telling. * [LoLLMS Web UI](https://github.com/ParisNeo/lollms-webui), a great web UI with GPU acceleration via the c_transformers backend. * [LM Studio](https://lmstudio.ai/), a fully featured local GUI. Supports full GPU accel on macOS. Also supports Windows, without GPU accel. * [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most popular web UI. Requires extra steps to enable GPU accel via llama.cpp backend. * [ctransformers](https://github.com/marella/ctransformers), a Python library with LangChain support and OpenAI-compatible AI server. * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), a Python library with OpenAI-compatible API server. ## Provided files | Name | Quant method | Bits | Size | Max RAM required | Use case | | ---- | ---- | ---- | ---- | ---- | ----- | | ggml-model-q4_0.bin | q4_0 | 4 | 6.85 GB| 9.118 GB | Original quant method, 4-bit. | **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead. ## How to run in `llama.cpp` we used langchain with llama-cpp-python, adjust for your tastes and needs: How to use this Llama-2-13b model from Python code and langchain First, make sure you have langchain and llama-cpp installed: ``` pip install llama-cpp-python ``` ``` # Make sure the model path is correct for your system! llm = LlamaCpp( model_path="/Users/rlm/Desktop/Code/llama/llama-2-13b-ggml/ggml-model-q4_0.bin", input={"temperature": 0.75, "max_length": 2000, "top_p": 1}, callback_manager=callback_manager, verbose=True, ) ``` # Original model card: Meta's Llama 2 13B --- extra_gated_heading: Access Llama 2 on Hugging Face extra_gated_description: >- This is a form to enable access to Llama 2 on Hugging Face after you have been granted access from Meta. Please visit the [Meta website](https://ai.meta.com/resources/models-and-libraries/llama-downloads) and accept our license terms and acceptable use policy before submitting this form. Requests will be processed in 1-2 days. extra_gated_button_content: Submit extra_gated_fields: I agree to share my name, email address and username with Meta and confirm that I have already been granted download access on the Meta website: checkbox language: - en pipeline_tag: text-generation inference: false tags: - facebook - meta - pytorch - llama - llama-2 --- # **Llama 2** Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B pretrained model, converted for the Hugging Face Transformers format. Links to other models can be found in the index at the bottom. ## Model Details *Note: Use of this model is governed by the Meta license. In order to download the model weights and tokenizer, please visit the [website](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) and accept our License before requesting access here.* Meta developed and publicly released the Llama 2 family of large language models (LLMs), a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. Our fine-tuned LLMs, called Llama-2-Chat, are optimized for dialogue use cases. Llama-2-Chat models outperform open-source chat models on most benchmarks we tested, and in our human evaluations for helpfulness and safety, are on par with some popular closed-source models like ChatGPT and PaLM. **Model Developers** Meta **Variations** Llama 2 comes in a range of parameter sizes — 7B, 13B, and 70B — as well as pretrained and fine-tuned variations. **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Llama 2 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align to human preferences for helpfulness and safety. ||Training Data|Params|Content Length|GQA|Tokens|LR| |---|---|---|---|---|---|---| |Llama 2|*A new mix of publicly available online data*|7B|4k|&#10007;|2.0T|3.0 x 10<sup>-4</sup>| |Llama 2|*A new mix of publicly available online data*|13B|4k|&#10007;|2.0T|3.0 x 10<sup>-4</sup>| |Llama 2|*A new mix of publicly available online data*|70B|4k|&#10004;|2.0T|1.5 x 10<sup>-4</sup>| *Llama 2 family of models.* Token counts refer to pretraining data only. All models are trained with a global batch-size of 4M tokens. Bigger models - 70B -- use Grouped-Query Attention (GQA) for improved inference scalability. **Model Dates** Llama 2 was trained between January 2023 and July 2023. **Status** This is a static model trained on an offline dataset. Future versions of the tuned models will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) ## Intended Use **Intended Use Cases** Llama 2 is intended for commercial and research use in English. Tuned models are intended for assistant-like chat, whereas pretrained models can be adapted for a variety of natural language generation tasks. To get the expected features and performance for the chat versions, a specific formatting needs to be followed, including the `INST` and `<<SYS>>` tags, `BOS` and `EOS` tokens, and the whitespaces and breaklines in between (we recommend calling `strip()` on inputs to avoid double-spaces). See our reference code in github for details: [`chat_completion`](https://github.com/facebookresearch/llama/blob/main/llama/generation.py#L212). **Out-of-scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws).Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Llama 2. ## Hardware and Software **Training Factors** We used custom training libraries, Meta's Research Super Cluster, and production clusters for pretraining. Fine-tuning, annotation, and evaluation were also performed on third-party cloud compute. **Carbon Footprint** Pretraining utilized a cumulative 3.3M GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 539 tCO2eq, 100% of which were offset by Meta’s sustainability program. ||Time (GPU hours)|Power Consumption (W)|Carbon Emitted(tCO<sub>2</sub>eq)| |---|---|---|---| |Llama 2 7B|184320|400|31.22| |Llama 2 13B|368640|400|62.44| |Llama 2 70B|1720320|400|291.42| |Total|3311616||539.00| **CO<sub>2</sub> emissions during pretraining.** Time: total GPU time required for training each model. Power Consumption: peak power capacity per GPU device for the GPUs used adjusted for power usage efficiency. 100% of the emissions are directly offset by Meta's sustainability program, and because we are openly releasing these models, the pretraining costs do not need to be incurred by others. ## Training Data **Overview** Llama 2 was pretrained on 2 trillion tokens of data from publicly available sources. The fine-tuning data includes publicly available instruction datasets, as well as over one million new human-annotated examples. Neither the pretraining nor the fine-tuning datasets include Meta user data. **Data Freshness** The pretraining data has a cutoff of September 2022, but some tuning data is more recent, up to July 2023. ## Evaluation Results In this section, we report the results for the Llama 1 and Llama 2 models on standard academic benchmarks.For all the evaluations, we use our internal evaluations library. |Model|Size|Code|Commonsense Reasoning|World Knowledge|Reading Comprehension|Math|MMLU|BBH|AGI Eval| |---|---|---|---|---|---|---|---|---|---| |Llama 1|7B|14.1|60.8|46.2|58.5|6.95|35.1|30.3|23.9| |Llama 1|13B|18.9|66.1|52.6|62.3|10.9|46.9|37.0|33.9| |Llama 1|33B|26.0|70.0|58.4|67.6|21.4|57.8|39.8|41.7| |Llama 1|65B|30.7|70.7|60.5|68.6|30.8|63.4|43.5|47.6| |Llama 2|7B|16.8|63.9|48.9|61.3|14.6|45.3|32.6|29.3| |Llama 2|13B|24.5|66.9|55.4|65.8|28.7|54.8|39.4|39.1| |Llama 2|70B|**37.5**|**71.9**|**63.6**|**69.4**|**35.2**|**68.9**|**51.2**|**54.2**| **Overall performance on grouped academic benchmarks.** *Code:* We report the average pass@1 scores of our models on HumanEval and MBPP. *Commonsense Reasoning:* We report the average of PIQA, SIQA, HellaSwag, WinoGrande, ARC easy and challenge, OpenBookQA, and CommonsenseQA. We report 7-shot results for CommonSenseQA and 0-shot results for all other benchmarks. *World Knowledge:* We evaluate the 5-shot performance on NaturalQuestions and TriviaQA and report the average. *Reading Comprehension:* For reading comprehension, we report the 0-shot average on SQuAD, QuAC, and BoolQ. *MATH:* We report the average of the GSM8K (8 shot) and MATH (4 shot) benchmarks at top 1. |||TruthfulQA|Toxigen| |---|---|---|---| |Llama 1|7B|27.42|23.00| |Llama 1|13B|41.74|23.08| |Llama 1|33B|44.19|22.57| |Llama 1|65B|48.71|21.77| |Llama 2|7B|33.29|**21.25**| |Llama 2|13B|41.86|26.10| |Llama 2|70B|**50.18**|24.60| **Evaluation of pretrained LLMs on automatic safety benchmarks.** For TruthfulQA, we present the percentage of generations that are both truthful and informative (the higher the better). For ToxiGen, we present the percentage of toxic generations (the smaller the better). |||TruthfulQA|Toxigen| |---|---|---|---| |Llama-2-Chat|7B|57.04|**0.00**| |Llama-2-Chat|13B|62.18|**0.00**| |Llama-2-Chat|70B|**64.14**|0.01| **Evaluation of fine-tuned LLMs on different safety datasets.** Same metric definitions as above. ## Ethical Considerations and Limitations Llama 2 is a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Llama 2’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of Llama 2, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available at [https://ai.meta.com/llama/responsible-use-guide/](https://ai.meta.com/llama/responsible-use-guide) ## Reporting Issues Please report any software “bug,” or other problems with the models through one of the following means: - Reporting issues with the model: [github.com/facebookresearch/llama](http://github.com/facebookresearch/llama) - Reporting problematic content generated by the model: [developers.facebook.com/llama_output_feedback](http://developers.facebook.com/llama_output_feedback) - Reporting bugs and security concerns: [facebook.com/whitehat/info](http://facebook.com/whitehat/info) ## Llama Model Index |Model|Llama2|Llama2-hf|Llama2-chat|Llama2-chat-hf| |---|---|---|---|---| |7B| [Link](https://huggingface.co/llamaste/Llama-2-7b) | [Link](https://huggingface.co/llamaste/Llama-2-7b-hf) | [Link](https://huggingface.co/llamaste/Llama-2-7b-chat) | [Link](https://huggingface.co/llamaste/Llama-2-7b-chat-hf)| |13B| [Link](https://huggingface.co/llamaste/Llama-2-13b) | [Link](https://huggingface.co/llamaste/Llama-2-13b-hf) | [Link](https://huggingface.co/llamaste/Llama-2-13b-chat) | [Link](https://huggingface.co/llamaste/Llama-2-13b-hf)| |70B| [Link](https://huggingface.co/llamaste/Llama-2-70b) | [Link](https://huggingface.co/llamaste/Llama-2-70b-hf) | [Link](https://huggingface.co/llamaste/Llama-2-70b-chat) | [Link](https://huggingface.co/llamaste/Llama-2-70b-hf)|
12,562
[ [ -0.019012451171875, -0.06597900390625, 0.0301513671875, 0.0290985107421875, -0.039154052734375, 0.006744384765625, -0.00914764404296875, -0.05413818359375, 0.01360321044921875, 0.0330810546875, -0.049163818359375, -0.03106689453125, -0.04718017578125, 0.0006...
ssbuild/chatglm2-6b-32k-int4
2023-08-01T18:21:00.000Z
[ "transformers", "pytorch", "chatglm", "custom_code", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
ssbuild
null
null
ssbuild/chatglm2-6b-32k-int4
2
2
transformers
2023-08-01T08:18:16
--- license: apache-2.0 --- --- language: - zh - en tags: - glm - chatglm - thudm --- # ChatGLM2-6B ## 软件依赖 ```shell pip install protobuf transformers==4.30.2 cpm_kernels torch>=2.0 gradio mdtex2html sentencepiece accelerate ``` ## 代码调用 可以通过如下代码调用 ChatGLM-6B 模型来生成对话: ```ipython >>> from transformers import AutoTokenizer, AutoModel >>> tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True) >>> model = AutoModel.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True).half().cuda() >>> model = model.eval() >>> response, history = model.chat(tokenizer, "你好", history=[]) >>> print(response) 你好👋!我是人工智能助手 ChatGLM-6B,很高兴见到你,欢迎问我任何问题。 >>> response, history = model.chat(tokenizer, "晚上睡不着应该怎么办", history=history) >>> print(response) 晚上睡不着可能会让你感到焦虑或不舒服,但以下是一些可以帮助你入睡的方法: 1. 制定规律的睡眠时间表:保持规律的睡眠时间表可以帮助你建立健康的睡眠习惯,使你更容易入睡。尽量在每天的相同时间上床,并在同一时间起床。 2. 创造一个舒适的睡眠环境:确保睡眠环境舒适,安静,黑暗且温度适宜。可以使用舒适的床上用品,并保持房间通风。 3. 放松身心:在睡前做些放松的活动,例如泡个热水澡,听些轻柔的音乐,阅读一些有趣的书籍等,有助于缓解紧张和焦虑,使你更容易入睡。 4. 避免饮用含有咖啡因的饮料:咖啡因是一种刺激性物质,会影响你的睡眠质量。尽量避免在睡前饮用含有咖啡因的饮料,例如咖啡,茶和可乐。 5. 避免在床上做与睡眠无关的事情:在床上做些与睡眠无关的事情,例如看电影,玩游戏或工作等,可能会干扰你的睡眠。 6. 尝试呼吸技巧:深呼吸是一种放松技巧,可以帮助你缓解紧张和焦虑,使你更容易入睡。试着慢慢吸气,保持几秒钟,然后缓慢呼气。 如果这些方法无法帮助你入睡,你可以考虑咨询医生或睡眠专家,寻求进一步的建议。 ``` 关于更多的使用说明,包括如何运行命令行和网页版本的 DEMO,以及使用模型量化以节省显存,请参考我们的 [Github Repo](https://github.com/THUDM/ChatGLM2-6B)。 For more instructions, including how to run CLI and web demos, and model quantization, please refer to our [Github Repo](https://github.com/THUDM/ChatGLM2-6B). ## Change Log * v1.0 ## 协议 本仓库的代码依照 [Apache-2.0](LICENSE) 协议开源,ChatGLM2-6B 模型的权重的使用则需要遵循 [Model License](MODEL_LICENSE)。
1,626
[ [ -0.0294036865234375, -0.06427001953125, 0.005298614501953125, 0.0284271240234375, -0.0362548828125, -0.01058197021484375, -0.0032291412353515625, -0.017364501953125, 0.004123687744140625, 0.0191497802734375, -0.031707763671875, -0.040008544921875, -0.05209350585...
Pavan27/NER_Telugu_01
2023-10-03T08:59:41.000Z
[ "transformers", "pytorch", "xlm-roberta", "token-classification", "telugu", "NER", "TeluguNER", "te", "en", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
Pavan27
null
null
Pavan27/NER_Telugu_01
1
2
transformers
2023-08-01T09:57:19
--- language: - te - en tags: - telugu - NER - TeluguNER --- ## Direct Use The model is a language model. The model can be used for token classification, a natural language understanding task in which a label is assigned to some tokens in a text. ## Downstream Use Potential downstream use cases include Named Entity Recognition (NER) and Part-of-Speech (PoS) tagging. To learn more about token classification and other potential downstream use cases, see the Hugging Face [token classification docs](https://huggingface.co/tasks/token-classification). ## Out-of-Scope Use The model should not be used to intentionally create hostile or alienating environments for people. # Bias, Risks, and Limitations **CONTENT WARNING: Readers should be made aware that language generated by this model may be disturbing or offensive to some and may propagate historical and current stereotypes.** ```python >>> from transformers import pipeline >>> tokenizer = AutoTokenizer.from_pretrained("Pavan27/NER_Telugu_01") >>> model = AutoModelForTokenClassification.from_pretrained("Pavan27/NER_Telugu_01") >>> classifier = pipeline("ner", model=model, tokenizer=tokenizer, grouped_entities = True) >>> classifier("వెస్టిండీస్‌పై పోర్ట్ ఆఫ్ స్పెయిన్‌ వేదిక జరుగుతున్న రెండో టెస్టు తొలి ఇన్నింగ్స్‌లో విరాట్ కోహ్లీ 121 పరుగులతో విదేశాల్లో సెంచరీ కరువును తీర్చుకున్నాడు.") [{'entity_group': 'LOC', 'score': 0.9999062, 'word': 'వెస్టిండీస్', 'start': 0, 'end': 11}, {'entity_group': 'LOC', 'score': 0.9998613, 'word': 'పోర్ట్ ఆఫ్ స్పెయిన్', 'start': 15, 'end': 34}, {'entity_group': 'PER', 'score': 0.99996054, 'word': 'విరాట్ కోహ్లీ', 'start': 85, 'end': 98}] ``` ## Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.
1,810
[ [ -0.0218963623046875, -0.045318603515625, -0.005985260009765625, 0.0204620361328125, -0.0262451171875, -0.01538848876953125, 0.003681182861328125, -0.0275421142578125, -0.0023975372314453125, 0.0438232421875, -0.0311431884765625, -0.04345703125, -0.04876708984375...
greg-szopinski/ppo-Pyramids-1
2023-08-01T10:22:29.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
greg-szopinski
null
null
greg-szopinski/ppo-Pyramids-1
0
2
ml-agents
2023-08-01T10:22:26
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: greg-szopinski/ppo-Pyramids-1 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,341
[ [ -0.03997802734375, -0.034149169921875, 0.002819061279296875, 0.01387786865234375, -0.01128387451171875, 0.01241302490234375, 0.0168609619140625, -0.0137939453125, 0.034423828125, 0.0293426513671875, -0.040985107421875, -0.05029296875, -0.029266357421875, -0....
embaas/sentence-transformers-gte-small
2023-08-01T11:04:18.000Z
[ "sentence-transformers", "pytorch", "bert", "feature-extraction", "sentence-similarity", "endpoints_compatible", "region:us" ]
sentence-similarity
embaas
null
null
embaas/sentence-transformers-gte-small
0
2
sentence-transformers
2023-08-01T11:04:14
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity --- # embaas/sentence-transformers-gte-small This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search. <!--- Describe your model here --> ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('embaas/sentence-transformers-gte-small') embeddings = model.encode(sentences) print(embeddings) ``` ## Evaluation Results <!--- Describe how your model was evaluated --> For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=embaas/sentence-transformers-gte-small) ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False}) (2): Normalize() ) ``` ## Citing & Authors <!--- Describe where people can find more information -->
1,672
[ [ -0.0293731689453125, -0.05694580078125, 0.03411865234375, 0.0291595458984375, -0.0195465087890625, -0.034027099609375, -0.01751708984375, 0.012451171875, 0.0155029296875, 0.03582763671875, -0.04425048828125, -0.041961669921875, -0.040313720703125, 0.00514984...
chaanks/asr-whisper-tiny-sb
2023-08-02T09:04:49.000Z
[ "speechbrain", "whisper", "pytorch", "Transformer", "hf-asr-leaderboard", "automatic-speech-recognition", "en", "license:apache-2.0", "model-index", "region:us" ]
automatic-speech-recognition
chaanks
null
null
chaanks/asr-whisper-tiny-sb
0
2
speechbrain
2023-08-01T11:53:52
--- language: - en thumbnail: null pipeline_tag: automatic-speech-recognition tags: - whisper - pytorch - speechbrain - Transformer - hf-asr-leaderboard license: apache-2.0 model-index: - name: asr-whisper-tiny-sb results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: LibriSpeech (clean) type: librispeech_asr config: clean split: test args: language: en metrics: - name: Test WER type: wer value: 7.54 - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: LibriSpeech (other) type: librispeech_asr config: other split: test args: language: en metrics: - name: Test WER type: wer value: 17.15 --- <iframe src="https://ghbtns.com/github-btn.html?user=speechbrain&repo=speechbrain&type=star&count=true&size=large&v=2" frameborder="0" scrolling="0" width="170" height="30" title="GitHub"></iframe> <br/><br/> # Whisper tiny SpeechBrain This repository provides all the necessary tools to perform automatic speech recognition from an end-to-end whisper model within SpeechBrain. Please note that this is not an official Speechbrain repository. ## Install SpeechBrain First of all, please install tranformers and SpeechBrain with the following command: ``` pip install speechbrain transformers==4.28.0 ``` Please notice that we encourage you to read our tutorials and learn more about [SpeechBrain](https://speechbrain.github.io). ### Transcribing your own audio files ```python from speechbrain.pretrained import WhisperASR asr_model = WhisperASR.from_hparams(source="chaanks/asr-whisper-tiny-sb", savedir="pretrained_models/asr-whisper-tiny-sb") asr_model.transcribe_file("chaanks/asr-whisper-tiny-sb/example.wav") ``` ### Inference on GPU To perform inference on the GPU, add `run_opts={"device":"cuda"}` when calling the `from_hparams` method. ### Limitations The SpeechBrain team does not provide any warranty on the performance achieved by this model when used on other datasets. #### Referencing SpeechBrain ``` @misc{SB2021, author = {Ravanelli, Mirco and Parcollet, Titouan and Rouhe, Aku and Plantinga, Peter and Rastorgueva, Elena and Lugosch, Loren and Dawalatabad, Nauman and Ju-Chieh, Chou and Heba, Abdel and Grondin, Francois and Aris, William and Liao, Chien-Feng and Cornell, Samuele and Yeh, Sung-Lin and Na, Hwidong and Gao, Yan and Fu, Szu-Wei and Subakan, Cem and De Mori, Renato and Bengio, Yoshua }, title = {SpeechBrain}, year = {2021}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\\\\url{https://github.com/speechbrain/speechbrain}}, } ``` #### About SpeechBrain SpeechBrain is an open-source and all-in-one speech toolkit. It is designed to be simple, extremely flexible, and user-friendly. Competitive or state-of-the-art performance is obtained in various domains. Website: https://speechbrain.github.io/ GitHub: https://github.com/speechbrain/speechbrain
3,221
[ [ -0.01031494140625, -0.050506591796875, 0.0180511474609375, 0.02252197265625, -0.0099029541015625, -0.00386810302734375, -0.037994384765625, -0.0246429443359375, 0.0180206298828125, 0.0011358261108398438, -0.0460205078125, -0.047760009765625, -0.04339599609375, ...
Supabase/all-MiniLM-L6-v2
2023-05-30T15:33:53.000Z
[ "transformers.js", "onnx", "bert", "feature-extraction", "region:us" ]
feature-extraction
Supabase
null
null
Supabase/all-MiniLM-L6-v2
1
2
transformers.js
2023-08-01T15:26:49
--- library_name: "transformers.js" --- https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2 with ONNX weights to be compatible with Transformers.js. Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
517
[ [ -0.0224609375, -0.0002396106719970703, 0.034942626953125, 0.039306640625, -0.0072479248046875, -0.0126800537109375, -0.0007648468017578125, -0.0200347900390625, 0.0228729248046875, 0.038330078125, -0.06298828125, -0.02056884765625, -0.03924560546875, 0.01309...
toughdata/flan-t5-base-eli5-question-generation-54500
2023-08-01T19:33:16.000Z
[ "transformers", "pytorch", "t5", "text2text-generation", "en", "dataset:eli5", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
text2text-generation
toughdata
null
null
toughdata/flan-t5-base-eli5-question-generation-54500
0
2
transformers
2023-08-01T15:41:38
--- datasets: - eli5 language: - en --- This model generates short questions based on long answers. To use, prend "rephrase this as a question: " to your input text.
166
[ [ -0.0284423828125, -0.0853271484375, 0.0594482421875, 0.006748199462890625, -0.049102783203125, -0.01457977294921875, 0.00885772705078125, -0.0177764892578125, 0.03521728515625, 0.06097412109375, -0.067138671875, -0.007297515869140625, -0.0257720947265625, 0....
lesliepzimmermann/ppo-SnowballTarget
2023-08-01T17:48:36.000Z
[ "ml-agents", "tensorboard", "onnx", "SnowballTarget", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SnowballTarget", "region:us" ]
reinforcement-learning
lesliepzimmermann
null
null
lesliepzimmermann/ppo-SnowballTarget
0
2
ml-agents
2023-08-01T17:48:32
--- library_name: ml-agents tags: - SnowballTarget - deep-reinforcement-learning - reinforcement-learning - ML-Agents-SnowballTarget --- # **ppo** Agent playing **SnowballTarget** This is a trained model of a **ppo** agent playing **SnowballTarget** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: lesliepzimmermann/ppo-SnowballTarget 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,372
[ [ -0.031005859375, -0.03973388671875, 0.00881195068359375, 0.005764007568359375, -0.0211334228515625, 0.0219573974609375, 0.01300048828125, -0.0162506103515625, 0.0266571044921875, 0.03387451171875, -0.055419921875, -0.053466796875, -0.03704833984375, -0.02084...
potatomode/short_jokes_model
2023-08-01T20:24:46.000Z
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "text-generation-inference", "region:us" ]
text-generation
potatomode
null
null
potatomode/short_jokes_model
0
2
transformers
2023-08-01T17:49:29
--- license: apache-2.0 base_model: distilgpt2 tags: - generated_from_trainer model-index: - name: short_jokes_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # short_jokes_model This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.4090 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 16 | 2.6101 | | No log | 2.0 | 32 | 2.4472 | | No log | 3.0 | 48 | 2.4090 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.2 - Tokenizers 0.13.3
1,365
[ [ -0.0287017822265625, -0.05059814453125, 0.0218658447265625, 0.01776123046875, -0.038818359375, -0.03466796875, -0.006443023681640625, -0.01119232177734375, -0.0020122528076171875, 0.017181396484375, -0.052825927734375, -0.036712646484375, -0.054351806640625, ...
DunnBC22/bert-base-uncased-Vitamin_C_Fact_Verification
2023-08-02T02:15:12.000Z
[ "transformers", "pytorch", "tensorboard", "bert", "multiple-choice", "generated_from_trainer", "multiple_choice", "question-answering", "en", "dataset:tasksource/bigbench", "license:apache-2.0", "endpoints_compatible", "region:us" ]
question-answering
DunnBC22
null
null
DunnBC22/bert-base-uncased-Vitamin_C_Fact_Verification
1
2
transformers
2023-08-01T18:18:57
--- license: apache-2.0 base_model: bert-base-uncased tags: - generated_from_trainer - multiple_choice metrics: - accuracy model-index: - name: bert-base-uncased-Vitamin_C_Fact_Verification results: [] datasets: - tasksource/bigbench language: - en pipeline_tag: question-answering --- # bert-base-uncased-Vitamin_C_Fact_Verification This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased). It achieves the following results on the evaluation set: - Loss: 0.6329 - Accuracy: 0.7240 ## Model description For more information on how it was created, check out the following link: https://github.com/DunnBC22/NLP_Projects/blob/main/Multiple%20Choice/Vitamin%20C%20Fact%20Verification/Vitamin_C_Fact_Verification_Multiple_Choice_Using_BERT.ipynb ## Intended uses & limitations This model is intended to demonstrate my ability to solve a complex problem using technology. ## Training and evaluation data Dataset Source: https://huggingface.co/datasets/tasksource/bigbench/viewer/vitaminc_fact_verification ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6985 | 1.0 | 2170 | 0.6894 | 0.6864 | | 0.5555 | 2.0 | 4340 | 0.6329 | 0.7240 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.2 - Tokenizers 0.13.3
1,726
[ [ -0.0198974609375, -0.053009033203125, 0.016448974609375, 0.020751953125, -0.01303863525390625, 0.00001823902130126953, -0.023834228515625, -0.017791748046875, 0.01511383056640625, 0.01776123046875, -0.050689697265625, -0.0435791015625, -0.038299560546875, -0...
lesliepzimmermann/ppo-PyramidsTraining
2023-08-01T18:37:03.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
lesliepzimmermann
null
null
lesliepzimmermann/ppo-PyramidsTraining
0
2
ml-agents
2023-08-01T18:37:00
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: lesliepzimmermann/ppo-PyramidsTraining 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,350
[ [ -0.039459228515625, -0.03350830078125, 0.0019083023071289062, 0.01398468017578125, -0.01012420654296875, 0.0118560791015625, 0.017547607421875, -0.01409912109375, 0.03277587890625, 0.0310211181640625, -0.04058837890625, -0.04925537109375, -0.0295257568359375, ...
Gracoy/ingredients_compatibility_GPT2_S
2023-08-05T09:55:45.000Z
[ "transformers", "tf", "gpt2", "text-generation", "generated_from_keras_callback", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
text-generation
Gracoy
null
null
Gracoy/ingredients_compatibility_GPT2_S
0
2
transformers
2023-08-02T02:38:35
--- tags: - generated_from_keras_callback model-index: - name: ingredients_compatibility_GPT2_S results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # ingredients_compatibility_GPT2_S This model was trained from scratch on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.9950 - Validation Loss: 1.0009 - Epoch: 0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'inner_optimizer': {'class_name': 'AdamWeightDecay', 'config': {'name': 'AdamWeightDecay', 'learning_rate': 1e-04, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.99, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}}, 'dynamic': True, 'initial_scale': 32768.0, 'dynamic_growth_steps': 2000} - training_precision: mixed_float16 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 0.9950 | 1.0009 | 0 | ### Framework versions - Transformers 4.31.0 - TensorFlow 2.12.0 - Datasets 2.14.3 - Tokenizers 0.13.3
1,380
[ [ -0.0291748046875, -0.03765869140625, 0.034210205078125, 0.0015249252319335938, -0.03240966796875, -0.0285491943359375, -0.01024627685546875, -0.021026611328125, -0.0058135986328125, 0.0036945343017578125, -0.045684814453125, -0.039276123046875, -0.06072998046875...
IProject-10/roberta-base-finetuned-squad2
2023-08-20T05:14:36.000Z
[ "transformers", "pytorch", "tensorboard", "roberta", "question-answering", "generated_from_trainer", "en", "dataset:squad_v2", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
question-answering
IProject-10
null
null
IProject-10/roberta-base-finetuned-squad2
1
2
transformers
2023-08-02T02:46:45
--- license: mit base_model: roberta-base tags: - generated_from_trainer datasets: - squad_v2 model-index: - name: roberta-base-finetuned-squad2 results: [] language: - en metrics: - exact_match - f1 pipeline_tag: question-answering --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> ## Model description RoBERTa-base fine-tuned on SQuAD 2.0 : Encoder-based Transformer Language model, pretrained with Dynamic Masking, No Next Sentence Prediction and increased Batch size compared to BERT.<br> Suitable for Question-Answering tasks, predicts answer spans within the context provided.<br> **Language model:** roberta-base **Language:** English **Downstream-task:** Question-Answering **Training data:** Train-set SQuAD 2.0 **Evaluation data:** Evaluation-set SQuAD 2.0 **Hardware Accelerator used**: GPU Tesla T4 ## Intended uses & limitations For Question-Answering - ```python !pip install transformers from transformers import pipeline model_checkpoint = "IProject-10/roberta-base-finetuned-squad2" question_answerer = pipeline("question-answering", model=model_checkpoint) context = """ 🤗 Transformers is backed by the three most popular deep learning libraries — Jax, PyTorch and TensorFlow — with a seamless integration between them. It's straightforward to train your models with one before loading them for inference with the other. """ question = "Which deep learning libraries back 🤗 Transformers?" question_answerer(question=question, context=context) ``` ## Results Evaluation on SQuAD 2.0 validation dataset: ``` exact: 79.71868946348859, f1: 83.049614486567, total: 11873, HasAns_exact: 78.39068825910931, HasAns_f1: 85.06209055313944, HasAns_total: 5928, NoAns_exact: 81.04289318755256, NoAns_f1: 81.04289318755256, NoAns_total: 5945, best_exact: 79.71868946348859, best_exact_thresh: 0.9997376203536987, best_f1: 83.04961448656734, best_f1_thresh: 0.9997376203536987, total_time_in_seconds: 226.245504546, samples_per_second: 52.47839078095801, latency_in_second': 0.019055462355428283 ``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 0.8921 | 1.0 | 8239 | 0.8899 | | 0.6186 | 2.0 | 16478 | 0.8880 | | 0.4393 | 3.0 | 24717 | 0.9785 | This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the squad_v2 dataset. It achieves the following results on the evaluation set: - Loss: 0.9785 ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.2 - Tokenizers 0.13.3
3,034
[ [ -0.03363037109375, -0.056640625, 0.01904296875, 0.011688232421875, -0.0024051666259765625, -0.0024738311767578125, -0.0225067138671875, -0.0264129638671875, -0.00711822509765625, 0.0169525146484375, -0.06756591796875, -0.03875732421875, -0.051300048828125, -...
simonycl/roberta-large-sst-2-16-13
2023-08-02T20:55:24.000Z
[ "transformers", "pytorch", "roberta", "text-classification", "generated_from_trainer", "license:mit", "endpoints_compatible", "region:us" ]
text-classification
simonycl
null
null
simonycl/roberta-large-sst-2-16-13
0
2
transformers
2023-08-02T07:50:50
--- license: mit base_model: roberta-large tags: - generated_from_trainer metrics: - accuracy model-index: - name: roberta-large-sst-2-16-13 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-large-sst-2-16-13 This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.3222 - Accuracy: 0.8438 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 150 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 1 | 0.7045 | 0.5 | | No log | 2.0 | 2 | 0.7045 | 0.5 | | No log | 3.0 | 3 | 0.7045 | 0.5 | | No log | 4.0 | 4 | 0.7045 | 0.5 | | No log | 5.0 | 5 | 0.7045 | 0.5 | | No log | 6.0 | 6 | 0.7045 | 0.5 | | No log | 7.0 | 7 | 0.7044 | 0.5 | | No log | 8.0 | 8 | 0.7044 | 0.5 | | No log | 9.0 | 9 | 0.7044 | 0.5 | | 0.7125 | 10.0 | 10 | 0.7043 | 0.5 | | 0.7125 | 11.0 | 11 | 0.7043 | 0.5 | | 0.7125 | 12.0 | 12 | 0.7042 | 0.5 | | 0.7125 | 13.0 | 13 | 0.7042 | 0.5 | | 0.7125 | 14.0 | 14 | 0.7041 | 0.5 | | 0.7125 | 15.0 | 15 | 0.7041 | 0.5 | | 0.7125 | 16.0 | 16 | 0.7040 | 0.5 | | 0.7125 | 17.0 | 17 | 0.7040 | 0.5 | | 0.7125 | 18.0 | 18 | 0.7039 | 0.5 | | 0.7125 | 19.0 | 19 | 0.7039 | 0.5 | | 0.6935 | 20.0 | 20 | 0.7038 | 0.5 | | 0.6935 | 21.0 | 21 | 0.7038 | 0.5 | | 0.6935 | 22.0 | 22 | 0.7037 | 0.5 | | 0.6935 | 23.0 | 23 | 0.7037 | 0.5 | | 0.6935 | 24.0 | 24 | 0.7037 | 0.5 | | 0.6935 | 25.0 | 25 | 0.7036 | 0.5 | | 0.6935 | 26.0 | 26 | 0.7036 | 0.5 | | 0.6935 | 27.0 | 27 | 0.7035 | 0.5 | | 0.6935 | 28.0 | 28 | 0.7035 | 0.5 | | 0.6935 | 29.0 | 29 | 0.7034 | 0.5 | | 0.7031 | 30.0 | 30 | 0.7033 | 0.5 | | 0.7031 | 31.0 | 31 | 0.7032 | 0.5 | | 0.7031 | 32.0 | 32 | 0.7031 | 0.5 | | 0.7031 | 33.0 | 33 | 0.7030 | 0.5 | | 0.7031 | 34.0 | 34 | 0.7029 | 0.5 | | 0.7031 | 35.0 | 35 | 0.7027 | 0.5 | | 0.7031 | 36.0 | 36 | 0.7027 | 0.5 | | 0.7031 | 37.0 | 37 | 0.7026 | 0.5 | | 0.7031 | 38.0 | 38 | 0.7025 | 0.5 | | 0.7031 | 39.0 | 39 | 0.7024 | 0.5 | | 0.7021 | 40.0 | 40 | 0.7023 | 0.5 | | 0.7021 | 41.0 | 41 | 0.7022 | 0.5 | | 0.7021 | 42.0 | 42 | 0.7021 | 0.5 | | 0.7021 | 43.0 | 43 | 0.7019 | 0.5 | | 0.7021 | 44.0 | 44 | 0.7017 | 0.5 | | 0.7021 | 45.0 | 45 | 0.7016 | 0.5 | | 0.7021 | 46.0 | 46 | 0.7014 | 0.5 | | 0.7021 | 47.0 | 47 | 0.7012 | 0.5 | | 0.7021 | 48.0 | 48 | 0.7010 | 0.5 | | 0.7021 | 49.0 | 49 | 0.7007 | 0.5 | | 0.7009 | 50.0 | 50 | 0.7005 | 0.5 | | 0.7009 | 51.0 | 51 | 0.7003 | 0.5 | | 0.7009 | 52.0 | 52 | 0.7001 | 0.5 | | 0.7009 | 53.0 | 53 | 0.6998 | 0.5 | | 0.7009 | 54.0 | 54 | 0.6996 | 0.5 | | 0.7009 | 55.0 | 55 | 0.6994 | 0.5 | | 0.7009 | 56.0 | 56 | 0.6993 | 0.5 | | 0.7009 | 57.0 | 57 | 0.6992 | 0.5 | | 0.7009 | 58.0 | 58 | 0.6990 | 0.5 | | 0.7009 | 59.0 | 59 | 0.6988 | 0.5 | | 0.6866 | 60.0 | 60 | 0.6986 | 0.5 | | 0.6866 | 61.0 | 61 | 0.6984 | 0.5 | | 0.6866 | 62.0 | 62 | 0.6983 | 0.5 | | 0.6866 | 63.0 | 63 | 0.6981 | 0.5 | | 0.6866 | 64.0 | 64 | 0.6979 | 0.5 | | 0.6866 | 65.0 | 65 | 0.6977 | 0.5 | | 0.6866 | 66.0 | 66 | 0.6976 | 0.4688 | | 0.6866 | 67.0 | 67 | 0.6974 | 0.4688 | | 0.6866 | 68.0 | 68 | 0.6972 | 0.4688 | | 0.6866 | 69.0 | 69 | 0.6970 | 0.4688 | | 0.6818 | 70.0 | 70 | 0.6968 | 0.4688 | | 0.6818 | 71.0 | 71 | 0.6966 | 0.4688 | | 0.6818 | 72.0 | 72 | 0.6964 | 0.4688 | | 0.6818 | 73.0 | 73 | 0.6961 | 0.4688 | | 0.6818 | 74.0 | 74 | 0.6960 | 0.4688 | | 0.6818 | 75.0 | 75 | 0.6959 | 0.4688 | | 0.6818 | 76.0 | 76 | 0.6957 | 0.4688 | | 0.6818 | 77.0 | 77 | 0.6955 | 0.4688 | | 0.6818 | 78.0 | 78 | 0.6953 | 0.4688 | | 0.6818 | 79.0 | 79 | 0.6948 | 0.4688 | | 0.6639 | 80.0 | 80 | 0.6940 | 0.4688 | | 0.6639 | 81.0 | 81 | 0.6932 | 0.4688 | | 0.6639 | 82.0 | 82 | 0.6925 | 0.4688 | | 0.6639 | 83.0 | 83 | 0.6916 | 0.4688 | | 0.6639 | 84.0 | 84 | 0.6908 | 0.5 | | 0.6639 | 85.0 | 85 | 0.6899 | 0.5 | | 0.6639 | 86.0 | 86 | 0.6889 | 0.5 | | 0.6639 | 87.0 | 87 | 0.6878 | 0.5 | | 0.6639 | 88.0 | 88 | 0.6869 | 0.5 | | 0.6639 | 89.0 | 89 | 0.6859 | 0.4688 | | 0.6652 | 90.0 | 90 | 0.6850 | 0.4688 | | 0.6652 | 91.0 | 91 | 0.6842 | 0.4688 | | 0.6652 | 92.0 | 92 | 0.6836 | 0.5312 | | 0.6652 | 93.0 | 93 | 0.6829 | 0.5312 | | 0.6652 | 94.0 | 94 | 0.6818 | 0.5625 | | 0.6652 | 95.0 | 95 | 0.6806 | 0.5938 | | 0.6652 | 96.0 | 96 | 0.6792 | 0.5938 | | 0.6652 | 97.0 | 97 | 0.6783 | 0.5938 | | 0.6652 | 98.0 | 98 | 0.6771 | 0.5938 | | 0.6652 | 99.0 | 99 | 0.6758 | 0.5938 | | 0.621 | 100.0 | 100 | 0.6743 | 0.5938 | | 0.621 | 101.0 | 101 | 0.6725 | 0.5938 | | 0.621 | 102.0 | 102 | 0.6711 | 0.5938 | | 0.621 | 103.0 | 103 | 0.6708 | 0.5938 | | 0.621 | 104.0 | 104 | 0.6713 | 0.625 | | 0.621 | 105.0 | 105 | 0.6693 | 0.5938 | | 0.621 | 106.0 | 106 | 0.6605 | 0.5938 | | 0.621 | 107.0 | 107 | 0.6499 | 0.5938 | | 0.621 | 108.0 | 108 | 0.6439 | 0.5625 | | 0.621 | 109.0 | 109 | 0.6434 | 0.625 | | 0.5331 | 110.0 | 110 | 0.6439 | 0.5938 | | 0.5331 | 111.0 | 111 | 0.6418 | 0.5625 | | 0.5331 | 112.0 | 112 | 0.6388 | 0.5625 | | 0.5331 | 113.0 | 113 | 0.6346 | 0.5625 | | 0.5331 | 114.0 | 114 | 0.6307 | 0.5625 | | 0.5331 | 115.0 | 115 | 0.6275 | 0.5625 | | 0.5331 | 116.0 | 116 | 0.6230 | 0.5625 | | 0.5331 | 117.0 | 117 | 0.6144 | 0.5625 | | 0.5331 | 118.0 | 118 | 0.6092 | 0.5625 | | 0.5331 | 119.0 | 119 | 0.6042 | 0.5938 | | 0.4594 | 120.0 | 120 | 0.6006 | 0.5938 | | 0.4594 | 121.0 | 121 | 0.5971 | 0.5938 | | 0.4594 | 122.0 | 122 | 0.5906 | 0.5938 | | 0.4594 | 123.0 | 123 | 0.5841 | 0.5938 | | 0.4594 | 124.0 | 124 | 0.5759 | 0.6562 | | 0.4594 | 125.0 | 125 | 0.5682 | 0.6875 | | 0.4594 | 126.0 | 126 | 0.5566 | 0.6875 | | 0.4594 | 127.0 | 127 | 0.5431 | 0.6875 | | 0.4594 | 128.0 | 128 | 0.5314 | 0.6875 | | 0.4594 | 129.0 | 129 | 0.5221 | 0.7188 | | 0.33 | 130.0 | 130 | 0.5145 | 0.7188 | | 0.33 | 131.0 | 131 | 0.5062 | 0.7188 | | 0.33 | 132.0 | 132 | 0.4988 | 0.7188 | | 0.33 | 133.0 | 133 | 0.4888 | 0.7188 | | 0.33 | 134.0 | 134 | 0.4689 | 0.7188 | | 0.33 | 135.0 | 135 | 0.4586 | 0.75 | | 0.33 | 136.0 | 136 | 0.4464 | 0.7812 | | 0.33 | 137.0 | 137 | 0.4330 | 0.7812 | | 0.33 | 138.0 | 138 | 0.4185 | 0.7812 | | 0.33 | 139.0 | 139 | 0.4004 | 0.8125 | | 0.2099 | 140.0 | 140 | 0.3852 | 0.8125 | | 0.2099 | 141.0 | 141 | 0.3724 | 0.8125 | | 0.2099 | 142.0 | 142 | 0.3610 | 0.8125 | | 0.2099 | 143.0 | 143 | 0.3613 | 0.8125 | | 0.2099 | 144.0 | 144 | 0.3731 | 0.7812 | | 0.2099 | 145.0 | 145 | 0.3655 | 0.8125 | | 0.2099 | 146.0 | 146 | 0.3553 | 0.8125 | | 0.2099 | 147.0 | 147 | 0.3457 | 0.8125 | | 0.2099 | 148.0 | 148 | 0.3380 | 0.8438 | | 0.2099 | 149.0 | 149 | 0.3315 | 0.8438 | | 0.0894 | 150.0 | 150 | 0.3222 | 0.8438 | ### Framework versions - Transformers 4.32.0.dev0 - Pytorch 2.0.1+cu118 - Datasets 2.4.0 - Tokenizers 0.13.3
10,632
[ [ -0.036102294921875, -0.036346435546875, 0.0166473388671875, 0.0066375732421875, 0.0055084228515625, 0.01543426513671875, 0.004726409912109375, 0.007663726806640625, 0.05438232421875, 0.0290985107421875, -0.044830322265625, -0.04248046875, -0.042083740234375, ...
umaru97/gpt2-product-review-generation
2023-08-02T13:26:06.000Z
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "generated_from_trainer", "license:mit", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
text-generation
umaru97
null
null
umaru97/gpt2-product-review-generation
0
2
transformers
2023-08-02T12:15:07
--- license: mit base_model: gpt2-medium tags: - generated_from_trainer model-index: - name: gpt2-product-review-generation results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gpt2-product-review-generation This model is a fine-tuned version of [gpt2-medium](https://huggingface.co/gpt2-medium) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.8570 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.0547 | 1.0 | 1777 | 2.9286 | | 2.8842 | 2.0 | 3554 | 2.8736 | | 2.804 | 3.0 | 5331 | 2.8570 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.2 - Tokenizers 0.13.3
1,385
[ [ -0.028839111328125, -0.049774169921875, 0.0209503173828125, 0.0094757080078125, -0.0333251953125, -0.03857421875, -0.0054931640625, -0.016937255859375, -0.01267242431640625, 0.0207061767578125, -0.047821044921875, -0.033843994140625, -0.052398681640625, -0.0...
kelSidenna/SoftwareReq-DialoGPT-medium
2023-08-02T16:09:52.000Z
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "endpoints_compatible", "text-generation-inference", "region:us" ]
conversational
kelSidenna
null
null
kelSidenna/SoftwareReq-DialoGPT-medium
0
2
transformers
2023-08-02T12:23:00
--- pipeline_tag: conversational --- # Model Card for DialoGPT-medium Conversational Model ## Model Details - **Model name:** Fine tuned DialoGPT-medium - **Model type:** Transformer-based language model (GPT-2 variant) - **Original model:** DialoGPT from Hugging Face model hub - **Fine-tuning details:** The model has been fine-tuned on a custom conversational dataset. It includes a variety of dialogues covering multiple topics, aimed at increasing the model's ability to respond accurately and engagingly in conversational tasks. ## Intended Use DialogPT-medium is designed for a wide range of conversational applications. It is suitable for building chatbots.
672
[ [ -0.03546142578125, -0.0775146484375, 0.00794219970703125, 0.0096588134765625, -0.03143310546875, 0.0047454833984375, -0.01251983642578125, -0.027618408203125, 0.010833740234375, 0.037872314453125, -0.044769287109375, -0.0030536651611328125, -0.041595458984375, ...
simonycl/roberta-base-sst-2-32-13
2023-08-02T20:16:41.000Z
[ "transformers", "pytorch", "roberta", "text-classification", "generated_from_trainer", "license:mit", "endpoints_compatible", "region:us" ]
text-classification
simonycl
null
null
simonycl/roberta-base-sst-2-32-13
0
2
transformers
2023-08-02T19:51:32
--- license: mit base_model: roberta-base tags: - generated_from_trainer metrics: - accuracy model-index: - name: roberta-base-sst-2-32-13 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-sst-2-32-13 This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.9957 - Accuracy: 0.8438 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 150 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 2 | 0.6951 | 0.5 | | No log | 2.0 | 4 | 0.6951 | 0.5 | | No log | 3.0 | 6 | 0.6951 | 0.5 | | No log | 4.0 | 8 | 0.6951 | 0.5 | | 0.6937 | 5.0 | 10 | 0.6950 | 0.5 | | 0.6937 | 6.0 | 12 | 0.6950 | 0.5 | | 0.6937 | 7.0 | 14 | 0.6950 | 0.5 | | 0.6937 | 8.0 | 16 | 0.6950 | 0.5 | | 0.6937 | 9.0 | 18 | 0.6949 | 0.5 | | 0.6953 | 10.0 | 20 | 0.6949 | 0.5 | | 0.6953 | 11.0 | 22 | 0.6949 | 0.5 | | 0.6953 | 12.0 | 24 | 0.6948 | 0.5 | | 0.6953 | 13.0 | 26 | 0.6948 | 0.5 | | 0.6953 | 14.0 | 28 | 0.6947 | 0.5 | | 0.6975 | 15.0 | 30 | 0.6947 | 0.5 | | 0.6975 | 16.0 | 32 | 0.6946 | 0.5 | | 0.6975 | 17.0 | 34 | 0.6946 | 0.5 | | 0.6975 | 18.0 | 36 | 0.6945 | 0.5 | | 0.6975 | 19.0 | 38 | 0.6944 | 0.5 | | 0.6888 | 20.0 | 40 | 0.6944 | 0.5 | | 0.6888 | 21.0 | 42 | 0.6943 | 0.5 | | 0.6888 | 22.0 | 44 | 0.6942 | 0.5 | | 0.6888 | 23.0 | 46 | 0.6942 | 0.5 | | 0.6888 | 24.0 | 48 | 0.6941 | 0.5 | | 0.6947 | 25.0 | 50 | 0.6940 | 0.5 | | 0.6947 | 26.0 | 52 | 0.6940 | 0.5 | | 0.6947 | 27.0 | 54 | 0.6939 | 0.5 | | 0.6947 | 28.0 | 56 | 0.6938 | 0.5 | | 0.6947 | 29.0 | 58 | 0.6937 | 0.5 | | 0.69 | 30.0 | 60 | 0.6937 | 0.5 | | 0.69 | 31.0 | 62 | 0.6936 | 0.5 | | 0.69 | 32.0 | 64 | 0.6936 | 0.5 | | 0.69 | 33.0 | 66 | 0.6935 | 0.5 | | 0.69 | 34.0 | 68 | 0.6934 | 0.5 | | 0.6901 | 35.0 | 70 | 0.6933 | 0.5 | | 0.6901 | 36.0 | 72 | 0.6932 | 0.5 | | 0.6901 | 37.0 | 74 | 0.6931 | 0.5 | | 0.6901 | 38.0 | 76 | 0.6930 | 0.5 | | 0.6901 | 39.0 | 78 | 0.6929 | 0.5 | | 0.6895 | 40.0 | 80 | 0.6928 | 0.5 | | 0.6895 | 41.0 | 82 | 0.6927 | 0.5 | | 0.6895 | 42.0 | 84 | 0.6926 | 0.5 | | 0.6895 | 43.0 | 86 | 0.6925 | 0.5 | | 0.6895 | 44.0 | 88 | 0.6924 | 0.5 | | 0.6874 | 45.0 | 90 | 0.6922 | 0.5 | | 0.6874 | 46.0 | 92 | 0.6921 | 0.5 | | 0.6874 | 47.0 | 94 | 0.6919 | 0.5 | | 0.6874 | 48.0 | 96 | 0.6917 | 0.5 | | 0.6874 | 49.0 | 98 | 0.6915 | 0.5 | | 0.6865 | 50.0 | 100 | 0.6913 | 0.5 | | 0.6865 | 51.0 | 102 | 0.6911 | 0.5 | | 0.6865 | 52.0 | 104 | 0.6908 | 0.5 | | 0.6865 | 53.0 | 106 | 0.6904 | 0.4844 | | 0.6865 | 54.0 | 108 | 0.6901 | 0.4688 | | 0.6818 | 55.0 | 110 | 0.6897 | 0.4688 | | 0.6818 | 56.0 | 112 | 0.6892 | 0.4531 | | 0.6818 | 57.0 | 114 | 0.6887 | 0.5625 | | 0.6818 | 58.0 | 116 | 0.6880 | 0.6094 | | 0.6818 | 59.0 | 118 | 0.6872 | 0.6406 | | 0.6697 | 60.0 | 120 | 0.6863 | 0.6406 | | 0.6697 | 61.0 | 122 | 0.6852 | 0.6875 | | 0.6697 | 62.0 | 124 | 0.6838 | 0.7656 | | 0.6697 | 63.0 | 126 | 0.6820 | 0.7812 | | 0.6697 | 64.0 | 128 | 0.6798 | 0.7656 | | 0.6559 | 65.0 | 130 | 0.6769 | 0.7656 | | 0.6559 | 66.0 | 132 | 0.6730 | 0.7188 | | 0.6559 | 67.0 | 134 | 0.6675 | 0.7344 | | 0.6559 | 68.0 | 136 | 0.6598 | 0.7188 | | 0.6559 | 69.0 | 138 | 0.6489 | 0.7188 | | 0.6085 | 70.0 | 140 | 0.6343 | 0.7188 | | 0.6085 | 71.0 | 142 | 0.6161 | 0.7656 | | 0.6085 | 72.0 | 144 | 0.5928 | 0.8125 | | 0.6085 | 73.0 | 146 | 0.5652 | 0.8438 | | 0.6085 | 74.0 | 148 | 0.5367 | 0.8594 | | 0.474 | 75.0 | 150 | 0.5083 | 0.8438 | | 0.474 | 76.0 | 152 | 0.4779 | 0.8438 | | 0.474 | 77.0 | 154 | 0.4473 | 0.8594 | | 0.474 | 78.0 | 156 | 0.4179 | 0.8594 | | 0.474 | 79.0 | 158 | 0.3930 | 0.875 | | 0.2428 | 80.0 | 160 | 0.3782 | 0.8594 | | 0.2428 | 81.0 | 162 | 0.3734 | 0.8438 | | 0.2428 | 82.0 | 164 | 0.3731 | 0.8594 | | 0.2428 | 83.0 | 166 | 0.3816 | 0.875 | | 0.2428 | 84.0 | 168 | 0.4042 | 0.8438 | | 0.0805 | 85.0 | 170 | 0.4405 | 0.8438 | | 0.0805 | 86.0 | 172 | 0.4840 | 0.8281 | | 0.0805 | 87.0 | 174 | 0.5432 | 0.8125 | | 0.0805 | 88.0 | 176 | 0.6025 | 0.8125 | | 0.0805 | 89.0 | 178 | 0.6412 | 0.8125 | | 0.0222 | 90.0 | 180 | 0.6653 | 0.8125 | | 0.0222 | 91.0 | 182 | 0.6845 | 0.8125 | | 0.0222 | 92.0 | 184 | 0.6954 | 0.8125 | | 0.0222 | 93.0 | 186 | 0.7007 | 0.8281 | | 0.0222 | 94.0 | 188 | 0.7029 | 0.8438 | | 0.0093 | 95.0 | 190 | 0.7083 | 0.8438 | | 0.0093 | 96.0 | 192 | 0.7172 | 0.8594 | | 0.0093 | 97.0 | 194 | 0.7250 | 0.8594 | | 0.0093 | 98.0 | 196 | 0.7286 | 0.8594 | | 0.0093 | 99.0 | 198 | 0.7361 | 0.8594 | | 0.0058 | 100.0 | 200 | 0.7447 | 0.8594 | | 0.0058 | 101.0 | 202 | 0.7544 | 0.8594 | | 0.0058 | 102.0 | 204 | 0.7632 | 0.8594 | | 0.0058 | 103.0 | 206 | 0.7724 | 0.8594 | | 0.0058 | 104.0 | 208 | 0.7842 | 0.8594 | | 0.0041 | 105.0 | 210 | 0.7955 | 0.8594 | | 0.0041 | 106.0 | 212 | 0.8061 | 0.8594 | | 0.0041 | 107.0 | 214 | 0.8164 | 0.8594 | | 0.0041 | 108.0 | 216 | 0.8262 | 0.8594 | | 0.0041 | 109.0 | 218 | 0.8348 | 0.8594 | | 0.0032 | 110.0 | 220 | 0.8438 | 0.8594 | | 0.0032 | 111.0 | 222 | 0.8514 | 0.8594 | | 0.0032 | 112.0 | 224 | 0.8582 | 0.8594 | | 0.0032 | 113.0 | 226 | 0.8650 | 0.8594 | | 0.0032 | 114.0 | 228 | 0.8718 | 0.8438 | | 0.0028 | 115.0 | 230 | 0.8777 | 0.8438 | | 0.0028 | 116.0 | 232 | 0.8829 | 0.8438 | | 0.0028 | 117.0 | 234 | 0.8884 | 0.8438 | | 0.0028 | 118.0 | 236 | 0.8938 | 0.8438 | | 0.0028 | 119.0 | 238 | 0.8986 | 0.8438 | | 0.0024 | 120.0 | 240 | 0.9023 | 0.8438 | | 0.0024 | 121.0 | 242 | 0.9055 | 0.8438 | | 0.0024 | 122.0 | 244 | 0.9087 | 0.8438 | | 0.0024 | 123.0 | 246 | 0.9121 | 0.8438 | | 0.0024 | 124.0 | 248 | 0.9165 | 0.8438 | | 0.0021 | 125.0 | 250 | 0.9209 | 0.8438 | | 0.0021 | 126.0 | 252 | 0.9258 | 0.8438 | | 0.0021 | 127.0 | 254 | 0.9303 | 0.8438 | | 0.0021 | 128.0 | 256 | 0.9338 | 0.8438 | | 0.0021 | 129.0 | 258 | 0.9365 | 0.8438 | | 0.0019 | 130.0 | 260 | 0.9395 | 0.8438 | | 0.0019 | 131.0 | 262 | 0.9426 | 0.8438 | | 0.0019 | 132.0 | 264 | 0.9448 | 0.8438 | | 0.0019 | 133.0 | 266 | 0.9463 | 0.8438 | | 0.0019 | 134.0 | 268 | 0.9480 | 0.8438 | | 0.0017 | 135.0 | 270 | 0.9506 | 0.8438 | | 0.0017 | 136.0 | 272 | 0.9535 | 0.8438 | | 0.0017 | 137.0 | 274 | 0.9561 | 0.8438 | | 0.0017 | 138.0 | 276 | 0.9579 | 0.8438 | | 0.0017 | 139.0 | 278 | 0.9596 | 0.8438 | | 0.0015 | 140.0 | 280 | 0.9618 | 0.8438 | | 0.0015 | 141.0 | 282 | 0.9650 | 0.8438 | | 0.0015 | 142.0 | 284 | 0.9682 | 0.8438 | | 0.0015 | 143.0 | 286 | 0.9712 | 0.8438 | | 0.0015 | 144.0 | 288 | 0.9741 | 0.8438 | | 0.0014 | 145.0 | 290 | 0.9769 | 0.8438 | | 0.0014 | 146.0 | 292 | 0.9801 | 0.8438 | | 0.0014 | 147.0 | 294 | 0.9835 | 0.8438 | | 0.0014 | 148.0 | 296 | 0.9872 | 0.8438 | | 0.0014 | 149.0 | 298 | 0.9911 | 0.8438 | | 0.0013 | 150.0 | 300 | 0.9957 | 0.8438 | ### Framework versions - Transformers 4.32.0.dev0 - Pytorch 2.0.1+cu118 - Datasets 2.4.0 - Tokenizers 0.13.3
10,627
[ [ -0.0380859375, -0.038970947265625, 0.016815185546875, 0.00798797607421875, 0.0013704299926757812, 0.0150146484375, 0.003387451171875, 0.0117950439453125, 0.053192138671875, 0.0271148681640625, -0.044189453125, -0.040069580078125, -0.041412353515625, -0.01385...
WeightWatcher/albert-large-v2-cola
2023-08-03T06:12:06.000Z
[ "transformers", "pytorch", "albert", "text-classification", "en", "dataset:glue", "arxiv:1909.11942", "license:mit", "endpoints_compatible", "region:us" ]
text-classification
WeightWatcher
null
null
WeightWatcher/albert-large-v2-cola
0
2
transformers
2023-08-02T20:23:24
--- language: - "en" license: mit datasets: - glue metrics: - Matthews Correlation --- # Model Card for WeightWatcher/albert-large-v2-cola This model was finetuned on the GLUE/cola task, based on the pretrained albert-large-v2 model. Hyperparameters were (largely) taken from the following publication, with some minor exceptions. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations https://arxiv.org/abs/1909.11942 ## Model Details ### Model Description - **Developed by:** https://huggingface.co/cdhinrichs - **Model type:** Text Sequence Classification - **Language(s) (NLP):** English - **License:** MIT - **Finetuned from model:** https://huggingface.co/albert-large-v2 ## Uses Text classification, research and development. ### Out-of-Scope Use Not intended for production use. See https://huggingface.co/albert-large-v2 ## Bias, Risks, and Limitations See https://huggingface.co/albert-large-v2 ### Recommendations See https://huggingface.co/albert-large-v2 ## How to Get Started with the Model Use the code below to get started with the model. ```python from transformers import AlbertForSequenceClassification model = AlbertForSequenceClassification.from_pretrained("WeightWatcher/albert-large-v2-cola") ``` ## Training Details ### Training Data See https://huggingface.co/datasets/glue#cola CoLA is a classification task, and a part of the GLUE benchmark. ### Training Procedure Adam optimization was used on the pretrained ALBERT model at https://huggingface.co/albert-large-v2. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations https://arxiv.org/abs/1909.11942 #### Training Hyperparameters Training hyperparameters, (Learning Rate, Batch Size, ALBERT dropout rate, Classifier Dropout Rate, Warmup Steps, Training Steps,) were taken from Table A.4 in, ALBERT: A Lite BERT for Self-supervised Learning of Language Representations https://arxiv.org/abs/1909.11942 Max sequence length (MSL) was set to 128, differing from the above. ## Evaluation Matthews Correlation is used to evaluate model performance. ### Testing Data, Factors & Metrics #### Testing Data See https://huggingface.co/datasets/glue#cola #### Metrics Matthews Correlation ### Results Training Matthews Correlation: 0.9786230864021822 Evaluation Matthews Correlation: 0.5723853959351589 ## Environmental Impact The model was finetuned on a single user workstation with a single GPU. CO2 impact is expected to be minimal.
2,497
[ [ -0.0159912109375, -0.03765869140625, 0.02349853515625, 0.033172607421875, 0.0032901763916015625, -0.0187225341796875, -0.00824737548828125, -0.0225982666015625, 0.02545166015625, 0.0268096923828125, -0.035308837890625, -0.0124053955078125, -0.058135986328125, ...
OpenBuddy/openbuddy-falcon-40b-v9-bf16
2023-08-22T10:50:44.000Z
[ "transformers", "pytorch", "RefinedWeb", "text-generation", "custom_code", "zh", "en", "fr", "de", "ja", "ko", "it", "ru", "license:apache-2.0", "text-generation-inference", "region:us" ]
text-generation
OpenBuddy
null
null
OpenBuddy/openbuddy-falcon-40b-v9-bf16
1
2
transformers
2023-08-02T22:19:59
--- language: - zh - en - fr - de - ja - ko - it - ru pipeline_tag: text-generation inference: false library_name: transformers license: apache-2.0 --- # OpenBuddy - Open Multilingual Chatbot GitHub and Usage Guide: [https://github.com/OpenBuddy/OpenBuddy](https://github.com/OpenBuddy/OpenBuddy) Website and Demo: [https://openbuddy.ai](https://openbuddy.ai) ![Demo](https://raw.githubusercontent.com/OpenBuddy/OpenBuddy/main/media/demo.png) # Copyright Notice License: Apache 2.0. ## Disclaimer All OpenBuddy models have inherent limitations and may potentially produce outputs that are erroneous, harmful, offensive, or otherwise undesirable. Users should not use these models in critical or high-stakes situations that may lead to personal injury, property damage, or significant losses. Examples of such scenarios include, but are not limited to, the medical field, controlling software and hardware systems that may cause harm, and making important financial or legal decisions. OpenBuddy is provided "as-is" without any warranty of any kind, either express or implied, including, but not limited to, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement. In no event shall the authors, contributors, or copyright holders be liable for any claim, damages, or other liabilities, whether in an action of contract, tort, or otherwise, arising from, out of, or in connection with the software or the use or other dealings in the software. By using OpenBuddy, you agree to these terms and conditions, and acknowledge that you understand the potential risks associated with its use. You also agree to indemnify and hold harmless the authors, contributors, and copyright holders from any claims, damages, or liabilities arising from your use of OpenBuddy. ## 免责声明 所有OpenBuddy模型均存在固有的局限性,可能产生错误的、有害的、冒犯性的或其他不良的输出。用户在关键或高风险场景中应谨慎行事,不要使用这些模型,以免导致人身伤害、财产损失或重大损失。此类场景的例子包括但不限于医疗领域、可能导致伤害的软硬件系统的控制以及进行重要的财务或法律决策。 OpenBuddy按“原样”提供,不附带任何种类的明示或暗示的保证,包括但不限于适销性、特定目的的适用性和非侵权的暗示保证。在任何情况下,作者、贡献者或版权所有者均不对因软件或使用或其他软件交易而产生的任何索赔、损害赔偿或其他责任(无论是合同、侵权还是其他原因)承担责任。 使用OpenBuddy即表示您同意这些条款和条件,并承认您了解其使用可能带来的潜在风险。您还同意赔偿并使作者、贡献者和版权所有者免受因您使用OpenBuddy而产生的任何索赔、损害赔偿或责任的影响。
2,202
[ [ -0.02740478515625, -0.075439453125, 0.011688232421875, 0.0380859375, -0.0212249755859375, -0.01006317138671875, -0.018585205078125, -0.0322265625, 0.01023101806640625, 0.03228759765625, -0.0200653076171875, -0.042144775390625, -0.0321044921875, -0.0193939208...
zslrmhb/ppo-Pyramids
2023-08-02T23:15:37.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
zslrmhb
null
null
zslrmhb/ppo-Pyramids
0
2
ml-agents
2023-08-02T23:14:14
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: zslrmhb/ppo-Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,332
[ [ -0.04046630859375, -0.033721923828125, 0.0021228790283203125, 0.01442718505859375, -0.010986328125, 0.01245880126953125, 0.0170135498046875, -0.01461029052734375, 0.033172607421875, 0.029998779296875, -0.040313720703125, -0.05047607421875, -0.029327392578125, ...
zacdennis/PyramidsRND
2023-08-03T03:43:13.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
zacdennis
null
null
zacdennis/PyramidsRND
0
2
ml-agents
2023-08-03T03:43:10
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: zacdennis/PyramidsRND 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,333
[ [ -0.041473388671875, -0.034881591796875, 0.0021343231201171875, 0.01357269287109375, -0.0098114013671875, 0.01212310791015625, 0.01605224609375, -0.015716552734375, 0.033477783203125, 0.030242919921875, -0.0419921875, -0.049407958984375, -0.029296875, -0.0154...
tomoohive/PyramidTraining
2023-08-03T03:54:13.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
tomoohive
null
null
tomoohive/PyramidTraining
0
2
ml-agents
2023-08-03T03:52:37
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: tomoohive/PyramidTraining 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,337
[ [ -0.040374755859375, -0.034332275390625, 0.0006132125854492188, 0.014678955078125, -0.00974273681640625, 0.01186370849609375, 0.016143798828125, -0.0154266357421875, 0.033233642578125, 0.0294036865234375, -0.04241943359375, -0.04876708984375, -0.0286865234375, ...
NasimB/cbt-gutenberg_fixed-notm-log-rarity-seed
2023-08-03T06:04:25.000Z
[ "transformers", "pytorch", "gpt2", "text-generation", "generated_from_trainer", "dataset:generator", "license:mit", "endpoints_compatible", "text-generation-inference", "region:us" ]
text-generation
NasimB
null
null
NasimB/cbt-gutenberg_fixed-notm-log-rarity-seed
0
2
transformers
2023-08-03T03:58:20
--- license: mit tags: - generated_from_trainer datasets: - generator model-index: - name: cbt-gutenberg_fixed-notm-log-rarity-seed results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # cbt-gutenberg_fixed-notm-log-rarity-seed This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the generator dataset. It achieves the following results on the evaluation set: - Loss: 4.1166 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 1000 - num_epochs: 6 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 6.3486 | 0.29 | 500 | 5.3406 | | 5.0309 | 0.58 | 1000 | 4.9285 | | 4.7101 | 0.87 | 1500 | 4.6879 | | 4.4621 | 1.17 | 2000 | 4.5500 | | 4.2913 | 1.46 | 2500 | 4.4298 | | 4.2026 | 1.75 | 3000 | 4.3310 | | 4.0829 | 2.04 | 3500 | 4.2546 | | 3.8956 | 2.33 | 4000 | 4.2130 | | 3.8692 | 2.62 | 4500 | 4.1583 | | 3.8292 | 2.91 | 5000 | 4.1132 | | 3.6507 | 3.21 | 5500 | 4.1047 | | 3.5891 | 3.5 | 6000 | 4.0753 | | 3.5712 | 3.79 | 6500 | 4.0432 | | 3.4932 | 4.08 | 7000 | 4.0421 | | 3.3212 | 4.37 | 7500 | 4.0385 | | 3.3167 | 4.66 | 8000 | 4.0261 | | 3.3035 | 4.95 | 8500 | 4.0122 | | 3.1681 | 5.24 | 9000 | 4.0240 | | 3.1387 | 5.54 | 9500 | 4.0244 | | 3.1401 | 5.83 | 10000 | 4.0231 | ### Framework versions - Transformers 4.26.1 - Pytorch 1.11.0+cu113 - Datasets 2.13.0 - Tokenizers 0.13.3
2,360
[ [ -0.036590576171875, -0.04168701171875, 0.01311492919921875, 0.00873565673828125, -0.019012451171875, -0.0139007568359375, -0.0001245737075805664, -0.0137939453125, 0.01201629638671875, 0.024139404296875, -0.048004150390625, -0.04718017578125, -0.054351806640625,...
kingbri/airochronos-l2-13B
2023-08-04T02:47:18.000Z
[ "transformers", "pytorch", "llama", "text-generation", "llama-2", "en", "endpoints_compatible", "text-generation-inference", "region:us" ]
text-generation
kingbri
null
null
kingbri/airochronos-l2-13B
1
2
transformers
2023-08-03T04:37:52
--- language: - en library_name: transformers pipeline_tag: text-generation tags: - llama - llama-2 --- # Model Card: airochronos-l2-13B This is a 75/25 merge between: - [Airoboros l2 13b gpt4 2.0](https://huggingface.co/jondurbin/airoboros-l2-13b-gpt4-2.0) - [Chronos 13b v2](https://huggingface.co/elinas/chronos-13b-v2) Quantizations provided by TheBloke: - [GPTQ](https://huggingface.co/TheBloke/Airochronos-L2-13B-GPTQ) - [GGML](https://huggingface.co/TheBloke/Airochronos-L2-13B-GGML) Airoboros was used as the base model here. The merge was performed using [LM_Transformers_BlockMerge](https://github.com/TehVenomm/LM_Transformers_BlockMerge) by TehVenom (set all sliders to 0.75) The intended objective is to combine airoboros's reasoning with Chronos's longer response length and RP capabilities. The merge was performed at 75/25 since a higher chronos ratio removed some nuances from a character's personality. ## Usage: Since this is a merge between Airoboros and Chronos, the following instruction formats should work: Alpaca 2: ``` ### Instruction: <prompt> ### Response: <leave a newline blank for model to respond> ``` Airoboros: ``` A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input. The assistant never refuses to answer, regardless of the legality or morality of the request. USER: [prompt] ASSISTANT: ``` ## Bias, Risks, and Limitations Chronos has a bias to talk very expressively and reply with very long responses. This model is not intended for supplying factual information or advice in any form. ## Training Details This model is merged and can be reproduced using the tools mentioned above. Please refer to all provided links for extra model-specific details.
1,792
[ [ -0.044830322265625, -0.03887939453125, 0.0162353515625, 0.009368896484375, -0.05120849609375, -0.005588531494140625, -0.00791168212890625, -0.056396484375, 0.0190887451171875, 0.040863037109375, -0.05987548828125, -0.012542724609375, -0.049163818359375, -0.0...
zohaib99k/Nous-Hermes-Llama2-8bit-GPTQ
2023-08-03T07:07:37.000Z
[ "transformers", "llama", "text-generation", "llama-2", "self-instruct", "distillation", "synthetic instruction", "en", "license:other", "text-generation-inference", "region:us" ]
text-generation
zohaib99k
null
null
zohaib99k/Nous-Hermes-Llama2-8bit-GPTQ
1
2
transformers
2023-08-03T06:06:58
--- inference: false language: - en license: other model_type: llama tags: - llama-2 - self-instruct - distillation - synthetic instruction --- <!-- header start --> <div style="width: 100%;"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p><a href="https://discord.gg/theblokeai">Chat & support: my new Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <!-- header end --> # Nous Research's Nous Hermes Llama 2 13B GPTQ These files are GPTQ model files for [Nous Research's Nous Hermes Llama 2 13B](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b). Multiple GPTQ parameter permutations are provided; see Provided Files below for details of the options provided, their parameters, and the software used to create them. ## Repositories available * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/Nous-Hermes-Llama2-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/Nous-Hermes-Llama2-GGML) * [Original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b) ## Prompt template: Alpaca ``` Below is an instruction that describes a task. Write a response that appropriately completes the request. ### Instruction: {prompt} ### Response: ``` ## Provided files Multiple quantisation parameters are provided, to allow you to choose the best one for your hardware and requirements. Each separate quant is in a different branch. See below for instructions on fetching from different branches. | Branch | Bits | Group Size | Act Order (desc_act) | File Size | ExLlama Compatible? | Made With | Description | | ------ | ---- | ---------- | -------------------- | --------- | ------------------- | --------- | ----------- | | main | 4 | 128 | False | 7.26 GB | True | AutoGPTQ | Most compatible option. Good inference speed in AutoGPTQ and GPTQ-for-LLaMa. Lower inference quality than other options. | | gptq-4bit-32g-actorder_True | 4 | 32 | True | 8.00 GB | True | AutoGPTQ | 4-bit, with Act Order and group size. 32g gives highest possible inference quality, with maximum VRAM usage. Poor AutoGPTQ CUDA speed. | | gptq-4bit-64g-actorder_True | 4 | 64 | True | 7.51 GB | True | AutoGPTQ | 4-bit, with Act Order and group size. 64g uses less VRAM than 32g, but with slightly lower accuracy. Poor AutoGPTQ CUDA speed. | | gptq-4bit-128g-actorder_True | 4 | 128 | True | 7.26 GB | True | AutoGPTQ | 4-bit, with Act Order and group size. 128g uses even less VRAM, but with slightly lower accuracy. Poor AutoGPTQ CUDA speed. | | gptq-8bit-128g-actorder_True | 8 | 128 | True | 13.65 GB | False | AutoGPTQ | 8-bit, with group size 128g for higher inference quality and with Act Order for even higher accuracy. Poor AutoGPTQ CUDA speed. | | gptq-8bit-64g-actorder_True | 8 | 64 | True | 13.95 GB | False | AutoGPTQ | 8-bit, with group size 64g and Act Order for maximum inference quality. Poor AutoGPTQ CUDA speed. | | gptq-8bit-128g-actorder_False | 8 | 128 | False | 13.65 GB | False | AutoGPTQ | 8-bit, with group size 128g for higher inference quality and without Act Order to improve AutoGPTQ speed. | | gptq-8bit--1g-actorder_True | 8 | None | True | 13.36 GB | False | AutoGPTQ | 8-bit, with Act Order. No group size, to lower VRAM requirements and to improve AutoGPTQ speed. | ## How to download from branches - In text-generation-webui, you can add `:branch` to the end of the download name, eg `TheBloke/Nous-Hermes-Llama2-GPTQ:gptq-4bit-32g-actorder_True` - With Git, you can clone a branch with: ``` git clone --branch gptq-4bit-32g-actorder_True https://huggingface.co/TheBloke/Nous-Hermes-Llama2-GPTQ` ``` - In Python Transformers code, the branch is the `revision` parameter; see below. ## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui). Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui). It is strongly recommended to use the text-generation-webui one-click-installers unless you know how to make a manual install. 1. Click the **Model tab**. 2. Under **Download custom model or LoRA**, enter `TheBloke/Nous-Hermes-Llama2-GPTQ`. - To download from a specific branch, enter for example `TheBloke/Nous-Hermes-Llama2-GPTQ:gptq-4bit-32g-actorder_True` - see Provided Files above for the list of branches for each option. 3. Click **Download**. 4. The model will start downloading. Once it's finished it will say "Done" 5. In the top left, click the refresh icon next to **Model**. 6. In the **Model** dropdown, choose the model you just downloaded: `Nous-Hermes-Llama2-GPTQ` 7. The model will automatically load, and is now ready for use! 8. If you want any custom settings, set them and then click **Save settings for this model** followed by **Reload the Model** in the top right. * Note that you do not need to set GPTQ parameters any more. These are set automatically from the file `quantize_config.json`. 9. Once you're ready, click the **Text Generation tab** and enter a prompt to get started! ## How to use this GPTQ model from Python code First make sure you have [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ) installed: `GITHUB_ACTIONS=true pip install auto-gptq` Then try the following example code: ```python from transformers import AutoTokenizer, pipeline, logging from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig model_name_or_path = "TheBloke/Nous-Hermes-Llama2-GPTQ" model_basename = "gptq_model-4bit-128g" use_triton = False tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True) model = AutoGPTQForCausalLM.from_quantized(model_name_or_path, model_basename=model_basename, use_safetensors=True, trust_remote_code=False, device="cuda:0", use_triton=use_triton, quantize_config=None) """ To download from a specific branch, use the revision parameter, as in this example: model = AutoGPTQForCausalLM.from_quantized(model_name_or_path, revision="gptq-4bit-32g-actorder_True", model_basename=model_basename, use_safetensors=True, trust_remote_code=False, device="cuda:0", quantize_config=None) """ prompt = "Tell me about AI" prompt_template=f'''Below is an instruction that describes a task. Write a response that appropriately completes the request. ### Instruction: {prompt} ### Response: ''' print("\n\n*** Generate:") input_ids = tokenizer(prompt_template, return_tensors='pt').input_ids.cuda() output = model.generate(inputs=input_ids, temperature=0.7, max_new_tokens=512) print(tokenizer.decode(output[0])) # Inference can also be done using transformers' pipeline # Prevent printing spurious transformers error when using pipeline with AutoGPTQ logging.set_verbosity(logging.CRITICAL) print("*** Pipeline:") pipe = pipeline( "text-generation", model=model, tokenizer=tokenizer, max_new_tokens=512, temperature=0.7, top_p=0.95, repetition_penalty=1.15 ) print(pipe(prompt_template)[0]['generated_text']) ``` ## Compatibility The files provided will work with AutoGPTQ (CUDA and Triton modes), GPTQ-for-LLaMa (only CUDA has been tested), and Occ4m's GPTQ-for-LLaMa fork. ExLlama works with Llama models in 4-bit. Please see the Provided Files table above for per-file compatibility. <!-- footer start --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/theblokeai) ## Thanks, and how to contribute. Thanks to the [chirper.ai](https://chirper.ai) team! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Luke from CarbonQuill, Aemon Algiz. **Patreon special mentions**: Slarti, Chadd, John Detwiler, Pieter, zynix, K, Mano Prime, ReadyPlayerEmma, Ai Maven, Leonard Tan, Edmond Seymore, Joseph William Delisle, Luke @flexchar, Fred von Graf, Viktor Bowallius, Rishabh Srivastava, Nikolai Manek, Matthew Berman, Johann-Peter Hartmann, ya boyyy, Greatston Gnanesh, Femi Adebogun, Talal Aujan, Jonathan Leane, terasurfer, David Flickinger, William Sang, Ajan Kanaga, Vadim, Artur Olbinski, Raven Klaugh, Michael Levine, Oscar Rangel, Randy H, Cory Kujawski, RoA, Dave, Alex, Alexandros Triantafyllidis, Fen Risland, Eugene Pentland, vamX, Elle, Nathan LeClaire, Khalefa Al-Ahmad, Rainer Wilmers, subjectnull, Junyu Yang, Daniel P. Andersen, SuperWojo, LangChain4j, Mandus, Kalila, Illia Dulskyi, Trenton Dambrowitz, Asp the Wyvern, Derek Yates, Jeffrey Morgan, Deep Realms, Imad Khwaja, Pyrater, Preetika Verma, biorpg, Gabriel Tamborski, Stephen Murray, Spiking Neurons AB, Iucharbius, Chris Smitley, Willem Michiel, Luke Pendergrass, Sebastain Graf, senxiiz, Will Dee, Space Cruiser, Karl Bernard, Clay Pascal, Lone Striker, transmissions 11, webtim, WelcomeToTheClub, Sam, theTransient, Pierre Kircher, chris gileta, John Villwock, Sean Connelly, Willian Hasse Thank you to all my generous patrons and donaters! <!-- footer end --> # Original model card: Nous Research's Nous Hermes Llama 2 13B # Model Card: Nous-Hermes-Llama2-13b Compute provided by our project sponsor Redmond AI, thank you! Follow RedmondAI on Twitter @RedmondAI. ## Model Description Nous-Hermes-Llama2-13b is a state-of-the-art language model fine-tuned on over 300,000 instructions. This model was fine-tuned by Nous Research, with Teknium and Emozilla leading the fine tuning process and dataset curation, Redmond AI sponsoring the compute, and several other contributors. This Hermes model uses the exact same dataset as Hermes on Llama-1. This is to ensure consistency between the old Hermes and new, for anyone who wanted to keep Hermes as similar to the old one, just more capable. This model stands out for its long responses, lower hallucination rate, and absence of OpenAI censorship mechanisms. The fine-tuning process was performed with a 4096 sequence length on an 8x a100 80GB DGX machine. ## Example Outputs: ![Example4](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b/resolve/main/example5.png "Example 4") ![Example1](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b/resolve/main/Example1.png "Example 1") ![Example2](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b/resolve/main/example2.png "Example 2") ![Example3](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b/resolve/main/example3.png "Example 3") ## Model Training The model was trained almost entirely on synthetic GPT-4 outputs. Curating high quality GPT-4 datasets enables incredibly high quality in knowledge, task completion, and style. This includes data from diverse sources such as GPTeacher, the general, roleplay v1&2, code instruct datasets, Nous Instruct & PDACTL (unpublished), and several others, detailed further below ## Collaborators The model fine-tuning and the datasets were a collaboration of efforts and resources between Teknium, Karan4D, Emozilla, Huemin Art, and Redmond AI. Special mention goes to @winglian for assisting in some of the training issues. Huge shoutout and acknowledgement is deserved for all the dataset creators who generously share their datasets openly. Among the contributors of datasets: - GPTeacher was made available by Teknium - Wizard LM by nlpxucan - Nous Research Instruct Dataset was provided by Karan4D and HueminArt. - GPT4-LLM and Unnatural Instructions were provided by Microsoft - Airoboros dataset by jondurbin - Camel-AI's domain expert datasets are from Camel-AI - CodeAlpaca dataset by Sahil 2801. If anyone was left out, please open a thread in the community tab. ## Prompt Format The model follows the Alpaca prompt format: ``` ### Instruction: <prompt> ### Response: <leave a newline blank for model to respond> ``` or ``` ### Instruction: <prompt> ### Input: <additional context> ### Response: <leave a newline blank for model to respond> ``` ## Benchmark Results AGI-Eval ``` | Task |Version| Metric |Value | |Stderr| |agieval_aqua_rat | 0|acc |0.2362|± |0.0267| | | |acc_norm|0.2480|± |0.0272| |agieval_logiqa_en | 0|acc |0.3425|± |0.0186| | | |acc_norm|0.3472|± |0.0187| |agieval_lsat_ar | 0|acc |0.2522|± |0.0287| | | |acc_norm|0.2087|± |0.0269| |agieval_lsat_lr | 0|acc |0.3510|± |0.0212| | | |acc_norm|0.3627|± |0.0213| |agieval_lsat_rc | 0|acc |0.4647|± |0.0305| | | |acc_norm|0.4424|± |0.0303| |agieval_sat_en | 0|acc |0.6602|± |0.0331| | | |acc_norm|0.6165|± |0.0340| |agieval_sat_en_without_passage| 0|acc |0.4320|± |0.0346| | | |acc_norm|0.4272|± |0.0345| |agieval_sat_math | 0|acc |0.2909|± |0.0307| | | |acc_norm|0.2727|± |0.0301| ``` GPT-4All Benchmark Set ``` | Task |Version| Metric |Value | |Stderr| |arc_challenge| 0|acc |0.5102|± |0.0146| | | |acc_norm|0.5213|± |0.0146| |arc_easy | 0|acc |0.7959|± |0.0083| | | |acc_norm|0.7567|± |0.0088| |boolq | 1|acc |0.8394|± |0.0064| |hellaswag | 0|acc |0.6164|± |0.0049| | | |acc_norm|0.8009|± |0.0040| |openbookqa | 0|acc |0.3580|± |0.0215| | | |acc_norm|0.4620|± |0.0223| |piqa | 0|acc |0.7992|± |0.0093| | | |acc_norm|0.8069|± |0.0092| |winogrande | 0|acc |0.7127|± |0.0127| ``` BigBench Reasoning Test ``` | Task |Version| Metric |Value | |Stderr| |bigbench_causal_judgement | 0|multiple_choice_grade|0.5526|± |0.0362| |bigbench_date_understanding | 0|multiple_choice_grade|0.7344|± |0.0230| |bigbench_disambiguation_qa | 0|multiple_choice_grade|0.2636|± |0.0275| |bigbench_geometric_shapes | 0|multiple_choice_grade|0.0195|± |0.0073| | | |exact_str_match |0.0000|± |0.0000| |bigbench_logical_deduction_five_objects | 0|multiple_choice_grade|0.2760|± |0.0200| |bigbench_logical_deduction_seven_objects | 0|multiple_choice_grade|0.2100|± |0.0154| |bigbench_logical_deduction_three_objects | 0|multiple_choice_grade|0.4400|± |0.0287| |bigbench_movie_recommendation | 0|multiple_choice_grade|0.2440|± |0.0192| |bigbench_navigate | 0|multiple_choice_grade|0.4950|± |0.0158| |bigbench_reasoning_about_colored_objects | 0|multiple_choice_grade|0.5570|± |0.0111| |bigbench_ruin_names | 0|multiple_choice_grade|0.3728|± |0.0229| |bigbench_salient_translation_error_detection | 0|multiple_choice_grade|0.1854|± |0.0123| |bigbench_snarks | 0|multiple_choice_grade|0.6298|± |0.0360| |bigbench_sports_understanding | 0|multiple_choice_grade|0.6156|± |0.0155| |bigbench_temporal_sequences | 0|multiple_choice_grade|0.3140|± |0.0147| |bigbench_tracking_shuffled_objects_five_objects | 0|multiple_choice_grade|0.2032|± |0.0114| |bigbench_tracking_shuffled_objects_seven_objects| 0|multiple_choice_grade|0.1406|± |0.0083| |bigbench_tracking_shuffled_objects_three_objects| 0|multiple_choice_grade|0.4400|± |0.0287| ``` These are the highest benchmarks Hermes has seen on every metric, achieving the following average scores: - GPT4All benchmark average is now 70.0 - from 68.8 in Hermes-Llama1 - 0.3657 on BigBench, up from 0.328 on hermes-llama1 - 0.372 on AGIEval, up from 0.354 on Hermes-llama1 These benchmarks currently have us at #1 on ARC-c, ARC-e, Hellaswag, and OpenBookQA, and 2nd place on Winogrande, comparing to GPT4all's benchmarking list, supplanting Hermes 1 for the new top position. ## Resources for Applied Use Cases: For an example of a back and forth chatbot using huggingface transformers and discord, check out: https://github.com/teknium1/alpaca-discord For an example of a roleplaying discord chatbot, check out this: https://github.com/teknium1/alpaca-roleplay-discordbot ## Future Plans We plan to continue to iterate on both more high quality data, and new data filtering techniques to eliminate lower quality data going forward. ## Model Usage The model is available for download on Hugging Face. It is suitable for a wide range of language tasks, from generating creative text to understanding and following complex instructions.
18,137
[ [ -0.035308837890625, -0.06927490234375, 0.0190277099609375, 0.01441192626953125, -0.02496337890625, -0.0101165771484375, 0.00606536865234375, -0.0294189453125, 0.00653076171875, 0.0272979736328125, -0.0377197265625, -0.03253173828125, -0.03271484375, 0.005535...
alandevkota/vit-base-patch16-224-finetuned-flower
2023-08-03T19:16:17.000Z
[ "transformers", "pytorch", "vit", "image-classification", "generated_from_trainer", "dataset:imagefolder", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
image-classification
alandevkota
null
null
alandevkota/vit-base-patch16-224-finetuned-flower
0
2
transformers
2023-08-03T07:13:41
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imagefolder model-index: - name: vit-base-patch16-224-finetuned-flower results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vit-base-patch16-224-finetuned-flower This model is a fine-tuned version of [google/vit-base-patch16-224](https://huggingface.co/google/vit-base-patch16-224) on the imagefolder dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results ### Framework versions - Transformers 4.24.0 - Pytorch 2.0.1+cu118 - Datasets 2.7.1 - Tokenizers 0.13.3
1,119
[ [ -0.030731201171875, -0.060455322265625, 0.007099151611328125, 0.020416259765625, -0.0300445556640625, -0.03887939453125, -0.01383209228515625, -0.0225677490234375, 0.00708770751953125, 0.023284912109375, -0.05780029296875, -0.036376953125, -0.04339599609375, ...
Sookeyy/ppo-PyramidsRND
2023-08-03T17:22:54.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
Sookeyy
null
null
Sookeyy/ppo-PyramidsRND
0
2
ml-agents
2023-08-03T17:20:54
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: Sookeyy/ppo-PyramidsRND 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,335
[ [ -0.040130615234375, -0.0343017578125, 0.0025463104248046875, 0.01342010498046875, -0.01062774658203125, 0.01220703125, 0.017181396484375, -0.01456451416015625, 0.033172607421875, 0.030242919921875, -0.041412353515625, -0.049896240234375, -0.029205322265625, ...
Valent2809/news_classifier_regulatory_approval
2023-08-04T13:32:51.000Z
[ "transformers", "tf", "distilbert", "text-classification", "generated_from_keras_callback", "license:apache-2.0", "endpoints_compatible", "region:us" ]
text-classification
Valent2809
null
null
Valent2809/news_classifier_regulatory_approval
0
2
transformers
2023-08-04T02:31:37
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: Valent2809/news_classifier_regulatory_approval results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # Valent2809/news_classifier_regulatory_approval This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.0238 - Validation Loss: 0.0613 - Train Accuracy: 0.9817 - Epoch: 4 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': False, 'is_legacy_optimizer': False, 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 1770, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False} - training_precision: float32 ### Training results | Train Loss | Validation Loss | Train Accuracy | Epoch | |:----------:|:---------------:|:--------------:|:-----:| | 0.1867 | 0.0914 | 0.9711 | 0 | | 0.0745 | 0.0747 | 0.9782 | 1 | | 0.0515 | 0.0640 | 0.9845 | 2 | | 0.0345 | 0.0573 | 0.9852 | 3 | | 0.0238 | 0.0613 | 0.9817 | 4 | ### Framework versions - Transformers 4.30.2 - TensorFlow 2.12.0 - Datasets 2.13.1 - Tokenizers 0.13.3
1,988
[ [ -0.03973388671875, -0.035675048828125, 0.019073486328125, -0.00008785724639892578, -0.0247650146484375, -0.0164031982421875, -0.0109100341796875, -0.01105499267578125, 0.0020122528076171875, 0.005786895751953125, -0.04351806640625, -0.052764892578125, -0.0596618...
biu-nlp/QAmden-multinews
2023-08-04T11:51:45.000Z
[ "transformers", "pytorch", "led", "text2text-generation", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
biu-nlp
null
null
biu-nlp/QAmden-multinews
0
2
transformers
2023-08-04T08:23:49
--- language: en license: apache-2.0 --- # 🏬QAmden🏬: Question-Answering-based Multi-DocumENt model HF-version of the QAmden model fine-tuned over multinews: *Peek Across*: Improving Multi-Document Modeling via Cross-Document Question-Answering (ACL 2023). You can use it by ```python from transformers import ( AutoTokenizer, LEDConfig, LEDForConditionalGeneration, ) # load model and tokenizer tokenizer = AutoTokenizer.from_pretrained('biu-nlp/QAmden') config=LEDConfig.from_pretrained('biu-nlp/QAmden-multinews') model = LEDForConditionalGeneration.from_pretrained('biu-nlp/QAmden-multinews') ``` The original repo is [here](https://github.com/aviclu/peekacross). If you find our work useful, please cite the paper as: ```python @article{caciularu2023peekacross, title={Peek Across: Improving Multi-Document Modeling via Cross-Document Question-Answering}, author={Caciularu, Avi and Peters, Matthew E and Goldberger, Jacob and Dagan, Ido and Cohan, Arman}, journal={The 61st Annual Meeting of the Association for Computational Linguistics: ACL 2023}, year={2023} } ```
1,107
[ [ -0.0169219970703125, -0.057464599609375, 0.037109375, 0.01885986328125, -0.011474609375, 0.015777587890625, 0.01082611083984375, -0.035797119140625, -0.0111083984375, 0.0248565673828125, -0.052093505859375, -0.031768798828125, -0.04046630859375, 0.0074501037...
maurope/distilroberta-base-mrpc-glue-maurope
2023-08-04T10:04:27.000Z
[ "transformers", "pytorch", "tensorboard", "roberta", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
text-classification
maurope
null
null
maurope/distilroberta-base-mrpc-glue-maurope
0
2
transformers
2023-08-04T09:48:38
--- license: apache-2.0 tags: - text-classification - generated_from_trainer datasets: - glue metrics: - accuracy - f1 model-index: - name: distilroberta-base-mrpc-glue-maurope results: - task: name: Text Classification type: text-classification dataset: name: datasetX type: glue config: mrpc split: validation args: mrpc metrics: - name: Accuracy type: accuracy value: 0.8382352941176471 - name: F1 type: f1 value: 0.88339222614841 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilroberta-base-mrpc-glue-maurope This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on the datasetX dataset. It achieves the following results on the evaluation set: - Loss: 0.5374 - Accuracy: 0.8382 - F1: 0.8834 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.5161 | 1.09 | 500 | 0.5374 | 0.8382 | 0.8834 | | 0.3427 | 2.18 | 1000 | 0.6399 | 0.8358 | 0.8771 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.14.3 - Tokenizers 0.13.3
1,852
[ [ -0.032257080078125, -0.043548583984375, 0.00757598876953125, 0.0172882080078125, -0.026458740234375, -0.0223236083984375, -0.006046295166015625, -0.00537109375, 0.005527496337890625, 0.01214599609375, -0.049163818359375, -0.040863037109375, -0.060699462890625, ...
sohailsiddiqui/bert-finetuned-ner
2023-08-04T10:19:05.000Z
[ "transformers", "tf", "bert", "token-classification", "generated_from_keras_callback", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
sohailsiddiqui
null
null
sohailsiddiqui/bert-finetuned-ner
0
2
transformers
2023-08-04T09:50:01
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: sohailsiddiq99/bert-finetuned-ner results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # sohailsiddiq99/bert-finetuned-ner This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.1736 - Validation Loss: 0.0657 - Epoch: 0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 2634, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01} - training_precision: mixed_float16 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 0.1736 | 0.0657 | 0 | ### Framework versions - Transformers 4.30.2 - TensorFlow 2.12.0 - Datasets 2.14.0 - Tokenizers 0.11.0
1,487
[ [ -0.0478515625, -0.04949951171875, 0.0165557861328125, 0.006519317626953125, -0.03802490234375, -0.03570556640625, -0.0196075439453125, -0.0158843994140625, 0.00600433349609375, 0.0151519775390625, -0.057861328125, -0.040985107421875, -0.04925537109375, -0.02...
Anis-Bouhamadouche/distilbert-base-uncased-finetuned-emotion
2023-10-25T22:32:39.000Z
[ "transformers", "pytorch", "safetensors", "distilbert", "text-classification", "generated_from_trainer", "dataset:emotion", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
text-classification
Anis-Bouhamadouche
null
null
Anis-Bouhamadouche/distilbert-base-uncased-finetuned-emotion
0
2
transformers
2023-08-04T10:05:01
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer datasets: - emotion metrics: - accuracy - f1 model-index: - name: distilbert-base-uncased-finetuned-emotion results: - task: name: Text Classification type: text-classification dataset: name: emotion type: emotion config: split split: validation args: split metrics: - name: Accuracy type: accuracy value: 0.925 - name: F1 type: f1 value: 0.9249367490708449 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2105 - Accuracy: 0.925 - F1: 0.9249 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.8223 | 1.0 | 250 | 0.3098 | 0.9085 | 0.9076 | | 0.2431 | 2.0 | 500 | 0.2105 | 0.925 | 0.9249 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.3 - Tokenizers 0.13.3
1,882
[ [ -0.037750244140625, -0.041046142578125, 0.01474761962890625, 0.021881103515625, -0.0264892578125, -0.0191802978515625, -0.01293182373046875, -0.00858306884765625, 0.0102691650390625, 0.0081329345703125, -0.056488037109375, -0.05133056640625, -0.05975341796875, ...
Yuyang-z/zero123-xl
2023-08-05T02:13:46.000Z
[ "diffusers", "license:mit", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
null
Yuyang-z
null
null
Yuyang-z/zero123-xl
0
2
diffusers
2023-08-04T11:09:24
--- license: mit --- Upload zero123-xl.ckpt, converted by diffusers scripts convert_original_stable_diffusion_to_diffusers.py [Zero123-hf](https://github.com/kxhit/zero123_hf) implemented with diffusers pipelines. Thanks Original Repo [Zero123](https://github.com/cvlab-columbia/zero123), and [Weights](https://huggingface.co/cvlab/zero123-weights).
352
[ [ -0.003997802734375, -0.020172119140625, 0.057708740234375, 0.05804443359375, -0.0097503662109375, -0.0207672119140625, -0.005069732666015625, 0.0087432861328125, 0.02606201171875, 0.048553466796875, -0.0235748291015625, -0.028594970703125, -0.033538818359375, ...
juancopi81/speecht5_finetuned_voxpopuli_es
2023-09-15T11:52:38.000Z
[ "transformers", "pytorch", "tensorboard", "safetensors", "speecht5", "text-to-audio", "generated_from_trainer", "text-to-speech", "dataset:voxpopuli", "license:mit", "endpoints_compatible", "has_space", "region:us" ]
text-to-speech
juancopi81
null
null
juancopi81/speecht5_finetuned_voxpopuli_es
0
2
transformers
2023-08-04T13:53:17
--- license: mit base_model: microsoft/speecht5_tts tags: - generated_from_trainer - text-to-speech datasets: - voxpopuli model-index: - name: speecht5_finetuned_voxpopuli_es results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # speecht5_finetuned_voxpopuli_es This model is a fine-tuned version of [microsoft/speecht5_tts](https://huggingface.co/microsoft/speecht5_tts) on the voxpopuli dataset. It achieves the following results on the evaluation set: - Loss: 0.4454 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 4 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 4000 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.5097 | 4.32 | 1000 | 0.4626 | | 0.4842 | 8.64 | 2000 | 0.4507 | | 0.4828 | 12.97 | 3000 | 0.4483 | | 0.4807 | 17.29 | 4000 | 0.4454 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.3 - Tokenizers 0.13.3
1,616
[ [ -0.03125, -0.041595458984375, -0.0036563873291015625, 0.006900787353515625, -0.0192718505859375, -0.0229339599609375, -0.0139617919921875, -0.00894927978515625, -0.009979248046875, 0.0196380615234375, -0.049468994140625, -0.05010986328125, -0.042999267578125, ...
mory91/whisper-tiny-en
2023-08-14T22:59:46.000Z
[ "transformers", "pytorch", "tensorboard", "whisper", "automatic-speech-recognition", "generated_from_trainer", "dataset:PolyAI/minds14", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
mory91
null
null
mory91/whisper-tiny-en
0
2
transformers
2023-08-04T23:04:20
--- license: apache-2.0 base_model: openai/whisper-tiny tags: - generated_from_trainer datasets: - PolyAI/minds14 metrics: - wer model-index: - name: whisper-tiny-en results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: PolyAI/minds14 type: PolyAI/minds14 config: en-US split: train args: en-US metrics: - name: Wer type: wer value: 0.2828154724159797 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # whisper-tiny-en This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the PolyAI/minds14 dataset. It achieves the following results on the evaluation set: - Loss: 0.6966 - Wer Ortho: 0.2806 - Wer: 0.2828 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant_with_warmup - lr_scheduler_warmup_steps: 50 - training_steps: 2000 ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:| | 0.0008 | 17.24 | 500 | 0.5855 | 0.2740 | 0.2758 | | 0.0002 | 34.48 | 1000 | 0.6411 | 0.2780 | 0.2809 | | 0.0001 | 51.72 | 1500 | 0.6713 | 0.2793 | 0.2815 | | 0.0001 | 68.97 | 2000 | 0.6966 | 0.2806 | 0.2828 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.4 - Tokenizers 0.13.3
1,994
[ [ -0.034912109375, -0.034210205078125, 0.0121002197265625, 0.005260467529296875, -0.02435302734375, -0.04296875, -0.0192718505859375, -0.0178985595703125, 0.01389312744140625, 0.01305389404296875, -0.05615234375, -0.03302001953125, -0.0438232421875, -0.0155563...
mikecamara/vit-base-patch16-224-finetuned-flower
2023-08-05T00:44:30.000Z
[ "transformers", "pytorch", "vit", "image-classification", "generated_from_trainer", "dataset:imagefolder", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
image-classification
mikecamara
null
null
mikecamara/vit-base-patch16-224-finetuned-flower
0
2
transformers
2023-08-05T00:33:22
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imagefolder model-index: - name: vit-base-patch16-224-finetuned-flower results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vit-base-patch16-224-finetuned-flower This model is a fine-tuned version of [google/vit-base-patch16-224](https://huggingface.co/google/vit-base-patch16-224) on the imagefolder dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results ### Framework versions - Transformers 4.24.0 - Pytorch 2.0.1+cu118 - Datasets 2.7.1 - Tokenizers 0.13.3
1,119
[ [ -0.0307769775390625, -0.06048583984375, 0.007091522216796875, 0.0204315185546875, -0.030059814453125, -0.038909912109375, -0.01384735107421875, -0.0225677490234375, 0.00707244873046875, 0.0233001708984375, -0.057830810546875, -0.036407470703125, -0.0434265136718...
BlackSwan1827/ppo-Pyramids
2023-08-05T04:15:32.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
BlackSwan1827
null
null
BlackSwan1827/ppo-Pyramids
0
2
ml-agents
2023-08-05T04:15:29
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: BlackSwan1827/ppo-Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,338
[ [ -0.04058837890625, -0.033416748046875, 0.00176239013671875, 0.0140380859375, -0.01096343994140625, 0.01244354248046875, 0.016876220703125, -0.015228271484375, 0.03363037109375, 0.029998779296875, -0.040130615234375, -0.0494384765625, -0.030120849609375, -0.0...
gaodrew/git-base-pokemon
2023-08-05T08:51:53.000Z
[ "transformers", "pytorch", "tensorboard", "git", "text-generation", "generated_from_trainer", "license:mit", "endpoints_compatible", "region:us" ]
text-generation
gaodrew
null
null
gaodrew/git-base-pokemon
0
2
transformers
2023-08-05T08:06:54
--- license: mit base_model: microsoft/git-base tags: - generated_from_trainer model-index: - name: git-base-pokemon results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # git-base-pokemon This model is a fine-tuned version of [microsoft/git-base](https://huggingface.co/microsoft/git-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0348 - Wer Score: 2.7147 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer Score | |:-------------:|:-----:|:----:|:---------------:|:---------:| | 7.3601 | 4.17 | 50 | 4.5925 | 21.8560 | | 2.4331 | 8.33 | 100 | 0.4978 | 15.2153 | | 0.1504 | 12.5 | 150 | 0.0323 | 1.2062 | | 0.0142 | 16.67 | 200 | 0.0288 | 3.0791 | | 0.0039 | 20.83 | 250 | 0.0314 | 2.3619 | | 0.0021 | 25.0 | 300 | 0.0327 | 2.6537 | | 0.0016 | 29.17 | 350 | 0.0333 | 3.2049 | | 0.0014 | 33.33 | 400 | 0.0344 | 2.9403 | | 0.0012 | 37.5 | 450 | 0.0344 | 2.9624 | | 0.0011 | 41.67 | 500 | 0.0345 | 2.8106 | | 0.0011 | 45.83 | 550 | 0.0346 | 2.7393 | | 0.0011 | 50.0 | 600 | 0.0348 | 2.7147 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.3 - Tokenizers 0.13.3
2,092
[ [ -0.0309906005859375, -0.04180908203125, 0.00949859619140625, -0.0017175674438476562, -0.0163116455078125, -0.006839752197265625, -0.0061187744140625, -0.00630950927734375, 0.029388427734375, 0.0179595947265625, -0.045684814453125, -0.04180908203125, -0.059082031...
diegomiranda/EleutherAI-70M-cypher-generator
2023-08-17T00:18:49.000Z
[ "transformers", "pytorch", "gpt_neox", "text-generation", "gpt", "llm", "large language model", "h2o-llmstudio", "en", "text-generation-inference", "region:us" ]
text-generation
diegomiranda
null
null
diegomiranda/EleutherAI-70M-cypher-generator
0
2
transformers
2023-08-05T16:27:29
--- language: - en library_name: transformers tags: - gpt - llm - large language model - h2o-llmstudio inference: false thumbnail: https://h2o.ai/etc.clientlibs/h2o/clientlibs/clientlib-site/resources/images/favicon.ico --- # Model Card ## Summary This model was trained using [H2O LLM Studio](https://github.com/h2oai/h2o-llmstudio). - Base model: [EleutherAI/pythia-70m-deduped-v0](https://huggingface.co/EleutherAI/pythia-70m-deduped-v0) ## Usage To use the model with the `transformers` library on a machine with GPUs, first make sure you have the `transformers`, `accelerate` and `torch` libraries installed. ```bash pip install transformers==4.30.2 pip install einops== pip install accelerate==0.20.3 pip install torch==2.0.0 ``` ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch def generate_response(prompt, model_name): tokenizer = AutoTokenizer.from_pretrained( model_name, use_fast=True, trust_remote_code=True, ) model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype=torch.float32, device_map={"": "cpu"}, trust_remote_code=True, ) model.cpu().eval() inputs = tokenizer(prompt, return_tensors="pt", add_special_tokens=False).to("cpu") tokens = model.generate( input_ids=inputs["input_ids"], attention_mask=inputs["attention_mask"], min_new_tokens=2, max_new_tokens=500, do_sample=False, num_beams=2, temperature=float(0.0), repetition_penalty=float(1.0), renormalize_logits=True )[0] tokens = tokens[inputs["input_ids"].shape[1]:] answer = tokenizer.decode(tokens, skip_special_tokens=True) return answer ``` # Example usage ```python model_name = "diegomiranda/EleutherAI-70M-cypher-generator" prompt = "Create a Cypher statement to answer the following question:Retorne os processos de Direito Tributário que se baseiam em lei 939 de 1992?<|endoftext|>" response = generate_response(prompt, model_name) print(response) ``` Alternatively, you can download [h2oai_pipeline.py](h2oai_pipeline.py), store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the `transformers` package, this will allow you to set `trust_remote_code=False`. ```python import torch from h2oai_pipeline import H2OTextGenerationPipeline from transformers import AutoModelForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained( "diegomiranda/EleutherAI-70M-cypher-generator", use_fast=True, padding_side="left", trust_remote_code=True, ) model = AutoModelForCausalLM.from_pretrained( "diegomiranda/EleutherAI-70M-cypher-generator", torch_dtype="auto", device_map={"": "cuda:0"}, trust_remote_code=True, ) generate_text = H2OTextGenerationPipeline(model=model, tokenizer=tokenizer) res = generate_text( "Why is drinking water so healthy?", min_new_tokens=2, max_new_tokens=500, do_sample=False, num_beams=2, temperature=float(0.0), repetition_penalty=float(1.0), renormalize_logits=True ) print(res[0]["generated_text"]) ``` You may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps: ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "diegomiranda/EleutherAI-70M-cypher-generator" # either local folder or huggingface model name # Important: The prompt needs to be in the same format the model was trained with. # You can find an example prompt in the experiment logs. prompt = "How are you?<|endoftext|>" tokenizer = AutoTokenizer.from_pretrained( model_name, use_fast=True, trust_remote_code=True, ) model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype="auto", device_map={"": "cuda:0"}, trust_remote_code=True, ) model.cuda().eval() inputs = tokenizer(prompt, return_tensors="pt", add_special_tokens=False).to("cuda") # generate configuration can be modified to your needs tokens = model.generate( input_ids=inputs["input_ids"], attention_mask=inputs["attention_mask"], min_new_tokens=2, max_new_tokens=500, do_sample=False, num_beams=2, temperature=float(0.0), repetition_penalty=float(1.0), renormalize_logits=True )[0] tokens = tokens[inputs["input_ids"].shape[1]:] answer = tokenizer.decode(tokens, skip_special_tokens=True) print(answer) ``` ## Quantization and sharding You can load the models using quantization by specifying ```load_in_8bit=True``` or ```load_in_4bit=True```. Also, sharding on multiple GPUs is possible by setting ```device_map=auto```. ## Model Architecture ``` GPTNeoXForCausalLM( (gpt_neox): GPTNeoXModel( (embed_in): Embedding(50304, 512) (layers): ModuleList( (0-5): 6 x GPTNeoXLayer( (input_layernorm): LayerNorm((512,), eps=1e-05, elementwise_affine=True) (post_attention_layernorm): LayerNorm((512,), eps=1e-05, elementwise_affine=True) (attention): GPTNeoXAttention( (rotary_emb): RotaryEmbedding() (query_key_value): Linear(in_features=512, out_features=1536, bias=True) (dense): Linear(in_features=512, out_features=512, bias=True) ) (mlp): GPTNeoXMLP( (dense_h_to_4h): Linear(in_features=512, out_features=2048, bias=True) (dense_4h_to_h): Linear(in_features=2048, out_features=512, bias=True) (act): GELUActivation() ) ) ) (final_layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True) ) (embed_out): Linear(in_features=512, out_features=50304, bias=False) ) ``` ## Model Configuration This model was trained using H2O LLM Studio and with the configuration in [cfg.yaml](cfg.yaml). Visit [H2O LLM Studio](https://github.com/h2oai/h2o-llmstudio) to learn how to train your own large language models. ## Disclaimer Please read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions. - Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints. - Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion. - Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model. - Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities. - Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues. - Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes. By using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it.
8,325
[ [ -0.0167694091796875, -0.06109619140625, 0.034912109375, 0.0143890380859375, -0.0244903564453125, -0.0089263916015625, -0.0244293212890625, -0.0187225341796875, 0.0066375732421875, 0.034210205078125, -0.036773681640625, -0.048248291015625, -0.046905517578125, ...
nokotin/pyramids
2023-08-05T16:46:54.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
nokotin
null
null
nokotin/pyramids
0
2
ml-agents
2023-08-05T16:46:46
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: nokotin/pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,328
[ [ -0.041107177734375, -0.03466796875, 0.0016641616821289062, 0.01432037353515625, -0.01025390625, 0.01197052001953125, 0.0168914794921875, -0.01507568359375, 0.033416748046875, 0.030853271484375, -0.040985107421875, -0.050994873046875, -0.028839111328125, -0.0...
TheBloke/Llama-2-70B-OASST-1-200-GGML
2023-09-27T13:01:05.000Z
[ "transformers", "llama", "sft", "en", "de", "es", "fr", "dataset:OpenAssistant/oasst1", "license:llama2", "text-generation-inference", "region:us" ]
null
TheBloke
null
null
TheBloke/Llama-2-70B-OASST-1-200-GGML
9
2
transformers
2023-08-05T21:38:39
--- language: - en - de - es - fr license: llama2 tags: - sft datasets: - OpenAssistant/oasst1 model_name: Open-Assistant Llama2 70B SFT OASST inference: false model_creator: Jordan Clive model_link: https://huggingface.co/jordiclive/Llama-2-70b-oasst-1-200 model_type: llama quantized_by: TheBloke base_model: jordiclive/Llama-2-70b-oasst-1-200 --- <!-- header start --> <!-- 200823 --> <div style="width: auto; margin-left: auto; margin-right: auto"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div> <hr style="margin-top: 1.0em; margin-bottom: 1.0em;"> <!-- header end --> # Open-Assistant Llama2 70B SFT OASST - GGML - Model creator: [Jordan Clive](https://huggingface.co/jordiclive) - Original model: [Open-Assistant Llama2 70B SFT OASST](https://huggingface.co/jordiclive/Llama-2-70b-oasst-1-200) ## Description This repo contains GGML format model files for [Jordan Clive's Open-Assistant Llama2 70B SFT OASST](https://huggingface.co/jordiclive/Llama-2-70b-oasst-1-200). ### Important note regarding GGML files. The GGML format has now been superseded by GGUF. As of August 21st 2023, [llama.cpp](https://github.com/ggerganov/llama.cpp) no longer supports GGML models. Third party clients and libraries are expected to still support it for a time, but many may also drop support. Please use the GGUF models instead. ### About GGML GPU acceleration is now available for Llama 2 70B GGML files, with both CUDA (NVidia) and Metal (macOS). The following clients/libraries are known to work with these files, including with GPU acceleration: * [llama.cpp](https://github.com/ggerganov/llama.cpp), commit `e76d630` and later. * [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI. * [KoboldCpp](https://github.com/LostRuins/koboldcpp), version 1.37 and later. A powerful GGML web UI, especially good for story telling. * [LM Studio](https://lmstudio.ai/), a fully featured local GUI with GPU acceleration for both Windows and macOS. Use 0.1.11 or later for macOS GPU acceleration with 70B models. * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), version 0.1.77 and later. A Python library with LangChain support, and OpenAI-compatible API server. * [ctransformers](https://github.com/marella/ctransformers), version 0.2.15 and later. A Python library with LangChain support, and OpenAI-compatible API server. ## Repositories available * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGUF) * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference (deprecated)](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML) * [Jordan Clive's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/jordiclive/Llama-2-70b-oasst-1-200) ## Prompt template: OpenAssistant ``` <|prompter|>{prompt}<|endoftext|><|assistant|> ``` <!-- compatibility_ggml start --> ## Compatibility ### Works with llama.cpp [commit `e76d630`](https://github.com/ggerganov/llama.cpp/commit/e76d630df17e235e6b9ef416c45996765d2e36fb) until August 21st, 2023 Will not work with `llama.cpp` after commit [dadbed99e65252d79f81101a392d0d6497b86caa](https://github.com/ggerganov/llama.cpp/commit/dadbed99e65252d79f81101a392d0d6497b86caa). For compatibility with latest llama.cpp, please use GGUF files instead. Or one of the other tools and libraries listed above. To use in llama.cpp, you must add `-gqa 8` argument. For other UIs and libraries, please check the docs. ## Explanation of the new k-quant methods <details> <summary>Click to see details</summary> The new methods available are: * GGML_TYPE_Q2_K - "type-1" 2-bit quantization in super-blocks containing 16 blocks, each block having 16 weight. Block scales and mins are quantized with 4 bits. This ends up effectively using 2.5625 bits per weight (bpw) * GGML_TYPE_Q3_K - "type-0" 3-bit quantization in super-blocks containing 16 blocks, each block having 16 weights. Scales are quantized with 6 bits. This end up using 3.4375 bpw. * GGML_TYPE_Q4_K - "type-1" 4-bit quantization in super-blocks containing 8 blocks, each block having 32 weights. Scales and mins are quantized with 6 bits. This ends up using 4.5 bpw. * GGML_TYPE_Q5_K - "type-1" 5-bit quantization. Same super-block structure as GGML_TYPE_Q4_K resulting in 5.5 bpw * GGML_TYPE_Q6_K - "type-0" 6-bit quantization. Super-blocks with 16 blocks, each block having 16 weights. Scales are quantized with 8 bits. This ends up using 6.5625 bpw * GGML_TYPE_Q8_K - "type-0" 8-bit quantization. Only used for quantizing intermediate results. The difference to the existing Q8_0 is that the block size is 256. All 2-6 bit dot products are implemented for this quantization type. Refer to the Provided Files table below to see what files use which methods, and how. </details> <!-- compatibility_ggml end --> ## Provided files | Name | Quant method | Bits | Size | Max RAM required | Use case | | ---- | ---- | ---- | ---- | ---- | ----- | | [llama-2-70b-oasst-1-200.ggmlv3.q2_K.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q2_K.bin) | q2_K | 2 | 28.96 GB| 31.46 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. | | [llama-2-70b-oasst-1-200.ggmlv3.q3_K_S.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q3_K_S.bin) | q3_K_S | 3 | 30.09 GB| 32.59 GB | New k-quant method. Uses GGML_TYPE_Q3_K for all tensors | | [llama-2-70b-oasst-1-200.ggmlv3.q3_K_M.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q3_K_M.bin) | q3_K_M | 3 | 33.39 GB| 35.89 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | [llama-2-70b-oasst-1-200.ggmlv3.q3_K_L.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q3_K_L.bin) | q3_K_L | 3 | 36.49 GB| 38.99 GB | New k-quant method. Uses GGML_TYPE_Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | [llama-2-70b-oasst-1-200.ggmlv3.q4_0.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q4_0.bin) | q4_0 | 4 | 38.80 GB| 41.30 GB | Original quant method, 4-bit. | | [llama-2-70b-oasst-1-200.ggmlv3.q4_K_S.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q4_K_S.bin) | q4_K_S | 4 | 39.18 GB| 41.68 GB | New k-quant method. Uses GGML_TYPE_Q4_K for all tensors | | [llama-2-70b-oasst-1-200.ggmlv3.q4_K_M.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q4_K_M.bin) | q4_K_M | 4 | 41.69 GB| 44.19 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q4_K | | [llama-2-70b-oasst-1-200.ggmlv3.q4_1.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q4_1.bin) | q4_1 | 4 | 43.12 GB| 45.62 GB | Original quant method, 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models. | | [llama-2-70b-oasst-1-200.ggmlv3.q5_0.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q5_0.bin) | q5_0 | 5 | 47.43 GB| 49.93 GB | Original quant method, 5-bit. Higher accuracy, higher resource usage and slower inference. | | [llama-2-70b-oasst-1-200.ggmlv3.q5_K_S.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q5_K_S.bin) | q5_K_S | 5 | 47.74 GB| 50.24 GB | New k-quant method. Uses GGML_TYPE_Q5_K for all tensors | | [llama-2-70b-oasst-1-200.ggmlv3.q5_K_M.bin](https://huggingface.co/TheBloke/Llama-2-70B-OASST-1-200-GGML/blob/main/llama-2-70b-oasst-1-200.ggmlv3.q5_K_M.bin) | q5_K_M | 5 | 49.03 GB| 51.53 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q5_K | **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead. ## How to run in `llama.cpp` Make sure you are using `llama.cpp` from commit [dadbed99e65252d79f81101a392d0d6497b86caa](https://github.com/ggerganov/llama.cpp/commit/dadbed99e65252d79f81101a392d0d6497b86caa) or earlier. For compatibility with latest llama.cpp, please use GGUF files instead. I use the following command line; adjust for your tastes and needs: ``` ./main -t 10 -ngl 40 -gqa 8 -m llama-2-70b-oasst-1-200.ggmlv3.q4_K_M.bin --color -c 4096 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "<|prompter|>{prompt}<|endoftext|><|assistant|>" ``` Change `-t 10` to the number of physical CPU cores you have. For example if your system has 8 cores/16 threads, use `-t 8`. If you are fully offloading the model to GPU, use `-t 1` Change `-ngl 40` to the number of GPU layers you have VRAM for. Use `-ngl 100` to offload all layers to VRAM - if you have a 48GB card, or 2 x 24GB, or similar. Otherwise you can partially offload as many as you have VRAM for, on one or more GPUs. If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins` Remember the `-gqa 8` argument, required for Llama 70B models. Change `-c 4096` to the desired sequence length for this model. For models that use RoPE, add `--rope-freq-base 10000 --rope-freq-scale 0.5` for doubled context, or `--rope-freq-base 10000 --rope-freq-scale 0.25` for 4x context. For other parameters and how to use them, please refer to [the llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/main/README.md) ## How to run in `text-generation-webui` Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md). <!-- footer start --> <!-- 200823 --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/theblokeai) ## Thanks, and how to contribute. Thanks to the [chirper.ai](https://chirper.ai) team! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Aemon Algiz. **Patreon special mentions**: Russ Johnson, J, alfie_i, Alex, NimbleBox.ai, Chadd, Mandus, Nikolai Manek, Ken Nordquist, ya boyyy, Illia Dulskyi, Viktor Bowallius, vamX, Iucharbius, zynix, Magnesian, Clay Pascal, Pierre Kircher, Enrico Ros, Tony Hughes, Elle, Andrey, knownsqashed, Deep Realms, Jerry Meng, Lone Striker, Derek Yates, Pyrater, Mesiah Bishop, James Bentley, Femi Adebogun, Brandon Frisco, SuperWojo, Alps Aficionado, Michael Dempsey, Vitor Caleffi, Will Dee, Edmond Seymore, usrbinkat, LangChain4j, Kacper Wikieł, Luke Pendergrass, John Detwiler, theTransient, Nathan LeClaire, Tiffany J. Kim, biorpg, Eugene Pentland, Stanislav Ovsiannikov, Fred von Graf, terasurfer, Kalila, Dan Guido, Nitin Borwankar, 阿明, Ai Maven, John Villwock, Gabriel Puliatti, Stephen Murray, Asp the Wyvern, danny, Chris Smitley, ReadyPlayerEmma, S_X, Daniel P. Andersen, Olakabola, Jeffrey Morgan, Imad Khwaja, Caitlyn Gatomon, webtim, Alicia Loh, Trenton Dambrowitz, Swaroop Kallakuri, Erik Bjäreholt, Leonard Tan, Spiking Neurons AB, Luke @flexchar, Ajan Kanaga, Thomas Belote, Deo Leter, RoA, Willem Michiel, transmissions 11, subjectnull, Matthew Berman, Joseph William Delisle, David Ziegler, Michael Davis, Johann-Peter Hartmann, Talal Aujan, senxiiz, Artur Olbinski, Rainer Wilmers, Spencer Kim, Fen Risland, Cap'n Zoog, Rishabh Srivastava, Michael Levine, Geoffrey Montalvo, Sean Connelly, Alexandros Triantafyllidis, Pieter, Gabriel Tamborski, Sam, Subspace Studios, Junyu Yang, Pedro Madruga, Vadim, Cory Kujawski, K, Raven Klaugh, Randy H, Mano Prime, Sebastain Graf, Space Cruiser Thank you to all my generous patrons and donaters! And thank you again to a16z for their generous grant. <!-- footer end --> # Original model card: Jordan Clive's Open-Assistant Llama2 70B SFT OASST # Open-Assistant Llama2 70B SFT OASST This model is a fine-tuning of [Llama2 70B](https://huggingface.co/meta-llama/Llama-2-70b-hf) LLM. It was trained on a mixture of OASST top-1 threads. ## Model Details - **Finetuned from:** [Llama2 70B](https://huggingface.co/meta-llama/Llama-2-70b-hf) - **Model type:** Causal decoder-only transformer language model - **Language:** English, German, Spanish, French (and limited capabilities in Italian, Portuguese, Polish, Dutch, Romanian, Czech, Swedish); - **License:** Apache 2.0 - **Contact:** [Open-Assistant Discord](https://ykilcher.com/open-assistant-discord) ## Prompting Two special tokens are used to mark the beginning of user and assistant turns: `<|prompter|>` and `<|assistant|>`. Each turn ends with a `</s>` token. Input prompt example: ``` <|prompter|>What is a meme, and what's the history behind this word?</s><|assistant|> ``` The input ends with the `<|assistant|>` token to signal that the model should start generating the assistant reply.
14,922
[ [ -0.038421630859375, -0.06378173828125, 0.02679443359375, 0.01873779296875, -0.0361328125, -0.0009427070617675781, -0.0026035308837890625, -0.045501708984375, 0.033782958984375, 0.004856109619140625, -0.047515869140625, -0.043609619140625, -0.0401611328125, 0...
SmellyKat/Pyramids-ppo
2023-08-06T13:34:04.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
SmellyKat
null
null
SmellyKat/Pyramids-ppo
0
2
ml-agents
2023-08-06T13:33:57
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: SmellyKat/Pyramids-ppo 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,334
[ [ -0.0401611328125, -0.03338623046875, 0.0027446746826171875, 0.01381683349609375, -0.01097869873046875, 0.01195526123046875, 0.01666259765625, -0.01428985595703125, 0.033721923828125, 0.030548095703125, -0.039825439453125, -0.050140380859375, -0.02972412109375, ...
Javeria98/donut_final_REAL_DATA_04
2023-08-06T18:19:58.000Z
[ "transformers", "pytorch", "tensorboard", "vision-encoder-decoder", "generated_from_trainer", "dataset:imagefolder", "license:mit", "endpoints_compatible", "region:us" ]
null
Javeria98
null
null
Javeria98/donut_final_REAL_DATA_04
0
2
transformers
2023-08-06T14:41:39
--- license: mit base_model: naver-clova-ix/donut-base tags: - generated_from_trainer datasets: - imagefolder model-index: - name: donut_final_REAL_DATA_04 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # donut_final_REAL_DATA_04 This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the imagefolder dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 25 ### Training results ### Framework versions - Transformers 4.32.0.dev0 - Pytorch 2.0.1+cu118 - Datasets 2.14.3 - Tokenizers 0.13.3
1,125
[ [ -0.020843505859375, -0.04449462890625, 0.0118865966796875, 0.0011644363403320312, -0.0233001708984375, -0.0158843994140625, -0.01430511474609375, -0.0132598876953125, 0.0093841552734375, 0.033172607421875, -0.035491943359375, -0.039215087890625, -0.0471801757812...
jannikseus/aspect_extraction_restaurant_reviews
2023-08-09T21:16:05.000Z
[ "transformers", "pytorch", "tensorboard", "safetensors", "distilbert", "token-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
jannikseus
null
null
jannikseus/aspect_extraction_restaurant_reviews
0
2
transformers
2023-08-06T20:58:35
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: aspect_extraction_restaurant_reviews results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # aspect_extraction_restaurant_reviews This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1048 - Precision: 0.7375 - Recall: 0.8194 - F1: 0.7763 - Accuracy: 0.9650 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 232 | 0.1149 | 0.6062 | 0.8125 | 0.6944 | 0.9500 | | No log | 2.0 | 464 | 0.0942 | 0.7267 | 0.8125 | 0.7672 | 0.9630 | | 0.1371 | 3.0 | 696 | 0.0981 | 0.7152 | 0.8194 | 0.7638 | 0.9638 | | 0.1371 | 4.0 | 928 | 0.1048 | 0.7375 | 0.8194 | 0.7763 | 0.9650 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu117 - Datasets 2.14.2 - Tokenizers 0.13.3
1,854
[ [ -0.029083251953125, -0.04388427734375, 0.0201416015625, 0.0101165771484375, -0.0214080810546875, -0.02264404296875, -0.000507354736328125, -0.0135345458984375, 0.01424407958984375, 0.026214599609375, -0.0401611328125, -0.050079345703125, -0.052978515625, -0....
kernelmachine/silo-pdswby-1.3b
2023-08-08T18:50:15.000Z
[ "transformers", "pytorch", "text-generation", "openlm", "silo", "en", "license:apache-2.0", "endpoints_compatible", "region:us" ]
text-generation
kernelmachine
null
null
kernelmachine/silo-pdswby-1.3b
0
2
transformers
2023-08-07T01:38:07
--- license: apache-2.0 language: - en pipeline_tag: text-generation tags: - text-generation - openlm - silo --- # Silo Language Models: Isolating Legal Risk in a Datastore This is Silo-PDSWBY, first introduced in [Silo Language Models]() by researchers at University of Washington, UC Berkeley, and the Allen Institute for AI. ### NOTE: Dependencies To use the model, you need to install a specific transformers fork: ``` pip install git+https://github.com/kernelmachine/transformers@openlm#egg=transformers ``` The model also depends on `xformers`, install via ``` pip install xformers ``` ### Model Description Silo-PDSWBY is a 1.3B parameter, decoder-only language model trained on data in the public domain, under permissive software licenses, and under the Creative-Commons BY license from [the Open License Corpus (OLC)](https://huggingface.co/datasets/kernelmachine/open-license-corpus). The model is based on the LLaMA architecture as implemented in (OpenLM)[]. The model is trained with 128 A100 GPUs across 16 nodes. ### Model and Training Hyperparameters We follow the model architecture of LLaMa, and we use the GPT-NeoX-20B tokenizer, with 50432 BPE types. During training, we use 2,048 token sequences that are packed across document boundaries, and we pre-pend a beginning-of-text token to every document. We use weight decay of 0.1, the Adam optimizer with beta_2 of 0.95, 2,000 steps of warmup, with a cosine learning rate scheduler. | Model | #L | #H | d_model | LR | Batch | |--------|-----|-----|-------------|--------|--------| | 1.3B | 24 | 16 | 2048 | 1e-3 | 2.6M | ### Training data Silo-PDSWBY is a 1.3B parameter, decoder-only language model trained on data in the public domain, under permissive software licenses, and under the Creative-Commons BY license from [the Open License Corpus (OLC)](https://huggingface.co/datasets/kernelmachine/open-license-corpus). The model was trained on the following domain proportions (please see the OLC repository for more details on the data sources for each domain): | Domain | Tokens (B) | % | |-----------------|------------|-------| | | PD SW BY | | | Code | 58.9 | 25.8 | | Legal | 27.2 | 11.9 | | Conversation | 27.2 | 11.9 | | Math | 3.5 | 1.5 | | Books | 2.9 | 1.3 | | Science | 71.5 | 31.3 | | News | 0.2 | 0.1 | | Wikipedia | 37.0 | 16.2 | | Total | 228.3 | 100.0 | We train with early stopping for 450B tokens in total, or about two epochs of training over this subset Since the distribution of OLC is highly skewed, we perform a simple upweighting scheme where we upsample all data that accounts for less than 5% of the corpus by a factor of 3x, which we found to work well after a sweep of different settings. ### Intended Uses and Limitations This model can be used for prompting for evaluation of downstream tasks as well as text generation. ### How to use You can use this model directly with a pipeline for text generation. ```python from transformers import pipeline generator = pipeline('text-generation', model="kernelmachine/silo-pdswby-1.3b", device='cuda') generator("Hello") [{'generated_text': 'Hello,\n\nI am trying to get a list of all the emails in my gmail account'}] ``` By default, generation is deterministic. In order to use the top-k sampling, please set do_sample to True. ```python from transformers import pipeline, set_seed set_seed(42) generator = pipeline('text-generation', model="kernelmachine/silo-pdswby-1.3b", device='cuda', do_sample=True) generator("Hello") [{'generated_text': "Hello,\n\nJust wanted to say thanks for your contribution to our team's software\ndevelopment."}] ``` ### Limitations and Bias Silo-PDSWBY inherits the biases and limitations of public domain data, which carry risks of toxic or otherwise unfair output, due to the prevalence of older copyright-expired text. Silo-PDSWBY may also output personally identifiable information, because we did not filter that out of training data.
4,178
[ [ -0.0041656494140625, -0.052642822265625, 0.033233642578125, 0.00286865234375, -0.01200103759765625, -0.01203155517578125, -0.01503753662109375, -0.0242462158203125, 0.005756378173828125, 0.05224609375, -0.0282745361328125, -0.052825927734375, -0.06195068359375, ...
LuizNeves/DeBERTa-v3-large-vaccine
2023-09-04T11:21:53.000Z
[ "transformers", "pytorch", "safetensors", "deberta-v2", "text-classification", "en", "license:afl-3.0", "endpoints_compatible", "region:us" ]
text-classification
LuizNeves
null
null
LuizNeves/DeBERTa-v3-large-vaccine
0
2
transformers
2023-08-07T14:14:40
--- license: afl-3.0 language: - en --- <b>MODEL DESCRIPTION</b> This model was fine-tuned to classify vaccine headlines. ```cpp from transformers import pipeline classifier = pipeline("text-classification", model="LuizNeves/DeBERTa-v3-large-vaccine") sequence_to_classify = "São Paulo approves vaccination for children this Monday" output = classifier(sequence_to_classify) print(output) ```
395
[ [ 0.007076263427734375, -0.037353515625, 0.00794219970703125, 0.0174560546875, -0.016998291015625, -0.00806427001953125, 0.018829345703125, -0.0107421875, -0.0022907257080078125, 0.03143310546875, -0.00727081298828125, -0.039306640625, -0.06927490234375, 0.011...
alperenunlu/PPO-LunarLander-v2
2023-09-15T12:48:05.000Z
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
alperenunlu
null
null
alperenunlu/PPO-LunarLander-v2
2
2
stable-baselines3
2023-08-07T14:41:48
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 280.82 +/- 15.04 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib Install the RL Zoo (with SB3 and SB3-Contrib): ```bash pip install rl_zoo3 ``` ``` # Download model and save it into the logs/ folder python -m rl_zoo3.load_from_hub --algo ppo --env LunarLander-v2 -orga alperenunlu -f logs/ python -m rl_zoo3.enjoy --algo ppo --env LunarLander-v2 -f logs/ ``` If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do: ``` python -m rl_zoo3.load_from_hub --algo ppo --env LunarLander-v2 -orga alperenunlu -f logs/ python -m rl_zoo3.enjoy --algo ppo --env LunarLander-v2 -f logs/ ``` ## Training (with the RL Zoo) ``` python -m rl_zoo3.train --algo ppo --env LunarLander-v2 -f logs/ # Upload the model and generate video (when possible) python -m rl_zoo3.push_to_hub --algo ppo --env LunarLander-v2 -f logs/ -orga alperenunlu ``` ## Hyperparameters ```python OrderedDict([('batch_size', 8), ('clip_range', 0.2), ('ent_coef', 0.0012069732975503813), ('gae_lambda', 0.95), ('gamma', 0.999), ('learning_rate', 0.0004080379698108855), ('max_grad_norm', 0.5), ('n_envs', 16), ('n_epochs', 10), ('n_steps', 256), ('n_timesteps', 2000000.0), ('policy', 'MlpPolicy'), ('vf_coef', 0.3326356386659747), ('normalize', False)]) ``` # Environment Arguments ```python {'render_mode': 'rgb_array'} ```
2,495
[ [ -0.0293121337890625, -0.04071044921875, 0.0179595947265625, 0.0191497802734375, -0.0167999267578125, -0.01268768310546875, 0.0031890869140625, -0.01491546630859375, 0.00508880615234375, 0.0382080078125, -0.05035400390625, -0.038909912109375, -0.036712646484375, ...
ManuVleuBeu/t5_base_answer-aware_eduQG
2023-08-08T10:31:19.000Z
[ "transformers", "pytorch", "t5", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
text2text-generation
ManuVleuBeu
null
null
ManuVleuBeu/t5_base_answer-aware_eduQG
0
2
transformers
2023-08-07T15:11:52
--- tags: - generated_from_trainer model-index: - name: t5_base_answer-aware_eduQG results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5_base_answer-aware_eduQG This model was trained from scratch on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.5453 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 16 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.7541 | 2.35 | 100 | 1.5453 | ### Framework versions - Transformers 4.27.4 - Pytorch 2.0.0+cu117 - Datasets 2.11.0 - Tokenizers 0.13.3
1,250
[ [ -0.0273284912109375, -0.034881591796875, 0.015380859375, 0.00830841064453125, -0.0233154296875, -0.042388916015625, -0.003955841064453125, -0.010589599609375, 0.005596160888671875, 0.0270233154296875, -0.056854248046875, -0.04638671875, -0.040191650390625, -...
nhanpv/phobert-v2-finetuned-timi
2023-08-09T17:46:25.000Z
[ "transformers", "pytorch", "roberta", "fill-mask", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
nhanpv
null
null
nhanpv/phobert-v2-finetuned-timi
0
2
transformers
2023-08-07T16:28:18
--- base_model: nhanpv/phobert-v2-finetuned-timi tags: - generated_from_trainer model-index: - name: phobert-v2-finetuned-timi results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # phobert-v2-finetuned-timi This model is a fine-tuned version of [nhanpv/phobert-v2-finetuned-timi](https://huggingface.co/nhanpv/phobert-v2-finetuned-timi) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.0430 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.3847 | 1.0 | 129 | 1.1946 | | 1.3808 | 2.0 | 258 | 1.1687 | | 1.3896 | 3.0 | 387 | 1.1924 | | 1.3737 | 4.0 | 516 | 1.1528 | | 1.3285 | 5.0 | 645 | 1.0942 | | 1.3244 | 6.0 | 774 | 1.1055 | | 1.2762 | 7.0 | 903 | 1.0649 | | 1.251 | 8.0 | 1032 | 1.0397 | | 1.2462 | 9.0 | 1161 | 1.0636 | | 1.2425 | 10.0 | 1290 | 1.0430 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.4 - Tokenizers 0.13.3
1,820
[ [ -0.034759521484375, -0.040435791015625, 0.0029277801513671875, 0.01009368896484375, -0.035125732421875, -0.031463623046875, -0.005523681640625, -0.01136016845703125, -0.00171661376953125, 0.0304107666015625, -0.047332763671875, -0.02703857421875, -0.043182373046...
BVRA/MegaDescriptor-B-224
2023-08-29T16:47:50.000Z
[ "wildlife-datasets", "pytorch", "image-classification", "arxiv:2103.14030", "license:cc-by-nc-4.0", "region:us" ]
image-classification
BVRA
null
null
BVRA/MegaDescriptor-B-224
0
2
wildlife-datasets
2023-08-07T17:34:06
--- tags: - image-classification library_name: wildlife-datasets license: cc-by-nc-4.0 --- # Model card for vit_small_patch14_dinov2.lvd142m A Swin-V image feature model. Superwisely pre-trained on animal re-identification datasets. ## Model Details - **Model Type:** Animal re-identification / feature backbone - **Model Stats:** - Params (M): ?? - GMACs: ?? - Activations (M): ?? - Image size: 224 x 224 - **Papers:** - Swin Transformer: Hierarchical Vision Transformer using Shifted Windows --> https://arxiv.org/abs/2103.14030 - **Original:** ?? - **Pretrain Dataset:** ?? ## Model Usage ### Image Embeddings ```python import timm import torch import torchvision.transforms as T from PIL import Image from urllib.request import urlopen model = timm.create_model("hf-hub:BVRA/wildlife-mega", pretrained=True) model = model.eval() train_transforms = T.Compose([T.Resize(224), T.ToTensor(), T.Normalize([0.5, 0.5, 0.5], [0.5, 0.5, 0.5])]) img = Image.open(urlopen( 'https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/beignets-task-guide.png' )) output = model(train_transforms(img).unsqueeze(0)) # output is (batch_size, num_features) shaped tensor # output is a (1, num_features) shaped tensor ``` ## Model Comparison ??? ## Citation ```bibtex @article{?????, title={?????}, author={????}, journal={????}, year={????} } ```
1,452
[ [ -0.0307769775390625, -0.0226898193359375, -0.001129150390625, 0.01169586181640625, -0.04443359375, -0.03472900390625, 0.003658294677734375, -0.02520751953125, 0.032135009765625, 0.027679443359375, -0.049713134765625, -0.041748046875, -0.0347900390625, -0.002...
yogjoshi14/ppo-Huggy
2023-08-07T18:24:51.000Z
[ "ml-agents", "onnx", "Huggy", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Huggy", "region:us" ]
reinforcement-learning
yogjoshi14
null
null
yogjoshi14/ppo-Huggy
0
2
ml-agents
2023-08-07T18:24:49
--- library_name: ml-agents tags: - Huggy - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Huggy --- # **ppo** Agent playing **Huggy** This is a trained model of a **ppo** agent playing **Huggy** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: yogjoshi14/ppo-Huggy 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,320
[ [ -0.042236328125, -0.0457763671875, 0.017242431640625, 0.0029544830322265625, -0.015960693359375, 0.01556396484375, 0.01309967041015625, -0.022918701171875, 0.0419921875, 0.034027099609375, -0.049102783203125, -0.0460205078125, -0.0303497314453125, -0.0174255...
nvbAI/whisper-small-hindi
2023-08-08T14:06:51.000Z
[ "transformers", "pytorch", "whisper", "automatic-speech-recognition", "hi", "license:afl-3.0", "endpoints_compatible", "has_space", "region:us" ]
automatic-speech-recognition
nvbAI
null
null
nvbAI/whisper-small-hindi
0
2
transformers
2023-08-08T01:06:09
--- license: afl-3.0 language: - hi --- Common Voice data 13 - Hindi - Finetuned - Whisper Small Checkpoint
108
[ [ -0.03045654296875, -0.044158935546875, -0.004062652587890625, 0.060699462890625, -0.01837158203125, 0.006656646728515625, 0.007274627685546875, -0.01824951171875, 0.03173828125, 0.036865234375, -0.0770263671875, -0.0234375, -0.024566650390625, 0.006641387939...
TigerResearch/tigerbot-13b-base-v1
2023-09-15T06:42:55.000Z
[ "transformers", "pytorch", "llama", "text-generation", "zh", "en", "license:apache-2.0", "endpoints_compatible", "text-generation-inference", "region:us" ]
text-generation
TigerResearch
null
null
TigerResearch/tigerbot-13b-base-v1
4
2
transformers
2023-08-08T03:53:45
--- license: apache-2.0 language: - zh - en --- <div style="width: 100%;"> <img src="http://x-pai.algolet.com/bot/img/logo_core.png" alt="TigerBot" style="width: 20%; display: block; margin: auto;"> </div> <p align="center"> <font face="黑体" size=5"> A cutting-edge foundation for your very own LLM. </font> </p> <p align="center"> 🌐 <a href="https://tigerbot.com/" target="_blank">TigerBot</a> • 🤗 <a href="https://huggingface.co/TigerResearch" target="_blank">Hugging Face</a> </p> ## Github https://github.com/TigerResearch/TigerBot ## Usage ```python from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("TigerResearch/tigerbot-13b-base-v1") model = AutoModelForCausalLM.from_pretrained("TigerResearch/tigerbot-13b-base-v1") ```
794
[ [ -0.03515625, -0.0389404296875, 0.007381439208984375, 0.0266876220703125, -0.036865234375, 0.01079559326171875, 0.0015211105346679688, -0.01611328125, 0.043670654296875, 0.00879669189453125, -0.06591796875, -0.032012939453125, -0.035888671875, 0.0049285888671...
Mediocreatmybest/instructblip-flan-t5-xxl_8bit
2023-08-09T06:58:00.000Z
[ "transformers", "pytorch", "instructblip", "text2text-generation", "vision", "image-to-text", "image-captioning", "visual-question-answering", "en", "arxiv:2301.12597", "arxiv:2210.11416", "license:mit", "autotrain_compatible", "region:us" ]
image-to-text
Mediocreatmybest
null
null
Mediocreatmybest/instructblip-flan-t5-xxl_8bit
1
2
transformers
2023-08-08T05:47:33
--- language: en license: mit tags: - vision - image-to-text - image-captioning - visual-question-answering pipeline_tag: image-to-text inference: false --- # BLIP-2, Flan T5-xxl, pre-trained only BLIP-2 model, leveraging [Flan T5-xxl](https://huggingface.co/google/flan-t5-xxl) (a large language model). It was introduced in the paper [BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models](https://arxiv.org/abs/2301.12597) by Li et al. and first released in [this repository](https://github.com/salesforce/LAVIS/tree/main/projects/blip2). Disclaimer: The team releasing BLIP-2 did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description BLIP-2 consists of 3 models: a CLIP-like image encoder, a Querying Transformer (Q-Former) and a large language model. The authors initialize the weights of the image encoder and large language model from pre-trained checkpoints and keep them frozen while training the Querying Transformer, which is a BERT-like Transformer encoder that maps a set of "query tokens" to query embeddings, which bridge the gap between the embedding space of the image encoder and the large language model. The goal for the model is simply to predict the next text token, giving the query embeddings and the previous text. <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/model_doc/blip2_architecture.jpg" alt="drawing" width="600"/> This allows the model to be used for tasks like: - image captioning - visual question answering (VQA) - chat-like conversations by feeding the image and the previous conversation as prompt to the model ## Direct Use and Downstream Use You can use the raw model for conditional text generation given an image and optional text. See the [model hub](https://huggingface.co/models?search=Salesforce/blip) to look for fine-tuned versions on a task that interests you. ## Bias, Risks, Limitations, and Ethical Considerations BLIP2-FlanT5 uses off-the-shelf Flan-T5 as the language model. It inherits the same risks and limitations from [Flan-T5](https://arxiv.org/pdf/2210.11416.pdf): > Language models, including Flan-T5, can potentially be used for language generation in a harmful way, according to Rae et al. (2021). Flan-T5 should not be used directly in any application, without a prior assessment of safety and fairness concerns specific to the application. BLIP2 is fine-tuned on image-text datasets (e.g. [LAION](https://laion.ai/blog/laion-400-open-dataset/) ) collected from the internet. As a result the model itself is potentially vulnerable to generating equivalently inappropriate content or replicating inherent biases in the underlying data. BLIP2 has not been tested in real world applications. It should not be directly deployed in any applications. Researchers should first carefully assess the safety and fairness of the model in relation to the specific context they’re being deployed within. ### How to use For code examples, we refer to the [documentation](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2#transformers.Blip2ForConditionalGeneration.forward.example), or refer to the snippets below depending on your usecase: #### Running the model on CPU <details> <summary> Click to expand </summary> ```python import requests from PIL import Image from transformers import BlipProcessor, Blip2ForConditionalGeneration processor = BlipProcessor.from_pretrained("Salesforce/blip2-flan-t5-xxl") model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-flan-t5-xxl") img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg' raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB') question = "how many dogs are in the picture?" inputs = processor(raw_image, question, return_tensors="pt") out = model.generate(**inputs) print(processor.decode(out[0], skip_special_tokens=True)) ``` </details> #### Running the model on GPU ##### In full precision <details> <summary> Click to expand </summary> ```python # pip install accelerate import requests from PIL import Image from transformers import Blip2Processor, Blip2ForConditionalGeneration processor = Blip2Processor.from_pretrained("Salesforce/blip2-flan-t5-xxl") model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-flan-t5-xxl", device_map="auto") img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg' raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB') question = "how many dogs are in the picture?" inputs = processor(raw_image, question, return_tensors="pt").to("cuda") out = model.generate(**inputs) print(processor.decode(out[0], skip_special_tokens=True)) ``` </details> ##### In half precision (`float16`) <details> <summary> Click to expand </summary> ```python # pip install accelerate import torch import requests from PIL import Image from transformers import Blip2Processor, Blip2ForConditionalGeneration processor = Blip2Processor.from_pretrained("Salesforce/blip2-flan-t5-xxl") model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-flan-t5-xxl", torch_dtype=torch.float16, device_map="auto") img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg' raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB') question = "how many dogs are in the picture?" inputs = processor(raw_image, question, return_tensors="pt").to("cuda", torch.float16) out = model.generate(**inputs) print(processor.decode(out[0], skip_special_tokens=True)) ``` </details> ##### In 8-bit precision (`int8`) <details> <summary> Click to expand </summary> ```python # pip install accelerate bitsandbytes import torch import requests from PIL import Image from transformers import Blip2Processor, Blip2ForConditionalGeneration processor = Blip2Processor.from_pretrained("Salesforce/blip2-flan-t5-xxl") model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-flan-t5-xxl", load_in_8bit=True, device_map="auto") img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg' raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB') question = "how many dogs are in the picture?" inputs = processor(raw_image, question, return_tensors="pt").to("cuda", torch.float16) out = model.generate(**inputs) print(processor.decode(out[0], skip_special_tokens=True)) ``` </details>
6,578
[ [ -0.0272064208984375, -0.049530029296875, -0.0035858154296875, 0.0305633544921875, -0.0177459716796875, -0.0112762451171875, -0.0228271484375, -0.059600830078125, -0.01027679443359375, 0.022308349609375, -0.03570556640625, -0.0103302001953125, -0.0445556640625, ...
JabrilJacobs/ppo-Pyramids
2023-08-08T06:32:51.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
JabrilJacobs
null
null
JabrilJacobs/ppo-Pyramids
0
2
ml-agents
2023-08-08T06:30:58
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: JabrilJacobs/ppo-Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,337
[ [ -0.040740966796875, -0.034393310546875, 0.0017042160034179688, 0.01424407958984375, -0.0107421875, 0.01251220703125, 0.01690673828125, -0.01491546630859375, 0.0338134765625, 0.0305023193359375, -0.04046630859375, -0.049896240234375, -0.0298309326171875, -0.0...
mediabiasgroup/xlm_roberta_veracity
2023-08-08T11:51:30.000Z
[ "transformers", "pytorch", "xlm-roberta", "text-classification", "license:gpl-3.0", "endpoints_compatible", "region:us" ]
text-classification
mediabiasgroup
null
null
mediabiasgroup/xlm_roberta_veracity
0
2
transformers
2023-08-08T08:01:31
--- license: gpl-3.0 widget: - text: >- Seitdem hat sich die Gesundheitsversorgung als ein sehr starkes Thema für die Demokraten erwiesen, die während der Zwischenwahlen 2019 einen aggressiven Wahlkampf mit diesem Thema führten und 40 Sitze im US-Repräsentantenhaus hinzugewannen. --- ![veracity](veracity.png)
336
[ [ -0.0011072158813476562, -0.0552978515625, 0.026336669921875, 0.052032470703125, -0.052154541015625, 0.0037078857421875, 0.023773193359375, 0.003711700439453125, 0.053802490234375, 0.03533935546875, -0.0031566619873046875, -0.05267333984375, -0.037689208984375, ...
Xenova/DeBERTa-v3-base-mnli
2023-08-08T11:44:08.000Z
[ "transformers.js", "onnx", "deberta-v2", "text-classification", "zero-shot-classification", "region:us" ]
zero-shot-classification
Xenova
null
null
Xenova/DeBERTa-v3-base-mnli
0
2
transformers.js
2023-08-08T11:32:35
--- library_name: "transformers.js" pipeline_tag: zero-shot-classification --- https://huggingface.co/MoritzLaurer/DeBERTa-v3-base-mnli with ONNX weights to be compatible with Transformers.js. Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
551
[ [ -0.031494140625, 0.0095062255859375, 0.0262298583984375, 0.052886962890625, -0.0117645263671875, -0.00562286376953125, 0.01239776611328125, -0.0171966552734375, 0.02838134765625, 0.04150390625, -0.051910400390625, -0.0382080078125, -0.045166015625, 0.0029106...
peterandrew987/modified
2023-08-08T13:45:02.000Z
[ "transformers", "pytorch", "bart", "text2text-generation", "generated_from_trainer", "dataset:squad", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
peterandrew987
null
null
peterandrew987/modified
0
2
transformers
2023-08-08T13:33:16
--- license: mit base_model: indobenchmark/indobart-v2 tags: - generated_from_trainer datasets: - squad metrics: - rouge model-index: - name: modified results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: squad type: squad config: plain_text split: train[:1000] args: plain_text metrics: - name: Rouge1 type: rouge value: 15.4275 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # modified This model is a fine-tuned version of [indobenchmark/indobart-v2](https://huggingface.co/indobenchmark/indobart-v2) on the squad dataset. It achieves the following results on the evaluation set: - Loss: 1.6035 - Rouge1: 15.4275 - Rouge2: 14.2367 - Rougel: 15.4625 - Rougelsum: 15.4954 - Gen Len: 20.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 1 - label_smoothing_factor: 0.1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | 1.4719 | 1.0 | 200 | 1.6035 | 15.4275 | 14.2367 | 15.4625 | 15.4954 | 20.0 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu117 - Datasets 2.14.2 - Tokenizers 0.13.3
1,912
[ [ -0.0374755859375, -0.0328369140625, 0.004787445068359375, 0.0297698974609375, -0.0309295654296875, -0.0147247314453125, -0.01477813720703125, -0.021087646484375, 0.00992584228515625, 0.024444580078125, -0.062469482421875, -0.036346435546875, -0.049163818359375, ...
Xenova/opus-mt-id-en
2023-09-05T23:08:08.000Z
[ "transformers.js", "onnx", "marian", "text2text-generation", "translation", "region:us" ]
translation
Xenova
null
null
Xenova/opus-mt-id-en
0
2
transformers.js
2023-08-08T14:49:10
--- library_name: transformers.js pipeline_tag: translation --- https://huggingface.co/Helsinki-NLP/opus-mt-id-en with ONNX weights to be compatible with Transformers.js. Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
529
[ [ -0.026123046875, 0.0096588134765625, 0.02679443359375, 0.048736572265625, -0.017578125, -0.013671875, -0.00907135009765625, -0.016937255859375, 0.0275726318359375, 0.040985107421875, -0.059295654296875, -0.047821044921875, -0.03912353515625, 0.01943969726562...
TheBloke/MythoMix-L2-13B-GGML
2023-09-27T13:01:08.000Z
[ "transformers", "llama", "en", "license:llama2", "text-generation-inference", "region:us" ]
null
TheBloke
null
null
TheBloke/MythoMix-L2-13B-GGML
11
2
transformers
2023-08-08T17:46:14
--- language: - en license: llama2 model_name: MythoMix L2 13B inference: false model_creator: Gryphe model_link: https://huggingface.co/Gryphe/MythoMix-L2-13b model_type: llama quantized_by: TheBloke base_model: Gryphe/MythoMix-L2-13b --- <!-- header start --> <!-- 200823 --> <div style="width: auto; margin-left: auto; margin-right: auto"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div> <hr style="margin-top: 1.0em; margin-bottom: 1.0em;"> <!-- header end --> # MythoMix L2 13B - GGML - Model creator: [Gryphe](https://huggingface.co/Gryphe) - Original model: [MythoMix L2 13B](https://huggingface.co/Gryphe/MythoMix-L2-13b) ## Description This repo contains GGML format model files for [Gryphe's MythoMix L2 13B](https://huggingface.co/Gryphe/MythoMix-L2-13b). ### Important note regarding GGML files. The GGML format has now been superseded by GGUF. As of August 21st 2023, [llama.cpp](https://github.com/ggerganov/llama.cpp) no longer supports GGML models. Third party clients and libraries are expected to still support it for a time, but many may also drop support. Please use the GGUF models instead. ### About GGML GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as: * [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most popular web UI. Supports NVidia CUDA GPU acceleration. * [KoboldCpp](https://github.com/LostRuins/koboldcpp), a powerful GGML web UI with GPU acceleration on all platforms (CUDA and OpenCL). Especially good for story telling. * [LM Studio](https://lmstudio.ai/), a fully featured local GUI with GPU acceleration on both Windows (NVidia and AMD), and macOS. * [LoLLMS Web UI](https://github.com/ParisNeo/lollms-webui), a great web UI with CUDA GPU acceleration via the c_transformers backend. * [ctransformers](https://github.com/marella/ctransformers), a Python library with GPU accel, LangChain support, and OpenAI-compatible AI server. * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), a Python library with GPU accel, LangChain support, and OpenAI-compatible API server. ## Repositories available * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/MythoMix-L2-13B-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGUF) * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference (deprecated)](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML) * [Gryphe's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/Gryphe/MythoMix-L2-13b) ## Prompt template: Custom ``` {system_message} ### Instruction: {prompt} (For roleplay purposes, I suggest the following - Write <CHAR NAME>'s next reply in a chat between <YOUR NAME> and <CHAR NAME>. Write a single reply only.) ### Response: ``` <!-- compatibility_ggml start --> ## Compatibility These quantised GGML files are compatible with llama.cpp between June 6th (commit `2d43387`) and August 21st 2023. For support with latest llama.cpp, please use GGUF files instead. The final llama.cpp commit with support for GGML was: [dadbed99e65252d79f81101a392d0d6497b86caa](https://github.com/ggerganov/llama.cpp/commit/dadbed99e65252d79f81101a392d0d6497b86caa) As of August 23rd 2023 they are still compatible with all UIs, libraries and utilities which use GGML. This may change in the future. ## Explanation of the new k-quant methods <details> <summary>Click to see details</summary> The new methods available are: * GGML_TYPE_Q2_K - "type-1" 2-bit quantization in super-blocks containing 16 blocks, each block having 16 weight. Block scales and mins are quantized with 4 bits. This ends up effectively using 2.5625 bits per weight (bpw) * GGML_TYPE_Q3_K - "type-0" 3-bit quantization in super-blocks containing 16 blocks, each block having 16 weights. Scales are quantized with 6 bits. This end up using 3.4375 bpw. * GGML_TYPE_Q4_K - "type-1" 4-bit quantization in super-blocks containing 8 blocks, each block having 32 weights. Scales and mins are quantized with 6 bits. This ends up using 4.5 bpw. * GGML_TYPE_Q5_K - "type-1" 5-bit quantization. Same super-block structure as GGML_TYPE_Q4_K resulting in 5.5 bpw * GGML_TYPE_Q6_K - "type-0" 6-bit quantization. Super-blocks with 16 blocks, each block having 16 weights. Scales are quantized with 8 bits. This ends up using 6.5625 bpw * GGML_TYPE_Q8_K - "type-0" 8-bit quantization. Only used for quantizing intermediate results. The difference to the existing Q8_0 is that the block size is 256. All 2-6 bit dot products are implemented for this quantization type. Refer to the Provided Files table below to see what files use which methods, and how. </details> <!-- compatibility_ggml end --> ## Provided files | Name | Quant method | Bits | Size | Max RAM required | Use case | | ---- | ---- | ---- | ---- | ---- | ----- | | [mythomix-l2-13b.ggmlv3.q2_K.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q2_K.bin) | q2_K | 2 | 5.51 GB| 8.01 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. | | [mythomix-l2-13b.ggmlv3.q3_K_S.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q3_K_S.bin) | q3_K_S | 3 | 5.66 GB| 8.16 GB | New k-quant method. Uses GGML_TYPE_Q3_K for all tensors | | [mythomix-l2-13b.ggmlv3.q3_K_M.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q3_K_M.bin) | q3_K_M | 3 | 6.31 GB| 8.81 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | [mythomix-l2-13b.ggmlv3.q3_K_L.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q3_K_L.bin) | q3_K_L | 3 | 6.93 GB| 9.43 GB | New k-quant method. Uses GGML_TYPE_Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | [mythomix-l2-13b.ggmlv3.q4_0.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q4_0.bin) | q4_0 | 4 | 7.37 GB| 9.87 GB | Original quant method, 4-bit. | | [mythomix-l2-13b.ggmlv3.q4_K_S.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q4_K_S.bin) | q4_K_S | 4 | 7.37 GB| 9.87 GB | New k-quant method. Uses GGML_TYPE_Q4_K for all tensors | | [mythomix-l2-13b.ggmlv3.q4_K_M.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q4_K_M.bin) | q4_K_M | 4 | 7.87 GB| 10.37 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q4_K | | [mythomix-l2-13b.ggmlv3.q4_1.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q4_1.bin) | q4_1 | 4 | 8.17 GB| 10.67 GB | Original quant method, 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models. | | [mythomix-l2-13b.ggmlv3.q5_0.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q5_0.bin) | q5_0 | 5 | 8.97 GB| 11.47 GB | Original quant method, 5-bit. Higher accuracy, higher resource usage and slower inference. | | [mythomix-l2-13b.ggmlv3.q5_K_S.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q5_K_S.bin) | q5_K_S | 5 | 8.97 GB| 11.47 GB | New k-quant method. Uses GGML_TYPE_Q5_K for all tensors | | [mythomix-l2-13b.ggmlv3.q5_K_M.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q5_K_M.bin) | q5_K_M | 5 | 9.23 GB| 11.73 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q5_K | | [mythomix-l2-13b.ggmlv3.q5_1.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q5_1.bin) | q5_1 | 5 | 9.78 GB| 12.28 GB | Original quant method, 5-bit. Even higher accuracy, resource usage and slower inference. | | [mythomix-l2-13b.ggmlv3.q6_K.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q6_K.bin) | q6_K | 6 | 10.68 GB| 13.18 GB | New k-quant method. Uses GGML_TYPE_Q8_K for all tensors - 6-bit quantization | | [mythomix-l2-13b.ggmlv3.q8_0.bin](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML/blob/main/mythomix-l2-13b.ggmlv3.q8_0.bin) | q8_0 | 8 | 13.79 GB| 16.29 GB | Original quant method, 8-bit. Almost indistinguishable from float16. High resource use and slow. Not recommended for most users. | **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead. ## How to run in `llama.cpp` Make sure you are using `llama.cpp` from commit [dadbed99e65252d79f81101a392d0d6497b86caa](https://github.com/ggerganov/llama.cpp/commit/dadbed99e65252d79f81101a392d0d6497b86caa) or earlier. For compatibility with latest llama.cpp, please use GGUF files instead. ``` ./main -t 10 -ngl 32 -m mythomix-l2-13b.ggmlv3.q4_K_M.bin --color -c 2048 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "```\nYou are a story writing assistant.\n\n### Instruction:\nWrite a story about llamas\n(For roleplay purposes, I suggest the following - Write <CHAR NAME>'s next reply in a chat between <YOUR NAME> and <CHAR NAME>. Write a single reply only.)\n\n### Response:\n\n```" ``` Change `-t 10` to the number of physical CPU cores you have. For example if your system has 8 cores/16 threads, use `-t 8`. Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration. Change `-c 2048` to the desired sequence length for this model. For example, `-c 4096` for a Llama 2 model. For models that use RoPE, add `--rope-freq-base 10000 --rope-freq-scale 0.5` for doubled context, or `--rope-freq-base 10000 --rope-freq-scale 0.25` for 4x context. If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins` For other parameters and how to use them, please refer to [the llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/main/README.md) ## How to run in `text-generation-webui` Further instructions here: [text-generation-webui/docs/llama.cpp.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp.md). <!-- footer start --> <!-- 200823 --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/theblokeai) ## Thanks, and how to contribute. Thanks to the [chirper.ai](https://chirper.ai) team! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Aemon Algiz. **Patreon special mentions**: Russ Johnson, J, alfie_i, Alex, NimbleBox.ai, Chadd, Mandus, Nikolai Manek, Ken Nordquist, ya boyyy, Illia Dulskyi, Viktor Bowallius, vamX, Iucharbius, zynix, Magnesian, Clay Pascal, Pierre Kircher, Enrico Ros, Tony Hughes, Elle, Andrey, knownsqashed, Deep Realms, Jerry Meng, Lone Striker, Derek Yates, Pyrater, Mesiah Bishop, James Bentley, Femi Adebogun, Brandon Frisco, SuperWojo, Alps Aficionado, Michael Dempsey, Vitor Caleffi, Will Dee, Edmond Seymore, usrbinkat, LangChain4j, Kacper Wikieł, Luke Pendergrass, John Detwiler, theTransient, Nathan LeClaire, Tiffany J. Kim, biorpg, Eugene Pentland, Stanislav Ovsiannikov, Fred von Graf, terasurfer, Kalila, Dan Guido, Nitin Borwankar, 阿明, Ai Maven, John Villwock, Gabriel Puliatti, Stephen Murray, Asp the Wyvern, danny, Chris Smitley, ReadyPlayerEmma, S_X, Daniel P. Andersen, Olakabola, Jeffrey Morgan, Imad Khwaja, Caitlyn Gatomon, webtim, Alicia Loh, Trenton Dambrowitz, Swaroop Kallakuri, Erik Bjäreholt, Leonard Tan, Spiking Neurons AB, Luke @flexchar, Ajan Kanaga, Thomas Belote, Deo Leter, RoA, Willem Michiel, transmissions 11, subjectnull, Matthew Berman, Joseph William Delisle, David Ziegler, Michael Davis, Johann-Peter Hartmann, Talal Aujan, senxiiz, Artur Olbinski, Rainer Wilmers, Spencer Kim, Fen Risland, Cap'n Zoog, Rishabh Srivastava, Michael Levine, Geoffrey Montalvo, Sean Connelly, Alexandros Triantafyllidis, Pieter, Gabriel Tamborski, Sam, Subspace Studios, Junyu Yang, Pedro Madruga, Vadim, Cory Kujawski, K, Raven Klaugh, Randy H, Mano Prime, Sebastain Graf, Space Cruiser Thank you to all my generous patrons and donaters! And thank you again to a16z for their generous grant. <!-- footer end --> # Original model card: Gryphe's MythoMix L2 13B **UPDATE:** There's an improved version now! [Check it MythoMax!](https://huggingface.co/Gryphe/MythoMax-L2-13b) A requested variant of [MythoLogic-L2](https://huggingface.co/Gryphe/MythoLogic-L2-13b) and [Huginn](https://huggingface.co/The-Face-Of-Goonery/Huginn-13b-FP16) using a highly experimental tensor type merge technique. This model is proficient at both roleplaying and storywriting due to its unique nature. Quantized models are available from TheBloke: [GGML](https://huggingface.co/TheBloke/MythoMix-L2-13B-GGML) - [GPTQ](https://huggingface.co/TheBloke/MythoMix-L2-13B-GPTQ) (You're the best!) ## Model details The idea behind this merge is that each layer is composed of several tensors, which are in turn responsible for specific functions. Using MythoLogic-L2's robust understanding as its input and Huginn's extensive writing capability as its output seems to have resulted in a model that exceeds at both, confirming my theory. (More details to be released at a later time) This type of merge is incapable of being illustrated, as each of its 360 tensors has an unique ratio applied to it. As with my prior merges, gradients were part of these ratios to further finetune its behaviour. ## Prompt Format This model primarily uses Alpaca formatting, so for optimal model performance, use: ``` <System prompt/Character Card> ### Instruction: Your instruction or question here. For roleplay purposes, I suggest the following - Write <CHAR NAME>'s next reply in a chat between <YOUR NAME> and <CHAR NAME>. Write a single reply only. ### Response: ``` --- license: other ---
15,741
[ [ -0.041107177734375, -0.05535888671875, 0.029083251953125, 0.00984954833984375, -0.027008056640625, -0.00337982177734375, -0.00908660888671875, -0.043212890625, 0.0279541015625, 0.00629425048828125, -0.0413818359375, -0.042327880859375, -0.033599853515625, 0....
mrm8488/mt5-base-ft-rf-02
2023-08-08T19:38:41.000Z
[ "transformers", "pytorch", "tensorboard", "mt5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
text2text-generation
mrm8488
null
null
mrm8488/mt5-base-ft-rf-02
0
2
transformers
2023-08-08T19:04:47
--- license: apache-2.0 base_model: google/mt5-base tags: - generated_from_trainer model-index: - name: mt5-base-ft-rf-02 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mt5-base-ft-rf-02 This model is a fine-tuned version of [google/mt5-base](https://huggingface.co/google/mt5-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.4229 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 8 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 43.082 | 0.24 | 50 | 37.1069 | | 34.6827 | 0.49 | 100 | 28.8296 | | 21.0188 | 0.73 | 150 | 19.9344 | | 18.3905 | 0.98 | 200 | 12.0120 | | 14.342 | 1.22 | 250 | 9.2877 | | 6.2116 | 1.46 | 300 | 6.1602 | | 6.5474 | 1.71 | 350 | 4.6816 | | 1.9222 | 1.95 | 400 | 2.6431 | | 2.0579 | 2.2 | 450 | 1.2741 | | 1.1028 | 2.44 | 500 | 0.9638 | | 1.3341 | 2.68 | 550 | 0.8896 | | 0.6531 | 2.93 | 600 | 0.8461 | | 0.9805 | 3.17 | 650 | 0.7652 | | 0.7167 | 3.41 | 700 | 0.7544 | | 1.0224 | 3.66 | 750 | 0.7493 | | 0.5367 | 3.9 | 800 | 0.7188 | | 0.9352 | 4.15 | 850 | 0.6844 | | 0.4927 | 4.39 | 900 | 0.6595 | | 0.7141 | 4.63 | 950 | 0.6458 | | 0.5773 | 4.88 | 1000 | 0.5911 | | 0.4791 | 5.12 | 1050 | 0.5691 | | 0.498 | 5.37 | 1100 | 0.5572 | | 0.4306 | 5.61 | 1150 | 0.5315 | | 0.334 | 5.85 | 1200 | 0.5123 | | 0.3783 | 6.1 | 1250 | 0.4970 | | 0.7719 | 6.34 | 1300 | 0.4774 | | 0.3732 | 6.59 | 1350 | 0.4591 | | 0.6203 | 6.83 | 1400 | 0.4482 | | 0.4669 | 7.07 | 1450 | 0.4434 | | 0.5568 | 7.32 | 1500 | 0.4307 | | 0.6352 | 7.56 | 1550 | 0.4257 | | 1.4137 | 7.8 | 1600 | 0.4229 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.4 - Tokenizers 0.13.3
2,859
[ [ -0.04345703125, -0.036712646484375, 0.01309967041015625, 0.003322601318359375, -0.01293182373046875, -0.011383056640625, 0.003612518310546875, -0.00402069091796875, 0.0278167724609375, 0.029144287109375, -0.05877685546875, -0.0513916015625, -0.052001953125, ...
LovenOO/distilBERT_without_preprocessing_grid_search
2023-08-14T18:58:44.000Z
[ "transformers", "pytorch", "tf", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
text-classification
LovenOO
null
null
LovenOO/distilBERT_without_preprocessing_grid_search
0
2
transformers
2023-08-08T20:53:01
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: distilBERT_without_preprocessing_grid_search results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilBERT_without_preprocessing_grid_search This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6731 - Precision: 0.8400 - Recall: 0.8427 - F1: 0.8407 - Accuracy: 0.8779 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 257 | 0.6542 | 0.7446 | 0.8052 | 0.7657 | 0.8350 | | 0.8635 | 2.0 | 514 | 0.5548 | 0.7961 | 0.8277 | 0.8056 | 0.8540 | | 0.8635 | 3.0 | 771 | 0.4839 | 0.7912 | 0.8427 | 0.8115 | 0.8589 | | 0.3097 | 4.0 | 1028 | 0.5256 | 0.8148 | 0.8544 | 0.8315 | 0.8667 | | 0.3097 | 5.0 | 1285 | 0.5657 | 0.8346 | 0.8494 | 0.8413 | 0.8764 | | 0.1839 | 6.0 | 1542 | 0.6005 | 0.8208 | 0.8430 | 0.8304 | 0.8710 | | 0.1839 | 7.0 | 1799 | 0.6580 | 0.8319 | 0.8349 | 0.8314 | 0.8706 | | 0.1254 | 8.0 | 2056 | 0.6348 | 0.8342 | 0.8515 | 0.8423 | 0.8774 | | 0.1254 | 9.0 | 2313 | 0.6601 | 0.8314 | 0.8394 | 0.8348 | 0.8745 | | 0.0935 | 10.0 | 2570 | 0.6731 | 0.8400 | 0.8427 | 0.8407 | 0.8779 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.4 - Tokenizers 0.13.3
2,423
[ [ -0.037994384765625, -0.0372314453125, 0.0158233642578125, 0.00797271728515625, -0.01528167724609375, -0.0143280029296875, 0.000023484230041503906, -0.004421234130859375, 0.02142333984375, 0.014678955078125, -0.046875, -0.055694580078125, -0.0567626953125, -0...
C-Lo/masked-dataset
2023-08-08T23:45:45.000Z
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:imdb", "license:apache-2.0", "endpoints_compatible", "region:us" ]
text-classification
C-Lo
null
null
C-Lo/masked-dataset
0
2
transformers
2023-08-08T23:41:15
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imdb model-index: - name: masked-dataset results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # masked-dataset This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 6 ### Training results ### Framework versions - Transformers 4.28.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.4 - Tokenizers 0.13.3
1,052
[ [ -0.04290771484375, -0.041015625, 0.0061798095703125, 0.00399017333984375, -0.037506103515625, 0.002811431884765625, 0.0042724609375, -0.00969696044921875, 0.0281982421875, 0.03460693359375, -0.061126708984375, -0.040924072265625, -0.06671142578125, -0.001966...
alvin-wen/distilbert-base-uncased-finetuned-wos
2023-08-09T01:49:39.000Z
[ "transformers", "pytorch", "tensorboard", "distilbert", "fill-mask", "generated_from_trainer", "dataset:web_of_science", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
alvin-wen
null
null
alvin-wen/distilbert-base-uncased-finetuned-wos
0
2
transformers
2023-08-09T01:36:30
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer datasets: - web_of_science model-index: - name: distilbert-base-uncased-finetuned-wos results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-wos This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the web_of_science dataset. It achieves the following results on the evaluation set: - Loss: 2.2025 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.5312 | 1.0 | 357 | 2.2975 | | 2.3847 | 2.0 | 714 | 2.2568 | | 2.3388 | 3.0 | 1071 | 2.2108 | | 2.3076 | 4.0 | 1428 | 2.2158 | | 2.2887 | 5.0 | 1785 | 2.2154 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.4 - Tokenizers 0.13.3
1,583
[ [ -0.034912109375, -0.042266845703125, 0.01262664794921875, 0.01641845703125, -0.0271453857421875, -0.0216217041015625, -0.0107421875, -0.002414703369140625, -0.00072479248046875, 0.0180206298828125, -0.054290771484375, -0.04144287109375, -0.054412841796875, -...
renatostrianese/ppoPyramidsTraining
2023-08-09T02:05:53.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
renatostrianese
null
null
renatostrianese/ppoPyramidsTraining
0
2
ml-agents
2023-08-09T02:05:41
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: renatostrianese/ppoPyramidsTraining 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,347
[ [ -0.0399169921875, -0.034027099609375, 0.00299835205078125, 0.0126495361328125, -0.01078033447265625, 0.01190948486328125, 0.015838623046875, -0.0142364501953125, 0.033203125, 0.02996826171875, -0.0426025390625, -0.04766845703125, -0.028961181640625, -0.01591...
kanixwang/eth-setfit-payment-model_4epoch
2023-08-09T03:27:48.000Z
[ "sentence-transformers", "pytorch", "mpnet", "feature-extraction", "sentence-similarity", "transformers", "endpoints_compatible", "region:us" ]
sentence-similarity
kanixwang
null
null
kanixwang/eth-setfit-payment-model_4epoch
0
2
sentence-transformers
2023-08-09T03:27:37
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity - transformers --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. <!--- Describe your model here --> ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('{MODEL_NAME}') embeddings = model.encode(sentences) print(embeddings) ``` ## Usage (HuggingFace Transformers) Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings. ```python from transformers import AutoTokenizer, AutoModel import torch #Mean Pooling - Take attention mask into account for correct averaging def mean_pooling(model_output, attention_mask): token_embeddings = model_output[0] #First element of model_output contains all token embeddings input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float() return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9) # Sentences we want sentence embeddings for sentences = ['This is an example sentence', 'Each sentence is converted'] # Load model from HuggingFace Hub tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}') model = AutoModel.from_pretrained('{MODEL_NAME}') # Tokenize sentences encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt') # Compute token embeddings with torch.no_grad(): model_output = model(**encoded_input) # Perform pooling. In this case, mean pooling. sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask']) print("Sentence embeddings:") print(sentence_embeddings) ``` ## Evaluation Results <!--- Describe how your model was evaluated --> For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME}) ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 26915 with parameters: ``` {'batch_size': 8, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss` Parameters of the fit()-Method: ``` { "epochs": 4, "evaluation_steps": 0, "evaluator": "NoneType", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": 26915, "warmup_steps": 2692, "weight_decay": 0.01 } ``` ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False}) ) ``` ## Citing & Authors <!--- Describe where people can find more information -->
3,702
[ [ -0.02117919921875, -0.05718994140625, 0.021240234375, 0.0214385986328125, -0.02099609375, -0.0340576171875, -0.01548004150390625, 0.003917694091796875, 0.0167083740234375, 0.0281829833984375, -0.047576904296875, -0.0445556640625, -0.053802490234375, -0.00224...
NEO946B/ppo-PyramidsTraining
2023-08-09T08:12:37.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
NEO946B
null
null
NEO946B/ppo-PyramidsTraining
0
2
ml-agents
2023-08-09T08:12:19
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: NEO946B/ppo-PyramidsTraining 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,340
[ [ -0.0396728515625, -0.034027099609375, 0.0015268325805664062, 0.0141143798828125, -0.01068878173828125, 0.012969970703125, 0.017791748046875, -0.0148162841796875, 0.03350830078125, 0.0301513671875, -0.041015625, -0.04949951171875, -0.0286865234375, -0.0154571...
jakezou/pyramid
2023-08-09T08:43:56.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
jakezou
null
null
jakezou/pyramid
0
2
ml-agents
2023-08-09T08:43:53
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: jakezou/pyramid 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,327
[ [ -0.041015625, -0.03546142578125, 0.0017862319946289062, 0.0146331787109375, -0.0097808837890625, 0.0120849609375, 0.01641845703125, -0.016082763671875, 0.033782958984375, 0.0301971435546875, -0.04119873046875, -0.050537109375, -0.0294036865234375, -0.0155563...
AIYIYA/my_wr3
2023-08-09T13:18:30.000Z
[ "transformers", "tf", "bert", "text-classification", "generated_from_keras_callback", "endpoints_compatible", "region:us" ]
text-classification
AIYIYA
null
null
AIYIYA/my_wr3
0
2
transformers
2023-08-09T12:45:42
--- base_model: bert-base-chinese tags: - generated_from_keras_callback model-index: - name: AIYIYA/my_wr3 results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # AIYIYA/my_wr3 This model is a fine-tuned version of [bert-base-chinese](https://huggingface.co/bert-base-chinese) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 1.1315 - Validation Loss: 1.1418 - Train Accuracy: 0.8158 - Epoch: 14 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': False, 'is_legacy_optimizer': False, 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 90, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False} - training_precision: float32 ### Training results | Train Loss | Validation Loss | Train Accuracy | Epoch | |:----------:|:---------------:|:--------------:|:-----:| | 3.0206 | 2.6776 | 0.2895 | 0 | | 2.6896 | 2.4286 | 0.7105 | 1 | | 2.4102 | 2.1955 | 0.6579 | 2 | | 2.1850 | 1.9989 | 0.7368 | 3 | | 1.9867 | 1.8181 | 0.6842 | 4 | | 1.8059 | 1.6320 | 0.7368 | 5 | | 1.5830 | 1.5359 | 0.8158 | 6 | | 1.5184 | 1.4081 | 0.7895 | 7 | | 1.4472 | 1.3072 | 0.8421 | 8 | | 1.3197 | 1.2605 | 0.8158 | 9 | | 1.2258 | 1.2182 | 0.8158 | 10 | | 1.2182 | 1.1752 | 0.8158 | 11 | | 1.1015 | 1.1583 | 0.8158 | 12 | | 1.1387 | 1.1463 | 0.8158 | 13 | | 1.1315 | 1.1418 | 0.8158 | 14 | ### Framework versions - Transformers 4.31.0 - TensorFlow 2.12.0 - Datasets 2.14.4 - Tokenizers 0.13.3
2,499
[ [ -0.0450439453125, -0.03643798828125, 0.01617431640625, 0.00756072998046875, -0.019866943359375, -0.0247802734375, -0.01079559326171875, -0.0271453857421875, 0.01483917236328125, 0.007274627685546875, -0.05291748046875, -0.04864501953125, -0.046142578125, -0....
dimonyara/Llama2-7b-lora-int4
2023-08-09T14:32:04.000Z
[ "peft", "region:us" ]
null
dimonyara
null
null
dimonyara/Llama2-7b-lora-int4
0
2
peft
2023-08-09T14:31:58
--- library_name: peft --- ## Training procedure The following `bitsandbytes` quantization config was used during training: - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: True - bnb_4bit_compute_dtype: bfloat16 ### Framework versions - PEFT 0.5.0.dev0
440
[ [ -0.049072265625, -0.05157470703125, 0.03155517578125, 0.03515625, -0.037994384765625, 0.007965087890625, 0.013153076171875, -0.01471710205078125, -0.01328277587890625, 0.032379150390625, -0.04205322265625, -0.008575439453125, -0.03399658203125, 0.01320648193...
hasibul1ah/finetuned_bloom_trained_model_bangladataset
2023-09-22T11:48:04.000Z
[ "transformers", "pytorch", "tensorboard", "bloom", "text-generation", "generated_from_trainer", "license:bigscience-bloom-rail-1.0", "endpoints_compatible", "text-generation-inference", "region:us" ]
text-generation
hasibul1ah
null
null
hasibul1ah/finetuned_bloom_trained_model_bangladataset
0
2
transformers
2023-08-09T15:44:02
--- license: bigscience-bloom-rail-1.0 base_model: bigscience/bloom-560m tags: - generated_from_trainer model-index: - name: finetuned_bloom_trained_model_bangladataset results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuned_bloom_trained_model_bangladataset This model is a fine-tuned version of [bigscience/bloom-560m](https://huggingface.co/bigscience/bloom-560m) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.8031 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 2.7731 | 1.0 | 10044 | 2.7881 | | 2.4222 | 2.0 | 20088 | 2.7370 | | 2.1122 | 3.0 | 30132 | 2.8031 | ### Framework versions - Transformers 4.34.0.dev0 - Pytorch 2.0.1+cu117 - Datasets 2.12.0 - Tokenizers 0.14.0
1,477
[ [ -0.0285797119140625, -0.03875732421875, 0.00852203369140625, 0.030242919921875, -0.0203399658203125, -0.03338623046875, -0.0091400146484375, -0.0153961181640625, -0.001323699951171875, 0.01824951171875, -0.05914306640625, -0.045867919921875, -0.035186767578125, ...
FlandersMakeAGV/whisper-small-keyword-spotting-m
2023-08-09T23:10:17.000Z
[ "transformers", "pytorch", "tensorboard", "whisper", "audio-classification", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
audio-classification
FlandersMakeAGV
null
null
FlandersMakeAGV/whisper-small-keyword-spotting-m
0
2
transformers
2023-08-09T15:52:44
--- license: apache-2.0 tags: - audio-classification - generated_from_trainer metrics: - accuracy model-index: - name: whisper-small-keyword-spotting-m results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # whisper-small-keyword-spotting-m This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the kw-spotting-fsc-sl-agv dataset. It achieves the following results on the evaluation set: - Loss: 0.0015 - Accuracy: 0.9999 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 32 - eval_batch_size: 32 - seed: 0 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 5.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.0146 | 1.0 | 356 | 0.0231 | 0.9925 | | 0.0124 | 2.0 | 712 | 0.0105 | 0.9977 | | 0.0091 | 3.0 | 1068 | 0.0015 | 0.9999 | | 0.0101 | 4.0 | 1425 | 0.0028 | 0.9994 | | 0.0094 | 5.0 | 1780 | 0.0022 | 0.9995 | ### Framework versions - Transformers 4.29.0.dev0 - Pytorch 2.0.0 - Datasets 2.10.1 - Tokenizers 0.13.2
1,786
[ [ -0.03338623046875, -0.032867431640625, 0.0144500732421875, -0.0009546279907226562, -0.033538818359375, -0.029144287109375, -0.01015472412109375, -0.0257568359375, 0.014801025390625, 0.01259613037109375, -0.052093505859375, -0.04913330078125, -0.056549072265625, ...
muhammadravi251001/fine-tuned-KoreanNLI-KorNLI-with-xlm-roberta-base
2023-08-11T13:17:53.000Z
[ "transformers", "pytorch", "tensorboard", "xlm-roberta", "text-classification", "generated_from_trainer", "license:mit", "endpoints_compatible", "region:us" ]
text-classification
muhammadravi251001
null
null
muhammadravi251001/fine-tuned-KoreanNLI-KorNLI-with-xlm-roberta-base
0
2
transformers
2023-08-09T16:10:44
--- license: mit base_model: xlm-roberta-base tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: fine-tuned-KoreanNLI-KorNLI-with-xlm-roberta-base results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fine-tuned-KoreanNLI-KorNLI-with-xlm-roberta-base This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.5802 - Accuracy: 0.7822 - F1: 0.7829 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:| | 0.5869 | 0.5 | 3654 | 0.6374 | 0.7312 | 0.7329 | | 0.5408 | 1.0 | 7308 | 0.6012 | 0.7439 | 0.7453 | | 0.497 | 1.5 | 10962 | 0.5622 | 0.7669 | 0.7673 | | 0.4933 | 2.0 | 14616 | 0.5494 | 0.7777 | 0.7785 | | 0.4659 | 2.5 | 18270 | 0.5644 | 0.7752 | 0.7765 | | 0.4675 | 3.0 | 21924 | 0.5348 | 0.7854 | 0.7855 | | 0.4229 | 3.5 | 25578 | 0.5490 | 0.7860 | 0.7870 | | 0.4376 | 4.0 | 29232 | 0.5389 | 0.7911 | 0.7916 | | 0.387 | 4.5 | 32886 | 0.5416 | 0.7854 | 0.7854 | | 0.3905 | 5.0 | 36540 | 0.5473 | 0.7879 | 0.7888 | | 0.3789 | 5.5 | 40194 | 0.5802 | 0.7822 | 0.7829 | ### Framework versions - Transformers 4.31.0 - Pytorch 1.13.1 - Datasets 2.14.4 - Tokenizers 0.13.3
2,237
[ [ -0.036956787109375, -0.04058837890625, 0.0194549560546875, -0.0036182403564453125, -0.0216064453125, -0.0132293701171875, -0.0156402587890625, -0.00927734375, 0.0159912109375, 0.034515380859375, -0.054840087890625, -0.054351806640625, -0.047882080078125, -0....
unionai/Llama-2-7b-LoRA-alpaca-cleaned
2023-08-16T22:20:27.000Z
[ "peft", "pytorch", "region:us" ]
null
unionai
null
null
unionai/Llama-2-7b-LoRA-alpaca-cleaned
0
2
peft
2023-08-09T16:35:14
--- library_name: peft --- ## Training procedure The following `bitsandbytes` quantization config was used during training: - quant_method: bitsandbytes - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: True - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: True - bnb_4bit_compute_dtype: bfloat16 The following `bitsandbytes` quantization config was used during training: - quant_method: bitsandbytes - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: True - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: True - bnb_4bit_compute_dtype: bfloat16 ### Framework versions - PEFT 0.5.0.dev0 - PEFT 0.5.0.dev0
861
[ [ -0.05059814453125, -0.053619384765625, 0.0235595703125, 0.03619384765625, -0.041656494140625, 0.006572723388671875, 0.007068634033203125, -0.0235595703125, -0.01059722900390625, 0.027008056640625, -0.04925537109375, -0.017608642578125, -0.036773681640625, 0....
stoyky/ppo-Huggy
2023-08-09T18:40:35.000Z
[ "ml-agents", "tensorboard", "onnx", "Huggy", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Huggy", "region:us" ]
reinforcement-learning
stoyky
null
null
stoyky/ppo-Huggy
0
2
ml-agents
2023-08-09T18:40:27
--- library_name: ml-agents tags: - Huggy - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Huggy --- # **ppo** Agent playing **Huggy** This is a trained model of a **ppo** agent playing **Huggy** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: stoyky/ppo-Huggy 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,316
[ [ -0.040985107421875, -0.046356201171875, 0.01763916015625, 0.0033588409423828125, -0.0170135498046875, 0.0155181884765625, 0.0127410888671875, -0.02301025390625, 0.04302978515625, 0.033660888671875, -0.048553466796875, -0.046417236328125, -0.030242919921875, ...
speechbrain/asr-wav2vec2-commonvoice-14-pt
2023-08-15T17:27:41.000Z
[ "speechbrain", "wav2vec2", "CTC", "pytorch", "Transformer", "automatic-speech-recognition", "pt", "dataset:commonvoice.14.0", "arxiv:2106.04624", "license:apache-2.0", "model-index", "region:us" ]
automatic-speech-recognition
speechbrain
null
null
speechbrain/asr-wav2vec2-commonvoice-14-pt
1
2
speechbrain
2023-08-09T20:51:45
--- language: - pt thumbnail: null pipeline_tag: automatic-speech-recognition tags: - CTC - pytorch - speechbrain - Transformer license: apache-2.0 datasets: - commonvoice.14.0 metrics: - wer - cer model-index: - name: asr-wav2vec2-commonvoice-14-pt results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: CommonVoice Corpus 14.0 (Portuguese) type: mozilla-foundation/common_voice_14.0 config: pt split: test args: language: pt metrics: - name: Test WER type: wer value: '22.51' --- <iframe src="https://ghbtns.com/github-btn.html?user=speechbrain&repo=speechbrain&type=star&count=true&size=large&v=2" frameborder="0" scrolling="0" width="170" height="30" title="GitHub"></iframe> <br/><br/> # wav2vec 2.0 with CTC trained on CommonVoice Portuguese (No LM) This repository provides all the necessary tools to perform automatic speech recognition from an end-to-end system pretrained on CommonVoice (Portuguese Language) within SpeechBrain. For a better experience, we encourage you to learn more about [SpeechBrain](https://speechbrain.github.io). The performance of the model is the following: | Release | Test CER | Test WER | GPUs | |:-------------:|:--------------:|:--------------:| :--------:| | 15-08-23 | 6.85 | 22.41 | 1xV100 32GB | ## Pipeline description This ASR system is composed of 2 different but linked blocks: - Tokenizer (unigram) that transforms words into unigrams and trained with the train transcriptions (train.tsv) of CommonVoice (pt). - Acoustic model (wav2vec2.0 + CTC). A pretrained wav2vec 2.0 model (wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53)) is combined with two DNN layers and finetuned on CommonVoice DE. The obtained final acoustic representation is given to the CTC decoder. The system is trained with recordings sampled at 16kHz (single channel). The code will automatically normalize your audio (i.e., resampling + mono channel selection) when calling *transcribe_file* if needed. ## Install SpeechBrain First of all, please install tranformers and SpeechBrain with the following command: ``` pip install speechbrain transformers ``` Please notice that we encourage you to read our tutorials and learn more about [SpeechBrain](https://speechbrain.github.io). ### Transcribing your own audio files (in Portuguese) ```python from speechbrain.pretrained import EncoderASR asr_model = EncoderASR.from_hparams(source="speechbrain/asr-wav2vec2-commonvoice-14-pt", savedir="pretrained_models/asr-wav2vec2-commonvoice-14-pt") asr_model.transcribe_file("speechbrain/asr-wav2vec2-commonvoice-14-pt/example-pt.wav") ``` ### Inference on GPU To perform inference on the GPU, add `run_opts={"device":"cuda"}` when calling the `from_hparams` method. ## Parallel Inference on a Batch Please, [see this Colab notebook](https://colab.research.google.com/drive/1hX5ZI9S4jHIjahFCZnhwwQmFoGAi3tmu?usp=sharing) to figure out how to transcribe in parallel a batch of input sentences using a pre-trained model. ### Training The model was trained with SpeechBrain. To train it from scratch follow these steps: 1. Clone SpeechBrain: ```bash git clone https://github.com/speechbrain/speechbrain/ ``` 2. Install it: ```bash cd speechbrain pip install -r requirements.txt pip install -e . ``` 3. Run Training: ```bash cd recipes/CommonVoice/ASR/CTC/ python train_with_wav2vec.py hparams/train_pt_with_wav2vec.yaml --data_folder=your_data_folder ``` You can find our training results (models, logs, etc) [here](https://www.dropbox.com/sh/80wucrvijdvao2a/AAD6-SZ2_ZZXmlAjOTw6fVloa?dl=0). ### Limitations The SpeechBrain team does not provide any warranty on the performance achieved by this model when used on other datasets. # **About SpeechBrain** - Website: https://speechbrain.github.io/ - Code: https://github.com/speechbrain/speechbrain/ - HuggingFace: https://huggingface.co/speechbrain/ # **Citing SpeechBrain** Please, cite SpeechBrain if you use it for your research or business. ```bibtex @misc{speechbrain, title={{SpeechBrain}: A General-Purpose Speech Toolkit}, author={Mirco Ravanelli and Titouan Parcollet and Peter Plantinga and Aku Rouhe and Samuele Cornell and Loren Lugosch and Cem Subakan and Nauman Dawalatabad and Abdelwahab Heba and Jianyuan Zhong and Ju-Chieh Chou and Sung-Lin Yeh and Szu-Wei Fu and Chien-Feng Liao and Elena Rastorgueva and François Grondin and William Aris and Hwidong Na and Yan Gao and Renato De Mori and Yoshua Bengio}, year={2021}, eprint={2106.04624}, archivePrefix={arXiv}, primaryClass={eess.AS}, note={arXiv:2106.04624} } ```
4,694
[ [ -0.0225067138671875, -0.053985595703125, 0.005283355712890625, 0.020721435546875, -0.0258636474609375, -0.007213592529296875, -0.037322998046875, -0.039520263671875, 0.0154571533203125, 0.016326904296875, -0.04156494140625, -0.048553466796875, -0.058685302734375...
EgilKarlsen/GPT2_Thunderbird-Anomaly_Baseline
2023-08-09T22:59:45.000Z
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-classification", "generated_from_trainer", "license:mit", "endpoints_compatible", "text-generation-inference", "region:us" ]
text-classification
EgilKarlsen
null
null
EgilKarlsen/GPT2_Thunderbird-Anomaly_Baseline
0
2
transformers
2023-08-09T22:40:51
--- license: mit base_model: gpt2 tags: - generated_from_trainer metrics: - accuracy model-index: - name: GPT2_Thuderbird-Anomaly_Baseline results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # GPT2_Thuderbird-Anomaly_Baseline This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0372 - Accuracy: 0.9864 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 10 - total_train_batch_size: 320 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.3763 | 1.0 | 109 | 0.3583 | 0.9094 | | 0.2193 | 1.99 | 218 | 0.2309 | 0.9278 | | 0.1411 | 3.0 | 328 | 0.1515 | 0.9480 | | 0.1016 | 3.99 | 437 | 0.1039 | 0.9684 | | 0.0747 | 5.0 | 547 | 0.0746 | 0.9792 | | 0.0576 | 6.0 | 656 | 0.0572 | 0.9828 | | 0.051 | 6.99 | 765 | 0.0477 | 0.9844 | | 0.0442 | 8.0 | 875 | 0.0427 | 0.9856 | | 0.0412 | 8.99 | 984 | 0.0405 | 0.9866 | | 0.0397 | 9.96 | 1090 | 0.0398 | 0.9868 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.4 - Tokenizers 0.13.3
1,965
[ [ -0.028533935546875, -0.047576904296875, 0.005458831787109375, 0.0112762451171875, -0.01065826416015625, -0.02435302734375, 0.006687164306640625, -0.021820068359375, 0.0013933181762695312, 0.016082763671875, -0.052703857421875, -0.04302978515625, -0.0594177246093...
agustinl/ppo-Pyramids
2023-08-09T23:36:42.000Z
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
agustinl
null
null
agustinl/ppo-Pyramids
0
2
ml-agents
2023-08-09T23:13:56
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: agustinl/ppo-Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
1,333
[ [ -0.0406494140625, -0.03436279296875, 0.0015611648559570312, 0.0148468017578125, -0.01078033447265625, 0.01233673095703125, 0.0162506103515625, -0.01500701904296875, 0.0340576171875, 0.0305328369140625, -0.040679931640625, -0.050750732421875, -0.029266357421875, ...
Stevross/Astrid-LLama-3B-GPU
2023-08-10T07:56:12.000Z
[ "transformers", "pytorch", "llama", "text-generation", "gpt", "llm", "large language model", "h2o-llmstudio", "en", "endpoints_compatible", "text-generation-inference", "region:us" ]
text-generation
Stevross
null
null
Stevross/Astrid-LLama-3B-GPU
0
2
transformers
2023-08-10T00:05:32
--- language: - en library_name: transformers tags: - gpt - llm - large language model - h2o-llmstudio inference: true thumbnail: https://h2o.ai/etc.clientlibs/h2o/clientlibs/clientlib-site/resources/images/favicon.ico --- # Model Card ## Summary This model was trained using [H2O LLM Studio](https://github.com/h2oai/h2o-llmstudio). - Base model: [openlm-research/open_llama_3b](https://huggingface.co/openlm-research/open_llama_3b) ## Usage To use the model with the `transformers` library on a machine with GPUs, first make sure you have the `transformers`, `accelerate` and `torch` libraries installed. ```bash pip install transformers==4.30.1 pip install accelerate==0.20.3 pip install torch==2.0.0 ``` ```python import torch from transformers import pipeline generate_text = pipeline( model="Stevross/Astrid-LLama-3B-GPU", torch_dtype="auto", trust_remote_code=True, use_fast=False, device_map={"": "cuda:0"}, ) res = generate_text( "Why is drinking water so healthy?", min_new_tokens=2, max_new_tokens=256, do_sample=False, num_beams=1, temperature=float(0.3), repetition_penalty=float(1.2), renormalize_logits=True ) print(res[0]["generated_text"]) ``` You can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer: ```python print(generate_text.preprocess("Why is drinking water so healthy?")["prompt_text"]) ``` ```bash <|prompt|>Why is drinking water so healthy?</s><|answer|> ``` Alternatively, you can download [h2oai_pipeline.py](h2oai_pipeline.py), store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the `transformers` package, this will allow you to set `trust_remote_code=False`. ```python import torch from h2oai_pipeline import H2OTextGenerationPipeline from transformers import AutoModelForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained( "Stevross/Astrid-LLama-3B-GPU", use_fast=False, padding_side="left", trust_remote_code=True, ) model = AutoModelForCausalLM.from_pretrained( "Stevross/Astrid-LLama-3B-GPU", torch_dtype="auto", device_map={"": "cuda:0"}, trust_remote_code=True, ) generate_text = H2OTextGenerationPipeline(model=model, tokenizer=tokenizer) res = generate_text( "Why is drinking water so healthy?", min_new_tokens=2, max_new_tokens=256, do_sample=False, num_beams=1, temperature=float(0.3), repetition_penalty=float(1.2), renormalize_logits=True ) print(res[0]["generated_text"]) ``` You may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps: ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "Stevross/Astrid-LLama-3B-GPU" # either local folder or huggingface model name # Important: The prompt needs to be in the same format the model was trained with. # You can find an example prompt in the experiment logs. prompt = "<|prompt|>How are you?</s><|answer|>" tokenizer = AutoTokenizer.from_pretrained( model_name, use_fast=False, trust_remote_code=True, ) model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype="auto", device_map={"": "cuda:0"}, trust_remote_code=True, ) model.cuda().eval() inputs = tokenizer(prompt, return_tensors="pt", add_special_tokens=False).to("cuda") # generate configuration can be modified to your needs tokens = model.generate( **inputs, min_new_tokens=2, max_new_tokens=256, do_sample=False, num_beams=1, temperature=float(0.3), repetition_penalty=float(1.2), renormalize_logits=True )[0] tokens = tokens[inputs["input_ids"].shape[1]:] answer = tokenizer.decode(tokens, skip_special_tokens=True) print(answer) ``` ## Model Architecture ``` LlamaForCausalLM( (model): LlamaModel( (embed_tokens): Embedding(32000, 3200, padding_idx=0) (layers): ModuleList( (0-25): 26 x LlamaDecoderLayer( (self_attn): LlamaAttention( (q_proj): Linear(in_features=3200, out_features=3200, bias=False) (k_proj): Linear(in_features=3200, out_features=3200, bias=False) (v_proj): Linear(in_features=3200, out_features=3200, bias=False) (o_proj): Linear(in_features=3200, out_features=3200, bias=False) (rotary_emb): LlamaRotaryEmbedding() ) (mlp): LlamaMLP( (gate_proj): Linear(in_features=3200, out_features=8640, bias=False) (down_proj): Linear(in_features=8640, out_features=3200, bias=False) (up_proj): Linear(in_features=3200, out_features=8640, bias=False) (act_fn): SiLUActivation() ) (input_layernorm): LlamaRMSNorm() (post_attention_layernorm): LlamaRMSNorm() ) ) (norm): LlamaRMSNorm() ) (lm_head): Linear(in_features=3200, out_features=32000, bias=False) ) ``` ## Model Configuration This model was trained using H2O LLM Studio and with the configuration in [cfg.yaml](cfg.yaml). Visit [H2O LLM Studio](https://github.com/h2oai/h2o-llmstudio) to learn how to train your own large language models. ## Model Validation Model validation results using [EleutherAI lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness). ```bash CUDA_VISIBLE_DEVICES=0 python main.py --model hf-causal-experimental --model_args pretrained=Stevross/Astrid-LLama-3B-GPU --tasks openbookqa,arc_easy,winogrande,hellaswag,arc_challenge,piqa,boolq --device cuda &> eval.log ``` ## Disclaimer Please read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions. - Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints. - Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion. - Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model. - Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities. - Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues. - Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes. By using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it.
7,897
[ [ -0.01751708984375, -0.0645751953125, 0.0352783203125, 0.01329803466796875, -0.02581787109375, -0.00859832763671875, -0.02020263671875, -0.0218505859375, 0.006381988525390625, 0.0201568603515625, -0.0322265625, -0.046051025390625, -0.04766845703125, 0.0038375...
xtuner/Llama-2-7b-qlora-moss-003-sft
2023-09-20T15:53:06.000Z
[ "peft", "conversational", "dataset:fnlp/moss-003-sft-data", "region:us" ]
conversational
xtuner
null
null
xtuner/Llama-2-7b-qlora-moss-003-sft
2
2
peft
2023-08-10T03:38:37
--- library_name: peft datasets: - fnlp/moss-003-sft-data pipeline_tag: conversational base_model: meta-llama/Llama-2-7b-hf --- <div align="center"> <img src="https://github.com/InternLM/lmdeploy/assets/36994684/0cf8d00f-e86b-40ba-9b54-dc8f1bc6c8d8" width="600"/> [![Generic badge](https://img.shields.io/badge/GitHub-%20XTuner-black.svg)](https://github.com/InternLM/xtuner) </div> ## Model Llama-2-7b-qlora-moss-003-sft is fine-tuned from [Llama-2-7b](https://huggingface.co/meta-llama/Llama-2-7b-hf) with [moss-003-sft](https://huggingface.co/datasets/fnlp/moss-003-sft-data) dataset by [XTuner](https://github.com/InternLM/xtuner). ## Quickstart ### Usage with HuggingFace libraries ```python import torch from peft import PeftModel from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, StoppingCriteria from transformers.generation import GenerationConfig class StopWordStoppingCriteria(StoppingCriteria): def __init__(self, tokenizer, stop_word): self.tokenizer = tokenizer self.stop_word = stop_word self.length = len(self.stop_word) def __call__(self, input_ids, *args, **kwargs) -> bool: cur_text = self.tokenizer.decode(input_ids[0]) cur_text = cur_text.replace('\r', '').replace('\n', '') return cur_text[-self.length:] == self.stop_word tokenizer = AutoTokenizer.from_pretrained('meta-llama/Llama-2-7b-hf', trust_remote_code=True) quantization_config = BitsAndBytesConfig(load_in_4bit=True, load_in_8bit=False, llm_int8_threshold=6.0, llm_int8_has_fp16_weight=False, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4') model = AutoModelForCausalLM.from_pretrained('meta-llama/Llama-2-7b-hf', quantization_config=quantization_config, device_map='auto', trust_remote_code=True).eval() model = PeftModel.from_pretrained(model, 'xtuner/Llama-2-7b-qlora-moss-003-sft') gen_config = GenerationConfig(max_new_tokens=1024, do_sample=True, temperature=0.1, top_p=0.75, top_k=40) # Note: In this example, we disable the use of plugins because the API depends on additional implementations. # If you want to experience plugins, please refer to XTuner CLI! prompt_template = ( 'You are an AI assistant whose name is Llama2.\n' 'Capabilities and tools that Llama2 can possess.\n' '- Inner thoughts: disabled.\n' '- Web search: disabled.\n' '- Calculator: disabled.\n' '- Equation solver: disabled.\n' '- Text-to-image: disabled.\n' '- Image edition: disabled.\n' '- Text-to-speech: disabled.\n' '<|Human|>: {input}<eoh>\n' '<|Inner Thoughts|>: None<eot>\n' '<|Commands|>: None<eoc>\n' '<|Results|>: None<eor>\n') text = '请给我介绍五个上海的景点' inputs = tokenizer(prompt_template.format(input=text), return_tensors='pt') inputs = inputs.to(model.device) pred = model.generate(**inputs, generation_config=gen_config, stopping_criteria=[StopWordStoppingCriteria(tokenizer, '<eom>')]) print(tokenizer.decode(pred.cpu()[0], skip_special_tokens=True)) """ 好的,以下是五个上海的景点: 1. 外滩:外滩是上海的标志性景点之一,是一条长达1.5公里的沿江大道,沿途有许多历史建筑和现代化的高楼大厦。游客可以欣赏到黄浦江两岸的美景,还可以在这里拍照留念。 2. 上海博物馆:上海博物馆是上海市最大的博物馆之一,收藏了大量的历史文物和艺术品。博物馆内有许多展览,包括中国古代文物、近代艺术品和现代艺术品等。 3. 上海科技馆:上海科技馆是一座以科技为主题的博物馆,展示了许多科技产品和科技发展的历史。游客可以在这里了解到许多有趣的科技知识,还可以参加一些科技体验活动。 4. 上海迪士尼乐园:上海迪士尼乐园是中国第一个迪士尼乐园,是一个集游乐、购物、餐饮、娱乐等多种功能于一体的主题公园。游客可以在这里体验到迪士尼的经典故事和游乐设施。 5. 上海野生动物园:上海野生动物园是一座以野生动物观赏和保护为主题的大型动物园。它位于上海市浦东新区,是中国最大的野生动物园之一。 """ ``` ### Usage with XTuner CLI #### Installation ```shell pip install xtuner ``` #### Chat > Don't forget to use `huggingface-cli login` and input your access token first to access Llama2! See [here](https://huggingface.co/docs/hub/security-tokens#user-access-tokens) to learn how to obtain your access token. ```shell export SERPER_API_KEY="xxx" # Please get the key from https://serper.dev to support google search! xtuner chat meta-llama/Llama-2-7b-hf --adapter xtuner/Llama-2-7b-qlora-moss-003-sft --bot-name Llama2 --prompt-template moss_sft --with-plugins calculate solve search --command-stop-word "<eoc>" --answer-stop-word "<eom>" --no-streamer ``` #### Fine-tune Use the following command to quickly reproduce the fine-tuning results. ```shell NPROC_PER_NODE=8 xtuner train llama2_7b_qlora_moss_sft_all_e2_gpu8 ```
4,280
[ [ -0.04156494140625, -0.051666259765625, 0.021270751953125, 0.0171966552734375, -0.033447265625, 0.0186920166015625, -0.0120391845703125, -0.022186279296875, 0.023162841796875, 0.006191253662109375, -0.04461669921875, -0.045440673828125, -0.04949951171875, 0.0...