Upload 12 files
Browse files- DataSet/Json/gradio.json +0 -0
- DataSet/Json/huggyfacefintuning.json +1 -0
- DataSet/Json/openchatkitreadme.json +1 -0
- DataSet/Json/train.json +0 -0
- DataSet/train.json +0 -0
- README.md +192 -0
- pytorch_model.bin.index.json +671 -0
- special_tokens_map.json +5 -0
- tokenizer.json +0 -0
- tokenizer_config.json +9 -0
- train.json +0 -0
DataSet/Json/gradio.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
DataSet/Json/huggyfacefintuning.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
[{"finetuning":"Fine-tune a pretrained model\n\nThere are significant benefits to using a pretrained model. It reduces computation costs, your carbon footprint, and allows you to use state-of-the-art models without having to train one from scratch. 🤗 Transformers provides access to thousands of pretrained models for a wide range of tasks. When you use a pretrained model, you train it on a dataset specific to your task. This is known as fine-tuning, an incredibly powerful training technique. In this tutorial, you will fine-tune a pretrained model with a deep learning framework of your choice:\n\nFine-tune a pretrained model with 🤗 Transformers Trainer.\nFine-tune a pretrained model in TensorFlow with Keras.\nFine-tune a pretrained model in native PyTorch.\nPrepare a dataset\n\nBefore you can fine-tune a pretrained model, download a dataset and prepare it for training. The previous tutorial showed you how to process data for training, and now you get an opportunity to put those skills to the test!\n\nBegin by loading the Yelp Reviews dataset:\n\nCopied\n>>> from datasets import load_dataset\n\n>>> dataset = load_dataset(\"yelp_review_full\")\n>>> dataset[\"train\"][100]\n{'label': 0,\n 'text': 'My expectations for McDonalds are t rarely high. But for one to still fail so spectacularly...that takes something special!\\\\nThe cashier took my friends\\'s order, then promptly ignored me. I had to force myself in front of a cashier who opened his register to wait on the person BEHIND me. I waited over five minutes for a gigantic order that included precisely one kid\\'s meal. After watching two people who ordered after me be handed their food, I asked where mine was. The manager started yelling at the cashiers for \\\\\"serving off their orders\\\\\" when they didn\\'t have their food. But neither cashier was anywhere near those controls, and the manager was the one serving food to customers and clearing the boards.\\\\nThe manager was rude when giving me my order. She didn\\'t make sure that I had everything ON MY RECEIPT, and never even had the decency to apologize that I felt I was getting poor service.\\\\nI\\'ve eaten at various McDonalds restaurants for over 30 years. I\\'ve worked at more than one location. I expect bad days, bad moods, and the occasional mistake. But I have yet to have a decent experience at this store. It will remain a place I avoid unless someone in my party needs to avoid illness from low blood sugar. Perhaps I should go back to the racially biased service of Steak n Shake instead!'}\n\nAs you now know, you need a tokenizer to process the text and include a padding and truncation strategy to handle any variable sequence lengths. To process your dataset in one step, use 🤗 Datasets map method to apply a preprocessing function over the entire dataset:\n\nCopied\n>>> from transformers import AutoTokenizer\n\n>>> tokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\n\n\n>>> def tokenize_function(examples):\n... return tokenizer(examples[\"text\"], padding=\"max_length\", truncation=True)\n\n\n>>> tokenized_datasets = dataset.map(tokenize_function, batched=True)\n\nIf you like, you can create a smaller subset of the full dataset to fine-tune on to reduce the time it takes:\n\nCopied\n>>> small_train_dataset = tokenized_datasets[\"train\"].shuffle(seed=42).select(range(1000))\n>>> small_eval_dataset = tokenized_datasets[\"test\"].shuffle(seed=42).select(range(1000))\nTrain\n\nAt this point, you should follow the section corresponding to the framework you want to use. You can use the links in the right sidebar to jump to the one you want - and if you want to hide all of the content for a given framework, just use the button at the top-right of that framework’s block!\n\nPytorch\nHide Pytorch content\nTrain with PyTorch Trainer\n\n🤗 Transformers provides a Trainer class optimized for training 🤗 Transformers models, making it easier to start training without manually writing your own training loop. The Trainer API supports a wide range of training options and features such as logging, gradient accumulation, and mixed precision.\n\nStart by loading your model and specify the number of expected labels. From the Yelp Review dataset card, you know there are five labels:\n\nCopied\n>>> from transformers import AutoModelForSequenceClassification\n\n>>> model = AutoModelForSequenceClassification.from_pretrained(\"bert-base-cased\", num_labels=5)\n\nYou will see a warning about some of the pretrained weights not being used and some weights being randomly initialized. Don’t worry, this is completely normal! The pretrained head of the BERT model is discarded, and replaced with a randomly initialized classification head. You will fine-tune this new model head on your sequence classification task, transferring the knowledge of the pretrained model to it.\n\nTraining hyperparameters\n\nNext, create a TrainingArguments class which contains all the hyperparameters you can tune as well as flags for activating different training options. For this tutorial you can start with the default training hyperparameters, but feel free to experiment with these to find your optimal settings.\n\nSpecify where to save the checkpoints from your training:\n\nCopied\n>>> from transformers import TrainingArguments\n\n>>> training_args = TrainingArguments(output_dir=\"test_trainer\")\nEvaluate\n\nTrainer does not automatically evaluate model performance during training. You’ll need to pass Trainer a function to compute and report metrics. The 🤗 Evaluate library provides a simple accuracy function you can load with the evaluate.load (see this quicktour for more information) function:\n\nCopied\n>>> import numpy as np\n>>> import evaluate\n\n>>> metric = evaluate.load(\"accuracy\")\n\nCall compute on metric to calculate the accuracy of your predictions. Before passing your predictions to compute, you need to convert the predictions to logits (remember all 🤗 Transformers models return logits):\n\nCopied\n>>> def compute_metrics(eval_pred):\n... logits, labels = eval_pred\n... predictions = np.argmax(logits, axis=-1)\n... return metric.compute(predictions=predictions, references=labels)\n\nIf you’d like to monitor your evaluation metrics during fine-tuning, specify the evaluation_strategy parameter in your training arguments to report the evaluation metric at the end of each epoch:\n\nCopied\n>>> from transformers import TrainingArguments, Trainer\n\n>>> training_args = TrainingArguments(output_dir=\"test_trainer\", evaluation_strategy=\"epoch\")\nTrainer\n\nCreate a Trainer object with your model, training arguments, training and test datasets, and evaluation function:\n\nCopied\n>>> trainer = Trainer(\n... model=model,\n... args=training_args,\n... train_dataset=small_train_dataset,\n... eval_dataset=small_eval_dataset,\n... compute_metrics=compute_metrics,\n... )\n\nThen fine-tune your model by calling train():\n\nCopied\n>>> trainer.train()\nTensorFlow\nHide TensorFlow content\nTrain a TensorFlow model with Keras\n\nYou can also train 🤗 Transformers models in TensorFlow with the Keras API!\n\nLoading data for Keras\n\nWhen you want to train a 🤗 Transformers model with the Keras API, you need to convert your dataset to a format that Keras understands. If your dataset is small, you can just convert the whole thing to NumPy arrays and pass it to Keras. Let’s try that first before we do anything more complicated.\n\nFirst, load a dataset. We’ll use the CoLA dataset from the GLUE benchmark, since it’s a simple binary text classification task, and just take the training split for now.\n\nCopied\nfrom datasets import load_dataset\n\ndataset = load_dataset(\"glue\", \"cola\")\ndataset = dataset[\"train\"] # Just take the training split for now\n\nNext, load a tokenizer and tokenize the data as NumPy arrays. Note that the labels are already a list of 0 and 1s, so we can just convert that directly to a NumPy array without tokenization!\n\nCopied\nfrom transformers import AutoTokenizer\n\ntokenizer = AutoTokenizer.from_pretrained(\"bert-base-cased\")\ntokenized_data = tokenizer(dataset[\"sentence\"], return_tensors=\"np\", padding=True)\n# Tokenizer returns a BatchEncoding, but we convert that to a dict for Keras\ntokenized_data = dict(tokenized_data)\n\nlabels = np.array(dataset[\"label\"]) # Label is already an array of 0 and 1\n\nFinally, load, compile, and fit the model:\n\nCopied\nfrom transformers import TFAutoModelForSequenceClassification\nfrom tensorflow.keras.optimizers import Adam\n\n# Load and compile our model\nmodel = TFAutoModelForSequenceClassification.from_pretrained(\"bert-base-cased\")\n# Lower learning rates are often better for fine-tuning transformers\nmodel.compile(optimizer=Adam(3e-5))\n\nmodel.fit(tokenized_data, labels)\n\nYou don’t have to pass a loss argument to your models when you compile() them! Hugging Face models automatically choose a loss that is appropriate for their task and model architecture if this argument is left blank. You can always override this by specifying a loss yourself if you want to!\n\nThis approach works great for smaller datasets, but for larger datasets, you might find it starts to become a problem. Why? Because the tokenized array and labels would have to be fully loaded into memory, and because NumPy doesn’t handle “jagged” arrays, so every tokenized sample would have to be padded to the length of the longest sample in the whole dataset. That’s going to make your array even bigger, and all those padding tokens will slow down training too!\n\nLoading data as a tf.data.Dataset\n\nIf you want to avoid slowing down training, you can load your data as a tf.data.Dataset instead. Although you can write your own tf.data pipeline if you want, we have two convenience methods for doing this:\n\nprepare_tf_dataset(): This is the method we recommend in most cases. Because it is a method on your model, it can inspect the model to automatically figure out which columns are usable as model inputs, and discard the others to make a simpler, more performant dataset.\nto_tf_dataset: This method is more low-level, and is useful when you want to exactly control how your dataset is created, by specifying exactly which columns and label_cols to include.\n\nBefore you can use prepare_tf_dataset(), you will need to add the tokenizer outputs to your dataset as columns, as shown in the following code sample:\n\nCopied\ndef tokenize_dataset(data):\n # Keys of the returned dictionary will be added to the dataset as columns\n return tokenizer(data[\"text\"])\n\n\ndataset = dataset.map(tokenize_dataset)\n\nRemember that Hugging Face datasets are stored on disk by default, so this will not inflate your memory usage! Once the columns have been added, you can stream batches from the dataset and add padding to each batch, which greatly reduces the number of padding tokens compared to padding the entire dataset.\n\nCopied\n>>> tf_dataset = model.prepare_tf_dataset(dataset, batch_size=16, shuffle=True, tokenizer=tokenizer)\n\nNote that in the code sample above, you need to pass the tokenizer to prepare_tf_dataset so it can correctly pad batches as they’re loaded. If all the samples in your dataset are the same length and no padding is necessary, you can skip this argument. If you need to do something more complex than just padding samples (e.g. corrupting tokens for masked language modelling), you can use the collate_fn argument instead to pass a function that will be called to transform the list of samples into a batch and apply any preprocessing you want. See our examples or notebooks to see this approach in action.\n\nOnce you’ve created a tf.data.Dataset, you can compile and fit the model as before:\n\nCopied\nmodel.compile(optimizer=Adam(3e-5))\n\nmodel.fit(tf_dataset)\nTrain in native PyTorch\nPytorch\nHide Pytorch content\n\nTrainer takes care of the training loop and allows you to fine-tune a model in a single line of code. For users who prefer to write their own training loop, you can also fine-tune a 🤗 Transformers model in native PyTorch.\n\nAt this point, you may need to restart your notebook or execute the following code to free some memory:\n\nCopied\ndel model\ndel trainer\ntorch.cuda.empty_cache()\n\nNext, manually postprocess tokenized_dataset to prepare it for training.\n\nRemove the text column because the model does not accept raw text as an input:\n\nCopied\n>>> tokenized_datasets = tokenized_datasets.remove_columns([\"text\"])\n\nRename the label column to labels because the model expects the argument to be named labels:\n\nCopied\n>>> tokenized_datasets = tokenized_datasets.rename_column(\"label\", \"labels\")\n\nSet the format of the dataset to return PyTorch tensors instead of lists:\n\nCopied\n>>> tokenized_datasets.set_format(\"torch\")\n\nThen create a smaller subset of the dataset as previously shown to speed up the fine-tuning:\n\nCopied\n>>> small_train_dataset = tokenized_datasets[\"train\"].shuffle(seed=42).select(range(1000))\n>>> small_eval_dataset = tokenized_datasets[\"test\"].shuffle(seed=42).select(range(1000))\nDataLoader\n\nCreate a DataLoader for your training and test datasets so you can iterate over batches of data:\n\nCopied\n>>> from torch.utils.data import DataLoader\n\n>>> train_dataloader = DataLoader(small_train_dataset, shuffle=True, batch_size=8)\n>>> eval_dataloader = DataLoader(small_eval_dataset, batch_size=8)\n\nLoad your model with the number of expected labels:\n\nCopied\n>>> from transformers import AutoModelForSequenceClassification\n\n>>> model = AutoModelForSequenceClassification.from_pretrained(\"bert-base-cased\", num_labels=5)\nOptimizer and learning rate scheduler\n\nCreate an optimizer and learning rate scheduler to fine-tune the model. Let’s use the AdamW optimizer from PyTorch:\n\nCopied\n>>> from torch.optim import AdamW\n\n>>> optimizer = AdamW(model.parameters(), lr=5e-5)\n\nCreate the default learning rate scheduler from Trainer:\n\nCopied\n>>> from transformers import get_scheduler\n\n>>> num_epochs = 3\n>>> num_training_steps = num_epochs * len(train_dataloader)\n>>> lr_scheduler = get_scheduler(\n... name=\"linear\", optimizer=optimizer, num_warmup_steps=0, num_training_steps=num_training_steps\n... )\n\nLastly, specify device to use a GPU if you have access to one. Otherwise, training on a CPU may take several hours instead of a couple of minutes.\n\nCopied\n>>> import torch\n\n>>> device = torch.device(\"cuda\") if torch.cuda.is_available() else torch.device(\"cpu\")\n>>> model.to(device)\n\nGet free access to a cloud GPU if you don’t have one with a hosted notebook like Colaboratory or SageMaker StudioLab.\n\nGreat, now you are ready to train! 🥳\n\nTraining loop\n\nTo keep track of your training progress, use the tqdm library to add a progress bar over the number of training steps:\n\nCopied\n>>> from tqdm.auto import tqdm\n\n>>> progress_bar = tqdm(range(num_training_steps))\n\n>>> model.train()\n>>> for epoch in range(num_epochs):\n... for batch in train_dataloader:\n... batch = {k: v.to(device) for k, v in batch.items()}\n... outputs = model(**batch)\n... loss = outputs.loss\n... loss.backward()\n\n... optimizer.step()\n... lr_scheduler.step()\n... optimizer.zero_grad()\n... progress_bar.update(1)\nEvaluate\n\nJust like how you added an evaluation function to Trainer, you need to do the same when you write your own training loop. But instead of calculating and reporting the metric at the end of each epoch, this time you’ll accumulate all the batches with add_batch and calculate the metric at the very end.\n\nCopied\n>>> import evaluate\n\n>>> metric = evaluate.load(\"accuracy\")\n>>> model.eval()\n>>> for batch in eval_dataloader:\n... batch = {k: v.to(device) for k, v in batch.items()}\n... with torch.no_grad():\n... outputs = model(**batch)\n\n... logits = outputs.logits\n... predictions = torch.argmax(logits, dim=-1)\n... metric.add_batch(predictions=predictions, references=batch[\"labels\"])\n\n>>> metric.compute()\nAdditional resources\n\nFor more fine-tuning examples, refer to:\n\n🤗 Transformers Examples includes scripts to train common NLP tasks in PyTorch and TensorFlow.\n\n🤗 Transformers Notebooks contains various notebooks on how to fine-tune a model for specific tasks in PyTorch and TensorFlow."}]
|
DataSet/Json/openchatkitreadme.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
[{"propertyName1":"OpenChatKit\n\nOpenChatKit provides a powerful, open-source base to create both specialized and general purpose chatbots for various applications. The kit includes an instruction-tuned 20 billion parameter language model, a 6 billion parameter moderation model, and an extensible retrieval system for including up-to-date responses from custom repositories. It was trained on the OIG-43M training dataset, which was a collaboration between Together, LAION, and Ontocord.ai. Much more than a model release, this is the beginning of an open source project. We are releasing a set of tools and processes for ongoing improvement with community contributions.\n\nIn this repo, you'll find code for:\n\nTraining an OpenChatKit model\nTesting inference using the model\nAugmenting the model with additional context from a retrieval index\nContents\nRequirements\nPre-trained Weights\nDatasets\nData Contributions\nPretrained Base Model\nTraining and Finetuning\n(Optional) 8bit Adam\nTrain GPT-NeoX-Chat-Base-20B\nConverting Weights to Huggingface Format\nInference\nMonitoring\nLoguru\nWeights & Biases\nExperimental: Retrieval-Augmented Models\nLicense\nCiting OpenChatKit\nAcknowledgements\nRequirements\n\nBefore you begin, you need to install PyTorch and other dependencies.\n\nInstall Miniconda from their website.\nCreate an environment called OpenChatKit using the environment.yml file at the root of this repo.\nconda env create -f environment.yml\n\nThis repo also uses Git LFS to manage some files. Install it using the instructions on their site then run:\n\ngit lfs install\nPre-trained Weights\n\nGPT-NeoXT-Chat-Base-20B is a 20B-parameter variant of GPT-NeoX, fine-tuned on conversational datasets. We are releasing pre-trained weights for this model as togethercomputer/GPT-NeoXT-Chat-Base-20B on Huggingface.\n\nMore details can be found on the model card for GPT-NeoXT-Chat-Base-20B on Huggingface.\n\nDatasets\n\nThe chat model was trained on the OIG dataset built by LAION, Together, and Ontocord.ai. To download the dataset from Huggingface run the command below from the root of the repo.\n\npython data/OIG/prepare.py\n\nOnce the command completes, the data will be in the data/OIG/files directory.\n\nData Contributions\n\nYou can help make this chat model better by contributing data! See the OpenDataHub repo for more details.\n\nPretrained Base Model\n\nAs mentioned above, the chat model is a fine-tuned variant of GPT-NeoX-20B from Eleuther AI. To download GPT-NeoX-20B and prepare it for fine tuning, run this command from the root of the repo.\n\npython pretrained/GPT-NeoX-20B/prepare.py\n\nThe weights for this model will be in the pretrained/GPT-NeoX-20B/EleutherAI_gpt-neox-20b.\n\nTraining and Finetuning\n(Optional) 8bit Adam\n\nTo use 8bit-adam during training, install the bitsandbytes package.\n\npip install bitsandbytes # optional, to use 8bit-adam\nTrain GPT-NeoX-Chat-Base-20B\n\nThe training/finetune_GPT-NeoXT-Chat-Base-20B.sh script configures and runs the training loop. After downloading the dataset and the base model, run:\n\nbash training/finetune_GPT-NeoXT-Chat-Base-20B.sh\n\nThe script launches 8 processes with a pipeline-parallel degree of 8 and a data-parallel degree of 1.\n\nAs the training loop runs, checkpoints are saved to the model_ckpts directory at the root of the repo.\n\nPlease see the training README for more details about customizing the training run.\n\nConverting Weights to Huggingface Format\n\nBefore you can use this model to perform inference, it must be converted to the Hugginface format.\n\nmkdir huggingface_models \\\n&& python tools/convert_to_hf_gptneox.py \\\n --ckpt-path model_ckpts/GPT-Neo-XT-Chat-Base-20B/checkpoint_5 \n --save-path /huggingface_models/GPT-NeoXT-Chat-Base-20B \n --n-stages 8 \n --n-layer-per-stage 6\nInference\n\nTo help you test the model, we provide a simple test command line test harness to interact with the bot.\n\npython inference/bot.py\n\nBy default the script will load the model named GPT-NeoXT-Chat-Base-20B model under the huggingface_models directory, but you can override that behavior by specifying --model.\n\nFor example, if you want to load the base model from our Huggingface, repo, you can run the following command which downloads the weights from HuggingFace.\n\npython inference/bot.py --model togethercomputer/GPT-NeoXT-Chat-Base-20B\n\nOnce the model has loaded, enter text at the prompt and the model will reply.\n\n$ python inference/bot.py \nLoading /home/csris/src/github.com/togethercomputer/OpenChatKit/inference/../huggingface_models/GPT-NeoXT-Chat-Base-20B to cuda:1...\nWelcome to OpenChatKit shell. Type /help or /? to list commands.\n\n>>> Hello.\nSetting `pad_token_id` to `eos_token_id`:0 for open-end generation.\nHello human.\n\n>>> \n\nCommands are prefixed with a /, and the /quit command exits.\n\nMonitoring\n\nBy default, the training script simply prints the loss as training proceeds, but it can also output metrics to a file using loguru or report them to Weights & Biases.\n\nLoguru\n\nAdd the flag --train-log-backend loguru to your training script to log to ./logs/file_{time}.log\n\nWeights & Biases\n\nTo use Weights & Biases, first login with your Weights & Biases token.\n\nwandb login\n\nAnd set --train-log-backend wandb in the training script to enable logging to Weights & Biases.\n\nExperimental: Retrieval-Augmented Models\n\nNote: Retrieval is still experimental.\n\nThe code in /retrieval implements a python package for querying a Faiss index of Wikipedia. The following steps explain how to use this index to augment queries in the test harness with context from the retriever.\n\nDownload the Wikipedia index.\npython data/wikipedia-3sentence-level-retrieval-index/prepare.py\nRun the bot with the --retrieval flag.\npython inference/bot.py --retrieval\n\nAfter starting, the bot will load both the chat model and the retrieval index, which takes a long time. Once the model and the index are loaded, all queries will be augmented with extra context.\n\n$ python inference/bot.py --retrieval\nLoading /OpenChatKit/inference/../huggingface_models/GPT-NeoXT-Chat-Base-20B to cuda:0...\nLoading retrieval index...\nWelcome to OpenChatKit shell. Type /help or /? to list commands.\n\n>>> Where is Zurich?\nSetting `pad_token_id` to `eos_token_id`:0 for open-end generation.\nWhere is Zurich?\nZurich is located in Switzerland.\n\n>>>\nLicense\n\nAll code in this repository was developed by Together Computer except where otherwise noted. Copyright (c) 2023, Together Computer. All rights reserved. The code is licensed under the Apache 2.0 license.\n\nCopyright 2023 Together Computer\n\nLicensed under the Apache License, Version 2.0 (the \"License\");\nyou may not use this file except in compliance with the License.\nYou may obtain a copy of the License at\n\n http://www.apache.org/licenses/LICENSE-2.0\n\nUnless required by applicable law or agreed to in writing, software\ndistributed under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for the specific language governing permissions and\nlimitations under the License.\n\n\nThis repository also contains code written by a number of other authors. Such contributions are marked and the relevant licensing is included where appropriate.\n\nFor full terms, see the LICENSE file. If you have any questions, comments, or concerns about licensing please contact us.\n\nCiting OpenChatKit\n@software{openchatkit,\n title = {{OpenChatKit: An Open Toolkit and Base Model for Dialogue-style Applications}},\n author = {Together Computer},\n url = {https://github.com/togethercomputer/OpenChatKit}\n month = {3},\n year = {2023},\n version = {0.15},\n}\nAcknowledgements\n\nOur model is a fine-tuned version of gpt-neox-20b, a large language model trained by Eleuther AI. We evaluated our model on HELM provided by the Center for Research on Foundation Models. And we collaborated with both CRFM and HazyResearch at Stanford to build this model.\n\nWe collaborated with LAION and Ontocord.ai to build the training data used to fine tune this model."}]
|
DataSet/Json/train.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
DataSet/train.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
README.md
ADDED
|
@@ -0,0 +1,192 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
---
|
| 6 |
+
|
| 7 |
+
***<p style="font-size: 24px">Feel free to try out our [OpenChatKit feedback app](https://huggingface.co/spaces/togethercomputer/OpenChatKit)!</p>***
|
| 8 |
+
|
| 9 |
+
# GPT-NeoXT-Chat-Base-20B
|
| 10 |
+
|
| 11 |
+
> TLDR: As part of OpenChatKit (codebase available [here](https://github.com/togethercomputer/OpenChaT)),
|
| 12 |
+
> GPT-NeoXT-Chat-Base-20B is a 20B parameter language model, fine-tuned from EleutherAI’s GPT-NeoX with over 40 million instructions on 100% carbon negative compute.
|
| 13 |
+
|
| 14 |
+
GPT-NeoXT-Chat-Base-20B is based on ElutherAI’s GPT-NeoX model, and is fine-tuned with data focusing on dialog-style interactions.
|
| 15 |
+
We focused the tuning on several tasks such as question answering, classification, extraction, and summarization.
|
| 16 |
+
We’ve fine-tuned the model with a collection of 43 million high-quality instructions.
|
| 17 |
+
Together partnered with LAION and Ontocord.ai, who both helped curate the dataset the model is based on.
|
| 18 |
+
You can read more about this process and the availability of this dataset in LAION’s blog post [here](https://laion.ai/blog/oig-dataset/).
|
| 19 |
+
|
| 20 |
+
## Model Details
|
| 21 |
+
- **Developed by**: Together Computer.
|
| 22 |
+
- **Model type**: Language Model
|
| 23 |
+
- **Language(s)**: English
|
| 24 |
+
- **License**: Apache 2.0
|
| 25 |
+
- **Model Description**: A 20B parameter open source chat model, fine-tuned from EleutherAI’s NeoX with over 40M instructions on 100% carbon negative compute
|
| 26 |
+
- **Resources for more information**: [GitHub Repository](https://github.com/togethercomputer/OpenChaT).
|
| 27 |
+
|
| 28 |
+
# Quick Start
|
| 29 |
+
|
| 30 |
+
```python
|
| 31 |
+
from transformers import pipeline
|
| 32 |
+
pipe = pipeline(model='togethercomputer/GPT-NeoXT-Chat-Base-20B')
|
| 33 |
+
pipe('''<human>: Hello!\n<bot>:''')
|
| 34 |
+
```
|
| 35 |
+
or
|
| 36 |
+
```python
|
| 37 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 38 |
+
tokenizer = AutoTokenizer.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B")
|
| 39 |
+
model = AutoModelForCausalLM.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B")
|
| 40 |
+
```
|
| 41 |
+
|
| 42 |
+
## Strengths of the model
|
| 43 |
+
|
| 44 |
+
There are several tasks that OpenChatKit excels at out of the box. This includes:
|
| 45 |
+
|
| 46 |
+
- Example 1: Summarization and question answering within context.
|
| 47 |
+
|
| 48 |
+
```markdown
|
| 49 |
+
**Summarize a long document into a single sentence and conduct question answering related to the document, with multiple rounds**
|
| 50 |
+
|
| 51 |
+
<human>: Last year, the travel industry saw a big rebound in demand — and that demand is showing no signs of slowing down this spring break travel season. Planes and hotels will be full, travelers will likely face long queues, cancellations, massive crowds and plenty of other travel nightmares. But perhaps the most frustrating thing you’ll have to prepare for this spring break is if your luggage goes missing. You’ve got to somehow track down your belongings that were in the airline’s care and then wait to be reunited — all while trying to enjoy the start of your long-awaited vacation. As far as travel nightmares go, lost luggage is up there as one of the worst.
|
| 52 |
+
|
| 53 |
+
To be honest up front, I am great at strategically losing things — or rather, losing stuff with style. (I hope you appreciate the “Toy Story” reference.) So when a product comes along that’s a perfect way to keep track of my belongings — especially when traveling — I’m adding it to my travel bag.
|
| 54 |
+
|
| 55 |
+
I’ve been a big fan of Apple’s $29 item tracker, dubbed AirTag. The long-rumored AirTag product finally launched in 2021, and while it’s not the first item tracker to exist, it has two key features that set it apart from the rest to make it the ultimate travel companion: access to the Find My network and Precision Finding technology.
|
| 56 |
+
|
| 57 |
+
The Federal Aviation Administration has even ruled that having an AirTag in your luggage is permitted and, in fact, not a safety hazard. The devices only contain 0.1 grams of lithium, which falls below the FAA’s limit.
|
| 58 |
+
|
| 59 |
+
“Luggage tracking devices powered by lithium metal cells that have 0.3 grams or less of lithium can be used on checked baggage,” the FAA said in a statement. “Apple AirTags meet this threshold; other luggage tracking devices may not.”
|
| 60 |
+
|
| 61 |
+
But, you may be wondering what makes them so special and why you should carry one in your checked luggage this spring break? Let’s explain why.
|
| 62 |
+
|
| 63 |
+
Summarize the above into a single paragraph.
|
| 64 |
+
|
| 65 |
+
<bot>: If you're traveling this spring break, you might want to consider investing in an item tracker to help you find your lost luggage. Apple's AirTag is the best option, according to a new report.
|
| 66 |
+
|
| 67 |
+
<human>: But isn't it unsafe to put AirTag in a luggage?
|
| 68 |
+
|
| 69 |
+
<bot>: The Federal Aviation Administration has ruled that having an AirTag in your luggage is permitted and, in fact, not a safety hazard.
|
| 70 |
+
|
| 71 |
+
<human>: Ah, why is that?
|
| 72 |
+
|
| 73 |
+
<bot>: The devices only contain 0.1 grams of lithium, which falls below the FAA’s limit.
|
| 74 |
+
```
|
| 75 |
+
|
| 76 |
+
- Example 2: Extraction.
|
| 77 |
+
|
| 78 |
+
```markdown
|
| 79 |
+
**Extract structured information from unstructured documents, as illustrated in the following example.**
|
| 80 |
+
|
| 81 |
+
<human>: Generate a table summarizing the options outlined in this email.
|
| 82 |
+
|
| 83 |
+
Team, we have to make a decision on the product plan for Goocrux v2.
|
| 84 |
+
|
| 85 |
+
The first option will take us 3 weeks of development, and will achieve performance of 80ms, but means that the UI has to change.
|
| 86 |
+
|
| 87 |
+
The second option will take us 4 weeks of development, and will achieve performance of 60ms, without any UI change.
|
| 88 |
+
|
| 89 |
+
The third option will take the longest -- 6 weeks of development, will achieve performance of 30ms, without any UI change.
|
| 90 |
+
|
| 91 |
+
A markdown table with four columns: (1) Option ID, (2) Development time, (3) Performance, (4) UI Change:
|
| 92 |
+
|
| 93 |
+
<bot>: #.
|
| 94 |
+
| ID | Option ID | Development time | Performance | UI Change |
|
| 95 |
+
| --------- | --------- | --------- | --------- | --------- |
|
| 96 |
+
| 1 | 1 | 3 weeks | 80ms | Yes |
|
| 97 |
+
| 2 | 2 | 4 weeks | 60ms | No |
|
| 98 |
+
| 3 | 3 | 6 weeks | 30ms | No |
|
| 99 |
+
```
|
| 100 |
+
|
| 101 |
+
- Example 3: Classification.
|
| 102 |
+
|
| 103 |
+
```markdown
|
| 104 |
+
**Classifying a sentence or paragraph into different categories.**
|
| 105 |
+
|
| 106 |
+
<human>: Classify the sentiment of the following sentence into Positive, Neutral, or Negative: Super excited about teaching Stanford’s first course on Large Language Models! Check the syllabus out here
|
| 107 |
+
<bot>: Positive
|
| 108 |
+
|
| 109 |
+
<human>: Classify the sentiment of the following sentence into Positive, Neutral, or Negative: How about the following sentence: It is raining outside and I feel so blue
|
| 110 |
+
<bot>: Negative
|
| 111 |
+
```
|
| 112 |
+
|
| 113 |
+
In addition, the model does well on few-shot prompts. For both classification and extraction, the model performs even better with few shots, as in most HELM tasks. [Contact us](https://www.together.xyz/contact) if you’re interested in trying few-shot prompts with the model.
|
| 114 |
+
|
| 115 |
+
## Weaknesses of the model
|
| 116 |
+
|
| 117 |
+
That said, there are several areas where we have more work to do, and we need your help! Some of these include:
|
| 118 |
+
|
| 119 |
+
- Knowledge-based closed question and answering: The chatbot may hallucinate and give incorrect results. Be sure to fact check, and if possible provide feedback with the corrected information.
|
| 120 |
+
- Coding tasks: The chatbot was not trained on a large enough corpus of source code to excel at writing code. We welcome contributions of additional datasets to improve this!
|
| 121 |
+
- Repetition: Sometimes the chatbot will repeat its response. We’re working to improve this, but in the meantime you can click the refresh button to start a new conversation.
|
| 122 |
+
- Context switching: If you change the topic in the middle of a conversation the chatbot often cannot make the switch automatically and will continue to give answers related to the prior topic.
|
| 123 |
+
- Creative writing and longer answers: The chatbot does not generate long, creative text such as an essay or story.
|
| 124 |
+
|
| 125 |
+
We are excited to work with you to address these weaknesses by getting your feedback, bolstering data sets, and improving accuracy.
|
| 126 |
+
|
| 127 |
+
# Uses
|
| 128 |
+
|
| 129 |
+
## Direct Use
|
| 130 |
+
|
| 131 |
+
The model is intended for research purposes. Possible research areas and tasks include
|
| 132 |
+
|
| 133 |
+
- Safe deployment of models which have the potential to generate harmful content.
|
| 134 |
+
- Probing and understanding the limitations and biases of dialogue models or language models.
|
| 135 |
+
- Generation of artworks and use in design and other artistic processes.
|
| 136 |
+
- Applications in educational or creative tools.
|
| 137 |
+
- Research on dialogue models or language models.
|
| 138 |
+
|
| 139 |
+
Excluded uses are described below.
|
| 140 |
+
|
| 141 |
+
### Misuse, Malicious Use, and Out-of-Scope Use
|
| 142 |
+
|
| 143 |
+
The OpenChatKit community provides GPT-NeoXT-Chat-Base-20B as an open source tool for building chatbots.
|
| 144 |
+
The community is not responsible for any misuse, malicious use, or out-of-scope use of the model.
|
| 145 |
+
It is the responsibility of the end user to ensure that the model is used in a responsible and ethical manner.
|
| 146 |
+
|
| 147 |
+
#### Out-of-Scope Use
|
| 148 |
+
|
| 149 |
+
GPT-NeoXT-Chat-Base-20B is designed for use in chatbot applications and may not perform well for other use cases outside of its intended scope.
|
| 150 |
+
For example, it may not be suitable for use in safety-critical applications or for making decisions that have a significant impact on individuals or society.
|
| 151 |
+
It is important to consider the limitations of the model and to only use it for its intended purpose.
|
| 152 |
+
|
| 153 |
+
#### Misuse and Malicious Use
|
| 154 |
+
|
| 155 |
+
GPT-NeoXT-Chat-Base-20B is designed for use in chatbot applications and should not be used for any other purpose.
|
| 156 |
+
Misuse of the model, such as using it to engage in illegal or unethical activities, is strictly prohibited and goes against the principles of the OpenChatKit community project.
|
| 157 |
+
|
| 158 |
+
Using the model to generate content that is cruel to individuals is a misuse of this model. This includes, but is not limited to:
|
| 159 |
+
|
| 160 |
+
- Generating fake news, misinformation, or propaganda
|
| 161 |
+
- Promoting hate speech, discrimination, or violence against individuals or groups
|
| 162 |
+
- Impersonating individuals or organizations without their consent
|
| 163 |
+
- Engaging in cyberbullying or harassment
|
| 164 |
+
- Defamatory content
|
| 165 |
+
- Spamming or scamming
|
| 166 |
+
- Sharing confidential or sensitive information without proper authorization
|
| 167 |
+
- Violating the terms of use of the model or the data used to train it
|
| 168 |
+
- Creating automated bots for malicious purposes such as spreading malware, phishing scams, or spamming
|
| 169 |
+
|
| 170 |
+
## Limitations
|
| 171 |
+
|
| 172 |
+
GPT-NeoXT-Chat-Base-20B, like other language model-based chatbots, has limitations that should be taken into consideration.
|
| 173 |
+
For example, the model may not always provide accurate or relevant answers, particularly for questions that are complex, ambiguous, or outside of its training data.
|
| 174 |
+
We therefore welcome contributions from individuals and organizations, and encourage collaboration towards creating a more robust and inclusive chatbot.
|
| 175 |
+
|
| 176 |
+
## Training
|
| 177 |
+
|
| 178 |
+
**Training Data**
|
| 179 |
+
|
| 180 |
+
Please refer to [togethercomputer/OpenDataHub](https://github.com/togethercomputer/OpenDataHub)
|
| 181 |
+
|
| 182 |
+
**Training Procedure**
|
| 183 |
+
|
| 184 |
+
- **Hardware:** 2 x 8 x A100 GPUs
|
| 185 |
+
- **Optimizer:** [8bit-AdamW](https://github.com/TimDettmers/bitsandbytes)
|
| 186 |
+
- **Gradient Accumulations**: 2
|
| 187 |
+
- **Batch:** 2 x 2 x 64 x 2048 = 524288 tokens
|
| 188 |
+
- **Learning rate:** warmup to 1e-6 for 100 steps and then kept constant
|
| 189 |
+
|
| 190 |
+
## Community
|
| 191 |
+
|
| 192 |
+
Join us on [Together Discord](https://discord.gg/6ZVDU8tTD4)
|
pytorch_model.bin.index.json
ADDED
|
@@ -0,0 +1,671 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 41293685880
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"embed_out.weight": "pytorch_model-00005-of-00005.bin",
|
| 7 |
+
"gpt_neox.embed_in.weight": "pytorch_model-00001-of-00005.bin",
|
| 8 |
+
"gpt_neox.final_layer_norm.bias": "pytorch_model-00005-of-00005.bin",
|
| 9 |
+
"gpt_neox.final_layer_norm.weight": "pytorch_model-00005-of-00005.bin",
|
| 10 |
+
"gpt_neox.layers.0.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 11 |
+
"gpt_neox.layers.0.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 12 |
+
"gpt_neox.layers.0.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 13 |
+
"gpt_neox.layers.0.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 14 |
+
"gpt_neox.layers.0.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 15 |
+
"gpt_neox.layers.0.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 16 |
+
"gpt_neox.layers.0.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 17 |
+
"gpt_neox.layers.0.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 18 |
+
"gpt_neox.layers.0.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 19 |
+
"gpt_neox.layers.0.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 20 |
+
"gpt_neox.layers.0.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 21 |
+
"gpt_neox.layers.0.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 22 |
+
"gpt_neox.layers.0.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 23 |
+
"gpt_neox.layers.0.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 24 |
+
"gpt_neox.layers.0.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 25 |
+
"gpt_neox.layers.1.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 26 |
+
"gpt_neox.layers.1.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 27 |
+
"gpt_neox.layers.1.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 28 |
+
"gpt_neox.layers.1.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 29 |
+
"gpt_neox.layers.1.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 30 |
+
"gpt_neox.layers.1.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 31 |
+
"gpt_neox.layers.1.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 32 |
+
"gpt_neox.layers.1.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 33 |
+
"gpt_neox.layers.1.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 34 |
+
"gpt_neox.layers.1.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 35 |
+
"gpt_neox.layers.1.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 36 |
+
"gpt_neox.layers.1.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 37 |
+
"gpt_neox.layers.1.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 38 |
+
"gpt_neox.layers.1.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 39 |
+
"gpt_neox.layers.1.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 40 |
+
"gpt_neox.layers.10.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 41 |
+
"gpt_neox.layers.10.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 42 |
+
"gpt_neox.layers.10.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 43 |
+
"gpt_neox.layers.10.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 44 |
+
"gpt_neox.layers.10.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 45 |
+
"gpt_neox.layers.10.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 46 |
+
"gpt_neox.layers.10.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 47 |
+
"gpt_neox.layers.10.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 48 |
+
"gpt_neox.layers.10.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 49 |
+
"gpt_neox.layers.10.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 50 |
+
"gpt_neox.layers.10.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 51 |
+
"gpt_neox.layers.10.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 52 |
+
"gpt_neox.layers.10.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 53 |
+
"gpt_neox.layers.10.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 54 |
+
"gpt_neox.layers.10.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 55 |
+
"gpt_neox.layers.11.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 56 |
+
"gpt_neox.layers.11.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 57 |
+
"gpt_neox.layers.11.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 58 |
+
"gpt_neox.layers.11.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 59 |
+
"gpt_neox.layers.11.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 60 |
+
"gpt_neox.layers.11.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 61 |
+
"gpt_neox.layers.11.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 62 |
+
"gpt_neox.layers.11.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 63 |
+
"gpt_neox.layers.11.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 64 |
+
"gpt_neox.layers.11.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 65 |
+
"gpt_neox.layers.11.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 66 |
+
"gpt_neox.layers.11.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 67 |
+
"gpt_neox.layers.11.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 68 |
+
"gpt_neox.layers.11.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 69 |
+
"gpt_neox.layers.11.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 70 |
+
"gpt_neox.layers.12.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 71 |
+
"gpt_neox.layers.12.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 72 |
+
"gpt_neox.layers.12.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 73 |
+
"gpt_neox.layers.12.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 74 |
+
"gpt_neox.layers.12.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 75 |
+
"gpt_neox.layers.12.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 76 |
+
"gpt_neox.layers.12.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 77 |
+
"gpt_neox.layers.12.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 78 |
+
"gpt_neox.layers.12.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 79 |
+
"gpt_neox.layers.12.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 80 |
+
"gpt_neox.layers.12.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 81 |
+
"gpt_neox.layers.12.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 82 |
+
"gpt_neox.layers.12.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 83 |
+
"gpt_neox.layers.12.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 84 |
+
"gpt_neox.layers.12.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 85 |
+
"gpt_neox.layers.13.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 86 |
+
"gpt_neox.layers.13.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 87 |
+
"gpt_neox.layers.13.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 88 |
+
"gpt_neox.layers.13.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 89 |
+
"gpt_neox.layers.13.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 90 |
+
"gpt_neox.layers.13.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 91 |
+
"gpt_neox.layers.13.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 92 |
+
"gpt_neox.layers.13.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 93 |
+
"gpt_neox.layers.13.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 94 |
+
"gpt_neox.layers.13.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 95 |
+
"gpt_neox.layers.13.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 96 |
+
"gpt_neox.layers.13.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 97 |
+
"gpt_neox.layers.13.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 98 |
+
"gpt_neox.layers.13.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 99 |
+
"gpt_neox.layers.13.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 100 |
+
"gpt_neox.layers.14.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 101 |
+
"gpt_neox.layers.14.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 102 |
+
"gpt_neox.layers.14.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 103 |
+
"gpt_neox.layers.14.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 104 |
+
"gpt_neox.layers.14.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 105 |
+
"gpt_neox.layers.14.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 106 |
+
"gpt_neox.layers.14.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 107 |
+
"gpt_neox.layers.14.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 108 |
+
"gpt_neox.layers.14.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 109 |
+
"gpt_neox.layers.14.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 110 |
+
"gpt_neox.layers.14.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 111 |
+
"gpt_neox.layers.14.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 112 |
+
"gpt_neox.layers.14.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 113 |
+
"gpt_neox.layers.14.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 114 |
+
"gpt_neox.layers.14.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 115 |
+
"gpt_neox.layers.15.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 116 |
+
"gpt_neox.layers.15.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 117 |
+
"gpt_neox.layers.15.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 118 |
+
"gpt_neox.layers.15.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 119 |
+
"gpt_neox.layers.15.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 120 |
+
"gpt_neox.layers.15.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 121 |
+
"gpt_neox.layers.15.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 122 |
+
"gpt_neox.layers.15.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 123 |
+
"gpt_neox.layers.15.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 124 |
+
"gpt_neox.layers.15.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 125 |
+
"gpt_neox.layers.15.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 126 |
+
"gpt_neox.layers.15.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 127 |
+
"gpt_neox.layers.15.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 128 |
+
"gpt_neox.layers.15.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 129 |
+
"gpt_neox.layers.15.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 130 |
+
"gpt_neox.layers.16.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 131 |
+
"gpt_neox.layers.16.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 132 |
+
"gpt_neox.layers.16.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 133 |
+
"gpt_neox.layers.16.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 134 |
+
"gpt_neox.layers.16.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 135 |
+
"gpt_neox.layers.16.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 136 |
+
"gpt_neox.layers.16.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 137 |
+
"gpt_neox.layers.16.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 138 |
+
"gpt_neox.layers.16.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 139 |
+
"gpt_neox.layers.16.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 140 |
+
"gpt_neox.layers.16.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 141 |
+
"gpt_neox.layers.16.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 142 |
+
"gpt_neox.layers.16.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 143 |
+
"gpt_neox.layers.16.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 144 |
+
"gpt_neox.layers.16.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 145 |
+
"gpt_neox.layers.17.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 146 |
+
"gpt_neox.layers.17.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 147 |
+
"gpt_neox.layers.17.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 148 |
+
"gpt_neox.layers.17.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 149 |
+
"gpt_neox.layers.17.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 150 |
+
"gpt_neox.layers.17.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 151 |
+
"gpt_neox.layers.17.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 152 |
+
"gpt_neox.layers.17.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 153 |
+
"gpt_neox.layers.17.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 154 |
+
"gpt_neox.layers.17.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 155 |
+
"gpt_neox.layers.17.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 156 |
+
"gpt_neox.layers.17.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 157 |
+
"gpt_neox.layers.17.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 158 |
+
"gpt_neox.layers.17.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 159 |
+
"gpt_neox.layers.17.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 160 |
+
"gpt_neox.layers.18.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 161 |
+
"gpt_neox.layers.18.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 162 |
+
"gpt_neox.layers.18.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 163 |
+
"gpt_neox.layers.18.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 164 |
+
"gpt_neox.layers.18.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 165 |
+
"gpt_neox.layers.18.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 166 |
+
"gpt_neox.layers.18.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 167 |
+
"gpt_neox.layers.18.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 168 |
+
"gpt_neox.layers.18.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 169 |
+
"gpt_neox.layers.18.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 170 |
+
"gpt_neox.layers.18.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 171 |
+
"gpt_neox.layers.18.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 172 |
+
"gpt_neox.layers.18.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 173 |
+
"gpt_neox.layers.18.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 174 |
+
"gpt_neox.layers.18.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 175 |
+
"gpt_neox.layers.19.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 176 |
+
"gpt_neox.layers.19.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 177 |
+
"gpt_neox.layers.19.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 178 |
+
"gpt_neox.layers.19.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 179 |
+
"gpt_neox.layers.19.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 180 |
+
"gpt_neox.layers.19.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 181 |
+
"gpt_neox.layers.19.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 182 |
+
"gpt_neox.layers.19.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 183 |
+
"gpt_neox.layers.19.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 184 |
+
"gpt_neox.layers.19.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 185 |
+
"gpt_neox.layers.19.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 186 |
+
"gpt_neox.layers.19.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 187 |
+
"gpt_neox.layers.19.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 188 |
+
"gpt_neox.layers.19.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 189 |
+
"gpt_neox.layers.19.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 190 |
+
"gpt_neox.layers.2.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 191 |
+
"gpt_neox.layers.2.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 192 |
+
"gpt_neox.layers.2.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 193 |
+
"gpt_neox.layers.2.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 194 |
+
"gpt_neox.layers.2.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 195 |
+
"gpt_neox.layers.2.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 196 |
+
"gpt_neox.layers.2.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 197 |
+
"gpt_neox.layers.2.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 198 |
+
"gpt_neox.layers.2.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 199 |
+
"gpt_neox.layers.2.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 200 |
+
"gpt_neox.layers.2.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 201 |
+
"gpt_neox.layers.2.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 202 |
+
"gpt_neox.layers.2.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 203 |
+
"gpt_neox.layers.2.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 204 |
+
"gpt_neox.layers.2.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 205 |
+
"gpt_neox.layers.20.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 206 |
+
"gpt_neox.layers.20.attention.dense.bias": "pytorch_model-00002-of-00005.bin",
|
| 207 |
+
"gpt_neox.layers.20.attention.dense.weight": "pytorch_model-00002-of-00005.bin",
|
| 208 |
+
"gpt_neox.layers.20.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 209 |
+
"gpt_neox.layers.20.attention.query_key_value.bias": "pytorch_model-00002-of-00005.bin",
|
| 210 |
+
"gpt_neox.layers.20.attention.query_key_value.weight": "pytorch_model-00002-of-00005.bin",
|
| 211 |
+
"gpt_neox.layers.20.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 212 |
+
"gpt_neox.layers.20.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 213 |
+
"gpt_neox.layers.20.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 214 |
+
"gpt_neox.layers.20.mlp.dense_4h_to_h.bias": "pytorch_model-00002-of-00005.bin",
|
| 215 |
+
"gpt_neox.layers.20.mlp.dense_4h_to_h.weight": "pytorch_model-00002-of-00005.bin",
|
| 216 |
+
"gpt_neox.layers.20.mlp.dense_h_to_4h.bias": "pytorch_model-00002-of-00005.bin",
|
| 217 |
+
"gpt_neox.layers.20.mlp.dense_h_to_4h.weight": "pytorch_model-00002-of-00005.bin",
|
| 218 |
+
"gpt_neox.layers.20.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 219 |
+
"gpt_neox.layers.20.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 220 |
+
"gpt_neox.layers.21.attention.bias": "pytorch_model-00002-of-00005.bin",
|
| 221 |
+
"gpt_neox.layers.21.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 222 |
+
"gpt_neox.layers.21.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 223 |
+
"gpt_neox.layers.21.attention.masked_bias": "pytorch_model-00002-of-00005.bin",
|
| 224 |
+
"gpt_neox.layers.21.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 225 |
+
"gpt_neox.layers.21.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 226 |
+
"gpt_neox.layers.21.attention.rotary_emb.inv_freq": "pytorch_model-00002-of-00005.bin",
|
| 227 |
+
"gpt_neox.layers.21.input_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 228 |
+
"gpt_neox.layers.21.input_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 229 |
+
"gpt_neox.layers.21.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 230 |
+
"gpt_neox.layers.21.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 231 |
+
"gpt_neox.layers.21.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 232 |
+
"gpt_neox.layers.21.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 233 |
+
"gpt_neox.layers.21.post_attention_layernorm.bias": "pytorch_model-00002-of-00005.bin",
|
| 234 |
+
"gpt_neox.layers.21.post_attention_layernorm.weight": "pytorch_model-00002-of-00005.bin",
|
| 235 |
+
"gpt_neox.layers.22.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 236 |
+
"gpt_neox.layers.22.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 237 |
+
"gpt_neox.layers.22.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 238 |
+
"gpt_neox.layers.22.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 239 |
+
"gpt_neox.layers.22.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 240 |
+
"gpt_neox.layers.22.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 241 |
+
"gpt_neox.layers.22.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 242 |
+
"gpt_neox.layers.22.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 243 |
+
"gpt_neox.layers.22.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 244 |
+
"gpt_neox.layers.22.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 245 |
+
"gpt_neox.layers.22.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 246 |
+
"gpt_neox.layers.22.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 247 |
+
"gpt_neox.layers.22.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 248 |
+
"gpt_neox.layers.22.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 249 |
+
"gpt_neox.layers.22.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 250 |
+
"gpt_neox.layers.23.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 251 |
+
"gpt_neox.layers.23.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 252 |
+
"gpt_neox.layers.23.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 253 |
+
"gpt_neox.layers.23.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 254 |
+
"gpt_neox.layers.23.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 255 |
+
"gpt_neox.layers.23.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 256 |
+
"gpt_neox.layers.23.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 257 |
+
"gpt_neox.layers.23.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 258 |
+
"gpt_neox.layers.23.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 259 |
+
"gpt_neox.layers.23.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 260 |
+
"gpt_neox.layers.23.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 261 |
+
"gpt_neox.layers.23.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 262 |
+
"gpt_neox.layers.23.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 263 |
+
"gpt_neox.layers.23.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 264 |
+
"gpt_neox.layers.23.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 265 |
+
"gpt_neox.layers.24.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 266 |
+
"gpt_neox.layers.24.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 267 |
+
"gpt_neox.layers.24.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 268 |
+
"gpt_neox.layers.24.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 269 |
+
"gpt_neox.layers.24.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 270 |
+
"gpt_neox.layers.24.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 271 |
+
"gpt_neox.layers.24.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 272 |
+
"gpt_neox.layers.24.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 273 |
+
"gpt_neox.layers.24.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 274 |
+
"gpt_neox.layers.24.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 275 |
+
"gpt_neox.layers.24.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 276 |
+
"gpt_neox.layers.24.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 277 |
+
"gpt_neox.layers.24.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 278 |
+
"gpt_neox.layers.24.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 279 |
+
"gpt_neox.layers.24.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 280 |
+
"gpt_neox.layers.25.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 281 |
+
"gpt_neox.layers.25.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 282 |
+
"gpt_neox.layers.25.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 283 |
+
"gpt_neox.layers.25.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 284 |
+
"gpt_neox.layers.25.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 285 |
+
"gpt_neox.layers.25.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 286 |
+
"gpt_neox.layers.25.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 287 |
+
"gpt_neox.layers.25.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 288 |
+
"gpt_neox.layers.25.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 289 |
+
"gpt_neox.layers.25.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 290 |
+
"gpt_neox.layers.25.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 291 |
+
"gpt_neox.layers.25.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 292 |
+
"gpt_neox.layers.25.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 293 |
+
"gpt_neox.layers.25.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 294 |
+
"gpt_neox.layers.25.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 295 |
+
"gpt_neox.layers.26.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 296 |
+
"gpt_neox.layers.26.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 297 |
+
"gpt_neox.layers.26.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 298 |
+
"gpt_neox.layers.26.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 299 |
+
"gpt_neox.layers.26.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 300 |
+
"gpt_neox.layers.26.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 301 |
+
"gpt_neox.layers.26.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 302 |
+
"gpt_neox.layers.26.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 303 |
+
"gpt_neox.layers.26.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 304 |
+
"gpt_neox.layers.26.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 305 |
+
"gpt_neox.layers.26.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 306 |
+
"gpt_neox.layers.26.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 307 |
+
"gpt_neox.layers.26.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 308 |
+
"gpt_neox.layers.26.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 309 |
+
"gpt_neox.layers.26.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 310 |
+
"gpt_neox.layers.27.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 311 |
+
"gpt_neox.layers.27.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 312 |
+
"gpt_neox.layers.27.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 313 |
+
"gpt_neox.layers.27.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 314 |
+
"gpt_neox.layers.27.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 315 |
+
"gpt_neox.layers.27.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 316 |
+
"gpt_neox.layers.27.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 317 |
+
"gpt_neox.layers.27.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 318 |
+
"gpt_neox.layers.27.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 319 |
+
"gpt_neox.layers.27.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 320 |
+
"gpt_neox.layers.27.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 321 |
+
"gpt_neox.layers.27.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 322 |
+
"gpt_neox.layers.27.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 323 |
+
"gpt_neox.layers.27.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 324 |
+
"gpt_neox.layers.27.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 325 |
+
"gpt_neox.layers.28.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 326 |
+
"gpt_neox.layers.28.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 327 |
+
"gpt_neox.layers.28.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 328 |
+
"gpt_neox.layers.28.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 329 |
+
"gpt_neox.layers.28.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 330 |
+
"gpt_neox.layers.28.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 331 |
+
"gpt_neox.layers.28.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 332 |
+
"gpt_neox.layers.28.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 333 |
+
"gpt_neox.layers.28.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 334 |
+
"gpt_neox.layers.28.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 335 |
+
"gpt_neox.layers.28.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 336 |
+
"gpt_neox.layers.28.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 337 |
+
"gpt_neox.layers.28.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 338 |
+
"gpt_neox.layers.28.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 339 |
+
"gpt_neox.layers.28.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 340 |
+
"gpt_neox.layers.29.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 341 |
+
"gpt_neox.layers.29.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 342 |
+
"gpt_neox.layers.29.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 343 |
+
"gpt_neox.layers.29.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 344 |
+
"gpt_neox.layers.29.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 345 |
+
"gpt_neox.layers.29.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 346 |
+
"gpt_neox.layers.29.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 347 |
+
"gpt_neox.layers.29.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 348 |
+
"gpt_neox.layers.29.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 349 |
+
"gpt_neox.layers.29.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 350 |
+
"gpt_neox.layers.29.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 351 |
+
"gpt_neox.layers.29.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 352 |
+
"gpt_neox.layers.29.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 353 |
+
"gpt_neox.layers.29.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 354 |
+
"gpt_neox.layers.29.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 355 |
+
"gpt_neox.layers.3.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 356 |
+
"gpt_neox.layers.3.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 357 |
+
"gpt_neox.layers.3.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 358 |
+
"gpt_neox.layers.3.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 359 |
+
"gpt_neox.layers.3.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 360 |
+
"gpt_neox.layers.3.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 361 |
+
"gpt_neox.layers.3.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 362 |
+
"gpt_neox.layers.3.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 363 |
+
"gpt_neox.layers.3.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 364 |
+
"gpt_neox.layers.3.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 365 |
+
"gpt_neox.layers.3.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 366 |
+
"gpt_neox.layers.3.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 367 |
+
"gpt_neox.layers.3.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 368 |
+
"gpt_neox.layers.3.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 369 |
+
"gpt_neox.layers.3.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 370 |
+
"gpt_neox.layers.30.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 371 |
+
"gpt_neox.layers.30.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 372 |
+
"gpt_neox.layers.30.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 373 |
+
"gpt_neox.layers.30.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 374 |
+
"gpt_neox.layers.30.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 375 |
+
"gpt_neox.layers.30.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 376 |
+
"gpt_neox.layers.30.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 377 |
+
"gpt_neox.layers.30.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 378 |
+
"gpt_neox.layers.30.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 379 |
+
"gpt_neox.layers.30.mlp.dense_4h_to_h.bias": "pytorch_model-00003-of-00005.bin",
|
| 380 |
+
"gpt_neox.layers.30.mlp.dense_4h_to_h.weight": "pytorch_model-00003-of-00005.bin",
|
| 381 |
+
"gpt_neox.layers.30.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 382 |
+
"gpt_neox.layers.30.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 383 |
+
"gpt_neox.layers.30.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 384 |
+
"gpt_neox.layers.30.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 385 |
+
"gpt_neox.layers.31.attention.bias": "pytorch_model-00003-of-00005.bin",
|
| 386 |
+
"gpt_neox.layers.31.attention.dense.bias": "pytorch_model-00003-of-00005.bin",
|
| 387 |
+
"gpt_neox.layers.31.attention.dense.weight": "pytorch_model-00003-of-00005.bin",
|
| 388 |
+
"gpt_neox.layers.31.attention.masked_bias": "pytorch_model-00003-of-00005.bin",
|
| 389 |
+
"gpt_neox.layers.31.attention.query_key_value.bias": "pytorch_model-00003-of-00005.bin",
|
| 390 |
+
"gpt_neox.layers.31.attention.query_key_value.weight": "pytorch_model-00003-of-00005.bin",
|
| 391 |
+
"gpt_neox.layers.31.attention.rotary_emb.inv_freq": "pytorch_model-00003-of-00005.bin",
|
| 392 |
+
"gpt_neox.layers.31.input_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 393 |
+
"gpt_neox.layers.31.input_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 394 |
+
"gpt_neox.layers.31.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 395 |
+
"gpt_neox.layers.31.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 396 |
+
"gpt_neox.layers.31.mlp.dense_h_to_4h.bias": "pytorch_model-00003-of-00005.bin",
|
| 397 |
+
"gpt_neox.layers.31.mlp.dense_h_to_4h.weight": "pytorch_model-00003-of-00005.bin",
|
| 398 |
+
"gpt_neox.layers.31.post_attention_layernorm.bias": "pytorch_model-00003-of-00005.bin",
|
| 399 |
+
"gpt_neox.layers.31.post_attention_layernorm.weight": "pytorch_model-00003-of-00005.bin",
|
| 400 |
+
"gpt_neox.layers.32.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 401 |
+
"gpt_neox.layers.32.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 402 |
+
"gpt_neox.layers.32.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 403 |
+
"gpt_neox.layers.32.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 404 |
+
"gpt_neox.layers.32.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 405 |
+
"gpt_neox.layers.32.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 406 |
+
"gpt_neox.layers.32.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 407 |
+
"gpt_neox.layers.32.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 408 |
+
"gpt_neox.layers.32.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 409 |
+
"gpt_neox.layers.32.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 410 |
+
"gpt_neox.layers.32.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 411 |
+
"gpt_neox.layers.32.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 412 |
+
"gpt_neox.layers.32.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 413 |
+
"gpt_neox.layers.32.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 414 |
+
"gpt_neox.layers.32.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 415 |
+
"gpt_neox.layers.33.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 416 |
+
"gpt_neox.layers.33.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 417 |
+
"gpt_neox.layers.33.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 418 |
+
"gpt_neox.layers.33.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 419 |
+
"gpt_neox.layers.33.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 420 |
+
"gpt_neox.layers.33.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 421 |
+
"gpt_neox.layers.33.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 422 |
+
"gpt_neox.layers.33.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 423 |
+
"gpt_neox.layers.33.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 424 |
+
"gpt_neox.layers.33.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 425 |
+
"gpt_neox.layers.33.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 426 |
+
"gpt_neox.layers.33.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 427 |
+
"gpt_neox.layers.33.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 428 |
+
"gpt_neox.layers.33.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 429 |
+
"gpt_neox.layers.33.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 430 |
+
"gpt_neox.layers.34.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 431 |
+
"gpt_neox.layers.34.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 432 |
+
"gpt_neox.layers.34.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 433 |
+
"gpt_neox.layers.34.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 434 |
+
"gpt_neox.layers.34.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 435 |
+
"gpt_neox.layers.34.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 436 |
+
"gpt_neox.layers.34.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 437 |
+
"gpt_neox.layers.34.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 438 |
+
"gpt_neox.layers.34.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 439 |
+
"gpt_neox.layers.34.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 440 |
+
"gpt_neox.layers.34.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 441 |
+
"gpt_neox.layers.34.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 442 |
+
"gpt_neox.layers.34.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 443 |
+
"gpt_neox.layers.34.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 444 |
+
"gpt_neox.layers.34.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 445 |
+
"gpt_neox.layers.35.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 446 |
+
"gpt_neox.layers.35.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 447 |
+
"gpt_neox.layers.35.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 448 |
+
"gpt_neox.layers.35.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 449 |
+
"gpt_neox.layers.35.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 450 |
+
"gpt_neox.layers.35.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 451 |
+
"gpt_neox.layers.35.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 452 |
+
"gpt_neox.layers.35.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 453 |
+
"gpt_neox.layers.35.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 454 |
+
"gpt_neox.layers.35.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 455 |
+
"gpt_neox.layers.35.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 456 |
+
"gpt_neox.layers.35.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 457 |
+
"gpt_neox.layers.35.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 458 |
+
"gpt_neox.layers.35.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 459 |
+
"gpt_neox.layers.35.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 460 |
+
"gpt_neox.layers.36.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 461 |
+
"gpt_neox.layers.36.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 462 |
+
"gpt_neox.layers.36.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 463 |
+
"gpt_neox.layers.36.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 464 |
+
"gpt_neox.layers.36.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 465 |
+
"gpt_neox.layers.36.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 466 |
+
"gpt_neox.layers.36.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 467 |
+
"gpt_neox.layers.36.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 468 |
+
"gpt_neox.layers.36.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 469 |
+
"gpt_neox.layers.36.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 470 |
+
"gpt_neox.layers.36.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 471 |
+
"gpt_neox.layers.36.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 472 |
+
"gpt_neox.layers.36.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 473 |
+
"gpt_neox.layers.36.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 474 |
+
"gpt_neox.layers.36.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 475 |
+
"gpt_neox.layers.37.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 476 |
+
"gpt_neox.layers.37.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 477 |
+
"gpt_neox.layers.37.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 478 |
+
"gpt_neox.layers.37.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 479 |
+
"gpt_neox.layers.37.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 480 |
+
"gpt_neox.layers.37.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 481 |
+
"gpt_neox.layers.37.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 482 |
+
"gpt_neox.layers.37.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 483 |
+
"gpt_neox.layers.37.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 484 |
+
"gpt_neox.layers.37.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 485 |
+
"gpt_neox.layers.37.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 486 |
+
"gpt_neox.layers.37.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 487 |
+
"gpt_neox.layers.37.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 488 |
+
"gpt_neox.layers.37.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 489 |
+
"gpt_neox.layers.37.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 490 |
+
"gpt_neox.layers.38.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 491 |
+
"gpt_neox.layers.38.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 492 |
+
"gpt_neox.layers.38.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 493 |
+
"gpt_neox.layers.38.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 494 |
+
"gpt_neox.layers.38.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 495 |
+
"gpt_neox.layers.38.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 496 |
+
"gpt_neox.layers.38.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 497 |
+
"gpt_neox.layers.38.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 498 |
+
"gpt_neox.layers.38.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 499 |
+
"gpt_neox.layers.38.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 500 |
+
"gpt_neox.layers.38.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 501 |
+
"gpt_neox.layers.38.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 502 |
+
"gpt_neox.layers.38.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 503 |
+
"gpt_neox.layers.38.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 504 |
+
"gpt_neox.layers.38.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 505 |
+
"gpt_neox.layers.39.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 506 |
+
"gpt_neox.layers.39.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 507 |
+
"gpt_neox.layers.39.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 508 |
+
"gpt_neox.layers.39.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 509 |
+
"gpt_neox.layers.39.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 510 |
+
"gpt_neox.layers.39.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 511 |
+
"gpt_neox.layers.39.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 512 |
+
"gpt_neox.layers.39.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 513 |
+
"gpt_neox.layers.39.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 514 |
+
"gpt_neox.layers.39.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 515 |
+
"gpt_neox.layers.39.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 516 |
+
"gpt_neox.layers.39.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 517 |
+
"gpt_neox.layers.39.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 518 |
+
"gpt_neox.layers.39.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 519 |
+
"gpt_neox.layers.39.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 520 |
+
"gpt_neox.layers.4.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 521 |
+
"gpt_neox.layers.4.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 522 |
+
"gpt_neox.layers.4.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 523 |
+
"gpt_neox.layers.4.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 524 |
+
"gpt_neox.layers.4.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 525 |
+
"gpt_neox.layers.4.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 526 |
+
"gpt_neox.layers.4.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 527 |
+
"gpt_neox.layers.4.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 528 |
+
"gpt_neox.layers.4.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 529 |
+
"gpt_neox.layers.4.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 530 |
+
"gpt_neox.layers.4.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 531 |
+
"gpt_neox.layers.4.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 532 |
+
"gpt_neox.layers.4.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 533 |
+
"gpt_neox.layers.4.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 534 |
+
"gpt_neox.layers.4.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 535 |
+
"gpt_neox.layers.40.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 536 |
+
"gpt_neox.layers.40.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 537 |
+
"gpt_neox.layers.40.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 538 |
+
"gpt_neox.layers.40.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 539 |
+
"gpt_neox.layers.40.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 540 |
+
"gpt_neox.layers.40.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 541 |
+
"gpt_neox.layers.40.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 542 |
+
"gpt_neox.layers.40.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 543 |
+
"gpt_neox.layers.40.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 544 |
+
"gpt_neox.layers.40.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 545 |
+
"gpt_neox.layers.40.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 546 |
+
"gpt_neox.layers.40.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 547 |
+
"gpt_neox.layers.40.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 548 |
+
"gpt_neox.layers.40.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 549 |
+
"gpt_neox.layers.40.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 550 |
+
"gpt_neox.layers.41.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 551 |
+
"gpt_neox.layers.41.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 552 |
+
"gpt_neox.layers.41.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 553 |
+
"gpt_neox.layers.41.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 554 |
+
"gpt_neox.layers.41.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 555 |
+
"gpt_neox.layers.41.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 556 |
+
"gpt_neox.layers.41.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 557 |
+
"gpt_neox.layers.41.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 558 |
+
"gpt_neox.layers.41.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 559 |
+
"gpt_neox.layers.41.mlp.dense_4h_to_h.bias": "pytorch_model-00004-of-00005.bin",
|
| 560 |
+
"gpt_neox.layers.41.mlp.dense_4h_to_h.weight": "pytorch_model-00004-of-00005.bin",
|
| 561 |
+
"gpt_neox.layers.41.mlp.dense_h_to_4h.bias": "pytorch_model-00004-of-00005.bin",
|
| 562 |
+
"gpt_neox.layers.41.mlp.dense_h_to_4h.weight": "pytorch_model-00004-of-00005.bin",
|
| 563 |
+
"gpt_neox.layers.41.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 564 |
+
"gpt_neox.layers.41.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 565 |
+
"gpt_neox.layers.42.attention.bias": "pytorch_model-00004-of-00005.bin",
|
| 566 |
+
"gpt_neox.layers.42.attention.dense.bias": "pytorch_model-00004-of-00005.bin",
|
| 567 |
+
"gpt_neox.layers.42.attention.dense.weight": "pytorch_model-00004-of-00005.bin",
|
| 568 |
+
"gpt_neox.layers.42.attention.masked_bias": "pytorch_model-00004-of-00005.bin",
|
| 569 |
+
"gpt_neox.layers.42.attention.query_key_value.bias": "pytorch_model-00004-of-00005.bin",
|
| 570 |
+
"gpt_neox.layers.42.attention.query_key_value.weight": "pytorch_model-00004-of-00005.bin",
|
| 571 |
+
"gpt_neox.layers.42.attention.rotary_emb.inv_freq": "pytorch_model-00004-of-00005.bin",
|
| 572 |
+
"gpt_neox.layers.42.input_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 573 |
+
"gpt_neox.layers.42.input_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 574 |
+
"gpt_neox.layers.42.mlp.dense_4h_to_h.bias": "pytorch_model-00005-of-00005.bin",
|
| 575 |
+
"gpt_neox.layers.42.mlp.dense_4h_to_h.weight": "pytorch_model-00005-of-00005.bin",
|
| 576 |
+
"gpt_neox.layers.42.mlp.dense_h_to_4h.bias": "pytorch_model-00005-of-00005.bin",
|
| 577 |
+
"gpt_neox.layers.42.mlp.dense_h_to_4h.weight": "pytorch_model-00005-of-00005.bin",
|
| 578 |
+
"gpt_neox.layers.42.post_attention_layernorm.bias": "pytorch_model-00004-of-00005.bin",
|
| 579 |
+
"gpt_neox.layers.42.post_attention_layernorm.weight": "pytorch_model-00004-of-00005.bin",
|
| 580 |
+
"gpt_neox.layers.43.attention.bias": "pytorch_model-00005-of-00005.bin",
|
| 581 |
+
"gpt_neox.layers.43.attention.dense.bias": "pytorch_model-00005-of-00005.bin",
|
| 582 |
+
"gpt_neox.layers.43.attention.dense.weight": "pytorch_model-00005-of-00005.bin",
|
| 583 |
+
"gpt_neox.layers.43.attention.masked_bias": "pytorch_model-00005-of-00005.bin",
|
| 584 |
+
"gpt_neox.layers.43.attention.query_key_value.bias": "pytorch_model-00005-of-00005.bin",
|
| 585 |
+
"gpt_neox.layers.43.attention.query_key_value.weight": "pytorch_model-00005-of-00005.bin",
|
| 586 |
+
"gpt_neox.layers.43.attention.rotary_emb.inv_freq": "pytorch_model-00005-of-00005.bin",
|
| 587 |
+
"gpt_neox.layers.43.input_layernorm.bias": "pytorch_model-00005-of-00005.bin",
|
| 588 |
+
"gpt_neox.layers.43.input_layernorm.weight": "pytorch_model-00005-of-00005.bin",
|
| 589 |
+
"gpt_neox.layers.43.mlp.dense_4h_to_h.bias": "pytorch_model-00005-of-00005.bin",
|
| 590 |
+
"gpt_neox.layers.43.mlp.dense_4h_to_h.weight": "pytorch_model-00005-of-00005.bin",
|
| 591 |
+
"gpt_neox.layers.43.mlp.dense_h_to_4h.bias": "pytorch_model-00005-of-00005.bin",
|
| 592 |
+
"gpt_neox.layers.43.mlp.dense_h_to_4h.weight": "pytorch_model-00005-of-00005.bin",
|
| 593 |
+
"gpt_neox.layers.43.post_attention_layernorm.bias": "pytorch_model-00005-of-00005.bin",
|
| 594 |
+
"gpt_neox.layers.43.post_attention_layernorm.weight": "pytorch_model-00005-of-00005.bin",
|
| 595 |
+
"gpt_neox.layers.5.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 596 |
+
"gpt_neox.layers.5.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 597 |
+
"gpt_neox.layers.5.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 598 |
+
"gpt_neox.layers.5.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 599 |
+
"gpt_neox.layers.5.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 600 |
+
"gpt_neox.layers.5.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 601 |
+
"gpt_neox.layers.5.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 602 |
+
"gpt_neox.layers.5.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 603 |
+
"gpt_neox.layers.5.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 604 |
+
"gpt_neox.layers.5.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 605 |
+
"gpt_neox.layers.5.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 606 |
+
"gpt_neox.layers.5.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 607 |
+
"gpt_neox.layers.5.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 608 |
+
"gpt_neox.layers.5.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 609 |
+
"gpt_neox.layers.5.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 610 |
+
"gpt_neox.layers.6.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 611 |
+
"gpt_neox.layers.6.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 612 |
+
"gpt_neox.layers.6.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 613 |
+
"gpt_neox.layers.6.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 614 |
+
"gpt_neox.layers.6.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 615 |
+
"gpt_neox.layers.6.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 616 |
+
"gpt_neox.layers.6.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 617 |
+
"gpt_neox.layers.6.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 618 |
+
"gpt_neox.layers.6.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 619 |
+
"gpt_neox.layers.6.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 620 |
+
"gpt_neox.layers.6.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 621 |
+
"gpt_neox.layers.6.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 622 |
+
"gpt_neox.layers.6.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 623 |
+
"gpt_neox.layers.6.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 624 |
+
"gpt_neox.layers.6.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 625 |
+
"gpt_neox.layers.7.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 626 |
+
"gpt_neox.layers.7.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 627 |
+
"gpt_neox.layers.7.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 628 |
+
"gpt_neox.layers.7.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 629 |
+
"gpt_neox.layers.7.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 630 |
+
"gpt_neox.layers.7.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 631 |
+
"gpt_neox.layers.7.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 632 |
+
"gpt_neox.layers.7.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 633 |
+
"gpt_neox.layers.7.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 634 |
+
"gpt_neox.layers.7.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 635 |
+
"gpt_neox.layers.7.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 636 |
+
"gpt_neox.layers.7.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 637 |
+
"gpt_neox.layers.7.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 638 |
+
"gpt_neox.layers.7.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 639 |
+
"gpt_neox.layers.7.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 640 |
+
"gpt_neox.layers.8.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 641 |
+
"gpt_neox.layers.8.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 642 |
+
"gpt_neox.layers.8.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 643 |
+
"gpt_neox.layers.8.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 644 |
+
"gpt_neox.layers.8.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 645 |
+
"gpt_neox.layers.8.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 646 |
+
"gpt_neox.layers.8.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 647 |
+
"gpt_neox.layers.8.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 648 |
+
"gpt_neox.layers.8.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 649 |
+
"gpt_neox.layers.8.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 650 |
+
"gpt_neox.layers.8.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 651 |
+
"gpt_neox.layers.8.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 652 |
+
"gpt_neox.layers.8.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 653 |
+
"gpt_neox.layers.8.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 654 |
+
"gpt_neox.layers.8.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 655 |
+
"gpt_neox.layers.9.attention.bias": "pytorch_model-00001-of-00005.bin",
|
| 656 |
+
"gpt_neox.layers.9.attention.dense.bias": "pytorch_model-00001-of-00005.bin",
|
| 657 |
+
"gpt_neox.layers.9.attention.dense.weight": "pytorch_model-00001-of-00005.bin",
|
| 658 |
+
"gpt_neox.layers.9.attention.masked_bias": "pytorch_model-00001-of-00005.bin",
|
| 659 |
+
"gpt_neox.layers.9.attention.query_key_value.bias": "pytorch_model-00001-of-00005.bin",
|
| 660 |
+
"gpt_neox.layers.9.attention.query_key_value.weight": "pytorch_model-00001-of-00005.bin",
|
| 661 |
+
"gpt_neox.layers.9.attention.rotary_emb.inv_freq": "pytorch_model-00001-of-00005.bin",
|
| 662 |
+
"gpt_neox.layers.9.input_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 663 |
+
"gpt_neox.layers.9.input_layernorm.weight": "pytorch_model-00001-of-00005.bin",
|
| 664 |
+
"gpt_neox.layers.9.mlp.dense_4h_to_h.bias": "pytorch_model-00001-of-00005.bin",
|
| 665 |
+
"gpt_neox.layers.9.mlp.dense_4h_to_h.weight": "pytorch_model-00001-of-00005.bin",
|
| 666 |
+
"gpt_neox.layers.9.mlp.dense_h_to_4h.bias": "pytorch_model-00001-of-00005.bin",
|
| 667 |
+
"gpt_neox.layers.9.mlp.dense_h_to_4h.weight": "pytorch_model-00001-of-00005.bin",
|
| 668 |
+
"gpt_neox.layers.9.post_attention_layernorm.bias": "pytorch_model-00001-of-00005.bin",
|
| 669 |
+
"gpt_neox.layers.9.post_attention_layernorm.weight": "pytorch_model-00001-of-00005.bin"
|
| 670 |
+
}
|
| 671 |
+
}
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": "<|endoftext|>",
|
| 3 |
+
"eos_token": "<|endoftext|>",
|
| 4 |
+
"unk_token": "<|endoftext|>"
|
| 5 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"bos_token": "<|endoftext|>",
|
| 4 |
+
"eos_token": "<|endoftext|>",
|
| 5 |
+
"name_or_path": "EleutherAI/gpt-neox-20b",
|
| 6 |
+
"special_tokens_map_file": "/root/.cache/huggingface/transformers/d9026dc928c47ac2a72d46fea7db959acc4bacac2176bf32be5c331604b77d32.3ae9ae72462581d20e36bc528e9c47bb30cd671bb21add40ca0b24a0be9fac22",
|
| 7 |
+
"tokenizer_class": "GPTNeoXTokenizer",
|
| 8 |
+
"unk_token": "<|endoftext|>"
|
| 9 |
+
}
|
train.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|