diff --git a/checkpoint-100/README.md b/checkpoint-100/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-100/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-100/adapter_config.json b/checkpoint-100/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-100/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-100/adapter_model.safetensors b/checkpoint-100/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6098c38a9ac78055f1a22d0fe10c936f003acecf --- /dev/null +++ b/checkpoint-100/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1538634ebad5a502111d8376fcfd8042082019d6f20303edf3c191ad88fb8e7d +size 864513616 diff --git a/checkpoint-100/optimizer.pt b/checkpoint-100/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..350f54f002039bebcd944f66fd3abed1655bbfb7 --- /dev/null +++ b/checkpoint-100/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d58538d13cdf7633b64315109d3df252c8fee19afd0851f60e3d7b5b928030b +size 170951068 diff --git a/checkpoint-100/rng_state.pth b/checkpoint-100/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..3321daf20be2ebcec62fe26f4795ce433e667805 --- /dev/null +++ b/checkpoint-100/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05f339ec0a4f181d98618eee851e34698dd7852756fd50a1bcbff3dabf357ac1 +size 14244 diff --git a/checkpoint-100/scheduler.pt b/checkpoint-100/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..a78ab991c19727efd8fa2d4cde9f89d5f2f176e4 --- /dev/null +++ b/checkpoint-100/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07a3e70071e1a897d7e8618fc22532e1736f236dff9efebc89955a118d3be184 +size 1064 diff --git a/checkpoint-100/trainer_state.json b/checkpoint-100/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..f41b143cb24225aa54e227da827172ccc13a50d9 --- /dev/null +++ b/checkpoint-100/trainer_state.json @@ -0,0 +1,81 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5555555555555556, + "eval_steps": 25, + "global_step": 100, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 4421038846771200.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-100/training_args.bin b/checkpoint-100/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-100/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-125/README.md b/checkpoint-125/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-125/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-125/adapter_config.json b/checkpoint-125/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-125/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-125/adapter_model.safetensors b/checkpoint-125/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..33750c095a845307c5f25ee92932419df7360dfb --- /dev/null +++ b/checkpoint-125/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce5419c957a02658bb73357282da12de7013097b9de6084946b80d3dec7e7c82 +size 864513616 diff --git a/checkpoint-125/optimizer.pt b/checkpoint-125/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..11b93918e69dfd6812dd1bd1e27be433310073b5 --- /dev/null +++ b/checkpoint-125/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7d3c55a778c091907bd1eda17107477c270c014eb52b60a142e77f8731d09a0 +size 170951068 diff --git a/checkpoint-125/rng_state.pth b/checkpoint-125/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c24ad071929698e7e420dc882d22d61708c411e4 --- /dev/null +++ b/checkpoint-125/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb3c15912971c6f0e8f9f20a19330f80be37c602e32ded8d9b1b4f806a52e930 +size 14244 diff --git a/checkpoint-125/scheduler.pt b/checkpoint-125/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..3a12d25c1ce955af8a0db27634eeab2c315cb629 --- /dev/null +++ b/checkpoint-125/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56f26711d2ed7de2f7cac452591abedcceadf8020f33756b0e131aa1bae9da48 +size 1064 diff --git a/checkpoint-125/trainer_state.json b/checkpoint-125/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..a50be09794b71aab454c5840a2f9193dfc5e5d9f --- /dev/null +++ b/checkpoint-125/trainer_state.json @@ -0,0 +1,96 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6944444444444444, + "eval_steps": 25, + "global_step": 125, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 5526298558464000.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-125/training_args.bin b/checkpoint-125/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-125/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-150/README.md b/checkpoint-150/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-150/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-150/adapter_config.json b/checkpoint-150/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-150/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-150/adapter_model.safetensors b/checkpoint-150/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ee3e2884603b511022e7e580aadcd2fc54a4e5fc --- /dev/null +++ b/checkpoint-150/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57c89778e2a05ae218d7a49858b2add3a615dff2965c64a2509e5853993aa9cb +size 864513616 diff --git a/checkpoint-150/optimizer.pt b/checkpoint-150/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..1e87a477eaaa4e62c7cb4bbbf2214e0a74d9f912 --- /dev/null +++ b/checkpoint-150/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f18c9587012dfdf2e5d57ca90160a25390a83609a7cee988c3dfb1dafd0774d0 +size 170951068 diff --git a/checkpoint-150/rng_state.pth b/checkpoint-150/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c418e7f8513d16a959126afb0ea6d97ee8c3980d --- /dev/null +++ b/checkpoint-150/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b29ce5444dd529a6ef68878d56a0c03d3aa040a7d184ca56615dd57649f1dd1 +size 14244 diff --git a/checkpoint-150/scheduler.pt b/checkpoint-150/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..db233ffaa26df9f955f6ae5b434321990bc361c3 --- /dev/null +++ b/checkpoint-150/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ed838ded083e57bed8ca10815dcbe6a1c6a9b6db88e3884f79c6b2b1b8a663f +size 1064 diff --git a/checkpoint-150/trainer_state.json b/checkpoint-150/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..18bf1f6671127d7744a72a727c8e992c90e33dac --- /dev/null +++ b/checkpoint-150/trainer_state.json @@ -0,0 +1,111 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8333333333333334, + "eval_steps": 25, + "global_step": 150, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 6631558270156800.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-150/training_args.bin b/checkpoint-150/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-150/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-175/README.md b/checkpoint-175/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-175/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-175/adapter_config.json b/checkpoint-175/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-175/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-175/adapter_model.safetensors b/checkpoint-175/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1914eba735d5868debcf4f57ec4ca3fa07b1e10a --- /dev/null +++ b/checkpoint-175/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad64edb0f42ff36f0bf6626f2d32cba755020de025916f4286f57c62018cbe94 +size 864513616 diff --git a/checkpoint-175/optimizer.pt b/checkpoint-175/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..6de69e0e3809de577dd0b906ae82b26bf3df1b29 --- /dev/null +++ b/checkpoint-175/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abc4f76052231034b73b01f55fd9a2632e0ea2d18f14b70f2538e9a50d37054e +size 170951068 diff --git a/checkpoint-175/rng_state.pth b/checkpoint-175/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..6b10de24c34cfa749be19c35f2fb2668d8513111 --- /dev/null +++ b/checkpoint-175/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a20fefd3b4bc00c03f3c26a1ff10dc6893cabd201d81335e1bca6ac164e34c6 +size 14244 diff --git a/checkpoint-175/scheduler.pt b/checkpoint-175/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..82d82f86c9517c35561d0d42534b82b29aa06ae9 --- /dev/null +++ b/checkpoint-175/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae8230235363aa37e78c9224e6d86c2b5867a3305e08df519e2eb9af1b4c1fec +size 1064 diff --git a/checkpoint-175/trainer_state.json b/checkpoint-175/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..c588210ae91f777c1dd4ad6f25b14df891cad83e --- /dev/null +++ b/checkpoint-175/trainer_state.json @@ -0,0 +1,126 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9722222222222222, + "eval_steps": 25, + "global_step": 175, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 7736817981849600.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-175/training_args.bin b/checkpoint-175/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-175/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-200/README.md b/checkpoint-200/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-200/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-200/adapter_config.json b/checkpoint-200/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-200/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-200/adapter_model.safetensors b/checkpoint-200/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..157b7865fd1e523ed9bb0f09508422a887431141 --- /dev/null +++ b/checkpoint-200/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:915903ccf5675b7735a5c5b8de53883cc027732e46817ae4bad0ab3fe2cbcea8 +size 864513616 diff --git a/checkpoint-200/optimizer.pt b/checkpoint-200/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e1ee537ed8c19ad28a2415cf7a4dc01ae83ecb38 --- /dev/null +++ b/checkpoint-200/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6c8b5652b745b83324b89cef1b19ec7b5584777c6f51365c03d787e56f79e4a +size 170951068 diff --git a/checkpoint-200/rng_state.pth b/checkpoint-200/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c207c2308637dfa7b2db241ebebb75e334ebc1aa --- /dev/null +++ b/checkpoint-200/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b71eee1f0ee6c414dadc4dc4b2c31fd33e2acc67bcc69175b5e180d9bab49102 +size 14244 diff --git a/checkpoint-200/scheduler.pt b/checkpoint-200/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..bda5a54f4eab4bf1a36c71f3fb634284a11994ea --- /dev/null +++ b/checkpoint-200/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e777218181480936e44c75d39a4b0e55e8ec1ce38b3edf1e9e798ac6258943ee +size 1064 diff --git a/checkpoint-200/trainer_state.json b/checkpoint-200/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..5b1153312084952dd685c66eb7c176693a701e7c --- /dev/null +++ b/checkpoint-200/trainer_state.json @@ -0,0 +1,141 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.1111111111111112, + "eval_steps": 25, + "global_step": 200, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 8819972499308544.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-200/training_args.bin b/checkpoint-200/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-200/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-225/README.md b/checkpoint-225/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-225/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-225/adapter_config.json b/checkpoint-225/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-225/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-225/adapter_model.safetensors b/checkpoint-225/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d3f74d7e25c8584a81813c0a57dce0112fdc6750 --- /dev/null +++ b/checkpoint-225/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ddedb2a9a53475ce1860efd8c6160d95f78f9062a1cee0d7a553b8fe386fb338 +size 864513616 diff --git a/checkpoint-225/optimizer.pt b/checkpoint-225/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e7c8e27748a6a408797f97a5a10697ff753c47cc --- /dev/null +++ b/checkpoint-225/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97a0f13a0bade985ce85eb820acd880a71bab47fe22bdacd9e22e3d0fa0bc324 +size 170951068 diff --git a/checkpoint-225/rng_state.pth b/checkpoint-225/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..dbfaa78a2d0cee30a39a143ed37874a1a7b6bd17 --- /dev/null +++ b/checkpoint-225/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7b6ef38ee7803854cf569dcaacbb01f182762add04bd95ca10192858c39cceb +size 14244 diff --git a/checkpoint-225/scheduler.pt b/checkpoint-225/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..d08c88acd5c7529cd099d920802aae0c5d0ba635 --- /dev/null +++ b/checkpoint-225/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:609ac197c961b2d66e1671a292842afd420d76b7a7b59df2d6f9c3c1f406758e +size 1064 diff --git a/checkpoint-225/trainer_state.json b/checkpoint-225/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..1ca2adbce065b5f4908dcfad099ee931002c2444 --- /dev/null +++ b/checkpoint-225/trainer_state.json @@ -0,0 +1,156 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.25, + "eval_steps": 25, + "global_step": 225, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 9925232211001344.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-225/training_args.bin b/checkpoint-225/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-225/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-25/README.md b/checkpoint-25/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-25/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-25/adapter_config.json b/checkpoint-25/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-25/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-25/adapter_model.safetensors b/checkpoint-25/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f7c6dbcb7748c5e13e25020dacc8c74290d1a10a --- /dev/null +++ b/checkpoint-25/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6a9ada10b2dd5fe6f38f2692286ec640a6560a0e2d0bd98ce59fb638d05ecff +size 864513616 diff --git a/checkpoint-25/optimizer.pt b/checkpoint-25/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5aef693fe8a86c287f4bc796a9baa5564481bbc4 --- /dev/null +++ b/checkpoint-25/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0448feecca26ab2034a46d04ba0d820e24110f43455f430e561953d252ce3b2b +size 170951068 diff --git a/checkpoint-25/rng_state.pth b/checkpoint-25/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..90910dd6ceb844d8a113ed9768a5fb86710115a1 --- /dev/null +++ b/checkpoint-25/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11123afbe0854f2fc88ebc4181fc3fab5ebb7f7466fd47a6755a61c90cb0bebb +size 14244 diff --git a/checkpoint-25/scheduler.pt b/checkpoint-25/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ad952b6e6180f76acb377cf8d4c2b650f166fe4a --- /dev/null +++ b/checkpoint-25/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:227fa63ed4e8dfe423d4951eefe1d1f061706662383ee49f22ddca422105550d +size 1064 diff --git a/checkpoint-25/trainer_state.json b/checkpoint-25/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..aeeac8923dc8967ed0fc108d0065ba25e25c0ab4 --- /dev/null +++ b/checkpoint-25/trainer_state.json @@ -0,0 +1,36 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.1388888888888889, + "eval_steps": 25, + "global_step": 25, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1105259711692800.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-25/training_args.bin b/checkpoint-25/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-25/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-250/README.md b/checkpoint-250/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-250/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-250/adapter_config.json b/checkpoint-250/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-250/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-250/adapter_model.safetensors b/checkpoint-250/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2d591b8290782ab0010f85b9539ac1dfb02d5e4f --- /dev/null +++ b/checkpoint-250/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:299763bce504ab6118bc14c8f18eb215e7ae94a1f029ff0a1552cd596eb4f1dc +size 864513616 diff --git a/checkpoint-250/optimizer.pt b/checkpoint-250/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7b45e9642fa154972a4cc69a6e1e08aa8630e770 --- /dev/null +++ b/checkpoint-250/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04c645920fd3b6ad76ac52352a2cb507b76966faa27acf78de833377de6dff62 +size 170951068 diff --git a/checkpoint-250/rng_state.pth b/checkpoint-250/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..bfc2bf5c2ffd2b36d1ae2341c3dd8c113c9a29cb --- /dev/null +++ b/checkpoint-250/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f44f0d7818ec3033e5a807c5ffcdd3b94f9e356fe41f02111b6b120b1f89898 +size 14244 diff --git a/checkpoint-250/scheduler.pt b/checkpoint-250/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..6382b63464e51902afd6e8356ccca56ec53acadc --- /dev/null +++ b/checkpoint-250/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38f00a4de19c33ee4368a98fb744b5e3aa66f5d2b268e74de7afb96e77c1dae7 +size 1064 diff --git a/checkpoint-250/trainer_state.json b/checkpoint-250/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..d3d342c50674cfece15b463ecdd9d2f76ac0ed3e --- /dev/null +++ b/checkpoint-250/trainer_state.json @@ -0,0 +1,171 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.3888888888888888, + "eval_steps": 25, + "global_step": 250, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1.1030491922694144e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-250/training_args.bin b/checkpoint-250/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-250/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-275/README.md b/checkpoint-275/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-275/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-275/adapter_config.json b/checkpoint-275/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-275/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-275/adapter_model.safetensors b/checkpoint-275/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8c23af02523440f982bc70413b0269c7dc8866ff --- /dev/null +++ b/checkpoint-275/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fbfcdafcbb2cb84d14517d636eeb433d1fd6f63bd1dd7c71f55e4468861caeb +size 864513616 diff --git a/checkpoint-275/optimizer.pt b/checkpoint-275/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4c7f6564e5664faeb46fc167e14ee2fcb8c2cf8a --- /dev/null +++ b/checkpoint-275/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37e3e178c48d51c0defee6027e0c0ccce215afb9770e7c8f64358f7b410714f5 +size 170951516 diff --git a/checkpoint-275/rng_state.pth b/checkpoint-275/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..1b9fcc22c29ef7233761e7659fb12443ba621814 --- /dev/null +++ b/checkpoint-275/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b597d591074338c3b6251084220199bc613bf9a1cc38f6f8955ce839e30a49cd +size 14244 diff --git a/checkpoint-275/scheduler.pt b/checkpoint-275/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..af310636243656c840c0f0b1708f86443eb29315 --- /dev/null +++ b/checkpoint-275/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba744590cb285966155abdc60ce0ec9af6a1fb9b559f9e2430ee9dee318c3743 +size 1064 diff --git a/checkpoint-275/trainer_state.json b/checkpoint-275/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..9ecd500d29a4e217775d449dded13d426d8ee3ae --- /dev/null +++ b/checkpoint-275/trainer_state.json @@ -0,0 +1,186 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.5277777777777777, + "eval_steps": 25, + "global_step": 275, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1.2135751634386944e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-275/training_args.bin b/checkpoint-275/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-275/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-300/README.md b/checkpoint-300/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-300/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-300/adapter_config.json b/checkpoint-300/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-300/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-300/adapter_model.safetensors b/checkpoint-300/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d0f3bcdb22e10f0b5b246bfe3eb51303bef7d1ed --- /dev/null +++ b/checkpoint-300/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:535e2c09feedde69306e8481991877876e22583a84217cd7d4cd3ee21d5d49ea +size 864513616 diff --git a/checkpoint-300/optimizer.pt b/checkpoint-300/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5b32a01bd4e9dae61116f049ba27f21196a91e32 --- /dev/null +++ b/checkpoint-300/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd06af59ee012a95815d4d89ac572bbb0e348e2211de1926e932fd05256849af +size 170951516 diff --git a/checkpoint-300/rng_state.pth b/checkpoint-300/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..7e548f8f6df4abf72acec1da624a62edaf96660b --- /dev/null +++ b/checkpoint-300/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca6b071744fe70504e5b9bb8246e8fdb8b45eb752fdd8aa58c3860a6b32373a6 +size 14244 diff --git a/checkpoint-300/scheduler.pt b/checkpoint-300/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..58957e0f6534e1c1fa71881508c74f04acbdc515 --- /dev/null +++ b/checkpoint-300/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9060ab5252239304342fe63869f4494603685d29a3a00e119d3fe8b98f9b5488 +size 1064 diff --git a/checkpoint-300/trainer_state.json b/checkpoint-300/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..abe57e24a123b882e0d8cacd8115c0f8cf15b38a --- /dev/null +++ b/checkpoint-300/trainer_state.json @@ -0,0 +1,201 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.6666666666666665, + "eval_steps": 25, + "global_step": 300, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + }, + { + "epoch": 1.67, + "grad_norm": 5.02877950668335, + "learning_rate": 1.0120240480961924e-05, + "loss": 0.2323, + "step": 300 + }, + { + "epoch": 1.67, + "eval_loss": 0.6742722392082214, + "eval_runtime": 33.9146, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 300 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1.3241011346079744e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-300/training_args.bin b/checkpoint-300/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-300/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-325/README.md b/checkpoint-325/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-325/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-325/adapter_config.json b/checkpoint-325/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-325/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-325/adapter_model.safetensors b/checkpoint-325/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..353239fd910cda60ac5d2b3265ce62ad2e0290f6 --- /dev/null +++ b/checkpoint-325/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2761c3d36b73a8f66fb3c31286f6641209a4dbf231980363bf0ec068997b8520 +size 864513616 diff --git a/checkpoint-325/optimizer.pt b/checkpoint-325/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..8e7a50740226326329e447d19dbe8aaf679da949 --- /dev/null +++ b/checkpoint-325/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f87ea08a0978bea38a337187053ddad44f8f789f71f10d50ec2bc6b8bc0e29a1 +size 170951516 diff --git a/checkpoint-325/rng_state.pth b/checkpoint-325/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..2361e2657c9be7b598116d7959982bd2b0bacf10 --- /dev/null +++ b/checkpoint-325/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2116e67fe58882647b0960a683fd88637a4eeb7130eadc4b95981a3ef7552c9f +size 14244 diff --git a/checkpoint-325/scheduler.pt b/checkpoint-325/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..6676d3ae8ef9c0d85dd1fdf9e3a834555b7d4f9e --- /dev/null +++ b/checkpoint-325/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3d28f4227c93108cbc537125dd9d04cb3cc663e312aa953a2fe8dd802c68a03 +size 1064 diff --git a/checkpoint-325/trainer_state.json b/checkpoint-325/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..4127fea08381adec5ad8ffc905cb3a2eee3d45af --- /dev/null +++ b/checkpoint-325/trainer_state.json @@ -0,0 +1,216 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.8055555555555556, + "eval_steps": 25, + "global_step": 325, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + }, + { + "epoch": 1.67, + "grad_norm": 5.02877950668335, + "learning_rate": 1.0120240480961924e-05, + "loss": 0.2323, + "step": 300 + }, + { + "epoch": 1.67, + "eval_loss": 0.6742722392082214, + "eval_runtime": 33.9146, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 300 + }, + { + "epoch": 1.81, + "grad_norm": 3.24204158782959, + "learning_rate": 8.867735470941884e-06, + "loss": 0.2426, + "step": 325 + }, + { + "epoch": 1.81, + "eval_loss": 0.6821221709251404, + "eval_runtime": 33.9704, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 325 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1.4346271057772544e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-325/training_args.bin b/checkpoint-325/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-325/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-350/README.md b/checkpoint-350/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-350/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-350/adapter_config.json b/checkpoint-350/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-350/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-350/adapter_model.safetensors b/checkpoint-350/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d40c0ef9120b61d55b72ce3f4f8d223c962f39f7 --- /dev/null +++ b/checkpoint-350/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f79ee288169b8ba9c40f486eb8d43c42f2506af2708dc440027ab0e2a6d9c33 +size 864513616 diff --git a/checkpoint-350/optimizer.pt b/checkpoint-350/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..91d9c3aad16a8a4e66565f55174aab033d2f1bbd --- /dev/null +++ b/checkpoint-350/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fce94f306cbc8604ef85835bab5356f5fee65b2329e7dd94033f5b3ad5a64ef6 +size 170951516 diff --git a/checkpoint-350/rng_state.pth b/checkpoint-350/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..792254fb39e7951923cf2f2f49760622f5664a16 --- /dev/null +++ b/checkpoint-350/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10b2d3206139d2321d5b2ff012447802bd0580e761741c85b697112fe0250053 +size 14244 diff --git a/checkpoint-350/scheduler.pt b/checkpoint-350/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ee842fab706a63d2cf749863e946aa2acf838708 --- /dev/null +++ b/checkpoint-350/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa51f4493826b6527e2df1b00cfe212dc4a2ca4ffed9a71de4f452635ab29d5c +size 1064 diff --git a/checkpoint-350/trainer_state.json b/checkpoint-350/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..4b12c1871a8c63c59a2d2da2ca78098100e35575 --- /dev/null +++ b/checkpoint-350/trainer_state.json @@ -0,0 +1,231 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.9444444444444444, + "eval_steps": 25, + "global_step": 350, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + }, + { + "epoch": 1.67, + "grad_norm": 5.02877950668335, + "learning_rate": 1.0120240480961924e-05, + "loss": 0.2323, + "step": 300 + }, + { + "epoch": 1.67, + "eval_loss": 0.6742722392082214, + "eval_runtime": 33.9146, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 300 + }, + { + "epoch": 1.81, + "grad_norm": 3.24204158782959, + "learning_rate": 8.867735470941884e-06, + "loss": 0.2426, + "step": 325 + }, + { + "epoch": 1.81, + "eval_loss": 0.6821221709251404, + "eval_runtime": 33.9704, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 325 + }, + { + "epoch": 1.94, + "grad_norm": 5.852811336517334, + "learning_rate": 7.615230460921845e-06, + "loss": 0.2487, + "step": 350 + }, + { + "epoch": 1.94, + "eval_loss": 0.6622934341430664, + "eval_runtime": 33.9863, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 350 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1.5451530769465344e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-350/training_args.bin b/checkpoint-350/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-350/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-375/README.md b/checkpoint-375/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-375/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-375/adapter_config.json b/checkpoint-375/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-375/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-375/adapter_model.safetensors b/checkpoint-375/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a8286c333017c7c5aa3a5a6173ed82b45e9d98e0 --- /dev/null +++ b/checkpoint-375/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a64eb70bc364faf76707b1d8e107a017b8238fc23c08181baa52c4e02df2dda +size 864513616 diff --git a/checkpoint-375/optimizer.pt b/checkpoint-375/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..3565c2840f4d5067e1268e86fc866a89cd4e3881 --- /dev/null +++ b/checkpoint-375/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:862b2660eb06f45dc9d0c269701ca1e4230c4445bb5ed949d806994582320978 +size 170951516 diff --git a/checkpoint-375/rng_state.pth b/checkpoint-375/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..d580b7474b41a3ef33513cfd8ca9e70bc7f59b01 --- /dev/null +++ b/checkpoint-375/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2da3568966f38b12e4941e65c936b03f00741cfc8cfbfabd4d506fa2db520536 +size 14244 diff --git a/checkpoint-375/scheduler.pt b/checkpoint-375/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..9e38488fd3eea6743cadca92a115a318d461f261 --- /dev/null +++ b/checkpoint-375/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dcc6df55b43423a8a66421578b11fc4e0685c1a862f8449bf6ef2c3faa9d7736 +size 1064 diff --git a/checkpoint-375/trainer_state.json b/checkpoint-375/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b7ce884f16a9b681f1928b48537b4c9db4f8c4a1 --- /dev/null +++ b/checkpoint-375/trainer_state.json @@ -0,0 +1,246 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0833333333333335, + "eval_steps": 25, + "global_step": 375, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + }, + { + "epoch": 1.67, + "grad_norm": 5.02877950668335, + "learning_rate": 1.0120240480961924e-05, + "loss": 0.2323, + "step": 300 + }, + { + "epoch": 1.67, + "eval_loss": 0.6742722392082214, + "eval_runtime": 33.9146, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 300 + }, + { + "epoch": 1.81, + "grad_norm": 3.24204158782959, + "learning_rate": 8.867735470941884e-06, + "loss": 0.2426, + "step": 325 + }, + { + "epoch": 1.81, + "eval_loss": 0.6821221709251404, + "eval_runtime": 33.9704, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 325 + }, + { + "epoch": 1.94, + "grad_norm": 5.852811336517334, + "learning_rate": 7.615230460921845e-06, + "loss": 0.2487, + "step": 350 + }, + { + "epoch": 1.94, + "eval_loss": 0.6622934341430664, + "eval_runtime": 33.9863, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 350 + }, + { + "epoch": 2.08, + "grad_norm": 2.1195895671844482, + "learning_rate": 6.362725450901804e-06, + "loss": 0.1755, + "step": 375 + }, + { + "epoch": 2.08, + "eval_loss": 0.7334879040718079, + "eval_runtime": 33.9725, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 375 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1.6534685286924288e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-375/training_args.bin b/checkpoint-375/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-375/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-400/README.md b/checkpoint-400/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-400/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-400/adapter_config.json b/checkpoint-400/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-400/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-400/adapter_model.safetensors b/checkpoint-400/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0ba36a8c4b2ce76d5987699a4b1b27c346dc7713 --- /dev/null +++ b/checkpoint-400/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9aa5d1024e7f46241d15b75893521fc3d698e1606f5f85531545ef47c2cd2d7 +size 864513616 diff --git a/checkpoint-400/optimizer.pt b/checkpoint-400/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..d62b80c0f126ed457714ecfab78b253e6ef9c6c7 --- /dev/null +++ b/checkpoint-400/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3b826863b5c7e883e5aacbf0fe85d64ef6791036c853875f284400a0c734128 +size 170951516 diff --git a/checkpoint-400/rng_state.pth b/checkpoint-400/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..b10fc8585fe46a219f9c348ea39c960e518daa84 --- /dev/null +++ b/checkpoint-400/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:501fb991faf4047be2218317144b4a81b87b4d59939ca1087e509bbd6af1e194 +size 14244 diff --git a/checkpoint-400/scheduler.pt b/checkpoint-400/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..49b4fd106e1813e073d96a9f62a3763cf0b4b568 --- /dev/null +++ b/checkpoint-400/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3ae489f5e1bc65ebdfe723bb65561dca563fbef27d7fb329ff3ef24abe47a65 +size 1064 diff --git a/checkpoint-400/trainer_state.json b/checkpoint-400/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..834c692d00cf243d8a034c45bb943285e0d8e7ab --- /dev/null +++ b/checkpoint-400/trainer_state.json @@ -0,0 +1,261 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.2222222222222223, + "eval_steps": 25, + "global_step": 400, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + }, + { + "epoch": 1.67, + "grad_norm": 5.02877950668335, + "learning_rate": 1.0120240480961924e-05, + "loss": 0.2323, + "step": 300 + }, + { + "epoch": 1.67, + "eval_loss": 0.6742722392082214, + "eval_runtime": 33.9146, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 300 + }, + { + "epoch": 1.81, + "grad_norm": 3.24204158782959, + "learning_rate": 8.867735470941884e-06, + "loss": 0.2426, + "step": 325 + }, + { + "epoch": 1.81, + "eval_loss": 0.6821221709251404, + "eval_runtime": 33.9704, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 325 + }, + { + "epoch": 1.94, + "grad_norm": 5.852811336517334, + "learning_rate": 7.615230460921845e-06, + "loss": 0.2487, + "step": 350 + }, + { + "epoch": 1.94, + "eval_loss": 0.6622934341430664, + "eval_runtime": 33.9863, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 350 + }, + { + "epoch": 2.08, + "grad_norm": 2.1195895671844482, + "learning_rate": 6.362725450901804e-06, + "loss": 0.1755, + "step": 375 + }, + { + "epoch": 2.08, + "eval_loss": 0.7334879040718079, + "eval_runtime": 33.9725, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 375 + }, + { + "epoch": 2.22, + "grad_norm": 2.2676243782043457, + "learning_rate": 5.110220440881764e-06, + "loss": 0.1357, + "step": 400 + }, + { + "epoch": 2.22, + "eval_loss": 0.8047094345092773, + "eval_runtime": 33.8765, + "eval_samples_per_second": 1.801, + "eval_steps_per_second": 0.236, + "step": 400 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1.7639944998617088e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-400/training_args.bin b/checkpoint-400/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-400/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-425/README.md b/checkpoint-425/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-425/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-425/adapter_config.json b/checkpoint-425/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-425/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-425/adapter_model.safetensors b/checkpoint-425/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f0909add825f1d11ff09a9ff1e58168884bb4a56 --- /dev/null +++ b/checkpoint-425/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36c826f8a7c3717da38f767994c4a8c6cdbafa6719930c454d9986955f22d70b +size 864513616 diff --git a/checkpoint-425/optimizer.pt b/checkpoint-425/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f567731e9848eaf132ccb122a7ead7958dcb0c32 --- /dev/null +++ b/checkpoint-425/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4df78665d0f7e854b4799a08e4ee5f508ecc831b6578e774651b16c8da6f8133 +size 170951516 diff --git a/checkpoint-425/rng_state.pth b/checkpoint-425/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..843023c8c409bc9213732f16757113c69b02dab9 --- /dev/null +++ b/checkpoint-425/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:579e59cdb61ccd8c2ed4d8dca4df3f86202d9386f6d72720c65000af85814df6 +size 14244 diff --git a/checkpoint-425/scheduler.pt b/checkpoint-425/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..74c87e29c9f151793c47c9f69773894e1ecac207 --- /dev/null +++ b/checkpoint-425/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c17795170d20cfe2d2ffe94ad5e676acc2b236e5e4599a01639c64bda7fd50cb +size 1064 diff --git a/checkpoint-425/trainer_state.json b/checkpoint-425/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..9b78e444a96d76ffd3b166654832c71c934015f7 --- /dev/null +++ b/checkpoint-425/trainer_state.json @@ -0,0 +1,276 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.361111111111111, + "eval_steps": 25, + "global_step": 425, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + }, + { + "epoch": 1.67, + "grad_norm": 5.02877950668335, + "learning_rate": 1.0120240480961924e-05, + "loss": 0.2323, + "step": 300 + }, + { + "epoch": 1.67, + "eval_loss": 0.6742722392082214, + "eval_runtime": 33.9146, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 300 + }, + { + "epoch": 1.81, + "grad_norm": 3.24204158782959, + "learning_rate": 8.867735470941884e-06, + "loss": 0.2426, + "step": 325 + }, + { + "epoch": 1.81, + "eval_loss": 0.6821221709251404, + "eval_runtime": 33.9704, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 325 + }, + { + "epoch": 1.94, + "grad_norm": 5.852811336517334, + "learning_rate": 7.615230460921845e-06, + "loss": 0.2487, + "step": 350 + }, + { + "epoch": 1.94, + "eval_loss": 0.6622934341430664, + "eval_runtime": 33.9863, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 350 + }, + { + "epoch": 2.08, + "grad_norm": 2.1195895671844482, + "learning_rate": 6.362725450901804e-06, + "loss": 0.1755, + "step": 375 + }, + { + "epoch": 2.08, + "eval_loss": 0.7334879040718079, + "eval_runtime": 33.9725, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 375 + }, + { + "epoch": 2.22, + "grad_norm": 2.2676243782043457, + "learning_rate": 5.110220440881764e-06, + "loss": 0.1357, + "step": 400 + }, + { + "epoch": 2.22, + "eval_loss": 0.8047094345092773, + "eval_runtime": 33.8765, + "eval_samples_per_second": 1.801, + "eval_steps_per_second": 0.236, + "step": 400 + }, + { + "epoch": 2.36, + "grad_norm": 2.1398653984069824, + "learning_rate": 3.857715430861724e-06, + "loss": 0.1517, + "step": 425 + }, + { + "epoch": 2.36, + "eval_loss": 0.8226982355117798, + "eval_runtime": 33.9677, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 425 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1.874520471030989e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-425/training_args.bin b/checkpoint-425/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-425/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-450/README.md b/checkpoint-450/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-450/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-450/adapter_config.json b/checkpoint-450/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-450/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-450/adapter_model.safetensors b/checkpoint-450/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..874407289399ca13832164d0876df62ca2d31577 --- /dev/null +++ b/checkpoint-450/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e37da1a09bb6ff5e0ce99f0bbdf55090b7e4f124b092ebe469e52b94a5af9d7 +size 864513616 diff --git a/checkpoint-450/optimizer.pt b/checkpoint-450/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..fcccc4f4521b6484e28ffef10c4d080e12bfb696 --- /dev/null +++ b/checkpoint-450/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:786dfbb670a01ddff6ca8810c6b562ba463e214e4ec83a19d62cc70492409246 +size 170951516 diff --git a/checkpoint-450/rng_state.pth b/checkpoint-450/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..7a44a76294b571622f4d21d5caac981f7779f87a --- /dev/null +++ b/checkpoint-450/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf04273c74d6760f8195454f2c2823ede2ccffbd1cb2ba5f04e1e5a866fdfd76 +size 14244 diff --git a/checkpoint-450/scheduler.pt b/checkpoint-450/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..07ffa743cc45ffb55372255d64033606955d2721 --- /dev/null +++ b/checkpoint-450/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e6a517bef6a958827ae7fdc6f21f44accdb2b79094cb2f681c416cc6f70667f +size 1064 diff --git a/checkpoint-450/trainer_state.json b/checkpoint-450/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..cf1b3d0a4e4d03b6e5b4699f636d0baa7d2baff7 --- /dev/null +++ b/checkpoint-450/trainer_state.json @@ -0,0 +1,291 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.5, + "eval_steps": 25, + "global_step": 450, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + }, + { + "epoch": 1.67, + "grad_norm": 5.02877950668335, + "learning_rate": 1.0120240480961924e-05, + "loss": 0.2323, + "step": 300 + }, + { + "epoch": 1.67, + "eval_loss": 0.6742722392082214, + "eval_runtime": 33.9146, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 300 + }, + { + "epoch": 1.81, + "grad_norm": 3.24204158782959, + "learning_rate": 8.867735470941884e-06, + "loss": 0.2426, + "step": 325 + }, + { + "epoch": 1.81, + "eval_loss": 0.6821221709251404, + "eval_runtime": 33.9704, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 325 + }, + { + "epoch": 1.94, + "grad_norm": 5.852811336517334, + "learning_rate": 7.615230460921845e-06, + "loss": 0.2487, + "step": 350 + }, + { + "epoch": 1.94, + "eval_loss": 0.6622934341430664, + "eval_runtime": 33.9863, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 350 + }, + { + "epoch": 2.08, + "grad_norm": 2.1195895671844482, + "learning_rate": 6.362725450901804e-06, + "loss": 0.1755, + "step": 375 + }, + { + "epoch": 2.08, + "eval_loss": 0.7334879040718079, + "eval_runtime": 33.9725, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 375 + }, + { + "epoch": 2.22, + "grad_norm": 2.2676243782043457, + "learning_rate": 5.110220440881764e-06, + "loss": 0.1357, + "step": 400 + }, + { + "epoch": 2.22, + "eval_loss": 0.8047094345092773, + "eval_runtime": 33.8765, + "eval_samples_per_second": 1.801, + "eval_steps_per_second": 0.236, + "step": 400 + }, + { + "epoch": 2.36, + "grad_norm": 2.1398653984069824, + "learning_rate": 3.857715430861724e-06, + "loss": 0.1517, + "step": 425 + }, + { + "epoch": 2.36, + "eval_loss": 0.8226982355117798, + "eval_runtime": 33.9677, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 425 + }, + { + "epoch": 2.5, + "grad_norm": 3.4427037239074707, + "learning_rate": 2.6052104208416833e-06, + "loss": 0.1596, + "step": 450 + }, + { + "epoch": 2.5, + "eval_loss": 0.7910816073417664, + "eval_runtime": 33.9419, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 450 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 1.985046442200269e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-450/training_args.bin b/checkpoint-450/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-450/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-475/README.md b/checkpoint-475/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-475/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-475/adapter_config.json b/checkpoint-475/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-475/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-475/adapter_model.safetensors b/checkpoint-475/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7b06b458394b47d872db914dcd55bd8bc43941d2 --- /dev/null +++ b/checkpoint-475/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6ca150f2863cf12db2c6b29dc50e27ac977b55ea4ff3f5ec1291db6a2bcbb53 +size 864513616 diff --git a/checkpoint-475/optimizer.pt b/checkpoint-475/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..1f4f6e028f6eae5c8a2bc3364a8490f74d721ce8 --- /dev/null +++ b/checkpoint-475/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f4b7dbe0f5769af14aeff7f5d2f12548b67d61fe7dc73e5b39fa9bef886c811 +size 170951516 diff --git a/checkpoint-475/rng_state.pth b/checkpoint-475/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..e198d4a514fdad9c2fda0a1121dcb25de7aeca10 --- /dev/null +++ b/checkpoint-475/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77902545769c6f0dd146b26c44d68301ea84f8c0f217f713966d2ca66223926e +size 14244 diff --git a/checkpoint-475/scheduler.pt b/checkpoint-475/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ab3f5d7ec01bb798f27175e5b4563b80d1ee3cec --- /dev/null +++ b/checkpoint-475/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3920014c37fe5d259018f2555b10c042dbe46d4c171a01af9350f59309cf0708 +size 1064 diff --git a/checkpoint-475/trainer_state.json b/checkpoint-475/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..2586d095c851c9c11b394e1c0f0cbaebb0b9de55 --- /dev/null +++ b/checkpoint-475/trainer_state.json @@ -0,0 +1,306 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.638888888888889, + "eval_steps": 25, + "global_step": 475, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + }, + { + "epoch": 1.67, + "grad_norm": 5.02877950668335, + "learning_rate": 1.0120240480961924e-05, + "loss": 0.2323, + "step": 300 + }, + { + "epoch": 1.67, + "eval_loss": 0.6742722392082214, + "eval_runtime": 33.9146, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 300 + }, + { + "epoch": 1.81, + "grad_norm": 3.24204158782959, + "learning_rate": 8.867735470941884e-06, + "loss": 0.2426, + "step": 325 + }, + { + "epoch": 1.81, + "eval_loss": 0.6821221709251404, + "eval_runtime": 33.9704, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 325 + }, + { + "epoch": 1.94, + "grad_norm": 5.852811336517334, + "learning_rate": 7.615230460921845e-06, + "loss": 0.2487, + "step": 350 + }, + { + "epoch": 1.94, + "eval_loss": 0.6622934341430664, + "eval_runtime": 33.9863, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 350 + }, + { + "epoch": 2.08, + "grad_norm": 2.1195895671844482, + "learning_rate": 6.362725450901804e-06, + "loss": 0.1755, + "step": 375 + }, + { + "epoch": 2.08, + "eval_loss": 0.7334879040718079, + "eval_runtime": 33.9725, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 375 + }, + { + "epoch": 2.22, + "grad_norm": 2.2676243782043457, + "learning_rate": 5.110220440881764e-06, + "loss": 0.1357, + "step": 400 + }, + { + "epoch": 2.22, + "eval_loss": 0.8047094345092773, + "eval_runtime": 33.8765, + "eval_samples_per_second": 1.801, + "eval_steps_per_second": 0.236, + "step": 400 + }, + { + "epoch": 2.36, + "grad_norm": 2.1398653984069824, + "learning_rate": 3.857715430861724e-06, + "loss": 0.1517, + "step": 425 + }, + { + "epoch": 2.36, + "eval_loss": 0.8226982355117798, + "eval_runtime": 33.9677, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 425 + }, + { + "epoch": 2.5, + "grad_norm": 3.4427037239074707, + "learning_rate": 2.6052104208416833e-06, + "loss": 0.1596, + "step": 450 + }, + { + "epoch": 2.5, + "eval_loss": 0.7910816073417664, + "eval_runtime": 33.9419, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 450 + }, + { + "epoch": 2.64, + "grad_norm": 2.103198289871216, + "learning_rate": 1.3527054108216433e-06, + "loss": 0.149, + "step": 475 + }, + { + "epoch": 2.64, + "eval_loss": 0.7897896766662598, + "eval_runtime": 33.9551, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 475 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 2.095572413369549e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-475/training_args.bin b/checkpoint-475/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-475/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-50/README.md b/checkpoint-50/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-50/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-50/adapter_config.json b/checkpoint-50/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-50/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-50/adapter_model.safetensors b/checkpoint-50/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cb19895e665f6e3b7792407b3fe251c637b6fbef --- /dev/null +++ b/checkpoint-50/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcd68cd11765567f5c9766fa76530afdac3a0e05b8e0001600aaf796678f8680 +size 864513616 diff --git a/checkpoint-50/optimizer.pt b/checkpoint-50/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..0347733297ade24d6205e53fce12af4c12b105eb --- /dev/null +++ b/checkpoint-50/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9dff483e10aceaef13edabdfb3151c497a9c9f5fd268c4394c38ad6587154978 +size 170951068 diff --git a/checkpoint-50/rng_state.pth b/checkpoint-50/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c2e65cee68b0a16f32cce730df75aac69998bee5 --- /dev/null +++ b/checkpoint-50/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:845b54db4851e68ff9a0c5979b186f5887c53b72b2fcd223cc7eb6980d5610dd +size 14244 diff --git a/checkpoint-50/scheduler.pt b/checkpoint-50/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..0aba4ec590bb6fa450d06962d892b9c398eba78f --- /dev/null +++ b/checkpoint-50/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3c941c25caa799ba368c16ad5688734d8076373efed8c4bad6447da68a82587 +size 1064 diff --git a/checkpoint-50/trainer_state.json b/checkpoint-50/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..8e79200ad9d357d7085fa8575f1dc554f5d76d82 --- /dev/null +++ b/checkpoint-50/trainer_state.json @@ -0,0 +1,51 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.2777777777777778, + "eval_steps": 25, + "global_step": 50, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 2210519423385600.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-50/training_args.bin b/checkpoint-50/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-50/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-500/README.md b/checkpoint-500/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-500/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-500/adapter_config.json b/checkpoint-500/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-500/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-500/adapter_model.safetensors b/checkpoint-500/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d28ef21df7ffd22603440119157debce3904bf46 --- /dev/null +++ b/checkpoint-500/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a3f436abd4b2db4e3b4bdf6ab28bd8dd8693c1f306fc05a522a3a766287aef4 +size 864513616 diff --git a/checkpoint-500/optimizer.pt b/checkpoint-500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..2fd7fd02c7641fccdc653fc957e9450d0aafcc24 --- /dev/null +++ b/checkpoint-500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07f4fd92f3c7c305295a38450acbc74087fff1d6da0bd1e41467adbef83ef19f +size 170951516 diff --git a/checkpoint-500/rng_state.pth b/checkpoint-500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..7c9013e9cecd86742b47c32e7186ee4861543db4 --- /dev/null +++ b/checkpoint-500/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48a4d954d72656e6e5043f2a147181ab0a62586ff718afd12c71233cce98db0c +size 14244 diff --git a/checkpoint-500/scheduler.pt b/checkpoint-500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..510b41c787e8d2e646299db28263e256bdee4121 --- /dev/null +++ b/checkpoint-500/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d3770c531e0261d92b01eddd237bfc9608bbec329625a63ac8e0f798d2d6018 +size 1064 diff --git a/checkpoint-500/trainer_state.json b/checkpoint-500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..91d24fd589e785af01c009d3ead86ba6b7a69e00 --- /dev/null +++ b/checkpoint-500/trainer_state.json @@ -0,0 +1,321 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.7777777777777777, + "eval_steps": 25, + "global_step": 500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + }, + { + "epoch": 0.56, + "grad_norm": 6.133337020874023, + "learning_rate": 2.0090180360721444e-05, + "loss": 0.4387, + "step": 100 + }, + { + "epoch": 0.56, + "eval_loss": 0.6550642251968384, + "eval_runtime": 33.9823, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 100 + }, + { + "epoch": 0.69, + "grad_norm": 7.314888954162598, + "learning_rate": 1.8837675350701404e-05, + "loss": 0.4692, + "step": 125 + }, + { + "epoch": 0.69, + "eval_loss": 0.6102645993232727, + "eval_runtime": 33.9789, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 125 + }, + { + "epoch": 0.83, + "grad_norm": 5.678727149963379, + "learning_rate": 1.7585170340681365e-05, + "loss": 0.4364, + "step": 150 + }, + { + "epoch": 0.83, + "eval_loss": 0.6225055456161499, + "eval_runtime": 34.038, + "eval_samples_per_second": 1.792, + "eval_steps_per_second": 0.235, + "step": 150 + }, + { + "epoch": 0.97, + "grad_norm": 5.900119304656982, + "learning_rate": 1.633266533066132e-05, + "loss": 0.4211, + "step": 175 + }, + { + "epoch": 0.97, + "eval_loss": 0.6140083074569702, + "eval_runtime": 33.9476, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 175 + }, + { + "epoch": 1.11, + "grad_norm": 5.989519119262695, + "learning_rate": 1.5080160320641284e-05, + "loss": 0.2941, + "step": 200 + }, + { + "epoch": 1.11, + "eval_loss": 0.6877502799034119, + "eval_runtime": 33.9316, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 200 + }, + { + "epoch": 1.25, + "grad_norm": 2.5047194957733154, + "learning_rate": 1.3827655310621242e-05, + "loss": 0.2562, + "step": 225 + }, + { + "epoch": 1.25, + "eval_loss": 0.7026296257972717, + "eval_runtime": 33.9018, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 225 + }, + { + "epoch": 1.39, + "grad_norm": 5.878956317901611, + "learning_rate": 1.2575150300601201e-05, + "loss": 0.183, + "step": 250 + }, + { + "epoch": 1.39, + "eval_loss": 0.7297132015228271, + "eval_runtime": 33.9801, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 250 + }, + { + "epoch": 1.53, + "grad_norm": 6.625136852264404, + "learning_rate": 1.1322645290581163e-05, + "loss": 0.245, + "step": 275 + }, + { + "epoch": 1.53, + "eval_loss": 0.7251414060592651, + "eval_runtime": 33.928, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 275 + }, + { + "epoch": 1.67, + "grad_norm": 5.02877950668335, + "learning_rate": 1.0120240480961924e-05, + "loss": 0.2323, + "step": 300 + }, + { + "epoch": 1.67, + "eval_loss": 0.6742722392082214, + "eval_runtime": 33.9146, + "eval_samples_per_second": 1.799, + "eval_steps_per_second": 0.236, + "step": 300 + }, + { + "epoch": 1.81, + "grad_norm": 3.24204158782959, + "learning_rate": 8.867735470941884e-06, + "loss": 0.2426, + "step": 325 + }, + { + "epoch": 1.81, + "eval_loss": 0.6821221709251404, + "eval_runtime": 33.9704, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 325 + }, + { + "epoch": 1.94, + "grad_norm": 5.852811336517334, + "learning_rate": 7.615230460921845e-06, + "loss": 0.2487, + "step": 350 + }, + { + "epoch": 1.94, + "eval_loss": 0.6622934341430664, + "eval_runtime": 33.9863, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 350 + }, + { + "epoch": 2.08, + "grad_norm": 2.1195895671844482, + "learning_rate": 6.362725450901804e-06, + "loss": 0.1755, + "step": 375 + }, + { + "epoch": 2.08, + "eval_loss": 0.7334879040718079, + "eval_runtime": 33.9725, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.235, + "step": 375 + }, + { + "epoch": 2.22, + "grad_norm": 2.2676243782043457, + "learning_rate": 5.110220440881764e-06, + "loss": 0.1357, + "step": 400 + }, + { + "epoch": 2.22, + "eval_loss": 0.8047094345092773, + "eval_runtime": 33.8765, + "eval_samples_per_second": 1.801, + "eval_steps_per_second": 0.236, + "step": 400 + }, + { + "epoch": 2.36, + "grad_norm": 2.1398653984069824, + "learning_rate": 3.857715430861724e-06, + "loss": 0.1517, + "step": 425 + }, + { + "epoch": 2.36, + "eval_loss": 0.8226982355117798, + "eval_runtime": 33.9677, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 425 + }, + { + "epoch": 2.5, + "grad_norm": 3.4427037239074707, + "learning_rate": 2.6052104208416833e-06, + "loss": 0.1596, + "step": 450 + }, + { + "epoch": 2.5, + "eval_loss": 0.7910816073417664, + "eval_runtime": 33.9419, + "eval_samples_per_second": 1.797, + "eval_steps_per_second": 0.236, + "step": 450 + }, + { + "epoch": 2.64, + "grad_norm": 2.103198289871216, + "learning_rate": 1.3527054108216433e-06, + "loss": 0.149, + "step": 475 + }, + { + "epoch": 2.64, + "eval_loss": 0.7897896766662598, + "eval_runtime": 33.9551, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 475 + }, + { + "epoch": 2.78, + "grad_norm": 5.470449447631836, + "learning_rate": 1.002004008016032e-07, + "loss": 0.1392, + "step": 500 + }, + { + "epoch": 2.78, + "eval_loss": 0.7936234474182129, + "eval_runtime": 33.9247, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 500 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 2.206098384538829e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-500/training_args.bin b/checkpoint-500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-500/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920 diff --git a/checkpoint-75/README.md b/checkpoint-75/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d3dc7cb85422278e4603b668e7f60cfac7188aa2 --- /dev/null +++ b/checkpoint-75/README.md @@ -0,0 +1,204 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +### Framework versions + +- PEFT 0.8.2 \ No newline at end of file diff --git a/checkpoint-75/adapter_config.json b/checkpoint-75/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bf321cbd266a87598c75e2d99dd5de27af0b6bc --- /dev/null +++ b/checkpoint-75/adapter_config.json @@ -0,0 +1,33 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 64, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "o_proj", + "down_proj", + "q_proj", + "gate_proj", + "v_proj", + "k_proj", + "up_proj", + "lm_head" + ], + "task_type": "CAUSAL_LM", + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-75/adapter_model.safetensors b/checkpoint-75/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5d5eff7a663cdeaed87608f9b3aad4d78a406bbf --- /dev/null +++ b/checkpoint-75/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32081940efa78a3265336e198736ee042e2115f7a12db0cf69a975196262cb35 +size 864513616 diff --git a/checkpoint-75/optimizer.pt b/checkpoint-75/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..97b800b00a85303900e3231123faa3f1852cf241 --- /dev/null +++ b/checkpoint-75/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:939713c282a02f9d05ec493d194bf7f09f13ee42fc75cdb3d5a8abe0e5af8495 +size 170951068 diff --git a/checkpoint-75/rng_state.pth b/checkpoint-75/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..eaea6ef41eb491ab59cae70a72e52f1e3a4933c1 --- /dev/null +++ b/checkpoint-75/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77b076f55da447ae13773fe2249d49ba2f62e8c2b957c6ca2dfebcb53eb85fb5 +size 14244 diff --git a/checkpoint-75/scheduler.pt b/checkpoint-75/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..2c390d47cba5395648fc89bb93832c2172bd5243 --- /dev/null +++ b/checkpoint-75/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56156dc02631344b6a7ba7f24300f0e0e931c66155f084f88188840e7b47b3f7 +size 1064 diff --git a/checkpoint-75/trainer_state.json b/checkpoint-75/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..d05bc63df802db61a36844ecfc379820d9816f24 --- /dev/null +++ b/checkpoint-75/trainer_state.json @@ -0,0 +1,66 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4166666666666667, + "eval_steps": 25, + "global_step": 75, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.14, + "grad_norm": 6.271986484527588, + "learning_rate": 2.3847695390781562e-05, + "loss": 0.9018, + "step": 25 + }, + { + "epoch": 0.14, + "eval_loss": 0.6543389558792114, + "eval_runtime": 33.9351, + "eval_samples_per_second": 1.798, + "eval_steps_per_second": 0.236, + "step": 25 + }, + { + "epoch": 0.28, + "grad_norm": 6.715949535369873, + "learning_rate": 2.2595190380761526e-05, + "loss": 0.6032, + "step": 50 + }, + { + "epoch": 0.28, + "eval_loss": 0.6594853401184082, + "eval_runtime": 33.9905, + "eval_samples_per_second": 1.795, + "eval_steps_per_second": 0.235, + "step": 50 + }, + { + "epoch": 0.42, + "grad_norm": 5.575837135314941, + "learning_rate": 2.1342685370741483e-05, + "loss": 0.5277, + "step": 75 + }, + { + "epoch": 0.42, + "eval_loss": 0.6358431577682495, + "eval_runtime": 33.9553, + "eval_samples_per_second": 1.796, + "eval_steps_per_second": 0.236, + "step": 75 + } + ], + "logging_steps": 25, + "max_steps": 500, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 25, + "total_flos": 3315779135078400.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-75/training_args.bin b/checkpoint-75/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2ac826f970e917ce8bd622621abbd65746c266be --- /dev/null +++ b/checkpoint-75/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35b29e367dde5956ae9c9403938a756abc9dfab88d6384703c11aac044f2bed2 +size 4920