--- base_model: - mllm-dev/gpt2_f_experiment_9 - mllm-dev/gpt2_f_experiment_2 - mllm-dev/gpt2_f_experiment_8 - mllm-dev/gpt2_f_experiment_6 - mllm-dev/gpt2_f_experiment_7 - mllm-dev/gpt2_f_experiment_4 - mllm-dev/gpt2_f_experiment_3 - mllm-dev/gpt2_f_experiment_5 - mllm-dev/gpt2_f_experiment_0 - mllm-dev/gpt2_f_experiment_1 library_name: transformers tags: - mergekit - merge --- # sean_test_merge_out This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [mllm-dev/gpt2_f_experiment_0](https://huggingface.co/mllm-dev/gpt2_f_experiment_0) as a base. ### Models Merged The following models were included in the merge: * [mllm-dev/gpt2_f_experiment_9](https://huggingface.co/mllm-dev/gpt2_f_experiment_9) * [mllm-dev/gpt2_f_experiment_2](https://huggingface.co/mllm-dev/gpt2_f_experiment_2) * [mllm-dev/gpt2_f_experiment_8](https://huggingface.co/mllm-dev/gpt2_f_experiment_8) * [mllm-dev/gpt2_f_experiment_6](https://huggingface.co/mllm-dev/gpt2_f_experiment_6) * [mllm-dev/gpt2_f_experiment_7](https://huggingface.co/mllm-dev/gpt2_f_experiment_7) * [mllm-dev/gpt2_f_experiment_4](https://huggingface.co/mllm-dev/gpt2_f_experiment_4) * [mllm-dev/gpt2_f_experiment_3](https://huggingface.co/mllm-dev/gpt2_f_experiment_3) * [mllm-dev/gpt2_f_experiment_5](https://huggingface.co/mllm-dev/gpt2_f_experiment_5) * [mllm-dev/gpt2_f_experiment_1](https://huggingface.co/mllm-dev/gpt2_f_experiment_1) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: model: path: mllm-dev/gpt2_f_experiment_0 dtype: float16 merge_method: ties parameters: normalize: 1.0 slices: - sources: - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_0 - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_1 parameters: density: 0.9 weight: 0.1 - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_2 parameters: density: 0.9 weight: 0.1 - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_3 parameters: density: 0.9 weight: 0.1 - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_4 parameters: density: 0.9 weight: 0.1 - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_5 parameters: density: 0.9 weight: 0.1 - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_6 parameters: density: 0.9 weight: 0.1 - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_7 parameters: density: 0.9 weight: 0.1 - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_8 parameters: density: 0.9 weight: 0.1 - layer_range: [0, 12] model: model: path: mllm-dev/gpt2_f_experiment_9 parameters: density: 0.9 weight: 0.1 ```