Update README.md
Browse files
README.md
CHANGED
|
@@ -17,10 +17,21 @@ LoRAs and fusion gates for our paper
|
|
| 17 |
|
| 18 |
We released all of our checkpoints used in [LoRA-Flow](https://aclanthology.org/2024.acl-long.695.pdf) which has been accepted to ACL 2024 main conference.
|
| 19 |
# Summary
|
| 20 |
-
> In this repo, we release LoRA and the gate of 7B models trained in our paper in HuggingFace format.
|
| 21 |
# Introduction
|
| 22 |
LoRA-Flow provides an efficient way to fuse different LoRA modules which can outperform existing methods significantly. The following picture shows our proposed method, we use layer-wise fusion gates to facilitate dynamic LoRA fusion, which project input hidden states of each layer into fusion weights. For more details, please refer to our paper.
|
| 23 |

|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
# Citation
|
| 25 |
if you find our repo is helpful, please cite the following
|
| 26 |
```bibtex
|
|
|
|
| 17 |
|
| 18 |
We released all of our checkpoints used in [LoRA-Flow](https://aclanthology.org/2024.acl-long.695.pdf) which has been accepted to ACL 2024 main conference.
|
| 19 |
# Summary
|
| 20 |
+
> In this repo, we release LoRA modules and the gate of 7B models trained in our paper in HuggingFace format.
|
| 21 |
# Introduction
|
| 22 |
LoRA-Flow provides an efficient way to fuse different LoRA modules which can outperform existing methods significantly. The following picture shows our proposed method, we use layer-wise fusion gates to facilitate dynamic LoRA fusion, which project input hidden states of each layer into fusion weights. For more details, please refer to our paper.
|
| 23 |

|
| 24 |
+
# Training Details
|
| 25 |
+
## LoRA modules Training
|
| 26 |
+
For language LoRA modules: we use the data 52K training examples respectively which from [Okapi](https://aclanthology.org/2023.emnlp-demo.28).
|
| 27 |
+
|
| 28 |
+
For math LoRA module: the training data for English math LoRA is constructed by [Metamath](https://arxiv.org/abs/2309.12284), which is comprised of 395K mathematical problems in English.
|
| 29 |
+
|
| 30 |
+
For code LoRA module: we train the English code LoRA with the Magicoder dataset [Magicoder](https://arxiv.org/abs/2312.02120), which consists of 186K code generation problems in English.
|
| 31 |
+
|
| 32 |
+
## Gate Training
|
| 33 |
+
We use gates to fuse different LoRA modules. We employ few-shot training and have released our training data for further details please refer to our GitHub.
|
| 34 |
+
|
| 35 |
# Citation
|
| 36 |
if you find our repo is helpful, please cite the following
|
| 37 |
```bibtex
|