Bowen232 commited on
Commit
b026360
1 Parent(s): 7f859cb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -1
README.md CHANGED
@@ -17,10 +17,21 @@ LoRAs and fusion gates for our paper
17
 
18
  We released all of our checkpoints used in [LoRA-Flow](https://aclanthology.org/2024.acl-long.695.pdf) which has been accepted to ACL 2024 main conference.
19
  # Summary
20
- > In this repo, we release LoRA and the gate of 7B models trained in our paper in HuggingFace format.
21
  # Introduction
22
  LoRA-Flow provides an efficient way to fuse different LoRA modules which can outperform existing methods significantly. The following picture shows our proposed method, we use layer-wise fusion gates to facilitate dynamic LoRA fusion, which project input hidden states of each layer into fusion weights. For more details, please refer to our paper.
23
  ![1.jpg](https://cdn-uploads.huggingface.co/production/uploads/64d99f6cd7e30889c6c477b4/ifiu1FTHilrmUkD4FKkgV.jpeg)
 
 
 
 
 
 
 
 
 
 
 
24
  # Citation
25
  if you find our repo is helpful, please cite the following
26
  ```bibtex
 
17
 
18
  We released all of our checkpoints used in [LoRA-Flow](https://aclanthology.org/2024.acl-long.695.pdf) which has been accepted to ACL 2024 main conference.
19
  # Summary
20
+ > In this repo, we release LoRA modules and the gate of 7B models trained in our paper in HuggingFace format.
21
  # Introduction
22
  LoRA-Flow provides an efficient way to fuse different LoRA modules which can outperform existing methods significantly. The following picture shows our proposed method, we use layer-wise fusion gates to facilitate dynamic LoRA fusion, which project input hidden states of each layer into fusion weights. For more details, please refer to our paper.
23
  ![1.jpg](https://cdn-uploads.huggingface.co/production/uploads/64d99f6cd7e30889c6c477b4/ifiu1FTHilrmUkD4FKkgV.jpeg)
24
+ # Training Details
25
+ ## LoRA modules Training
26
+ For language LoRA modules: we use the data 52K training examples respectively which from [Okapi](https://aclanthology.org/2023.emnlp-demo.28).
27
+
28
+ For math LoRA module: the training data for English math LoRA is constructed by [Metamath](https://arxiv.org/abs/2309.12284), which is comprised of 395K mathematical problems in English.
29
+
30
+ For code LoRA module: we train the English code LoRA with the Magicoder dataset [Magicoder](https://arxiv.org/abs/2312.02120), which consists of 186K code generation problems in English.
31
+
32
+ ## Gate Training
33
+ We use gates to fuse different LoRA modules. We employ few-shot training and have released our training data for further details please refer to our GitHub.
34
+
35
  # Citation
36
  if you find our repo is helpful, please cite the following
37
  ```bibtex