Bowen232 commited on
Commit
4827ef5
1 Parent(s): ec910a7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -16
README.md CHANGED
@@ -19,21 +19,21 @@ We released all of our checkpoints used in [LoRA-Flow](https://aclanthology.org/
19
  # Summary
20
  > In this repo, we release LoRA modules and the gate of 7B models trained in our paper in HuggingFace format.
21
  # Introduction
22
- LoRA-Flow provides an efficient way to fuse different LoRA modules which can outperform existing methods significantly. The following picture shows our proposed method, we use layer-wise fusion gates to facilitate dynamic LoRA fusion, which project input hidden states of each layer into fusion weights. For more details, please refer to our paper.
23
  ![1.jpg](https://cdn-uploads.huggingface.co/production/uploads/64d99f6cd7e30889c6c477b4/ifiu1FTHilrmUkD4FKkgV.jpeg)
24
- # Training Details
25
- ## LoRA modules Training
26
- For language LoRA modules: we use the 52K training samples respectively which from [Okapi](https://aclanthology.org/2023.emnlp-demo.28).
27
 
28
- For math LoRA module: the training data for English math LoRA is constructed by [Metamath](https://arxiv.org/abs/2309.12284), which is comprised of 395K mathematical problems in English.
29
 
30
- For code LoRA module: we train the English code LoRA with the Magicoder dataset [Magicoder](https://arxiv.org/abs/2312.02120), which consists of 186K code generation problems in English.
31
 
32
- ## Gate Training
33
- We use gates to fuse different LoRA modules. We employ few-shot training and have released our training data for further details please refer to our GitHub.
34
 
35
- # Result
36
- We have released the result for LoRAs and LoRA-Flow
37
 
38
  | **Method** | | **MGSM (Math)** | | | | **HumanEval (Code)** | | | |
39
  |-----------------------|-------|-------------------------------|---------|---------|---------|----------------------------------|---------|---------|---------|
@@ -47,7 +47,6 @@ We have released the result for LoRAs and LoRA-Flow
47
 
48
 
49
  # Citation
50
- if you find our repo is helpful, please cite the following
51
  ```bibtex
52
  @inproceedings{wang-etal-2024-lora-flow,
53
  title = "LoRA-Flow: Dynamic LoRA Fusion for Large Language Models in Generative Tasks",
@@ -58,9 +57,6 @@ if you find our repo is helpful, please cite the following
58
  Chen, Yun and
59
  Liu, Zhiyuan and
60
  Sun, Maosong",
61
- editor = "Ku, Lun-Wei and
62
- Martins, Andre and
63
- Srikumar, Vivek",
64
  booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
65
  month = aug,
66
  year = "2024",
@@ -68,8 +64,7 @@ if you find our repo is helpful, please cite the following
68
  publisher = "Association for Computational Linguistics",
69
  url = "https://aclanthology.org/2024.acl-long.695",
70
  doi = "10.18653/v1/2024.acl-long.695",
71
- pages = "12871--12882",
72
- abstract = "LoRA employs lightweight modules to customize large language models (LLMs) for each downstream task or domain, where different learned additional modules represent diverse skills. Combining existing LoRAs to address new tasks can enhance the reusability of learned LoRAs, particularly beneficial for tasks with limited annotated data. Most prior works on LoRA combination primarily rely on task-level weights for each involved LoRA, making different examples and tokens share the same LoRA weights. However, in generative tasks, different tokens may necessitate diverse skills to manage. Taking the Chinese math task as an example, understanding the problem description may depend more on the Chinese LoRA, while the calculation part may rely more on the math LoRA. To this end, we propose LoRA-Flow, which utilizes dynamic weights to adjust the impact of different LoRAs. The weights at each step are determined by a fusion gate with extremely few parameters, which can be learned with only 200 training examples. Experiments across six generative tasks demonstrate that our method consistently outperforms baselines with task-level fusion weights. This underscores the necessity of introducing dynamic fusion weights for LoRA combination.",
73
  }
74
  ```
75
  <!-- [LoRA-Flow: Dynamic LoRA Fusion for Large Language Models in Generative Tasks](https://aclanthology.org/2024.acl-long.695) -->
 
19
  # Summary
20
  > In this repo, we release LoRA modules and the gate of 7B models trained in our paper in HuggingFace format.
21
  # Introduction
22
+ LoRA-Flow provides an efficient way to fuse different LoRA modules. The following picture shows our proposed method, we use layer-wise fusion gates to facilitate dynamic LoRA fusion, which project input hidden states of each layer into fusion weights. For more details, please refer to our paper.
23
  ![1.jpg](https://cdn-uploads.huggingface.co/production/uploads/64d99f6cd7e30889c6c477b4/ifiu1FTHilrmUkD4FKkgV.jpeg)
24
+ # Training Data
25
+ ## Data used for LoRA modules
26
+ For the language LoRA modules: we use the 52K training samples from [Okapi](https://aclanthology.org/2023.emnlp-demo.28) for each language, respectively.
27
 
28
+ For the math LoRA module: we use [Metamath](https://arxiv.org/abs/2309.12284) that is comprised of 395K mathematical problems and the corresponding solutions in English.
29
 
30
+ For the code LoRA module: we use the Magicoder dataset [Magicoder](https://arxiv.org/abs/2312.02120), which consists of 186K code generation problems and the corresponding solutions in English.
31
 
32
+ ## Data used for gates
33
+ We use gates to fuse different LoRA modules. We employ few-shot training and have released our training data. For more details, please refer to our GitHub.
34
 
35
+ # Results
36
+ We have released the results for LoRAs and LoRA-Flow
37
 
38
  | **Method** | | **MGSM (Math)** | | | | **HumanEval (Code)** | | | |
39
  |-----------------------|-------|-------------------------------|---------|---------|---------|----------------------------------|---------|---------|---------|
 
47
 
48
 
49
  # Citation
 
50
  ```bibtex
51
  @inproceedings{wang-etal-2024-lora-flow,
52
  title = "LoRA-Flow: Dynamic LoRA Fusion for Large Language Models in Generative Tasks",
 
57
  Chen, Yun and
58
  Liu, Zhiyuan and
59
  Sun, Maosong",
 
 
 
60
  booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
61
  month = aug,
62
  year = "2024",
 
64
  publisher = "Association for Computational Linguistics",
65
  url = "https://aclanthology.org/2024.acl-long.695",
66
  doi = "10.18653/v1/2024.acl-long.695",
67
+ pages = "12871--12882"
 
68
  }
69
  ```
70
  <!-- [LoRA-Flow: Dynamic LoRA Fusion for Large Language Models in Generative Tasks](https://aclanthology.org/2024.acl-long.695) -->