Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,40 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- hotpotqa/hotpot_qa
|
5 |
+
metrics:
|
6 |
+
- f1
|
7 |
+
---
|
8 |
+
|
9 |
+
# LLoCO: Learning Long Contexts Offline
|
10 |
+
[**Paper**](https://arxiv.org/abs/2404.07979) | [**Code**](https://github.com/jeffreysijuntan/lloco)
|
11 |
+
|
12 |
+
Lloco-7b-hqa is the LoRA adaptor checkpoint finetuned from [AutoCompressor-Llama-2-7b-6k](https://huggingface.co/princeton-nlp/AutoCompressor-Llama-2-7b-6k/) and [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf)
|
13 |
+
using the **LLoCO** method in [LLoCO: Learning Long Contexts Offline](https://arxiv.org/abs/2404.07979). It is instruction-tuned on the HotpotQA training set.
|
14 |
+
|
15 |
+
**LLoCO** enables LLMs to process long-context efficiently by learning contexts offline through context compression and in-domain parameter-efficient finetuning with LoRA. This approach extends the effective context window of a 4k token LLaMA2-7B model to handle up to 128k tokens, while using
|
16 |
+
30x fewer tokens and achieving up to 7.62x inference speed-up.
|
17 |
+
|
18 |
+
## Released LoRA Checkpoint
|
19 |
+
| Model | LoRA Rank | Dataset | Link |
|
20 |
+
|:----------------|-----------|-------------|--------------------------------------------------------|
|
21 |
+
| Lloco-7b-quality| 8 | QuALITY | [link](https://huggingface.co/xiuyul/Lloco-7b-quality/)|
|
22 |
+
| Lloco-7b-qasper | 8 | Qasper | [link](https://huggingface.co/xiuyul/Lloco-7b-qasper/) |
|
23 |
+
| Lloco-7b-qmsum | 8 | QMSum | [link](https://huggingface.co/xiuyul/Lloco-7b-qmsum/) |
|
24 |
+
| Lloco-7b-nqa | 8 | NarrativeQA | [link](https://huggingface.co/xiuyul/Lloco-7b-nqa/) |
|
25 |
+
| Lloco-7b-hqa | 8 | HotpotQA | [link](https://huggingface.co/xiuyul/Lloco-7b-hqa/) |
|
26 |
+
|
27 |
+
## Citation
|
28 |
+
If you find this project useful, please consider citing:
|
29 |
+
|
30 |
+
```
|
31 |
+
@article{tan2024lloco,
|
32 |
+
title={LLoCO: Learning Long Contexts Offline},
|
33 |
+
author={Tan, Sijun and Li, Xiuyu and Patil, Shishir and Wu, Ziyang and Zhang, Tianjun and Keutzer, Kurt and Gonzalez, Joseph E and Popa, Raluca Ada},
|
34 |
+
journal={arXiv preprint arXiv:2404.07979},
|
35 |
+
year={2024}
|
36 |
+
}
|
37 |
+
```
|
38 |
+
|
39 |
+
## Evaluation
|
40 |
+
Check out [LLoCO: Learning Long Contexts Offline](https://arxiv.org/abs/2404.07979) for evaluation results on various long-context tasks such as long document question answering and summarization.
|