Update README.md
Browse files
README.md
CHANGED
@@ -5,11 +5,11 @@ language:
|
|
5 |
---
|
6 |
# Leia-Swallow-7B
|
7 |
|
8 |
-
LEIA is a training technique for autoregressive LLMs
|
9 |
This model is constructed by applying LEIA to Swallow, a Japanese-English bilingual LLM based on LLaMA 2.
|
10 |
-
The model achieves
|
11 |
|
12 |
-
Please refer to our paper or blog post (in Japanese) for further technical details
|
13 |
|
14 |
- [LEIA: Facilitating Cross-Lingual Knowledge Transfer in Language Models with Entity-based Data Augmentation](https://arxiv.org/abs/2402.11485) (arxiv.org)
|
15 |
- [LEIA: 言語間転移学習でLLMを賢くする新しい方法](#) (zenn.dev)
|
@@ -38,5 +38,5 @@ For further details of this experiment, please refer to [our paper](https://arxi
|
|
38 |
|
39 |
## Contributors
|
40 |
|
41 |
-
Ikuya Yamada (Studio Ousia, RIKEN)
|
42 |
-
Ryokan Ri (LY Corporation, SB Intuitions)
|
|
|
5 |
---
|
6 |
# Leia-Swallow-7B
|
7 |
|
8 |
+
LEIA is a training technique for autoregressive LLMs that effectively improves their performance in languages other than English by enhancing cross-lingual knowledge transfer from English to a target language.
|
9 |
This model is constructed by applying LEIA to Swallow, a Japanese-English bilingual LLM based on LLaMA 2.
|
10 |
+
The model achieves enhanced performance on six Japanese question-answering benchmarks, as reported below.
|
11 |
|
12 |
+
Please refer to our paper or blog post (in Japanese) for further technical details.
|
13 |
|
14 |
- [LEIA: Facilitating Cross-Lingual Knowledge Transfer in Language Models with Entity-based Data Augmentation](https://arxiv.org/abs/2402.11485) (arxiv.org)
|
15 |
- [LEIA: 言語間転移学習でLLMを賢くする新しい方法](#) (zenn.dev)
|
|
|
38 |
|
39 |
## Contributors
|
40 |
|
41 |
+
- Ikuya Yamada (Studio Ousia, RIKEN)
|
42 |
+
- Ryokan Ri (LY Corporation, SB Intuitions)
|