Update README.md
Browse files
README.md
CHANGED
@@ -6,7 +6,7 @@ tags:
|
|
6 |
- gpt-neox
|
7 |
- KoAlpaca
|
8 |
model-index:
|
9 |
-
- name: KoAlpaca-Polyglot-
|
10 |
results: []
|
11 |
language:
|
12 |
- ko
|
@@ -16,11 +16,9 @@ pipeline_tag: text-generation
|
|
16 |
---
|
17 |
|
18 |
|
19 |
-
# KoAlpaca-Polyglot-
|
20 |
|
21 |
-
This model is a fine-tuned version of [EleutherAI/polyglot-ko-
|
22 |
-
|
23 |
-
Detail Codes are available at [KoAlpaca Github Repository](https://github.com/Beomi/KoAlpaca)
|
24 |
|
25 |
|
26 |
## Training procedure
|
@@ -31,8 +29,8 @@ The following hyperparameters were used during training:
|
|
31 |
- learning_rate: 5e-05
|
32 |
- train_batch_size: 1
|
33 |
- seed: 42
|
34 |
-
- distributed_type: multi-GPU (
|
35 |
-
- num_devices:
|
36 |
- gradient_accumulation_steps: 64
|
37 |
- total_train_batch_size: 256
|
38 |
- total_eval_batch_size: 32
|
|
|
6 |
- gpt-neox
|
7 |
- KoAlpaca
|
8 |
model-index:
|
9 |
+
- name: KoAlpaca-Polyglot-5.8B
|
10 |
results: []
|
11 |
language:
|
12 |
- ko
|
|
|
16 |
---
|
17 |
|
18 |
|
19 |
+
# KoAlpaca-Polyglot-5.8B (v1.1b)
|
20 |
|
21 |
+
This model is a fine-tuned version of [EleutherAI/polyglot-ko-5.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b) on a KoAlpaca Dataset v1.1b
|
|
|
|
|
22 |
|
23 |
|
24 |
## Training procedure
|
|
|
29 |
- learning_rate: 5e-05
|
30 |
- train_batch_size: 1
|
31 |
- seed: 42
|
32 |
+
- distributed_type: multi-GPU (A40 40G)
|
33 |
+
- num_devices: 8
|
34 |
- gradient_accumulation_steps: 64
|
35 |
- total_train_batch_size: 256
|
36 |
- total_eval_batch_size: 32
|