Pankaj Mathur
commited on
Commit
•
ad636d6
1
Parent(s):
8889e72
Update README.md
Browse files
README.md
CHANGED
@@ -4,14 +4,20 @@ language:
|
|
4 |
- en
|
5 |
library_name: adapter-transformers
|
6 |
---
|
7 |
-
# alpaca_orca_open_llama: An Open_LLaMA-3B model trained on Alpaca dataset using Orca Research paper approaches
|
8 |
|
9 |
|
10 |
-
# Dataset
|
11 |
|
12 |
-
We train OpenLLaMa-3B model
|
13 |
|
14 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
The training configurations are provided in the table below.
|
17 |
|
@@ -84,4 +90,4 @@ Next Goals:
|
|
84 |
2) Try bigger OpenLLaMA models 7B and 13B
|
85 |
3) Try better GPU for training, couldn't get 8xA100 (40GB), I guess they are in hot demand now.
|
86 |
4) Provide more options for Text generation UI. (may be https://github.com/oobabooga/text-generation-webui)
|
87 |
-
6) Provide 4bit quantized model (may be TheBloke can help here)
|
|
|
4 |
- en
|
5 |
library_name: adapter-transformers
|
6 |
---
|
7 |
+
# alpaca_orca_open_llama: An Open_LLaMA-3B model trained on custom Alpaca dataset using Orca Research paper approaches
|
8 |
|
9 |
|
10 |
+
# Dataset
|
11 |
|
12 |
+
We train OpenLLaMa-3B model on custom explained tuned Alpaca dataset (~52K) created using approaches from [Orca Research Paper](https://arxiv.org/abs/2306.02707).
|
13 |
|
14 |
+
We leverage all of the 15 system instructions provided in [Orca Research Paper](https://arxiv.org/abs/2306.02707) to generate custom Alpaca dataset, in contrast to vanilla instruction tuning approaches used by original [Alpaca research paper](https://crfm.stanford.edu/2023/03/13/alpaca.html).
|
15 |
+
|
16 |
+
This helps student model aka [alpaca_orca_open_llama_3b](psmathur/alpaca_orca_open_llama_3b) to learn ***thought*** process from teacher model, which is ChatGPT (gpt-3.5-turbo-0301 version).
|
17 |
+
|
18 |
+
Please pay attention how the **System** prompt is added before each *instruction* in below example usage.
|
19 |
+
|
20 |
+
# Training
|
21 |
|
22 |
The training configurations are provided in the table below.
|
23 |
|
|
|
90 |
2) Try bigger OpenLLaMA models 7B and 13B
|
91 |
3) Try better GPU for training, couldn't get 8xA100 (40GB), I guess they are in hot demand now.
|
92 |
4) Provide more options for Text generation UI. (may be https://github.com/oobabooga/text-generation-webui)
|
93 |
+
6) Provide 4bit GGML/GPTQ quantized model (may be [TheBloke](https://huggingface.co/TheBloke) can help here)
|