DopeorNope
commited on
Commit
•
90a961e
1
Parent(s):
831fc99
Upload README.md
Browse files
README.md
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
- ko
|
5 |
+
datasets:
|
6 |
+
- DopeorNope/combined
|
7 |
+
library_name: transformers
|
8 |
+
pipeline_tag: text-generation
|
9 |
+
license: cc-by-nc-sa-4.0
|
10 |
+
---
|
11 |
+
**(주)미디어그룹사람과숲과 (주)마커의 LLM 연구 컨소시엄에서 개발된 모델입니다**
|
12 |
+
**The license is `cc-by-nc-sa-4.0`.**
|
13 |
+
|
14 |
+
# **COLA3-7B : Lamm2 7B 베이스 모델을 IA3방식으로 Fine tuning한 모델**
|
15 |
+
|
16 |
+
** IA3방식에 대한 디테일 정보: [K(G)OAT](https://github.com/Marker-Inc-Korea/K-G-OAT)**
|
17 |
+
|
18 |
+
## Model Details
|
19 |
+
|
20 |
+
**Model Developers** Seungyoo-Lee (DopeorNope)
|
21 |
+
|
22 |
+
**Input** Models input text only.
|
23 |
+
|
24 |
+
**Output** Models generate text only.
|
25 |
+
|
26 |
+
**Model Architecture**
|
27 |
+
|
28 |
+
KO-Platypus2-7B-ex is an auto-regressive language model based on the LLaMA2 transformer architecture.
|
29 |
+
|
30 |
+
**Base Model**
|
31 |
+
[kyujinpy/KO-Platypus2-7B-ex](https://huggingface.co/kyujinpy/KO-Platypus2-7B-ex)
|
32 |
+
|
33 |
+
**Training Dataset**
|
34 |
+
|
35 |
+
[Eng_Kor_COT_combined](https://huggingface.co/datasets/DopeorNope/Eng_Kor_COT_combined) was used for finetuning.
|
36 |
+
|
37 |
+
|
38 |
+
I used A5000 GPU 24GB x2 desktop for training.
|
39 |
+
|
40 |
+
|
41 |
+
|
42 |
+
### Limitations and bias
|
43 |
+
|
44 |
+
Llama 2 and fine-tuned variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Llama 2 and any fine-tuned varient's potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of Llama 2 variants, developers should perform safety testing and tuning tailored to their specific applications of the model.
|
45 |
+
|
46 |
+
Please see the Responsible Use Guide available at https://ai.meta.com/llama/responsible-use-guide/
|
47 |
+
|
48 |
+
### Citations
|
49 |
+
```bibtex
|
50 |
+
@article{platypus2023,
|
51 |
+
title={Platypus: Quick, Cheap, and Powerful Refinement of LLMs},
|
52 |
+
author={Ariel N. Lee and Cole J. Hunter and Nataniel Ruiz},
|
53 |
+
booktitle={arXiv preprint arxiv:2308.07317},
|
54 |
+
year={2023}
|
55 |
+
}
|
56 |
+
```
|
57 |
+
```bibtex
|
58 |
+
@misc{touvron2023llama,
|
59 |
+
title={Llama 2: Open Foundation and Fine-Tuned Chat Models},
|
60 |
+
author={Hugo Touvron and Louis Martin and Kevin Stone and Peter Albert and Amjad Almahairi and Yasmine Babaei and Nikolay Bashlykov year={2023},
|
61 |
+
eprint={2307.09288},
|
62 |
+
archivePrefix={arXiv},
|
63 |
+
}
|
64 |
+
```
|
65 |
+
```bibtex
|
66 |
+
@inproceedings{
|
67 |
+
hu2022lora,
|
68 |
+
title={Lo{RA}: Low-Rank Adaptation of Large Language Models},
|
69 |
+
author={Edward J Hu and Yelong Shen and Phillip Wallis and Zeyuan Allen-Zhu and Yuanzhi Li and Shean Wang and Lu Wang and Weizhu Chen},
|
70 |
+
booktitle={International Conference on Learning Representations},
|
71 |
+
year={2022},
|
72 |
+
url={https://openreview.net/forum?id=nZeVKeeFYf9}
|
73 |
+
}
|
74 |
+
```
|