File size: 5,157 Bytes
66fa5a6 9702c2f 5745a62 9702c2f 2a16f77 69ed144 66fa5a6 6065260 66fa5a6 6065260 a7b3c53 1eb10b7 185e8f6 66fa5a6 1029356 66fa5a6 1029356 5295464 5410e75 1029356 66fa5a6 53bea96 fc25979 0334bb4 2c03338 7e191fe 5295464 66fa5a6 25bdaae 65b45a7 25bdaae 3269cc5 65b45a7 66fa5a6 ac6d9f2 cdac047 ac6d9f2 1b1355b ac6d9f2 cdac047 ac6d9f2 66fa5a6 ba49d06 66fa5a6 ed69f62 87b9c95 66fa5a6 87b9c95 66fa5a6 795d67d 66fa5a6 4c445f6 66fa5a6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 |
---
base_model:
- Bllossom/llama-3-Korean-Bllossom-70B
library_name: transformers
tags:
- mergekit
- merge
---
ππ Huggingface Upload Issue
Maximum individual file size is 50.0GB to upload huggingface.
To clear it, Split the file into part_aa, part_ab, part_ac chunks as my "Practical Idea".
After you download this repo on your folder path, command like this.
```
cat part_* > llama-3-korean-bllossom-120b-Q4_K_M.gguf
```
I thought uploading it as a GGUF rather than a simple original file was for your benefit, so I'm uploading it like this even if it takes a bit of trouble.
```
Perhaps this will be the first GGUF model to upload such a large GGUF file of over 50GB to huggingface?
Sometimes we need to use a trick called chunks.
```
Please wait to upload.....
### π°π· About the JayLee "AsianSoul"
```
"A leader who can make you rich π΅ !!!"
"Prove yourself with actual results, not just saying I know more than you!!!"
```
<a href="https://ibb.co/4g2SJVM"><img src="https://i.ibb.co/PzMWt64/Screenshot-2024-05-18-at-11-08-12-PM.png" alt="Screenshot-2024-05-18-at-11-08-12-PM" border="0"></a>
### About this model
This is a 128B model based on [Bllossom/llama-3-Korean-Bllossom-70B](https://huggingface.co/Bllossom/llama-3-Korean-Bllossom-70B)
β I started this Korean 120B model merge while drinking an iced Americano at Starbucks referring to other [Cognitive Computations 120B](https://huggingface.co/cognitivecomputations/MegaDolphin-120b).
If you walk around Starbucks in Seoul, Korea, you will see someone creating a merge and an application based on it.
At that time, please come up to me and say "hello".
"Also, if you want to create the Application project you want and provide me with support, I will create the entire architecture for you whatever it is."
ποΈ I am a person whose goal is to turn the great results created by great genius scientists & groups around the world into profitable ones.
```
My role model is J. Robert Oppenheimer!!!
J. Robert Oppenheimer is highly regarded for his ability to gather and lead a team of brilliant scientists, merging their diverse expertise and efforts towards a common goal.
```
[Learn more about J. Robert Oppenheimer](https://en.wikipedia.org/wiki/J._Robert_Oppenheimer).
I hope this 120B is a helpful model for your future.
```
π Collaboration is always welcome π
π You can't beat these giant corporations & groups alone and you can never become rich.
Now we have to come together.
People who can actually become rich together, let's collaborate with me.!!! πΈ
```
```
About Bllossom/llama-3-Korean-Bllossom-70B
- Full model released in Korean over 100GB by Blossom team
- First in Korean! Expansion of Korean vocabulary to over 30,000 words
- Capable of processing Korean context that is approximately 25% longer than Llama3
- Connecting Korean-English knowledge using the Korean-English Parallel Corpus (pre-study)
- Fine tuning using data produced by linguists considering Korean culture and language
- Reinforcement learning
π°οΈ About asiansoul/llama-3-Korean-Bllossom-120B-GGUF
- Q4_K_M : On a GPU with 68GB / more OR a CPU with 68G / more memory
- More Quantization ones i hope to upload, but your computer won't be able to handle it then. you know what i mean!!
```
### Models Merged
The following models were included in the merge:
* [Bllossom/llama-3-Korean-Bllossom-70B](https://huggingface.co/Bllossom/llama-3-Korean-Bllossom-70B)
### Ollama
Check the information indicated above and run it when your computer is ready.
Otherwise, your computer will freeze.
* Create
```
ollama create Bllossom -f ./Modelfile_Q4_K_M
```
* MODELFILE (you can change this for your preference)
```
FROM ./llama-3-korean-bllossom-120b-Q4_K_M.gguf
TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|>
{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
{{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|>
{{ .Response }}<|eot_id|>"""
SYSTEM """
μΉμ ν μ±λ΄μΌλ‘μ μλλ°©μ μμ²μ μ΅λν μμΈνκ³ μΉμ νκ² λ΅νμ. λͺ¨λ λλ΅μ νκ΅μ΄(Korean)μΌλ‘ λλ΅ν΄μ€.
"""
PARAMETER num_keep 24
PARAMETER temperature 0.6
PARAMETER top_p 0.9
PARAMETER num_predict 2048
PARAMETER stop "<|start_header_id|>"
PARAMETER stop "<|end_header_id|>"
PARAMETER stop "<|eot_id|>"
```
### Configuration
The following YAML configuration was used to produce this model:
```yaml
slices:
- sources:
- layer_range: [0, 20]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [10, 30]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [20, 40]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [30, 50]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [40, 60]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [50, 70]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [60, 80]
model: Bllossom/llama-3-Korean-Bllossom-70B
merge_method: passthrough
dtype: float16
``` |