Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ language:
|
|
10 |
|
11 |
Original model [elyza/ELYZA-japanese-Llama-2-7b-instruct](https://huggingface.co/elyza/ELYZA-japanese-Llama-2-7b-instruct) which is based on Meta's "Llama 2" and has undergone additional pre-training in Japanese instruction.
|
12 |
|
13 |
-
This model is a AWQ quantized(miniaturized to 3.89GB) version of the original model(13.48GB).
|
14 |
|
15 |
## Model Details
|
16 |
|
@@ -26,10 +26,12 @@ But this model has better ability to follow instructions than the previous [GPTQ
|
|
26 |
|
27 |
## Sample Script
|
28 |
|
|
|
|
|
29 |
[AWQ version Colab sample A100 only](https://github.com/webbigdata-jp/python_sample/blob/main/ELYZA_japanese_Llama_2_7b_instruct_AWQ_sample.ipynb)
|
30 |
|
31 |
|
32 |
-
|
33 |
|
34 |
install Library.
|
35 |
```
|
@@ -95,7 +97,7 @@ Output
|
|
95 |
γ΅γͺγ£γγΌγ―γͺγ©γγ―γγ«θΉζ©γεΎγ«γγγ
|
96 |
```
|
97 |
|
98 |
-
|
99 |
|
100 |
This model is based on the work of the following people:
|
101 |
|
@@ -129,10 +131,11 @@ This model is based on the work of the following people:
|
|
129 |
```
|
130 |
|
131 |
|
132 |
-
|
133 |
- **This Quantization work was done by :** [webbigdata](https://webbigdata.jp/)
|
134 |
|
135 |
|
136 |
-
|
|
|
137 |
[mit-han-lab/llm-awq](https://github.com/mit-han-lab/llm-awq)
|
138 |
[casper-hansen/AutoAWQ](https://github.com/casper-hansen/AutoAWQ)
|
|
|
10 |
|
11 |
Original model [elyza/ELYZA-japanese-Llama-2-7b-instruct](https://huggingface.co/elyza/ELYZA-japanese-Llama-2-7b-instruct) which is based on Meta's "Llama 2" and has undergone additional pre-training in Japanese instruction.
|
12 |
|
13 |
+
This model is a [AWQ](https://arxiv.org/abs/2306.00978) quantized(miniaturized to 3.89GB) version of the original model(13.48GB).
|
14 |
|
15 |
## Model Details
|
16 |
|
|
|
26 |
|
27 |
## Sample Script
|
28 |
|
29 |
+
### Colab
|
30 |
+
|
31 |
[AWQ version Colab sample A100 only](https://github.com/webbigdata-jp/python_sample/blob/main/ELYZA_japanese_Llama_2_7b_instruct_AWQ_sample.ipynb)
|
32 |
|
33 |
|
34 |
+
### local PC
|
35 |
|
36 |
install Library.
|
37 |
```
|
|
|
97 |
γ΅γͺγ£γγΌγ―γͺγ©γγ―γγ«θΉζ©γεΎγ«γγγ
|
98 |
```
|
99 |
|
100 |
+
## Citations
|
101 |
|
102 |
This model is based on the work of the following people:
|
103 |
|
|
|
131 |
```
|
132 |
|
133 |
|
134 |
+
## about this work
|
135 |
- **This Quantization work was done by :** [webbigdata](https://webbigdata.jp/)
|
136 |
|
137 |
|
138 |
+
## See also
|
139 |
+
[AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration](https://arxiv.org/abs/2306.00978)
|
140 |
[mit-han-lab/llm-awq](https://github.com/mit-han-lab/llm-awq)
|
141 |
[casper-hansen/AutoAWQ](https://github.com/casper-hansen/AutoAWQ)
|