dahara1 commited on
Commit
5afa541
β€’
1 Parent(s): 8434384

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -5
README.md CHANGED
@@ -10,7 +10,7 @@ language:
10
 
11
  Original model [elyza/ELYZA-japanese-Llama-2-7b-instruct](https://huggingface.co/elyza/ELYZA-japanese-Llama-2-7b-instruct) which is based on Meta's "Llama 2" and has undergone additional pre-training in Japanese instruction.
12
 
13
- This model is a AWQ quantized(miniaturized to 3.89GB) version of the original model(13.48GB).
14
 
15
  ## Model Details
16
 
@@ -26,10 +26,12 @@ But this model has better ability to follow instructions than the previous [GPTQ
26
 
27
  ## Sample Script
28
 
 
 
29
  [AWQ version Colab sample A100 only](https://github.com/webbigdata-jp/python_sample/blob/main/ELYZA_japanese_Llama_2_7b_instruct_AWQ_sample.ipynb)
30
 
31
 
32
- for local PC script.
33
 
34
  install Library.
35
  ```
@@ -95,7 +97,7 @@ Output
95
  ちγͺっしーはγƒͺγƒ©γƒƒγ‚―γƒžγ«θˆΉζ©‹γ‚’εΎŒγ«γ™γ‚‹γ‚ˆ
96
  ```
97
 
98
- ### Citations
99
 
100
  This model is based on the work of the following people:
101
 
@@ -129,10 +131,11 @@ This model is based on the work of the following people:
129
  ```
130
 
131
 
132
- ### about this work
133
  - **This Quantization work was done by :** [webbigdata](https://webbigdata.jp/)
134
 
135
 
136
- ### See also
 
137
  [mit-han-lab/llm-awq](https://github.com/mit-han-lab/llm-awq)
138
  [casper-hansen/AutoAWQ](https://github.com/casper-hansen/AutoAWQ)
 
10
 
11
  Original model [elyza/ELYZA-japanese-Llama-2-7b-instruct](https://huggingface.co/elyza/ELYZA-japanese-Llama-2-7b-instruct) which is based on Meta's "Llama 2" and has undergone additional pre-training in Japanese instruction.
12
 
13
+ This model is a [AWQ](https://arxiv.org/abs/2306.00978) quantized(miniaturized to 3.89GB) version of the original model(13.48GB).
14
 
15
  ## Model Details
16
 
 
26
 
27
  ## Sample Script
28
 
29
+ ### Colab
30
+
31
  [AWQ version Colab sample A100 only](https://github.com/webbigdata-jp/python_sample/blob/main/ELYZA_japanese_Llama_2_7b_instruct_AWQ_sample.ipynb)
32
 
33
 
34
+ ### local PC
35
 
36
  install Library.
37
  ```
 
97
  ちγͺっしーはγƒͺγƒ©γƒƒγ‚―γƒžγ«θˆΉζ©‹γ‚’εΎŒγ«γ™γ‚‹γ‚ˆ
98
  ```
99
 
100
+ ## Citations
101
 
102
  This model is based on the work of the following people:
103
 
 
131
  ```
132
 
133
 
134
+ ## about this work
135
  - **This Quantization work was done by :** [webbigdata](https://webbigdata.jp/)
136
 
137
 
138
+ ## See also
139
+ [AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration](https://arxiv.org/abs/2306.00978)
140
  [mit-han-lab/llm-awq](https://github.com/mit-han-lab/llm-awq)
141
  [casper-hansen/AutoAWQ](https://github.com/casper-hansen/AutoAWQ)