dahara1 commited on
Commit
05f743f
1 Parent(s): 83ca26f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -4,12 +4,15 @@ language:
4
  - ja
5
  ---
6
 
 
 
 
7
  ```
8
  import torch
9
  from transformers import AutoModelForCausalLM, AutoTokenizer, AwqConfig
10
 
11
  model_id = "dahara1/llama3.1-8b-Instruct-awq"
12
- # Perplexity: 9.864517211914062
13
 
14
  quantization_config = AwqConfig(
15
  bits=4,
 
4
  - ja
5
  ---
6
 
7
+ 量子化時に日本語と中国語を多めに追加しているため、[hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4](hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4)より日本語データを使って計測したPerplexityが良い事がわかっています
8
+
9
+
10
  ```
11
  import torch
12
  from transformers import AutoModelForCausalLM, AutoTokenizer, AwqConfig
13
 
14
  model_id = "dahara1/llama3.1-8b-Instruct-awq"
15
+
16
 
17
  quantization_config = AwqConfig(
18
  bits=4,