iproskurina commited on
Commit
6ae06d4
1 Parent(s): c7619ba

AutoGPTQ model for mistralai/Mistral-7B-v0.3: 4bits, gr128, desc_act=False

Browse files
Files changed (3) hide show
  1. config.json +13 -0
  2. model.safetensors +2 -2
  3. quantize_config.json +3 -1
config.json CHANGED
@@ -16,6 +16,19 @@
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  "rms_norm_eps": 1e-05,
20
  "rope_theta": 1000000.0,
21
  "sliding_window": null,
 
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
19
+ "quantization_config": {
20
+ "bits": 4,
21
+ "checkpoint_format": "gptq",
22
+ "damp_percent": 0.01,
23
+ "desc_act": false,
24
+ "group_size": 128,
25
+ "model_file_base_name": null,
26
+ "model_name_or_path": null,
27
+ "quant_method": "gptq",
28
+ "static_groups": false,
29
+ "sym": true,
30
+ "true_sequential": true
31
+ },
32
  "rms_norm_eps": 1e-05,
33
  "rope_theta": 1000000.0,
34
  "sliding_window": null,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55e32da462993acadd3128119160a3015dbdc2e618755707c1849929e707b66a
3
- size 4171245128
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55f1c2e111e27c4322a743eeb609e3455e93d8617b0562738d0b678b1822dec0
3
+ size 4168468392
quantize_config.json CHANGED
@@ -7,5 +7,7 @@
7
  "sym": true,
8
  "true_sequential": true,
9
  "model_name_or_path": null,
10
- "model_file_base_name": null
 
 
11
  }
 
7
  "sym": true,
8
  "true_sequential": true,
9
  "model_name_or_path": null,
10
+ "model_file_base_name": null,
11
+ "quant_method": "gptq",
12
+ "checkpoint_format": "gptq"
13
  }