use 8-bit weights
Browse files- __pycache__/handler.cpython-310.pyc +0 -0
- config.json +7 -7
- generation_config.json +1 -1
- model.safetensors +2 -2
__pycache__/handler.cpython-310.pyc
ADDED
Binary file (1.58 kB). View file
|
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "bart-large-cnn",
|
3 |
"_num_labels": 3,
|
4 |
"activation_dropout": 0.0,
|
5 |
"activation_function": "gelu",
|
@@ -53,9 +53,9 @@
|
|
53 |
"pad_token_id": 1,
|
54 |
"prefix": " ",
|
55 |
"quantization_config": {
|
56 |
-
"_load_in_4bit":
|
57 |
-
"_load_in_8bit":
|
58 |
-
"bnb_4bit_compute_dtype": "
|
59 |
"bnb_4bit_quant_storage": "uint8",
|
60 |
"bnb_4bit_quant_type": "fp4",
|
61 |
"bnb_4bit_use_double_quant": false,
|
@@ -63,8 +63,8 @@
|
|
63 |
"llm_int8_has_fp16_weight": false,
|
64 |
"llm_int8_skip_modules": null,
|
65 |
"llm_int8_threshold": 6.0,
|
66 |
-
"load_in_4bit":
|
67 |
-
"load_in_8bit":
|
68 |
"quant_method": "bitsandbytes"
|
69 |
},
|
70 |
"scale_embedding": false,
|
@@ -79,7 +79,7 @@
|
|
79 |
}
|
80 |
},
|
81 |
"torch_dtype": "float16",
|
82 |
-
"transformers_version": "4.
|
83 |
"use_cache": true,
|
84 |
"vocab_size": 50264
|
85 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "booksouls/bart-large-cnn",
|
3 |
"_num_labels": 3,
|
4 |
"activation_dropout": 0.0,
|
5 |
"activation_function": "gelu",
|
|
|
53 |
"pad_token_id": 1,
|
54 |
"prefix": " ",
|
55 |
"quantization_config": {
|
56 |
+
"_load_in_4bit": false,
|
57 |
+
"_load_in_8bit": true,
|
58 |
+
"bnb_4bit_compute_dtype": "float32",
|
59 |
"bnb_4bit_quant_storage": "uint8",
|
60 |
"bnb_4bit_quant_type": "fp4",
|
61 |
"bnb_4bit_use_double_quant": false,
|
|
|
63 |
"llm_int8_has_fp16_weight": false,
|
64 |
"llm_int8_skip_modules": null,
|
65 |
"llm_int8_threshold": 6.0,
|
66 |
+
"load_in_4bit": false,
|
67 |
+
"load_in_8bit": true,
|
68 |
"quant_method": "bitsandbytes"
|
69 |
},
|
70 |
"scale_embedding": false,
|
|
|
79 |
}
|
80 |
},
|
81 |
"torch_dtype": "float16",
|
82 |
+
"transformers_version": "4.40.2",
|
83 |
"use_cache": true,
|
84 |
"vocab_size": 50264
|
85 |
}
|
generation_config.json
CHANGED
@@ -12,5 +12,5 @@
|
|
12 |
"no_repeat_ngram_size": 3,
|
13 |
"num_beams": 4,
|
14 |
"pad_token_id": 1,
|
15 |
-
"transformers_version": "4.
|
16 |
}
|
|
|
12 |
"no_repeat_ngram_size": 3,
|
13 |
"num_beams": 4,
|
14 |
"pad_token_id": 1,
|
15 |
+
"transformers_version": "4.40.2"
|
16 |
}
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b840ef94392ee418f91f8bb77983b81f1c55bb9a0599f856bb5416bd5002c222
|
3 |
+
size 461542680
|