Upload folder using huggingface_hub (#1)
Browse files- acc789cf2ec45a3b4238f63bf644a3cf7133af0deb1e9f9b5285855b381c9cac (92bee968e4d4a8baede89171016f2922bf38cb3b)
- cef3134eb6b0e19e7675a9ab7d3bca7f53ab4079437aece2848dba3a60cf2fa8 (7ae221777ae5e3bd06d7406c95ef512c419c89bc)
- a1823acc47ddb9447ec1689c14194703dd9785355bbbaab51a8cf02c18e54484 (d858ee9061b9700854994709785f208795855616)
- a19ae5aac19222fa6ff7ef7c3b31fc6ab5ea40e1b8cbfa896fd5e26f2f371600 (215e09c51ba38a463e6a457de2558a557c634ea0)
- 0a0c0f0c815824378dce02a9cfddf5d45a82383354ec4954970688c1e594f6c1 (73736a02f3a1c5abc84800e796bc69318d665c3e)
- 6e492f439d227448bce5c6e08f5d1303e5ef9ac0b4c5d8ba3534852f86bfd53b (88d28e67fbf67d3eea14319894a3dfcee31250de)
- 479d2523ed6a857b2398817881743a755a33c67a099e54b0e5567f11c5dfc129 (ac1534c8d48b7ca715e5cc1f62d9ed43ce3c4786)
- 1a8fcb57781e7284f08d68f4a96805cefedc8da5f7568419aad1d87f8d911b6a (981649ebe9e71cb849b0c8c3d9b52fb3eb0dec51)
- 2e0d3bdda95109e9b01efbdce2a5e1285a9e9362afad0489f54be135412ed9d9 (ad5cbb2a757511a0b258127204278e3e2c3f1a4d)
- dc8a2a106533b6e36536c9a52a5b6a72c206e75f77de402aecf27351c8dc8e2a (a66cc17a11fca339a966b4ca230a5847925cd695)
- 7e72910f0bdf7172d91f85559798ce54677e97b29735043b2002c28367207b1a (27b08be5582a6d94432c41870dda57ba6fdeaa52)
- 4f3ebcac0c740918e29355bcb9cf81299033636407b6bdfb4c7fae9bdc2ebb29 (833aeb125170466cab7184f337bf5f4b6a5b5e25)
- ce2fdbb2e652915d17cfbaeef9b9faa865b343be61952c0ee29dde94bc14c572 (915b35c254a3e33cda41b8dca857f229bd10aa26)
- a3956a11d4d8216163bb2e8ae2740c7293461e56293fb3f3f803beaa2a9d80a7 (c77e6f60c259913ead5366d0eed60abc48a059e7)
- 46b466877d14fb2c7f90ec28aea889c1dcffdd6f6fdf53c10a4af001af6404cd (e611733b1f6580807ce5694141a5e8a280efb5ad)
- added_tokens.json +5 -0
- config.json +56 -0
- generation_config.json +8 -0
- model-00001-of-00015.safetensors +3 -0
- model-00002-of-00015.safetensors +3 -0
- model-00003-of-00015.safetensors +3 -0
- model-00004-of-00015.safetensors +3 -0
- model-00005-of-00015.safetensors +3 -0
- model-00006-of-00015.safetensors +3 -0
- model-00007-of-00015.safetensors +3 -0
- model-00008-of-00015.safetensors +3 -0
- model-00009-of-00015.safetensors +3 -0
- model-00010-of-00015.safetensors +3 -0
- model-00011-of-00015.safetensors +3 -0
- model-00012-of-00015.safetensors +3 -0
- model-00013-of-00015.safetensors +3 -0
- model-00014-of-00015.safetensors +3 -0
- model-00015-of-00015.safetensors +3 -0
- model.safetensors.index.json +0 -0
- special_tokens_map.json +34 -0
- tokenizer_config.json +60 -0
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"<|im_end|>": 100279,
|
3 |
+
"<|im_start|>": 100278,
|
4 |
+
"<|pad|>": 100277
|
5 |
+
}
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "SinclairSchneider/dbrx-instruct-quantization-fixed",
|
3 |
+
"architectures": [
|
4 |
+
"DbrxForCausalLM"
|
5 |
+
],
|
6 |
+
"attn_config": {
|
7 |
+
"clip_qkv": 8,
|
8 |
+
"kv_n_heads": 8,
|
9 |
+
"model_type": "",
|
10 |
+
"rope_theta": 500000
|
11 |
+
},
|
12 |
+
"auto_map": {
|
13 |
+
"AutoConfig": "SinclairSchneider/dbrx-instruct-quantization-fixed--configuration_dbrx.DbrxConfig",
|
14 |
+
"AutoModelForCausalLM": "SinclairSchneider/dbrx-instruct-quantization-fixed--modeling_dbrx.DbrxForCausalLM"
|
15 |
+
},
|
16 |
+
"d_model": 6144,
|
17 |
+
"emb_pdrop": 0.0,
|
18 |
+
"ffn_config": {
|
19 |
+
"ffn_hidden_size": 10752,
|
20 |
+
"model_type": "",
|
21 |
+
"moe_jitter_eps": 0,
|
22 |
+
"moe_loss_weight": 0.05,
|
23 |
+
"moe_num_experts": 16,
|
24 |
+
"moe_top_k": 4
|
25 |
+
},
|
26 |
+
"initializer_range": 0.02,
|
27 |
+
"max_seq_len": 32768,
|
28 |
+
"model_type": "dbrx",
|
29 |
+
"n_heads": 48,
|
30 |
+
"n_layers": 40,
|
31 |
+
"output_router_logits": false,
|
32 |
+
"quantization_config": {
|
33 |
+
"_load_in_4bit": true,
|
34 |
+
"_load_in_8bit": false,
|
35 |
+
"bnb_4bit_compute_dtype": "bfloat16",
|
36 |
+
"bnb_4bit_quant_storage": "uint8",
|
37 |
+
"bnb_4bit_quant_type": "fp4",
|
38 |
+
"bnb_4bit_use_double_quant": true,
|
39 |
+
"llm_int8_enable_fp32_cpu_offload": false,
|
40 |
+
"llm_int8_has_fp16_weight": false,
|
41 |
+
"llm_int8_skip_modules": [
|
42 |
+
"lm_head"
|
43 |
+
],
|
44 |
+
"llm_int8_threshold": 6.0,
|
45 |
+
"load_in_4bit": true,
|
46 |
+
"load_in_8bit": false,
|
47 |
+
"quant_method": "bitsandbytes"
|
48 |
+
},
|
49 |
+
"resid_pdrop": 0.0,
|
50 |
+
"router_aux_loss_coef": 0.05,
|
51 |
+
"tie_word_embeddings": false,
|
52 |
+
"torch_dtype": "bfloat16",
|
53 |
+
"transformers_version": "4.39.2",
|
54 |
+
"use_cache": true,
|
55 |
+
"vocab_size": 100352
|
56 |
+
}
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"eos_token_id": [
|
4 |
+
100257,
|
5 |
+
100279
|
6 |
+
],
|
7 |
+
"transformers_version": "4.39.2"
|
8 |
+
}
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39a7d2e79067a4038e3d3e2e868fb4c637a79e1e12feb2780cc93c9b07b48900
|
3 |
+
size 4982218709
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6caefabcbef424f5c164bba37f24f4c9a45ccd5ba27772956026c69fce1f1028
|
3 |
+
size 4975991632
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b394b566d55a32eb8796ba3c4ee9fed687ca8fab8eb0c7e3b0759fdb285f190b
|
3 |
+
size 4975991636
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:726f672000ccd78e4552920e703363efcb4e857086fb357d41df25e52e0b808f
|
3 |
+
size 4975991957
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba3347d50ac5f64e8ee58f8145d0e0ec534e320569d8c44320ff07b4ef3452eb
|
3 |
+
size 4975992536
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:749045e516bd5673583de3a8da1aa2a5580c4fedf369e84796b97104f62390b7
|
3 |
+
size 4975992539
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4b7e2a409b2e4586b8f07dd381db414634196f7ee167bea1b0f2967659a980fa
|
3 |
+
size 4998646326
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a2531b0fa4f0a13f2fee54ceac1a68321cb3afd8cf53f5c1cabad4dd48655ec7
|
3 |
+
size 4975980133
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2b4dfcd4fbf2358f2020f28536d38040e6cf55bf6d93f9871112e9f9609ca39e
|
3 |
+
size 4975992531
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:69e875475cae60d267509f02547c87cb27665c12d39e3677664cd298ff54068f
|
3 |
+
size 4975992531
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba2f01c3c5e6732fd200e7f9d04855eeec00e34708cd923eecd0f93dd379100c
|
3 |
+
size 4975992531
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c260cb7890346ed6364486b826863e3e99d45dcee9b32a0751f51e24f2c76b9a
|
3 |
+
size 4975992528
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a09f7c97c3e9c9f2aaa8cf41d8c7d8a2c0cb13d26539e190b8fc9564970f9243
|
3 |
+
size 4975992516
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cbb846a1be7dd94c54cc7ee72cfd1ccd88d5a8ae95f3b849ded8beda2ec0ec6d
|
3 |
+
size 3771747555
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0c09710cf02b44be76007e8ddc836d335f3e5f008bf247431d55c4afbc0a3b17
|
3 |
+
size 1233125504
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"<|im_start|>",
|
4 |
+
"<|im_end|>"
|
5 |
+
],
|
6 |
+
"bos_token": {
|
7 |
+
"content": "<|endoftext|>",
|
8 |
+
"lstrip": false,
|
9 |
+
"normalized": false,
|
10 |
+
"rstrip": false,
|
11 |
+
"single_word": false
|
12 |
+
},
|
13 |
+
"eos_token": {
|
14 |
+
"content": "<|endoftext|>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false
|
19 |
+
},
|
20 |
+
"pad_token": {
|
21 |
+
"content": "<|pad|>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": false,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false
|
26 |
+
},
|
27 |
+
"unk_token": {
|
28 |
+
"content": "<|endoftext|>",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false
|
33 |
+
}
|
34 |
+
}
|
@@ -0,0 +1,60 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": false,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"add_prefix_space": false,
|
5 |
+
"added_tokens_decoder": {
|
6 |
+
"100257": {
|
7 |
+
"content": "<|endoftext|>",
|
8 |
+
"lstrip": false,
|
9 |
+
"normalized": false,
|
10 |
+
"rstrip": false,
|
11 |
+
"single_word": false,
|
12 |
+
"special": true
|
13 |
+
},
|
14 |
+
"100277": {
|
15 |
+
"content": "<|pad|>",
|
16 |
+
"lstrip": false,
|
17 |
+
"normalized": false,
|
18 |
+
"rstrip": false,
|
19 |
+
"single_word": false,
|
20 |
+
"special": true
|
21 |
+
},
|
22 |
+
"100278": {
|
23 |
+
"content": "<|im_start|>",
|
24 |
+
"lstrip": false,
|
25 |
+
"normalized": false,
|
26 |
+
"rstrip": false,
|
27 |
+
"single_word": false,
|
28 |
+
"special": true
|
29 |
+
},
|
30 |
+
"100279": {
|
31 |
+
"content": "<|im_end|>",
|
32 |
+
"lstrip": false,
|
33 |
+
"normalized": false,
|
34 |
+
"rstrip": false,
|
35 |
+
"single_word": false,
|
36 |
+
"special": true
|
37 |
+
}
|
38 |
+
},
|
39 |
+
"additional_special_tokens": [
|
40 |
+
"<|im_start|>",
|
41 |
+
"<|im_end|>"
|
42 |
+
],
|
43 |
+
"auto_map": {
|
44 |
+
"AutoTokenizer": [
|
45 |
+
"SinclairSchneider/dbrx-instruct-quantization-fixed--tiktoken.TiktokenTokenizerWrapper",
|
46 |
+
null
|
47 |
+
]
|
48 |
+
},
|
49 |
+
"bos_token": "<|endoftext|>",
|
50 |
+
"clean_up_tokenization_spaces": true,
|
51 |
+
"encoding_name": null,
|
52 |
+
"eos_token": "<|endoftext|>",
|
53 |
+
"errors": "replace",
|
54 |
+
"model_max_length": 1000000000000000019884624838656,
|
55 |
+
"model_name": "gpt-4",
|
56 |
+
"pad_token": "<|pad|>",
|
57 |
+
"tokenizer_class": "TiktokenTokenizerWrapper",
|
58 |
+
"unk_token": "<|endoftext|>",
|
59 |
+
"use_default_system_prompt": true
|
60 |
+
}
|