BoscoTheDog commited on
Commit
abbabc6
1 Parent(s): 5b6894b

Upload 16 files

Browse files
.gitattributes CHANGED
@@ -34,3 +34,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
37
+ bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
38
+ bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
39
+ bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
41
+ bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
44
+ bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
45
+ bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6cfcf9db1cc45f14c7db74856e97251ed98338c8366c3cd891bbd9d338bbe3f
3
+ size 136873984
bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44b5cf45c53448183f74725bc601a63473db506343a015687a2471ed922dcc71
3
+ size 184733696
bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1b6366513a443a7875784db81788a3bb0878826775cb10cc86c2adefc7d534d
3
+ size 171102208
bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b039aadc918d722018104f6c2dcdfca97c796fb5b1d53f3c13943d6104bea5cb
3
+ size 155537408
bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12bb45f7ec46727849544e0f6d680ee760b786c3cba60ea313b3e0141dc05237
3
+ size 204716032
bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9557d6557c76ff210b9a6a7b3a9d3236a17b8aeffdb03b8cb56458a530433b3e
3
+ size 196032512
bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c33409ff79b55dc0de5090c7158f5f989360fb72870705375ff3b9fadbbf9f1
3
+ size 237123584
bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdd2709ed80314ccc64e8838be48f508d4e1f7266dab92efb08f7bd50fc3b894
3
+ size 232110080
bitnet-mistral.0.2-330m-v0.2-grokfast-v2.9.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2d467d2b994d709f3d396f0bfd5ba3ef1de0b3abe4e3bcf99ac42c560709eb6
3
+ size 271556608
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "liminerity/Bitnet-Mistral.0.2-330m-v0.2-grokfast-v3",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 1024,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 1024,
13
+ "max_position_embeddings": 4096,
14
+ "model_type": "mistral",
15
+ "num_attention_heads": 16,
16
+ "num_hidden_layers": 36,
17
+ "num_key_value_heads": 16,
18
+ "rms_norm_eps": 1e-05,
19
+ "rope_theta": 1000000.0,
20
+ "sliding_window": 4096,
21
+ "tie_word_embeddings": false,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.41.2",
24
+ "use_cache": true,
25
+ "vocab_size": 32000
26
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90640e0d759643f9a5dd32af9136a0da8cb439c9873492e7838046d43c6e3496
3
+ size 1319444792
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "legacy": true,
35
+ "model_max_length": 1000000000000000019884624838656,
36
+ "pad_token": "</s>",
37
+ "sp_model_kwargs": {},
38
+ "spaces_between_special_tokens": false,
39
+ "tokenizer_class": "LlamaTokenizer",
40
+ "unk_token": "<unk>",
41
+ "use_default_system_prompt": false
42
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ca4371762b1c47e2464638bc76ecf4a30b8df2aa1a9866a221278231c149d36
3
+ size 5240