morriszms commited on
Commit
5502a34
โ€ข
1 Parent(s): adf4399

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Llama-160M-Chat-v1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Llama-160M-Chat-v1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Llama-160M-Chat-v1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Llama-160M-Chat-v1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Llama-160M-Chat-v1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Llama-160M-Chat-v1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Llama-160M-Chat-v1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Llama-160M-Chat-v1-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Llama-160M-Chat-v1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Llama-160M-Chat-v1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Llama-160M-Chat-v1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Llama-160M-Chat-v1-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
Llama-160M-Chat-v1-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5d01e1b59ad9d17998ac24be8e995e715afdaf0582c5c17736659ba40fa9537
3
+ size 70504352
Llama-160M-Chat-v1-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b96e923e811b9fbdc8043489eb42f20a24f86984b6594acfc2f740d449bc8fd6
3
+ size 91137440
Llama-160M-Chat-v1-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28dfcfe05e6930e9f5ba0a36d32ddf515df44861e98d6770e2babc85c04e9e89
3
+ size 85976480
Llama-160M-Chat-v1-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe314a5b9318928d0fb1805bea2b2fa8593df9bf9f436827c3b9f83d64c65c0f
3
+ size 80188832
Llama-160M-Chat-v1-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e80a493e95f764ead26dc575f694a3cfe2bf44bbf5858590e9e44f6948e61056
3
+ size 98493344
Llama-160M-Chat-v1-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e163d0ee23796e393b46ceaedeaa86c2fb487f1f1f36f8b5f301ef2d1aa8b704
3
+ size 103055264
Llama-160M-Chat-v1-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3c5a19044d1dbb9f5d76975b691667fe76a43db230dd0ca0d7bc244d11a6f73
3
+ size 99083168
Llama-160M-Chat-v1-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c8e60c09d7ec610095b2cb6e0cd485d7db26a9c01776016a81b7e6ebd948592
3
+ size 115721120
Llama-160M-Chat-v1-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ed008bacbc518c62c5bc92e092a62b38572d1c864f2c89b9d30ff3b5b6eff60
3
+ size 118071200
Llama-160M-Chat-v1-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a6ea35f6efc3aee9a0434b8d414495995788c8e718654cd9042e19d82caa690
3
+ size 115721120
Llama-160M-Chat-v1-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a61ebd03a8bf40460f86d994016c612e58948734948bc3aedd47188f22776b06
3
+ size 134025632
Llama-160M-Chat-v1-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7a1d3ba79b2851a7da502a14b665bd77c85b703d3ed23592c8f77318905c9ab
3
+ size 173356448
README.md ADDED
@@ -0,0 +1,347 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ tags:
6
+ - text-generation
7
+ - TensorBlock
8
+ - GGUF
9
+ base_model: Felladrin/Llama-160M-Chat-v1
10
+ datasets:
11
+ - ehartford/wizard_vicuna_70k_unfiltered
12
+ - totally-not-an-llm/EverythingLM-data-V3
13
+ - Open-Orca/SlimOrca-Dedup
14
+ - databricks/databricks-dolly-15k
15
+ - THUDM/webglm-qa
16
+ widget:
17
+ - messages:
18
+ - role: system
19
+ content: You are a helpful assistant, who answers with empathy.
20
+ - role: user
21
+ content: Got a question for you!
22
+ - role: assistant
23
+ content: Sure! What's it?
24
+ - role: user
25
+ content: Why do you love cats so much!? ๐Ÿˆ
26
+ - messages:
27
+ - role: system
28
+ content: You are a helpful assistant who answers user's questions with empathy.
29
+ - role: user
30
+ content: Who is Mona Lisa?
31
+ - messages:
32
+ - role: system
33
+ content: You are a helpful assistant who provides concise responses.
34
+ - role: user
35
+ content: Heya!
36
+ - role: assistant
37
+ content: Hi! How may I help you today?
38
+ - role: user
39
+ content: I need to build a simple website. Where should I start learning about
40
+ web development?
41
+ - messages:
42
+ - role: user
43
+ content: Invited some friends to come home today. Give me some ideas for games
44
+ to play with them!
45
+ - messages:
46
+ - role: system
47
+ content: You are a helpful assistant who answers user's questions with details
48
+ and curiosity.
49
+ - role: user
50
+ content: What are some potential applications for quantum computing?
51
+ - messages:
52
+ - role: system
53
+ content: You are a helpful assistant who gives creative responses.
54
+ - role: user
55
+ content: Write the specs of a game about mages in a fantasy world.
56
+ - messages:
57
+ - role: system
58
+ content: You are a helpful assistant who answers user's questions with details.
59
+ - role: user
60
+ content: Tell me about the pros and cons of social media.
61
+ - messages:
62
+ - role: system
63
+ content: You are a helpful assistant who answers user's questions with confidence.
64
+ - role: user
65
+ content: What is a dog?
66
+ - role: assistant
67
+ content: A dog is a four-legged, domesticated animal that is a member of the class
68
+ Mammalia, which includes all mammals. Dogs are known for their loyalty, playfulness,
69
+ and ability to be trained for various tasks. They are also used for hunting,
70
+ herding, and as service animals.
71
+ - role: user
72
+ content: What is the color of an apple?
73
+ inference:
74
+ parameters:
75
+ max_new_tokens: 250
76
+ penalty_alpha: 0.5
77
+ top_k: 4
78
+ repetition_penalty: 1.01
79
+ model-index:
80
+ - name: Llama-160M-Chat-v1
81
+ results:
82
+ - task:
83
+ type: text-generation
84
+ name: Text Generation
85
+ dataset:
86
+ name: AI2 Reasoning Challenge (25-Shot)
87
+ type: ai2_arc
88
+ config: ARC-Challenge
89
+ split: test
90
+ args:
91
+ num_few_shot: 25
92
+ metrics:
93
+ - type: acc_norm
94
+ value: 24.74
95
+ name: normalized accuracy
96
+ source:
97
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
98
+ name: Open LLM Leaderboard
99
+ - task:
100
+ type: text-generation
101
+ name: Text Generation
102
+ dataset:
103
+ name: HellaSwag (10-Shot)
104
+ type: hellaswag
105
+ split: validation
106
+ args:
107
+ num_few_shot: 10
108
+ metrics:
109
+ - type: acc_norm
110
+ value: 35.29
111
+ name: normalized accuracy
112
+ source:
113
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
114
+ name: Open LLM Leaderboard
115
+ - task:
116
+ type: text-generation
117
+ name: Text Generation
118
+ dataset:
119
+ name: MMLU (5-Shot)
120
+ type: cais/mmlu
121
+ config: all
122
+ split: test
123
+ args:
124
+ num_few_shot: 5
125
+ metrics:
126
+ - type: acc
127
+ value: 26.13
128
+ name: accuracy
129
+ source:
130
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
131
+ name: Open LLM Leaderboard
132
+ - task:
133
+ type: text-generation
134
+ name: Text Generation
135
+ dataset:
136
+ name: TruthfulQA (0-shot)
137
+ type: truthful_qa
138
+ config: multiple_choice
139
+ split: validation
140
+ args:
141
+ num_few_shot: 0
142
+ metrics:
143
+ - type: mc2
144
+ value: 44.16
145
+ source:
146
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
147
+ name: Open LLM Leaderboard
148
+ - task:
149
+ type: text-generation
150
+ name: Text Generation
151
+ dataset:
152
+ name: Winogrande (5-shot)
153
+ type: winogrande
154
+ config: winogrande_xl
155
+ split: validation
156
+ args:
157
+ num_few_shot: 5
158
+ metrics:
159
+ - type: acc
160
+ value: 51.3
161
+ name: accuracy
162
+ source:
163
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
164
+ name: Open LLM Leaderboard
165
+ - task:
166
+ type: text-generation
167
+ name: Text Generation
168
+ dataset:
169
+ name: GSM8k (5-shot)
170
+ type: gsm8k
171
+ config: main
172
+ split: test
173
+ args:
174
+ num_few_shot: 5
175
+ metrics:
176
+ - type: acc
177
+ value: 0.0
178
+ name: accuracy
179
+ source:
180
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
181
+ name: Open LLM Leaderboard
182
+ - task:
183
+ type: text-generation
184
+ name: Text Generation
185
+ dataset:
186
+ name: IFEval (0-Shot)
187
+ type: HuggingFaceH4/ifeval
188
+ args:
189
+ num_few_shot: 0
190
+ metrics:
191
+ - type: inst_level_strict_acc and prompt_level_strict_acc
192
+ value: 15.75
193
+ name: strict accuracy
194
+ source:
195
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
196
+ name: Open LLM Leaderboard
197
+ - task:
198
+ type: text-generation
199
+ name: Text Generation
200
+ dataset:
201
+ name: BBH (3-Shot)
202
+ type: BBH
203
+ args:
204
+ num_few_shot: 3
205
+ metrics:
206
+ - type: acc_norm
207
+ value: 3.17
208
+ name: normalized accuracy
209
+ source:
210
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
211
+ name: Open LLM Leaderboard
212
+ - task:
213
+ type: text-generation
214
+ name: Text Generation
215
+ dataset:
216
+ name: MATH Lvl 5 (4-Shot)
217
+ type: hendrycks/competition_math
218
+ args:
219
+ num_few_shot: 4
220
+ metrics:
221
+ - type: exact_match
222
+ value: 0.0
223
+ name: exact match
224
+ source:
225
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
226
+ name: Open LLM Leaderboard
227
+ - task:
228
+ type: text-generation
229
+ name: Text Generation
230
+ dataset:
231
+ name: GPQA (0-shot)
232
+ type: Idavidrein/gpqa
233
+ args:
234
+ num_few_shot: 0
235
+ metrics:
236
+ - type: acc_norm
237
+ value: 1.01
238
+ name: acc_norm
239
+ source:
240
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
241
+ name: Open LLM Leaderboard
242
+ - task:
243
+ type: text-generation
244
+ name: Text Generation
245
+ dataset:
246
+ name: MuSR (0-shot)
247
+ type: TAUR-Lab/MuSR
248
+ args:
249
+ num_few_shot: 0
250
+ metrics:
251
+ - type: acc_norm
252
+ value: 3.17
253
+ name: acc_norm
254
+ source:
255
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
256
+ name: Open LLM Leaderboard
257
+ - task:
258
+ type: text-generation
259
+ name: Text Generation
260
+ dataset:
261
+ name: MMLU-PRO (5-shot)
262
+ type: TIGER-Lab/MMLU-Pro
263
+ config: main
264
+ split: test
265
+ args:
266
+ num_few_shot: 5
267
+ metrics:
268
+ - type: acc
269
+ value: 1.51
270
+ name: accuracy
271
+ source:
272
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Felladrin/Llama-160M-Chat-v1
273
+ name: Open LLM Leaderboard
274
+ ---
275
+
276
+ <div style="width: auto; margin-left: auto; margin-right: auto">
277
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
278
+ </div>
279
+ <div style="display: flex; justify-content: space-between; width: 100%;">
280
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
281
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
282
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
283
+ </p>
284
+ </div>
285
+ </div>
286
+
287
+ ## Felladrin/Llama-160M-Chat-v1 - GGUF
288
+
289
+ This repo contains GGUF format model files for [Felladrin/Llama-160M-Chat-v1](https://huggingface.co/Felladrin/Llama-160M-Chat-v1).
290
+
291
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
292
+
293
+ <div style="text-align: left; margin: 20px 0;">
294
+ <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
295
+ Run them on the TensorBlock client using your local machine โ†—
296
+ </a>
297
+ </div>
298
+
299
+ ## Prompt template
300
+
301
+ ```
302
+ <|im_start|>system
303
+ {system_prompt}<|im_end|>
304
+ <|im_start|>user
305
+ {prompt}<|im_end|>
306
+ <|im_start|>assistant
307
+ ```
308
+
309
+ ## Model file specification
310
+
311
+ | Filename | Quant type | File Size | Description |
312
+ | -------- | ---------- | --------- | ----------- |
313
+ | [Llama-160M-Chat-v1-Q2_K.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q2_K.gguf) | Q2_K | 0.066 GB | smallest, significant quality loss - not recommended for most purposes |
314
+ | [Llama-160M-Chat-v1-Q3_K_S.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q3_K_S.gguf) | Q3_K_S | 0.075 GB | very small, high quality loss |
315
+ | [Llama-160M-Chat-v1-Q3_K_M.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q3_K_M.gguf) | Q3_K_M | 0.080 GB | very small, high quality loss |
316
+ | [Llama-160M-Chat-v1-Q3_K_L.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q3_K_L.gguf) | Q3_K_L | 0.085 GB | small, substantial quality loss |
317
+ | [Llama-160M-Chat-v1-Q4_0.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q4_0.gguf) | Q4_0 | 0.092 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
318
+ | [Llama-160M-Chat-v1-Q4_K_S.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q4_K_S.gguf) | Q4_K_S | 0.092 GB | small, greater quality loss |
319
+ | [Llama-160M-Chat-v1-Q4_K_M.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q4_K_M.gguf) | Q4_K_M | 0.096 GB | medium, balanced quality - recommended |
320
+ | [Llama-160M-Chat-v1-Q5_0.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q5_0.gguf) | Q5_0 | 0.108 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
321
+ | [Llama-160M-Chat-v1-Q5_K_S.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q5_K_S.gguf) | Q5_K_S | 0.108 GB | large, low quality loss - recommended |
322
+ | [Llama-160M-Chat-v1-Q5_K_M.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q5_K_M.gguf) | Q5_K_M | 0.110 GB | large, very low quality loss - recommended |
323
+ | [Llama-160M-Chat-v1-Q6_K.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q6_K.gguf) | Q6_K | 0.125 GB | very large, extremely low quality loss |
324
+ | [Llama-160M-Chat-v1-Q8_0.gguf](https://huggingface.co/tensorblock/Llama-160M-Chat-v1-GGUF/blob/main/Llama-160M-Chat-v1-Q8_0.gguf) | Q8_0 | 0.161 GB | very large, extremely low quality loss - not recommended |
325
+
326
+
327
+ ## Downloading instruction
328
+
329
+ ### Command line
330
+
331
+ Firstly, install Huggingface Client
332
+
333
+ ```shell
334
+ pip install -U "huggingface_hub[cli]"
335
+ ```
336
+
337
+ Then, downoad the individual model file the a local directory
338
+
339
+ ```shell
340
+ huggingface-cli download tensorblock/Llama-160M-Chat-v1-GGUF --include "Llama-160M-Chat-v1-Q2_K.gguf" --local-dir MY_LOCAL_DIR
341
+ ```
342
+
343
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
344
+
345
+ ```shell
346
+ huggingface-cli download tensorblock/Llama-160M-Chat-v1-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
347
+ ```