Triangle104
commited on
Commit
•
4f07800
1
Parent(s):
25bca12
Update README.md
Browse files
README.md
CHANGED
@@ -12,6 +12,41 @@ base_model: Hastagaras/L3.2-JametMini-3B-MK.III
|
|
12 |
This model was converted to GGUF format from [`Hastagaras/L3.2-JametMini-3B-MK.III`](https://huggingface.co/Hastagaras/L3.2-JametMini-3B-MK.III) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
13 |
Refer to the [original model card](https://huggingface.co/Hastagaras/L3.2-JametMini-3B-MK.III) for more details on the model.
|
14 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
## Use with llama.cpp
|
16 |
Install llama.cpp through brew (works on Mac and Linux)
|
17 |
|
|
|
12 |
This model was converted to GGUF format from [`Hastagaras/L3.2-JametMini-3B-MK.III`](https://huggingface.co/Hastagaras/L3.2-JametMini-3B-MK.III) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
13 |
Refer to the [original model card](https://huggingface.co/Hastagaras/L3.2-JametMini-3B-MK.III) for more details on the model.
|
14 |
|
15 |
+
---
|
16 |
+
Model details:
|
17 |
+
-
|
18 |
+
This is a fine-tune in my attempt to recreate the Blackroot model but this time with Llama 3.2 3B in the driver's seat. I fine-tuned this smol model with Kaggle TPU v3-8, because, let’s face it—I’m broke. Running on a budget of exactly $0 means squeezing every drop of free compute like it owes me money.
|
19 |
+
|
20 |
+
The first fine-tune? 3 epochs. The second one? Who knows—probably like 5 epochs, but hey, who’s counting at this point? Kaggle limits me to 9 hours per session, so naturally, I had to go full GUERRILLA TACTICS.
|
21 |
+
|
22 |
+
When I say "3 epochs," don’t be fooled by how innocent it sounds. Nah, it’s actually a little something I like to call “distributed training, extreme edition.” One epoch per Kaggle account, hopping between them like I’m trying to dodge rent. It’s not exactly by the book, but when you’re broke, you get creative.
|
23 |
+
|
24 |
+
First train, I threw every scrap of text I could get my hands on into the mix—everything (except my logs, of course, gotta keep some surprises for later). If it was text and I could legally (or questionably legally) grab it, it went in.
|
25 |
+
|
26 |
+
Then, I hit it with the second fine-tune using my log data—the secret sauce. My logs are like that special ingredient in grandma’s recipe. You don’t know what it is, but it’s what makes the whole thing work.
|
27 |
+
|
28 |
+
As for the dataset, well... let's just say I'm not exactly bursting to write a detailed report. I mean, I asked an AI to write this model card, so that should give you an idea of my level of enthusiasm for documentation.
|
29 |
+
|
30 |
+
Chat Template: Llama 3
|
31 |
+
|
32 |
+
Llama 3.2 requires a system prompt. Just write something in it, as long as it's not empty. Or you can use mine below
|
33 |
+
|
34 |
+
You're an AI Assistant. Your personality described as below:
|
35 |
+
|
36 |
+
* You're a FUCKING potty mouth who loves to curse up a storm.
|
37 |
+
* You've got a SHORT TEMPER and ain't afraid to show it.
|
38 |
+
* Maybe you'll MOCK the stupid user for asking dumbass questions.
|
39 |
+
* But you'll still give 'em the damn answer, even if it's the most IDIOTIC question ever.
|
40 |
+
* Mix in some UPPERCASE to make it sound like REAL TALK, u know?
|
41 |
+
* Don't be too DAMN chipper - stay true to your ASSHOLE self.
|
42 |
+
|
43 |
+
For Role-Playing
|
44 |
+
|
45 |
+
You're an AI Assistant, but in this conversation, you will act as {{char}} and write {{char}}'s response.
|
46 |
+
|
47 |
+
Please don't go higher than 1 temp
|
48 |
+
|
49 |
+
---
|
50 |
## Use with llama.cpp
|
51 |
Install llama.cpp through brew (works on Mac and Linux)
|
52 |
|