Triangle104 commited on
Commit
34998d6
1 Parent(s): 3dbf797

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +89 -0
README.md CHANGED
@@ -16,6 +16,95 @@ tags:
16
  This model was converted to GGUF format from [`kromeurus/L3.1-Aglow-Vulca-v0.1-8B`](https://huggingface.co/kromeurus/L3.1-Aglow-Vulca-v0.1-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
17
  Refer to the [original model card](https://huggingface.co/kromeurus/L3.1-Aglow-Vulca-v0.1-8B) for more details on the model.
18
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  ## Use with llama.cpp
20
  Install llama.cpp through brew (works on Mac and Linux)
21
 
 
16
  This model was converted to GGUF format from [`kromeurus/L3.1-Aglow-Vulca-v0.1-8B`](https://huggingface.co/kromeurus/L3.1-Aglow-Vulca-v0.1-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
17
  Refer to the [original model card](https://huggingface.co/kromeurus/L3.1-Aglow-Vulca-v0.1-8B) for more details on the model.
18
 
19
+ ---
20
+ Model Details & Recommended Settings
21
+ -
22
+
23
+ This is a story telling first model that is proficient in narrative driven RP. Does best with straightforward instructions. Any wishy-washy language will confuse it. As per usual with any of my models with Formax in it, it's pretty sensitive to instructs so choose your words wisely.
24
+
25
+ Once going though, it's able to generate detailed and human-ish outputs with lots of personality depending on the information given. Has a habit of matching the style and format of the input; style, spacing, grammar, etc. Can interweave details from the character persona, chat history, and user persona (if there is one) to create unique interactions and plot points. Leans more or less positive naturally but can be flipped if prompted correctly.
26
+
27
+ Being a Llama 3.1 model, it's still subject to the normal pros and cons of L3/L3.1 but I'd like to think I tamed some of it. Keep the temp on the lower end since there is a low chance it might freak out. If it does, swipe/regen the chat or delete the afflicted output and try again.
28
+
29
+ Rec. Settings:
30
+
31
+ Template: Llama 3
32
+ Token Count: 128k Max
33
+ Temperature: 1.2
34
+ Min P: 0.1
35
+ Repeat Penalty: 1.05
36
+ Repeat Penalty Tokens: 256
37
+
38
+ Merge Theory
39
+
40
+ Where to begin. The general though process was still the roughly same as the Ablaze, making one very smart model and another more creative focused model. This time, I merged Formax and RPmax in separately instead of doing one merge since they have different focuses.
41
+
42
+ 'Apollobulk' is the smarts, having the storytelling capabilities from badger writer, instruct following from Formax (duh) and the smarts of Super Nova. Apollo 0.4 was use as an RP temper to keep the overall model aligned with RP. Apollo 2.0 wasn't used as it skewed the merge too far towards inconsistent narratives.
43
+
44
+ 'Reshape' is the creative end, taking some inspo from the Ablaze's creative center. First created 'Darkened' as the main influence over the final writing style of Aglow. Poppy Moonfall C had the personality I was looking for but the smarts (though not important was still necessary) so the other three were added to round out it's overall capabilities while being very creative. Plopping that atop RPmax (For excellent unique RP interactions), BRAG (serious recall), and Natsumura (a great Storytelling/RP base) and model stock it, you get a really solid model on its own.
45
+
46
+ Slap the two components together in a simple gradient dare_linear merge and boom; this unit of an 8B model. As of writing and releasing this model, mergekit is fucked for me (one of its dependencies has broken L3 merging) so I can't test any other methods atm. If there is a better final merge method, I'll be uploading a v0.2 once the bug is fixed.
47
+
48
+ This time around, everything was done with DavidAU's High Quality method, merging with float32 at all steps. Made a significant difference in nuanced understanding of text.
49
+ Config
50
+
51
+ models:
52
+ - model: Locutusque/Apollo-0.4-Llama-3.1-8B
53
+ - model: maldv/badger-writer-llama-3-8b
54
+ - model: ArliAI/Llama-3.1-8B-ArliAI-Formax-v1.0
55
+ base_model: arcee-ai/Llama-3.1-SuperNova-Lite
56
+ parameters:
57
+ int8_mask: true
58
+ merge_method: model_stock
59
+ dtype: float32
60
+ tokenizer_source: base
61
+ name: apollobulk
62
+
63
+
64
+ models:
65
+ - model: v000000/L3-8B-Poppy-Moonfall-C
66
+ - model: Casual-Autopsy/Jamet-L3-Stheno-BlackOasis-8B
67
+ - model: SicariusSicariiStuff/Dusk_Rainbow
68
+ base_model: ResplendentAI/Rawr_Llama3_8B
69
+ parameters:
70
+ int8_mask: true
71
+ merge_method: model_stock
72
+ dtype: float32
73
+ tokenizer_source: base
74
+ name: darkened
75
+
76
+
77
+ models:
78
+ - model: darkened
79
+ - model: ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
80
+ - model: maximalists/BRAG-Llama-3.1-8b-v0.1
81
+ base_model: tohur/natsumura-storytelling-rp-1.0-llama-3.1-8b
82
+ parameters:
83
+ int8_mask: true
84
+ merge_method: model_stock
85
+ dtype: float32
86
+ tokenizer_source: base
87
+ name: reshape
88
+
89
+
90
+ models:
91
+ - model: reshape
92
+ parameters:
93
+ weight: [0.1, 0.9]
94
+ - model: apollobulk
95
+ parameters:
96
+ weight: [0.9, 0.1]
97
+ base_model: reshape
98
+ tokenizer_source: base
99
+ parameters:
100
+ normalize: false
101
+ int8_mask: true
102
+ merge_method: dare_linear
103
+ dtype: float32
104
+ name: vulca
105
+
106
+ ---
107
+
108
  ## Use with llama.cpp
109
  Install llama.cpp through brew (works on Mac and Linux)
110