Triangle104 commited on
Commit
a7d78fb
1 Parent(s): 47ac82e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +72 -0
README.md CHANGED
@@ -12,6 +12,78 @@ base_model: NeverSleep/Lumimaid-v0.2-12B
12
  This model was converted to GGUF format from [`NeverSleep/Lumimaid-v0.2-12B`](https://huggingface.co/NeverSleep/Lumimaid-v0.2-12B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
  Refer to the [original model card](https://huggingface.co/NeverSleep/Lumimaid-v0.2-12B) for more details on the model.
14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  ## Use with llama.cpp
16
  Install llama.cpp through brew (works on Mac and Linux)
17
 
 
12
  This model was converted to GGUF format from [`NeverSleep/Lumimaid-v0.2-12B`](https://huggingface.co/NeverSleep/Lumimaid-v0.2-12B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
  Refer to the [original model card](https://huggingface.co/NeverSleep/Lumimaid-v0.2-12B) for more details on the model.
14
 
15
+ ---
16
+ Model details:
17
+ -
18
+ This model is based on: Mistral-Nemo-Instruct-2407
19
+
20
+ Wandb: https://wandb.ai/undis95/Lumi-Mistral-Nemo?nw=nwuserundis95
21
+
22
+ NOTE: As explained on Mistral-Nemo-Instruct-2407 repo, it's recommended to use a low temperature, please experiment!
23
+
24
+ Lumimaid 0.1 -> 0.2 is a HUGE step up dataset wise.
25
+
26
+ As some people have told us our models are sloppy, Ikari decided to say fuck it and literally nuke all chats out with most slop.
27
+
28
+ Our dataset stayed the same since day one, we added data over time, cleaned them, and repeat. After not releasing model for a while because we were never satisfied, we think it's time to come back!
29
+ Prompt template: Mistral
30
+
31
+ <s>[INST] {input} [/INST] {output}</s>
32
+
33
+ Credits:
34
+
35
+ Undi
36
+ IkariDev
37
+
38
+ Training data we used to make our dataset:
39
+
40
+ Epiculous/Gnosis
41
+ ChaoticNeutrals/Luminous_Opus
42
+ ChaoticNeutrals/Synthetic-Dark-RP
43
+ ChaoticNeutrals/Synthetic-RP
44
+ Gryphe/Sonnet3.5-SlimOrcaDedupCleaned
45
+ Gryphe/Opus-WritingPrompts
46
+ meseca/writing-opus-6k
47
+ meseca/opus-instruct-9k
48
+ PJMixers/grimulkan_theory-of-mind-ShareGPT
49
+ NobodyExistsOnTheInternet/ToxicQAFinal
50
+ Undi95/toxic-dpo-v0.1-sharegpt
51
+ cgato/SlimOrcaDedupCleaned
52
+ kalomaze/Opus_Instruct_25k
53
+ Doctor-Shotgun/no-robots-sharegpt
54
+ Norquinal/claude_multiround_chat_30k
55
+ nothingiisreal/Claude-3-Opus-Instruct-15K
56
+ All the Aesirs dataset, cleaned, unslopped
57
+ All le luminae dataset, cleaned, unslopped
58
+ Small part of Airoboros reduced
59
+
60
+ We sadly didn't find the sources of the following, DM us if you recognize your set !
61
+
62
+ Opus_Instruct-v2-6.5K-Filtered-v2-sharegpt
63
+ claude_sharegpt_trimmed
64
+ CapybaraPure_Decontaminated-ShareGPT_reduced
65
+
66
+ Datasets credits:
67
+
68
+ Epiculous
69
+ ChaoticNeutrals
70
+ Gryphe
71
+ meseca
72
+ PJMixers
73
+ NobodyExistsOnTheInternet
74
+ cgato
75
+ kalomaze
76
+ Doctor-Shotgun
77
+ Norquinal
78
+ nothingiisreal
79
+
80
+ Others
81
+
82
+ Undi: If you want to support us, you can here.
83
+
84
+ IkariDev: Visit my retro/neocities style website please kek
85
+
86
+ ---
87
  ## Use with llama.cpp
88
  Install llama.cpp through brew (works on Mac and Linux)
89