Spaces:
Running
on
Zero
Running
on
Zero
Upload 2 files
Browse files- llmdolphin.py +7 -0
- requirements.txt +1 -1
llmdolphin.py
CHANGED
@@ -14,9 +14,15 @@ llm_models = {
|
|
14 |
"L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
15 |
"Llama-3.1-8B-EZO-1.1-it.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-EZO-1.1-it-GGUF", MessagesFormatterType.MISTRAL],
|
16 |
"MN-12B-Starcannon-v1.i1-Q4_K_M.gguf": ["mradermacher/MN-12B-Starcannon-v1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
17 |
"Lumimaid-Magnum-12B.i1-Q4_K_M.gguf": ["mradermacher/Lumimaid-Magnum-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
|
|
|
|
|
|
18 |
"mini-magnum-12b-v1.1.i1-Q4_K_M.gguf": ["mradermacher/mini-magnum-12b-v1.1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
19 |
"MN-12B-Celeste-V1.9-Q4_K_M.gguf": ["bartowski/MN-12B-Celeste-V1.9-GGUF", MessagesFormatterType.CHATML],
|
|
|
20 |
"L3-Rhaenys-8B.i1-Q5_K_M.gguf": ["mradermacher/L3-Rhaenys-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
21 |
"Llama-3.1-8b-Uncensored-Dare.i1-Q4_K_M.gguf": ["mradermacher/Llama-3.1-8b-Uncensored-Dare-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
22 |
"Eros_Scribe-10.7b-v3.Q4_K_M.gguf": ["mradermacher/Eros_Scribe-10.7b-v3-GGUF", MessagesFormatterType.MISTRAL],
|
@@ -110,6 +116,7 @@ llm_models = {
|
|
110 |
#"": ["", MessagesFormatterType.MISTRAL],
|
111 |
#"": ["", MessagesFormatterType.ALPACA],
|
112 |
#"": ["", MessagesFormatterType.OPEN_CHAT],
|
|
|
113 |
}
|
114 |
llm_formats = {
|
115 |
"MISTRAL": MessagesFormatterType.MISTRAL,
|
|
|
14 |
"L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
15 |
"Llama-3.1-8B-EZO-1.1-it.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-EZO-1.1-it-GGUF", MessagesFormatterType.MISTRAL],
|
16 |
"MN-12B-Starcannon-v1.i1-Q4_K_M.gguf": ["mradermacher/MN-12B-Starcannon-v1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
17 |
+
"MN-12B-Starcannon-v2.i1-Q4_K_M.gguf": ["mradermacher/MN-12B-Starcannon-v2-i1-GGUF", MessagesFormatterType.CHATML],
|
18 |
"Lumimaid-Magnum-12B.i1-Q4_K_M.gguf": ["mradermacher/Lumimaid-Magnum-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
19 |
+
"Nemo-12B-Marlin-v1.i1-Q4_K_M.gguf": ["mradermacher/Nemo-12B-Marlin-v1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
20 |
+
"DarkIdol-Llama-3.1-8B-Instruct-1.2-Uncensored-Q5_K_M.gguf": ["bartowski/DarkIdol-Llama-3.1-8B-Instruct-1.2-Uncensored-GGUF", MessagesFormatterType.LLAMA_3],
|
21 |
+
"Llama-3-Swallow-8B-Instruct-v0.1.Q5_K_M.gguf": ["YukiTomita-CC/Llama-3-Swallow-8B-Instruct-v0.1-IMat-GGUF_dolly-15k-ja-prompt", MessagesFormatterType.ALPACA],
|
22 |
+
"natsumura-storytelling-rp-1.0-llama-3.1-8B.Q5_K_M.gguf": ["tohur/natsumura-storytelling-rp-1.0-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
23 |
"mini-magnum-12b-v1.1.i1-Q4_K_M.gguf": ["mradermacher/mini-magnum-12b-v1.1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
24 |
"MN-12B-Celeste-V1.9-Q4_K_M.gguf": ["bartowski/MN-12B-Celeste-V1.9-GGUF", MessagesFormatterType.CHATML],
|
25 |
+
"Llama-3.1-Techne-RP-8b-v1.i1-Q5_K_M.gguf": ["mradermacher/Llama-3.1-Techne-RP-8b-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
26 |
"L3-Rhaenys-8B.i1-Q5_K_M.gguf": ["mradermacher/L3-Rhaenys-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
27 |
"Llama-3.1-8b-Uncensored-Dare.i1-Q4_K_M.gguf": ["mradermacher/Llama-3.1-8b-Uncensored-Dare-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
28 |
"Eros_Scribe-10.7b-v3.Q4_K_M.gguf": ["mradermacher/Eros_Scribe-10.7b-v3-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
116 |
#"": ["", MessagesFormatterType.MISTRAL],
|
117 |
#"": ["", MessagesFormatterType.ALPACA],
|
118 |
#"": ["", MessagesFormatterType.OPEN_CHAT],
|
119 |
+
#"": ["", MessagesFormatterType.CHATML],
|
120 |
}
|
121 |
llm_formats = {
|
122 |
"MISTRAL": MessagesFormatterType.MISTRAL,
|
requirements.txt
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
huggingface_hub
|
2 |
scikit-build-core
|
3 |
-
https://github.com/abetlen/llama-cpp-python/releases/download/v0.2.
|
4 |
git+https://github.com/Maximilian-Winter/llama-cpp-agent
|
5 |
pybind11>=2.12
|
6 |
torch==2.2.0
|
|
|
1 |
huggingface_hub
|
2 |
scikit-build-core
|
3 |
+
https://github.com/abetlen/llama-cpp-python/releases/download/v0.2.85-cu124/llama_cpp_python-0.2.85-cp310-cp310-linux_x86_64.whl
|
4 |
git+https://github.com/Maximilian-Winter/llama-cpp-agent
|
5 |
pybind11>=2.12
|
6 |
torch==2.2.0
|