Spaces:
Running
on
Zero
Running
on
Zero
gokaygokay
commited on
Commit
•
2037b55
1
Parent(s):
f712bed
Update app.py
Browse files
app.py
CHANGED
@@ -18,13 +18,6 @@ hf_hub_download(
|
|
18 |
local_dir = "./models"
|
19 |
)
|
20 |
|
21 |
-
# Add new model download
|
22 |
-
hf_hub_download(
|
23 |
-
repo_id="sm54/Artificium-llama3.1-8B-001-Q8_0-GGUF",
|
24 |
-
filename="artificium-llama3.1-8b-001-q8_0.gguf",
|
25 |
-
local_dir = "./models"
|
26 |
-
)
|
27 |
-
|
28 |
|
29 |
|
30 |
def get_messages_formatter_type(model_name):
|
@@ -105,20 +98,18 @@ def respond(
|
|
105 |
outputs += output
|
106 |
yield outputs
|
107 |
|
108 |
-
description = """<p
|
109 |
-
<a href="https://huggingface.co/mattshumer/Reflection-Llama-3.1-70B" target="_blank">[Reflection 70B]</a>
|
110 |
-
|
111 |
-
<a href="https://huggingface.co/
|
112 |
-
|
113 |
-
</p>
|
114 |
"""
|
115 |
|
116 |
demo = gr.ChatInterface(
|
117 |
respond,
|
118 |
additional_inputs=[
|
119 |
gr.Dropdown([
|
120 |
-
'Reflection-Llama-3.1-70B-Q3_K_M.gguf'
|
121 |
-
'artificium-llama3.1-8b-001-q8_0.gguf' # Add new model to dropdown
|
122 |
],
|
123 |
value="Reflection-Llama-3.1-70B-Q3_K_M.gguf",
|
124 |
label="Model"
|
|
|
18 |
local_dir = "./models"
|
19 |
)
|
20 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
|
22 |
|
23 |
def get_messages_formatter_type(model_name):
|
|
|
98 |
outputs += output
|
99 |
yield outputs
|
100 |
|
101 |
+
description = """<p><center>
|
102 |
+
<a href="https://huggingface.co/mattshumer/Reflection-Llama-3.1-70B" target="_blank">[Reflection Llama 3.1 70B Model Page]</a>
|
103 |
+
|
104 |
+
<a href="https://huggingface.co/bartowski/Reflection-Llama-3.1-70B-GGUF" target="_blank">[70B Model GGUF]</a>
|
105 |
+
</center></p>
|
|
|
106 |
"""
|
107 |
|
108 |
demo = gr.ChatInterface(
|
109 |
respond,
|
110 |
additional_inputs=[
|
111 |
gr.Dropdown([
|
112 |
+
'Reflection-Llama-3.1-70B-Q3_K_M.gguf'
|
|
|
113 |
],
|
114 |
value="Reflection-Llama-3.1-70B-Q3_K_M.gguf",
|
115 |
label="Model"
|