Request to add seallm.preset.json for LM-studio to work properly
Browse filesGreat thanks
@LoneStriker
for creating gguf for our model. We would like to add this seallm.preset.json so that LM-studio can load the chat template more correctly.
- seallm.preset.json +14 -0
seallm.preset.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"name": "SeaLLM",
|
3 |
+
"inference_params": {
|
4 |
+
"input_prefix": "</s><|im_start|>user\n",
|
5 |
+
"input_suffix": "</s><|im_start|>assistant\n",
|
6 |
+
"antiprompt": [
|
7 |
+
"<|im_start|>",
|
8 |
+
"</s>"
|
9 |
+
],
|
10 |
+
"pre_prompt_prefix": "<|im_start|>system\n",
|
11 |
+
"pre_prompt_suffix": "",
|
12 |
+
"pre_prompt": "You are a helpful assistant."
|
13 |
+
}
|
14 |
+
}
|