GGUF
multilingual
sea
Inference Endpoints
conversational

Correct template for use with this model

#1
by saknarak - opened

I try to use preset with LM Studio but model generate few token and then stop generate

I also convert to ollama with following Modelfile, but got the same results
there is no problem for v2 model

FROM seallm-7b-v2.5.Q4_K_M.gguf
TEMPLATE """<|im_start|>system
{{ .System }}<eos>
<|im_start|>user
{{ .Prompt }}<eos>
<|im_start|>assistant
"""
SYSTEM """You are a helpful assistant."""
PARAMETER stop "<eos>"
PARAMETER stop "<|im_start|>"

Please suggest me

image.png

SeaLLMs - Language Models for Southeast Asian Languages org

@saknarak If you're using llama.cpp, there are few BIG notes:

  1. <bos> should always at the beginning, so in case ollama doesn't add it implicitly, you may try adding it with template
TEMPLATE """<bos><|im_start|>system
{{ .System }}<eos>
<|im_start|>user
{{ .Prompt }}<eos>
<|im_start|>assistant
"""
  1. You must set --repeat-penalty 1.0 for llama.cpp, old versions of llama.cpp set it to 1.1, which break gemma model.

  2. What was the max new tokens used in ollama? Should set it to 1024.

  3. If you have tried LM-studio with seallm-7b-v2.5-chatml.Q4_K_M.gguf using the ChatML format instead?

let me know if you still run into problem.

Thank you for useful information.

https://github.com/ollama/ollama/blob/main/docs/modelfile.md
But still no luck :(

This is what I have try

  1. with/without
  2. set repeat_penalty to 1
  3. set num_predict to 1024
  4. change gguf model to seallm-7b-v2.5-chatml.Q4_K_M.gguf
    All of 1-4 in combinations, none of them is work.

Modelfile

FROM ./Publisher/Repository/seallm-7b-v2.5.Q4_K_M.gguf
TEMPLATE """<bos><|im_start|>system
{{ .System }}<eos>
<|im_start|>user
{{ .Prompt }}<eos>
<|im_start|>assistant
"""
SYSTEM """You are a helpful assistant."""
PARAMETER stop "<eos>"
PARAMETER stop "<|im_start|>"
PARAMETER temperature 1
PARAMETER repeat_penalty 1
PARAMETER num_predict 1024

image.png

seallm-7b-v2.5-chatml.Q4_K_M.gguf is the same result

Only SeaLLM 2.0 is working.

Sign up or log in to comment