bad_llm_dpov03-gguf / Modelfile
ebowwa's picture
Upload Modelfile
99bf212 verified
raw
history blame contribute delete
496 Bytes
FROM bad_llm_dpov03-gguf-unsloth.Q4_K_M.gguf
# sets the context window size to 4096, this controls how many tokens the LLM can use as context to generate the next token
PARAMETER num_ctx 4096
PARAMETER temperature 0.7
TEMPLATE """
{{ if .System }}<|start_header_id|>system<|end_header_id|>
{{ .System }}<|eot_id|>
{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
{{ .Prompt }}<|eot_id|>
{{ end }}<|start_header_id|>assistant<|end_header_id|>
{{ .Response }}<|eot_id|>
"""