eve / llm.py
ChandimaPrabath's picture
init
af3518b
raw
history blame contribute delete
954 Bytes
from openai import OpenAI
from dotenv import load_dotenv
import os
from utils import read_config
load_dotenv()
BASE_URL = os.getenv("LLM_BASE_URL")
# Initialize the OpenAI API with your API key
client = OpenAI(
api_key="koboldcpp",
base_url=BASE_URL
)
def pre_process():
# Read the config each time pre_process is called
config = read_config()
system_prompt = config['llm']['system_prompt']
char = config['llm']['char']
system_prompt = system_prompt.replace("{char}", char)
return system_prompt
# Function to handle messages and stream response from OpenAI
def generate_llm(prompt):
system_prompt = pre_process()
response = client.chat.completions.create(
model="koboldcpp/HF_SPACE_Tiefighter-13B",
messages=[
{"role": "system", "content": system_prompt},
{"role": "user", "content": prompt}
]
)
msg = response.choices[0].message.content
return msg