Spaces:
Sleeping
Sleeping
# import os | |
# import sys, os | |
# import traceback | |
# from dotenv import load_dotenv | |
# load_dotenv() | |
# import os, io | |
# sys.path.insert( | |
# 0, os.path.abspath("../..") | |
# ) # Adds the parent directory to the system path | |
# import pytest | |
# import litellm | |
# from litellm import embedding, completion, text_completion, completion_cost | |
# from langchain.chat_models import ChatLiteLLM | |
# from langchain.prompts.chat import ( | |
# ChatPromptTemplate, | |
# SystemMessagePromptTemplate, | |
# AIMessagePromptTemplate, | |
# HumanMessagePromptTemplate, | |
# ) | |
# from langchain.schema import AIMessage, HumanMessage, SystemMessage | |
# def test_chat_gpt(): | |
# try: | |
# chat = ChatLiteLLM(model="gpt-3.5-turbo", max_tokens=10) | |
# messages = [ | |
# HumanMessage( | |
# content="what model are you" | |
# ) | |
# ] | |
# resp = chat(messages) | |
# print(resp) | |
# except Exception as e: | |
# pytest.fail(f"Error occurred: {e}") | |
# # test_chat_gpt() | |
# def test_claude(): | |
# try: | |
# chat = ChatLiteLLM(model="claude-2", max_tokens=10) | |
# messages = [ | |
# HumanMessage( | |
# content="what model are you" | |
# ) | |
# ] | |
# resp = chat(messages) | |
# print(resp) | |
# except Exception as e: | |
# pytest.fail(f"Error occurred: {e}") | |
# # test_claude() | |
# def test_palm(): | |
# try: | |
# chat = ChatLiteLLM(model="palm/chat-bison", max_tokens=10) | |
# messages = [ | |
# HumanMessage( | |
# content="what model are you" | |
# ) | |
# ] | |
# resp = chat(messages) | |
# print(resp) | |
# except Exception as e: | |
# pytest.fail(f"Error occurred: {e}") | |
# # test_palm() | |
# # def test_openai_with_params(): | |
# # try: | |
# # api_key = os.environ["OPENAI_API_KEY"] | |
# # os.environ.pop("OPENAI_API_KEY") | |
# # print("testing openai with params") | |
# # llm = ChatLiteLLM( | |
# # model="gpt-3.5-turbo", | |
# # openai_api_key=api_key, | |
# # # Prefer using None which is the default value, endpoint could be empty string | |
# # openai_api_base= None, | |
# # max_tokens=20, | |
# # temperature=0.5, | |
# # request_timeout=10, | |
# # model_kwargs={ | |
# # "frequency_penalty": 0, | |
# # "presence_penalty": 0, | |
# # }, | |
# # verbose=True, | |
# # max_retries=0, | |
# # ) | |
# # messages = [ | |
# # HumanMessage( | |
# # content="what model are you" | |
# # ) | |
# # ] | |
# # resp = llm(messages) | |
# # print(resp) | |
# # except Exception as e: | |
# # pytest.fail(f"Error occurred: {e}") | |
# # test_openai_with_params() | |