Spaces:
Sleeping
Sleeping
import sys | |
import os | |
import io | |
sys.path.insert(0, os.path.abspath("../..")) | |
from litellm import completion | |
import litellm | |
litellm.success_callback = ["langsmith"] | |
litellm.set_verbose = True | |
import time | |
def test_langsmith_logging(): | |
try: | |
response = completion( | |
model="claude-instant-1.2", | |
messages=[{"role": "user", "content": "what llm are u"}], | |
max_tokens=10, | |
temperature=0.2, | |
) | |
print(response) | |
except Exception as e: | |
print(e) | |
# test_langsmith_logging() | |
def test_langsmith_logging_with_metadata(): | |
try: | |
response = completion( | |
model="gpt-3.5-turbo", | |
messages=[{"role": "user", "content": "what llm are u"}], | |
max_tokens=10, | |
temperature=0.2, | |
metadata={ | |
"run_name": "litellmRUN", | |
"project_name": "litellm-completion", | |
}, | |
) | |
print(response) | |
except Exception as e: | |
print(e) | |
# test_langsmith_logging_with_metadata() | |
def test_langsmith_logging_with_streaming_and_metadata(): | |
try: | |
response = completion( | |
model="gpt-3.5-turbo", | |
messages=[{"role": "user", "content": "what llm are u"}], | |
max_tokens=10, | |
temperature=0.2, | |
metadata={ | |
"run_name": "litellmRUN", | |
"project_name": "litellm-completion", | |
}, | |
stream=True, | |
) | |
for chunk in response: | |
continue | |
except Exception as e: | |
print(e) | |
test_langsmith_logging_with_streaming_and_metadata() | |