Added LiteLLM to the stack
This commit is contained in:
@@ -0,0 +1,90 @@
|
||||
# import os
|
||||
# import sys, os
|
||||
# import traceback
|
||||
# from dotenv import load_dotenv
|
||||
|
||||
# load_dotenv()
|
||||
# import os, io
|
||||
|
||||
# sys.path.insert(
|
||||
# 0, os.path.abspath("../..")
|
||||
# ) # Adds the parent directory to the system path
|
||||
# import pytest
|
||||
# import litellm
|
||||
# from litellm import embedding, completion, text_completion, completion_cost
|
||||
|
||||
# from langchain.chat_models import ChatLiteLLM
|
||||
# from langchain.prompts.chat import (
|
||||
# ChatPromptTemplate,
|
||||
# SystemMessagePromptTemplate,
|
||||
# AIMessagePromptTemplate,
|
||||
# HumanMessagePromptTemplate,
|
||||
# )
|
||||
# from langchain.schema import AIMessage, HumanMessage, SystemMessage
|
||||
|
||||
# def test_chat_gpt():
|
||||
# try:
|
||||
# chat = ChatLiteLLM(model="gpt-3.5-turbo", max_tokens=10)
|
||||
# messages = [
|
||||
# HumanMessage(
|
||||
# content="what model are you"
|
||||
# )
|
||||
# ]
|
||||
# resp = chat(messages)
|
||||
|
||||
# print(resp)
|
||||
# except Exception as e:
|
||||
# pytest.fail(f"Error occurred: {e}")
|
||||
|
||||
# # test_chat_gpt()
|
||||
|
||||
|
||||
# def test_claude():
|
||||
# try:
|
||||
# chat = ChatLiteLLM(model="claude-2", max_tokens=10)
|
||||
# messages = [
|
||||
# HumanMessage(
|
||||
# content="what model are you"
|
||||
# )
|
||||
# ]
|
||||
# resp = chat(messages)
|
||||
|
||||
# print(resp)
|
||||
# except Exception as e:
|
||||
# pytest.fail(f"Error occurred: {e}")
|
||||
|
||||
# # test_claude()
|
||||
|
||||
|
||||
# # def test_openai_with_params():
|
||||
# # try:
|
||||
# # api_key = os.environ["OPENAI_API_KEY"]
|
||||
# # os.environ.pop("OPENAI_API_KEY")
|
||||
# # print("testing openai with params")
|
||||
# # llm = ChatLiteLLM(
|
||||
# # model="gpt-3.5-turbo",
|
||||
# # openai_api_key=api_key,
|
||||
# # # Prefer using None which is the default value, endpoint could be empty string
|
||||
# # openai_api_base= None,
|
||||
# # max_tokens=20,
|
||||
# # temperature=0.5,
|
||||
# # request_timeout=10,
|
||||
# # model_kwargs={
|
||||
# # "frequency_penalty": 0,
|
||||
# # "presence_penalty": 0,
|
||||
# # },
|
||||
# # verbose=True,
|
||||
# # max_retries=0,
|
||||
# # )
|
||||
# # messages = [
|
||||
# # HumanMessage(
|
||||
# # content="what model are you"
|
||||
# # )
|
||||
# # ]
|
||||
# # resp = llm(messages)
|
||||
|
||||
# # print(resp)
|
||||
# # except Exception as e:
|
||||
# # pytest.fail(f"Error occurred: {e}")
|
||||
|
||||
# # test_openai_with_params()
|
Reference in New Issue
Block a user