From d06f0062910b9345cdd1aaf4b5ed69e9721cc36c Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Wed, 8 Nov 2023 19:29:57 -0800 Subject: [PATCH] v0 litellm --- llmzoo/eval/eval_gpt_review_all.py | 3 ++- llmzoo/eval/prompt_turbo.py | 3 ++- 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/llmzoo/eval/eval_gpt_review_all.py b/llmzoo/eval/eval_gpt_review_all.py index 34cff04..6a297fa 100644 --- a/llmzoo/eval/eval_gpt_review_all.py +++ b/llmzoo/eval/eval_gpt_review_all.py @@ -6,13 +6,14 @@ import backoff import numpy as np import openai +import litellm import ray @ray.remote(num_cpus=4) @backoff.on_exception(backoff.expo, openai.error.RateLimitError) def get_eval(content: str, max_tokens: int): - response = openai.ChatCompletion.create( + response = litellm.completion( model='gpt-3.5-turbo', messages=[{ 'role': 'system', diff --git a/llmzoo/eval/prompt_turbo.py b/llmzoo/eval/prompt_turbo.py index b426d81..f5bc6bf 100644 --- a/llmzoo/eval/prompt_turbo.py +++ b/llmzoo/eval/prompt_turbo.py @@ -3,6 +3,7 @@ import backoff import openai +import litellm import ray import shortuuid @@ -12,7 +13,7 @@ @ray.remote(num_cpus=4) @backoff.on_exception(backoff.expo, openai.error.RateLimitError) def completions_with_backoff(**kwargs): - return openai.ChatCompletion.create(**kwargs) + return litellm.completion(**kwargs) if __name__ == '__main__':