make compatible with remote openai-compatible servers
This commit is contained in:
parent
95aae1ba85
commit
b3dfaddd43
|
@ -410,11 +410,13 @@ class TransformersModel(Model):
|
|||
|
||||
|
||||
class LiteLLMModel(Model):
|
||||
def __init__(self, model_id="anthropic/claude-3-5-sonnet-20240620"):
|
||||
def __init__(self, model_id="anthropic/claude-3-5-sonnet-20240620", api_base=None, api_key=None):
|
||||
super().__init__()
|
||||
self.model_id = model_id
|
||||
# IMPORTANT - Set this to TRUE to add the function to the prompt for Non OpenAI LLMs
|
||||
litellm.add_function_to_prompt = True
|
||||
self.api_base = api_base
|
||||
self.api_key = api_key
|
||||
|
||||
def __call__(
|
||||
self,
|
||||
|
@ -432,6 +434,8 @@ class LiteLLMModel(Model):
|
|||
messages=messages,
|
||||
stop=stop_sequences,
|
||||
max_tokens=max_tokens,
|
||||
api_base=self.api_base,
|
||||
api_key=self.api_key,
|
||||
)
|
||||
self.last_input_token_count = response.usage.prompt_tokens
|
||||
self.last_output_token_count = response.usage.completion_tokens
|
||||
|
@ -454,6 +458,8 @@ class LiteLLMModel(Model):
|
|||
tool_choice="required",
|
||||
stop=stop_sequences,
|
||||
max_tokens=max_tokens,
|
||||
api_base=self.api_base,
|
||||
api_key=self.api_key,
|
||||
)
|
||||
tool_calls = response.choices[0].message.tool_calls[0]
|
||||
self.last_input_token_count = response.usage.prompt_tokens
|
||||
|
|
Loading…
Reference in New Issue