fix ollama integration by forcing flat messages (#406)
This commit is contained in:
parent
183869de04
commit
6d72ea7283
|
@ -690,6 +690,7 @@ class LiteLLMModel(Model):
|
|||
api_base=self.api_base,
|
||||
api_key=self.api_key,
|
||||
convert_images_to_image_urls=True,
|
||||
flatten_messages_as_text=self.model_id.startswith("ollama"),
|
||||
custom_role_conversions=self.custom_role_conversions,
|
||||
**kwargs,
|
||||
)
|
||||
|
|
Loading…
Reference in New Issue