diff --git a/llama_index/multi_modal_llms/ollama.py b/llama_index/multi_modal_llms/ollama.py index 5bff2ba72a2d610e75b5dd2d1163a659f3b74e16..4d56c1327473ae91289c9880d66b1b74f6a262f7 100644 --- a/llama_index/multi_modal_llms/ollama.py +++ b/llama_index/multi_modal_llms/ollama.py @@ -106,7 +106,9 @@ class OllamaMultiModal(MultiModalLLM): import ollama ollama_messages = _messages_to_dicts(messages) - response = ollama.chat(model=self.model, messages=ollama_messages, stream=False) + response = ollama.chat( + model=self.model, messages=ollama_messages, stream=False, **kwargs + ) return ChatResponse( message=ChatMessage( content=response["message"]["content"], @@ -124,7 +126,9 @@ class OllamaMultiModal(MultiModalLLM): import ollama ollama_messages = _messages_to_dicts(messages) - response = ollama.chat(model=self.model, messages=ollama_messages, stream=True) + response = ollama.chat( + model=self.model, messages=ollama_messages, stream=True, **kwargs + ) text = "" for chunk in response: if "done" in chunk and chunk["done"]: @@ -161,6 +165,7 @@ class OllamaMultiModal(MultiModalLLM): images=image_documents_to_base64(image_documents), stream=False, options=self._model_kwargs, + **kwargs, ) return CompletionResponse( text=response["response"], @@ -184,6 +189,7 @@ class OllamaMultiModal(MultiModalLLM): images=image_documents_to_base64(image_documents), stream=True, options=self._model_kwargs, + **kwargs, ) text = "" for chunk in response: