diff --git a/llama-index-integrations/llms/llama-index-llms-ollama/llama_index/llms/ollama/base.py b/llama-index-integrations/llms/llama-index-llms-ollama/llama_index/llms/ollama/base.py index f0848a89dff0269336e8db3019bc44ced0aff6e8..80079faeed82d2ee4e0b08c5a4885b957f3ae7be 100644 --- a/llama-index-integrations/llms/llama-index-llms-ollama/llama_index/llms/ollama/base.py +++ b/llama-index-integrations/llms/llama-index-llms-ollama/llama_index/llms/ollama/base.py @@ -147,48 +147,6 @@ class Ollama(CustomLLM): additional_kwargs=get_additional_kwargs(raw, ("message",)), ) - @llm_chat_callback() - async def achat( - self, messages: Sequence[ChatMessage], **kwargs: Any - ) -> ChatResponse: - payload = { - "model": self.model, - "messages": [ - { - "role": message.role.value, - "content": message.content, - **message.additional_kwargs, - } - for message in messages - ], - "options": self._model_kwargs, - "stream": False, - **kwargs, - } - - if self.json_mode: - payload["format"] = "json" - - async with httpx.AsyncClient(timeout=Timeout(self.request_timeout)) as client: - response = await client.post( - url=f"{self.base_url}/api/chat", - json=payload, - ) - response.raise_for_status() - raw = response.json() - message = raw["message"] - return ChatResponse( - message=ChatMessage( - content=message.get("content"), - role=MessageRole(message.get("role")), - additional_kwargs=get_additional_kwargs( - message, ("content", "role") - ), - ), - raw=raw, - additional_kwargs=get_additional_kwargs(raw, ("message",)), - ) - @llm_chat_callback() def stream_chat( self, messages: Sequence[ChatMessage], **kwargs: Any diff --git a/llama-index-integrations/llms/llama-index-llms-ollama/pyproject.toml b/llama-index-integrations/llms/llama-index-llms-ollama/pyproject.toml index 6f086a98816d4d4dfd6139a04eb2be825e03762b..da6ac04d601ce76cdb932ec745454a454db0a28c 100644 --- a/llama-index-integrations/llms/llama-index-llms-ollama/pyproject.toml +++ b/llama-index-integrations/llms/llama-index-llms-ollama/pyproject.toml @@ -27,7 +27,7 @@ exclude = ["**/BUILD"] license = "MIT" name = "llama-index-llms-ollama" readme = "README.md" -version = "0.1.3" +version = "0.1.4" [tool.poetry.dependencies] python = ">=3.8.1,<4.0"