diff --git a/llama-index-integrations/llms/llama-index-llms-mistralai/llama_index/llms/mistralai/base.py b/llama-index-integrations/llms/llama-index-llms-mistralai/llama_index/llms/mistralai/base.py
index 05977726e8d1af562d72ca177253092626020542..78f9f58bc042d25d131762bc4493f8f92eff4f0c 100644
--- a/llama-index-integrations/llms/llama-index-llms-mistralai/llama_index/llms/mistralai/base.py
+++ b/llama-index-integrations/llms/llama-index-llms-mistralai/llama_index/llms/mistralai/base.py
@@ -173,7 +173,7 @@ class MistralAI(LLM):
     @llm_chat_callback()
     def chat(self, messages: Sequence[ChatMessage], **kwargs: Any) -> ChatResponse:
         # convert messages to mistral ChatMessage
-        from mistralai.client import ChatMessage as mistral_chatmessage
+        from mistralai.models.chat_completion import ChatMessage as mistral_chatmessage
 
         messages = [
             mistral_chatmessage(role=x.role, content=x.content) for x in messages
@@ -199,7 +199,7 @@ class MistralAI(LLM):
         self, messages: Sequence[ChatMessage], **kwargs: Any
     ) -> ChatResponseGen:
         # convert messages to mistral ChatMessage
-        from mistralai.client import ChatMessage as mistral_chatmessage
+        from mistralai.models.chat_completion import ChatMessage as mistral_chatmessage
 
         messages = [
             mistral_chatmessage(role=message.role, content=message.content)
@@ -237,7 +237,7 @@ class MistralAI(LLM):
         self, messages: Sequence[ChatMessage], **kwargs: Any
     ) -> ChatResponse:
         # convert messages to mistral ChatMessage
-        from mistralai.client import ChatMessage as mistral_chatmessage
+        from mistralai.models.chat_completion import ChatMessage as mistral_chatmessage
 
         messages = [
             mistral_chatmessage(role=message.role, content=message.content)
@@ -264,7 +264,7 @@ class MistralAI(LLM):
         self, messages: Sequence[ChatMessage], **kwargs: Any
     ) -> ChatResponseAsyncGen:
         # convert messages to mistral ChatMessage
-        from mistralai.client import ChatMessage as mistral_chatmessage
+        from mistralai.models.chat_completion import ChatMessage as mistral_chatmessage
 
         messages = [
             mistral_chatmessage(role=x.role, content=x.content) for x in messages
diff --git a/llama-index-integrations/llms/llama-index-llms-mistralai/pyproject.toml b/llama-index-integrations/llms/llama-index-llms-mistralai/pyproject.toml
index d538db4ae37c240b50bcb4b91d99ef219abc7f60..1069342f727afd05446c04d7b7ef477fd2a2c139 100644
--- a/llama-index-integrations/llms/llama-index-llms-mistralai/pyproject.toml
+++ b/llama-index-integrations/llms/llama-index-llms-mistralai/pyproject.toml
@@ -27,7 +27,7 @@ exclude = ["**/BUILD"]
 license = "MIT"
 name = "llama-index-llms-mistralai"
 readme = "README.md"
-version = "0.1.5"
+version = "0.1.6"
 
 [tool.poetry.dependencies]
 python = ">=3.9,<4.0"