Skip to content
Snippets Groups Projects
Unverified Commit cd70d3b4 authored by George He's avatar George He Committed by GitHub
Browse files

Improve error logs for llamacloud indices (#17827)

parent 7cd9f2a9
No related branches found
No related tags found
No related merge requests found
...@@ -12,6 +12,7 @@ from typing import Any, List, Optional, Sequence, Type ...@@ -12,6 +12,7 @@ from typing import Any, List, Optional, Sequence, Type
from urllib.parse import quote_plus from urllib.parse import quote_plus
from llama_cloud import ( from llama_cloud import (
ManagedIngestionStatusResponse,
PipelineCreate, PipelineCreate,
PipelineCreateEmbeddingConfig, PipelineCreateEmbeddingConfig,
PipelineCreateTransformConfig, PipelineCreateTransformConfig,
...@@ -165,31 +166,41 @@ class LlamaCloudIndex(BaseManagedIndex): ...@@ -165,31 +166,41 @@ class LlamaCloudIndex(BaseManagedIndex):
self, self,
verbose: bool = False, verbose: bool = False,
raise_on_partial_success: bool = False, raise_on_partial_success: bool = False,
) -> None: sleep_interval: float = 0.5,
) -> Optional[ManagedIngestionStatusResponse]:
if sleep_interval < 0.5:
# minimum sleep interval at 0.5 seconds to prevent rate-limiting
sleep_interval = 0.5
if verbose: if verbose:
print("Syncing pipeline: ", end="") print(f"Syncing pipeline {self.pipeline.id}: ", end="")
is_done = False is_done = False
status_response: Optional[ManagedIngestionStatusResponse] = None
while not is_done: while not is_done:
status = self._client.pipelines.get_pipeline_status( status_response = self._client.pipelines.get_pipeline_status(
pipeline_id=self.pipeline.id pipeline_id=self.pipeline.id
).status )
status = status_response.status
if status == ManagedIngestionStatus.ERROR or ( if status == ManagedIngestionStatus.ERROR or (
raise_on_partial_success raise_on_partial_success
and status == ManagedIngestionStatus.PARTIAL_SUCCESS and status == ManagedIngestionStatus.PARTIAL_SUCCESS
): ):
raise ValueError(f"Pipeline ingestion failed for {self.pipeline.id}") error_details = status_response.json()
raise ValueError(
f"Pipeline ingestion failed for {self.pipeline.id}. Error details: {error_details}"
)
elif status in [ elif status in [
ManagedIngestionStatus.NOT_STARTED, ManagedIngestionStatus.NOT_STARTED,
ManagedIngestionStatus.IN_PROGRESS, ManagedIngestionStatus.IN_PROGRESS,
]: ]:
if verbose: if verbose:
print(".", end="") print(".", end="")
time.sleep(0.5) time.sleep(sleep_interval)
else: else:
is_done = True is_done = True
if verbose: if verbose:
print("Done!") print("Done!")
return status_response
def _wait_for_file_ingestion( def _wait_for_file_ingestion(
self, self,
......
...@@ -34,7 +34,7 @@ exclude = ["**/BUILD"] ...@@ -34,7 +34,7 @@ exclude = ["**/BUILD"]
license = "MIT" license = "MIT"
name = "llama-index-indices-managed-llama-cloud" name = "llama-index-indices-managed-llama-cloud"
readme = "README.md" readme = "README.md"
version = "0.6.4" version = "0.6.5"
[tool.poetry.dependencies] [tool.poetry.dependencies]
python = ">=3.9,<4.0" python = ">=3.9,<4.0"
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment