From 94ec6ad806a5e348bad6672cc514be5b021763cb Mon Sep 17 00:00:00 2001 From: Bill Metangmo <25366207+billmetangmo@users.noreply.github.com> Date: Fri, 8 Mar 2024 04:05:14 +0100 Subject: [PATCH] fix: allow guideline evaluator with bedrock (#11719) --- llama-index-core/llama_index/core/evaluation/base.py | 10 ++++++++-- 1 file changed, 8 insertions(+), 2 deletions(-) diff --git a/llama-index-core/llama_index/core/evaluation/base.py b/llama-index-core/llama_index/core/evaluation/base.py index d95ae2772..8be36063b 100644 --- a/llama-index-core/llama_index/core/evaluation/base.py +++ b/llama-index-core/llama_index/core/evaluation/base.py @@ -95,8 +95,14 @@ class BaseEvaluator(PromptMixin): Subclasses can override this method to provide custom evaluation logic and take in additional arguments. """ - return asyncio.run( - self.aevaluate_response(query=query, response=response, **kwargs) + response_str: Optional[str] = None + contexts: Optional[Sequence[str]] = None + if response is not None: + response_str = response.response + contexts = [node.get_content() for node in response.source_nodes] + + return self.evaluate( + query=query, response=response_str, contexts=contexts, **kwargs ) async def aevaluate_response( -- GitLab