From 94ec6ad806a5e348bad6672cc514be5b021763cb Mon Sep 17 00:00:00 2001
From: Bill Metangmo <25366207+billmetangmo@users.noreply.github.com>
Date: Fri, 8 Mar 2024 04:05:14 +0100
Subject: [PATCH] fix: allow guideline evaluator with bedrock (#11719)

---
 llama-index-core/llama_index/core/evaluation/base.py | 10 ++++++++--
 1 file changed, 8 insertions(+), 2 deletions(-)

diff --git a/llama-index-core/llama_index/core/evaluation/base.py b/llama-index-core/llama_index/core/evaluation/base.py
index d95ae2772..8be36063b 100644
--- a/llama-index-core/llama_index/core/evaluation/base.py
+++ b/llama-index-core/llama_index/core/evaluation/base.py
@@ -95,8 +95,14 @@ class BaseEvaluator(PromptMixin):
         Subclasses can override this method to provide custom evaluation logic and
         take in additional arguments.
         """
-        return asyncio.run(
-            self.aevaluate_response(query=query, response=response, **kwargs)
+        response_str: Optional[str] = None
+        contexts: Optional[Sequence[str]] = None
+        if response is not None:
+            response_str = response.response
+            contexts = [node.get_content() for node in response.source_nodes]
+
+        return self.evaluate(
+            query=query, response=response_str, contexts=contexts, **kwargs
         )
 
     async def aevaluate_response(
-- 
GitLab