From cf4892ff3924ddb25fecf42baf1535924d99b9b0 Mon Sep 17 00:00:00 2001
From: Joel Gotsch <joel.gotsch@gmail.com>
Date: Fri, 29 Mar 2024 23:35:45 +0100
Subject: [PATCH] Fix Token Counting of Completions for async calls (#12421)

---
 llama-index-core/llama_index/core/llms/callbacks.py | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/llama-index-core/llama_index/core/llms/callbacks.py b/llama-index-core/llama_index/core/llms/callbacks.py
index d63a0327f1..e3602007d1 100644
--- a/llama-index-core/llama_index/core/llms/callbacks.py
+++ b/llama-index-core/llama_index/core/llms/callbacks.py
@@ -260,7 +260,7 @@ def llm_completion_callback() -> Callable:
                         CBEventType.LLM,
                         payload={
                             EventPayload.PROMPT: args[0],
-                            EventPayload.RESPONSE: f_return_val,
+                            EventPayload.COMPLETION: f_return_val,
                         },
                         event_id=event_id,
                     )
-- 
GitLab