From cf4892ff3924ddb25fecf42baf1535924d99b9b0 Mon Sep 17 00:00:00 2001 From: Joel Gotsch <joel.gotsch@gmail.com> Date: Fri, 29 Mar 2024 23:35:45 +0100 Subject: [PATCH] Fix Token Counting of Completions for async calls (#12421) --- llama-index-core/llama_index/core/llms/callbacks.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama-index-core/llama_index/core/llms/callbacks.py b/llama-index-core/llama_index/core/llms/callbacks.py index d63a0327f1..e3602007d1 100644 --- a/llama-index-core/llama_index/core/llms/callbacks.py +++ b/llama-index-core/llama_index/core/llms/callbacks.py @@ -260,7 +260,7 @@ def llm_completion_callback() -> Callable: CBEventType.LLM, payload={ EventPayload.PROMPT: args[0], - EventPayload.RESPONSE: f_return_val, + EventPayload.COMPLETION: f_return_val, }, event_id=event_id, ) -- GitLab