Skip to content
Snippets Groups Projects
Commit e755a632 authored by Elliot Kang's avatar Elliot Kang
Browse files

fixed example based on new interface

parent 29c6b62b
No related branches found
No related tags found
No related merge requests found
...@@ -7,7 +7,9 @@ import { ...@@ -7,7 +7,9 @@ import {
import { ChatMessage, MessageType, OpenAI } from "../packages/core/src/llm/LLM"; import { ChatMessage, MessageType, OpenAI } from "../packages/core/src/llm/LLM";
async function main() { async function main() {
const query: string = "Where is Istanbul?"; const query: string = `
Where is Istanbul?
`;
const llm = new OpenAI({ model: "gpt-3.5-turbo", temperature: 0.1 }); const llm = new OpenAI({ model: "gpt-3.5-turbo", temperature: 0.1 });
const message: ChatMessage = { content: query, role: "user" }; const message: ChatMessage = { content: query, role: "user" };
...@@ -19,26 +21,14 @@ async function main() { ...@@ -19,26 +21,14 @@ async function main() {
//GPT 3.5 Turbo uses CL100K_Base encodings, check your LLM to see which tokenizer it uses. //GPT 3.5 Turbo uses CL100K_Base encodings, check your LLM to see which tokenizer it uses.
const encoding = tiktoken.getEncoding("cl100k_base"); const encoding = tiktoken.getEncoding("cl100k_base");
const callback: CallbackManager = new CallbackManager(); //Stream Complete
callback.onLLMStream = (callback_response) => { //Note: Setting streaming flag to true or false will auto-set your return type to
//Token text //either an AsyncGenerator or a Response.
const text = callback_response.token.choices[0].delta.content // Omitting the streaming flag automatically sets streaming to false
? callback_response.token.choices[0].delta.content
: "";
//Increment total number of tokens
total_tokens += encoding.encode(text).length;
};
llm.callbackManager = callback;
//Create a dummy event to trigger our Stream Callback // const stream2 = await llm.chat([message], undefined);
const dummy_event: Event = { const stream = await llm.complete(query, undefined, true);
id: "something",
type: "intermediate" as EventType,
};
//Stream Complete
const stream = llm.stream_complete(query, dummy_event);
for await (const part of stream) { for await (const part of stream) {
//This only gives you the string part of a stream //This only gives you the string part of a stream
...@@ -49,13 +39,11 @@ async function main() { ...@@ -49,13 +39,11 @@ async function main() {
const correct_total_tokens: number = const correct_total_tokens: number =
encoding.encode(accumulated_result).length; encoding.encode(accumulated_result).length;
console.log(accumulated_result);
//Check if our stream token counter works //Check if our stream token counter works
console.log( console.log(
`Output token total using tokenizer on accumulated output: ${correct_total_tokens}`, `Output token total using tokenizer on accumulated output: ${correct_total_tokens}`,
); );
console.log(
`Output token total using tokenizer on stream output: ${total_tokens}`,
);
} }
main(); main();
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment