From 5c5f4c1c84c7a2acc51309d865ab81afbaae4747 Mon Sep 17 00:00:00 2001 From: Marcus Schiesser <mail@marcusschiesser.de> Date: Mon, 1 Apr 2024 13:52:07 +0800 Subject: [PATCH] Revert "feat: support calculate llama 2 tokens (#676)" This reverts commit 041acd11fe913593516ba4782684dc5f5b7b6f5f. --- .changeset/khaki-books-type.md | 6 ------ packages/core/package.json | 5 ++--- packages/core/src/llm/LLM.ts | 4 +--- packages/edge/package.json | 5 ++--- pnpm-lock.yaml | 10 ---------- 5 files changed, 5 insertions(+), 25 deletions(-) delete mode 100644 .changeset/khaki-books-type.md diff --git a/.changeset/khaki-books-type.md b/.changeset/khaki-books-type.md deleted file mode 100644 index 4954897a9..000000000 --- a/.changeset/khaki-books-type.md +++ /dev/null @@ -1,6 +0,0 @@ ---- -"llamaindex": patch -"@llamaindex/edge": patch ---- - -feat: support calculate Llama2 model tokens size diff --git a/packages/core/package.json b/packages/core/package.json index 7f83c8c69..1261452d1 100644 --- a/packages/core/package.json +++ b/packages/core/package.json @@ -25,7 +25,6 @@ "chromadb": "~1.7.3", "cohere-ai": "^7.7.5", "js-tiktoken": "^1.0.10", - "llama-tokenizer-js": "^1.2.1", "lodash": "^4.17.21", "magic-bytes.js": "^1.10.0", "mammoth": "^1.6.0", @@ -42,8 +41,8 @@ "rake-modified": "^1.0.8", "replicate": "^0.25.2", "string-strip-html": "^13.4.6", - "wikipedia": "^2.1.2", - "wink-nlp": "^1.14.3" + "wink-nlp": "^1.14.3", + "wikipedia": "^2.1.2" }, "devDependencies": { "@swc/cli": "^0.3.9", diff --git a/packages/core/src/llm/LLM.ts b/packages/core/src/llm/LLM.ts index 2f6c69d5a..6970cab33 100644 --- a/packages/core/src/llm/LLM.ts +++ b/packages/core/src/llm/LLM.ts @@ -8,7 +8,6 @@ import { type StreamCallbackResponse, } from "../callbacks/CallbackManager.js"; -import llamaTokenizer from "llama-tokenizer-js"; import type { ChatCompletionMessageParam } from "openai/resources/index.js"; import type { LLMOptions } from "portkey-ai"; import { Tokenizers, globalsHelper } from "../GlobalsHelper.js"; @@ -416,8 +415,7 @@ export class LlamaDeuce extends BaseLLM { } tokens(messages: ChatMessage[]): number { - return llamaTokenizer.encode(this.mapMessagesToPrompt(messages).prompt) - .length; + throw new Error("Method not implemented."); } get metadata() { diff --git a/packages/edge/package.json b/packages/edge/package.json index afba58f29..8edf4ba89 100644 --- a/packages/edge/package.json +++ b/packages/edge/package.json @@ -24,7 +24,6 @@ "chromadb": "~1.7.3", "cohere-ai": "^7.7.5", "js-tiktoken": "^1.0.10", - "llama-tokenizer-js": "^1.2.1", "lodash": "^4.17.21", "magic-bytes.js": "^1.10.0", "mammoth": "^1.6.0", @@ -41,8 +40,8 @@ "rake-modified": "^1.0.8", "replicate": "^0.25.2", "string-strip-html": "^13.4.6", - "wikipedia": "^2.1.2", - "wink-nlp": "^1.14.3" + "wink-nlp": "^1.14.3", + "wikipedia": "^2.1.2" }, "engines": { "node": ">=18.0.0" diff --git a/pnpm-lock.yaml b/pnpm-lock.yaml index 9b6a7a666..4b69ec290 100644 --- a/pnpm-lock.yaml +++ b/pnpm-lock.yaml @@ -228,9 +228,6 @@ importers: js-tiktoken: specifier: ^1.0.10 version: 1.0.10 - llama-tokenizer-js: - specifier: ^1.2.1 - version: 1.2.1 lodash: specifier: ^4.17.21 version: 4.17.21 @@ -376,9 +373,6 @@ importers: js-tiktoken: specifier: ^1.0.10 version: 1.0.10 - llama-tokenizer-js: - specifier: ^1.2.1 - version: 1.2.1 lodash: specifier: ^4.17.21 version: 4.17.21 @@ -9809,10 +9803,6 @@ packages: wrap-ansi: 9.0.0 dev: true - /llama-tokenizer-js@1.2.1: - resolution: {integrity: sha512-SEVVc++cXR0D0Wv30AzMVWzPCAKM701vZYU31h5lCTIn4k5cfZpJ070YDcb2nPq2Ts3xgu44L19wIrq1z/XjXQ==} - dev: false - /load-yaml-file@0.2.0: resolution: {integrity: sha512-OfCBkGEw4nN6JLtgRidPX6QxjBQGQf72q3si2uvqyFEMbycSFFHwAZeXx6cJgFM9wmLrf9zBwCP3Ivqa+LLZPw==} engines: {node: '>=6'} -- GitLab