diff --git a/server/utils/AiProviders/openRouter/models.js b/server/utils/AiProviders/openRouter/models.js index c920b88a412c640cb9a3b94946f58a9ece9ba8e3..4c2d7946cdd3d85102f2197bcaf19cf85318d492 100644 --- a/server/utils/AiProviders/openRouter/models.js +++ b/server/utils/AiProviders/openRouter/models.js @@ -1,10 +1,4 @@ const MODELS = { - "nousresearch/nous-capybara-34b": { - id: "nousresearch/nous-capybara-34b", - name: "Nous: Capybara 34B", - organization: "Nousresearch", - maxLength: 32768, - }, "openrouter/auto": { id: "openrouter/auto", name: "Auto (best for prompt)", @@ -21,6 +15,12 @@ const MODELS = { id: "mistralai/mistral-7b-instruct:free", name: "Mistral 7B Instruct (free)", organization: "Mistralai", + maxLength: 32768, + }, + "openchat/openchat-7b:free": { + id: "openchat/openchat-7b:free", + name: "OpenChat 3.5 (free)", + organization: "Openchat", maxLength: 8192, }, "gryphe/mythomist-7b:free": { @@ -45,13 +45,7 @@ const MODELS = { id: "google/gemma-7b-it:free", name: "Google: Gemma 7B (free)", organization: "Google", - maxLength: 8000, - }, - "jondurbin/bagel-34b": { - id: "jondurbin/bagel-34b", - name: "Bagel 34B v0.2", - organization: "Jondurbin", - maxLength: 8000, + maxLength: 8192, }, "jebcarter/psyfighter-13b": { id: "jebcarter/psyfighter-13b", @@ -65,54 +59,12 @@ const MODELS = { organization: "Koboldai", maxLength: 4096, }, - "neversleep/noromaid-mixtral-8x7b-instruct": { - id: "neversleep/noromaid-mixtral-8x7b-instruct", - name: "Noromaid Mixtral 8x7B Instruct", - organization: "Neversleep", - maxLength: 8000, - }, - "nousresearch/nous-hermes-llama2-13b": { - id: "nousresearch/nous-hermes-llama2-13b", - name: "Nous: Hermes 13B", - organization: "Nousresearch", - maxLength: 4096, - }, - "meta-llama/codellama-34b-instruct": { - id: "meta-llama/codellama-34b-instruct", - name: "Meta: CodeLlama 34B Instruct", - organization: "Meta-llama", - maxLength: 8192, - }, - "phind/phind-codellama-34b": { - id: "phind/phind-codellama-34b", - name: "Phind: CodeLlama 34B v2", - organization: "Phind", - maxLength: 4096, - }, "intel/neural-chat-7b": { id: "intel/neural-chat-7b", name: "Neural Chat 7B v3.1", organization: "Intel", maxLength: 4096, }, - "mistralai/mixtral-8x7b-instruct": { - id: "mistralai/mixtral-8x7b-instruct", - name: "Mistral: Mixtral 8x7B Instruct", - organization: "Mistralai", - maxLength: 32768, - }, - "nousresearch/nous-hermes-2-mixtral-8x7b-dpo": { - id: "nousresearch/nous-hermes-2-mixtral-8x7b-dpo", - name: "Nous: Hermes 2 Mixtral 8x7B DPO", - organization: "Nousresearch", - maxLength: 32000, - }, - "nousresearch/nous-hermes-2-mixtral-8x7b-sft": { - id: "nousresearch/nous-hermes-2-mixtral-8x7b-sft", - name: "Nous: Hermes 2 Mixtral 8x7B SFT", - organization: "Nousresearch", - maxLength: 32000, - }, "haotian-liu/llava-13b": { id: "haotian-liu/llava-13b", name: "Llava 13B", @@ -143,30 +95,12 @@ const MODELS = { organization: "Pygmalionai", maxLength: 8192, }, - "undi95/remm-slerp-l2-13b-6k": { - id: "undi95/remm-slerp-l2-13b-6k", - name: "ReMM SLERP 13B 6k", - organization: "Undi95", - maxLength: 6144, - }, - "gryphe/mythomax-l2-13b": { - id: "gryphe/mythomax-l2-13b", - name: "MythoMax 13B", - organization: "Gryphe", - maxLength: 4096, - }, "xwin-lm/xwin-lm-70b": { id: "xwin-lm/xwin-lm-70b", name: "Xwin 70B", organization: "Xwin-lm", maxLength: 8192, }, - "gryphe/mythomax-l2-13b-8k": { - id: "gryphe/mythomax-l2-13b-8k", - name: "MythoMax 13B 8k", - organization: "Gryphe", - maxLength: 8192, - }, "alpindale/goliath-120b": { id: "alpindale/goliath-120b", name: "Goliath 120B", @@ -185,15 +119,27 @@ const MODELS = { organization: "Gryphe", maxLength: 32768, }, + "sophosympatheia/midnight-rose-70b": { + id: "sophosympatheia/midnight-rose-70b", + name: "Midnight Rose 70B", + organization: "Sophosympatheia", + maxLength: 4096, + }, + "undi95/remm-slerp-l2-13b:extended": { + id: "undi95/remm-slerp-l2-13b:extended", + name: "ReMM SLERP 13B (extended)", + organization: "Undi95", + maxLength: 6144, + }, "mancer/weaver": { id: "mancer/weaver", name: "Mancer: Weaver (alpha)", organization: "Mancer", maxLength: 8000, }, - "nousresearch/nous-hermes-llama2-70b": { - id: "nousresearch/nous-hermes-llama2-70b", - name: "Nous: Hermes 70B", + "nousresearch/nous-hermes-llama2-13b": { + id: "nousresearch/nous-hermes-llama2-13b", + name: "Nous: Hermes 13B", organization: "Nousresearch", maxLength: 4096, }, @@ -203,12 +149,24 @@ const MODELS = { organization: "Nousresearch", maxLength: 4096, }, + "meta-llama/codellama-34b-instruct": { + id: "meta-llama/codellama-34b-instruct", + name: "Meta: CodeLlama 34B Instruct", + organization: "Meta-llama", + maxLength: 8192, + }, "codellama/codellama-70b-instruct": { id: "codellama/codellama-70b-instruct", name: "Meta: CodeLlama 70B Instruct", organization: "Codellama", maxLength: 2048, }, + "phind/phind-codellama-34b": { + id: "phind/phind-codellama-34b", + name: "Phind: CodeLlama 34B v2", + organization: "Phind", + maxLength: 4096, + }, "teknium/openhermes-2-mistral-7b": { id: "teknium/openhermes-2-mistral-7b", name: "OpenHermes 2 Mistral 7B", @@ -227,12 +185,6 @@ const MODELS = { organization: "Undi95", maxLength: 4096, }, - "undi95/toppy-m-7b": { - id: "undi95/toppy-m-7b", - name: "Toppy M 7B", - organization: "Undi95", - maxLength: 4096, - }, "openrouter/cinematika-7b": { id: "openrouter/cinematika-7b", name: "Cinematika 7B (alpha)", @@ -271,7 +223,7 @@ const MODELS = { }, "mistralai/mixtral-8x7b": { id: "mistralai/mixtral-8x7b", - name: "Mistral: Mixtral 8x7B (base)", + name: "Mixtral 8x7B (base)", organization: "Mistralai", maxLength: 32768, }, @@ -281,6 +233,12 @@ const MODELS = { organization: "Nousresearch", maxLength: 4096, }, + "nousresearch/nous-hermes-2-mixtral-8x7b-sft": { + id: "nousresearch/nous-hermes-2-mixtral-8x7b-sft", + name: "Nous: Hermes 2 Mixtral 8x7B SFT", + organization: "Nousresearch", + maxLength: 32000, + }, "nousresearch/nous-hermes-2-mistral-7b-dpo": { id: "nousresearch/nous-hermes-2-mistral-7b-dpo", name: "Nous: Hermes 2 Mistral 7B DPO", @@ -303,7 +261,7 @@ const MODELS = { id: "openai/gpt-3.5-turbo", name: "OpenAI: GPT-3.5 Turbo", organization: "Openai", - maxLength: 4095, + maxLength: 16385, }, "openai/gpt-3.5-turbo-0125": { id: "openai/gpt-3.5-turbo-0125", @@ -335,9 +293,15 @@ const MODELS = { organization: "Openai", maxLength: 16385, }, + "openai/gpt-4-turbo": { + id: "openai/gpt-4-turbo", + name: "OpenAI: GPT-4 Turbo", + organization: "Openai", + maxLength: 128000, + }, "openai/gpt-4-turbo-preview": { id: "openai/gpt-4-turbo-preview", - name: "OpenAI: GPT-4 Turbo (preview)", + name: "OpenAI: GPT-4 Turbo Preview", organization: "Openai", maxLength: 128000, }, @@ -373,7 +337,7 @@ const MODELS = { }, "openai/gpt-4-vision-preview": { id: "openai/gpt-4-vision-preview", - name: "OpenAI: GPT-4 Vision (preview)", + name: "OpenAI: GPT-4 Vision", organization: "Openai", maxLength: 128000, }, @@ -387,37 +351,43 @@ const MODELS = { id: "google/palm-2-chat-bison", name: "Google: PaLM 2 Chat", organization: "Google", - maxLength: 36864, + maxLength: 25804, }, "google/palm-2-codechat-bison": { id: "google/palm-2-codechat-bison", name: "Google: PaLM 2 Code Chat", organization: "Google", - maxLength: 28672, + maxLength: 20070, }, "google/palm-2-chat-bison-32k": { id: "google/palm-2-chat-bison-32k", name: "Google: PaLM 2 Chat 32k", organization: "Google", - maxLength: 131072, + maxLength: 91750, }, "google/palm-2-codechat-bison-32k": { id: "google/palm-2-codechat-bison-32k", name: "Google: PaLM 2 Code Chat 32k", organization: "Google", - maxLength: 131072, + maxLength: 91750, }, "google/gemini-pro": { id: "google/gemini-pro", - name: "Google: Gemini Pro (preview)", + name: "Google: Gemini Pro 1.0", organization: "Google", - maxLength: 131040, + maxLength: 91728, }, "google/gemini-pro-vision": { id: "google/gemini-pro-vision", - name: "Google: Gemini Pro Vision (preview)", + name: "Google: Gemini Pro Vision 1.0", organization: "Google", - maxLength: 65536, + maxLength: 45875, + }, + "google/gemini-pro-1.5": { + id: "google/gemini-pro-1.5", + name: "Google: Gemini Pro 1.5 (preview)", + organization: "Google", + maxLength: 2800000, }, "perplexity/pplx-70b-online": { id: "perplexity/pplx-70b-online", @@ -443,18 +413,96 @@ const MODELS = { organization: "Perplexity", maxLength: 4096, }, + "perplexity/sonar-small-chat": { + id: "perplexity/sonar-small-chat", + name: "Perplexity: Sonar 7B", + organization: "Perplexity", + maxLength: 16384, + }, + "perplexity/sonar-medium-chat": { + id: "perplexity/sonar-medium-chat", + name: "Perplexity: Sonar 8x7B", + organization: "Perplexity", + maxLength: 16384, + }, + "perplexity/sonar-small-online": { + id: "perplexity/sonar-small-online", + name: "Perplexity: Sonar 7B Online", + organization: "Perplexity", + maxLength: 12000, + }, + "perplexity/sonar-medium-online": { + id: "perplexity/sonar-medium-online", + name: "Perplexity: Sonar 8x7B Online", + organization: "Perplexity", + maxLength: 12000, + }, + "fireworks/mixtral-8x22b-instruct-preview": { + id: "fireworks/mixtral-8x22b-instruct-preview", + name: "Fireworks Mixtral 8x22B Instruct OH (preview)", + organization: "Fireworks", + maxLength: 8192, + }, + "anthropic/claude-3-opus": { + id: "anthropic/claude-3-opus", + name: "Anthropic: Claude 3 Opus", + organization: "Anthropic", + maxLength: 200000, + }, + "anthropic/claude-3-sonnet": { + id: "anthropic/claude-3-sonnet", + name: "Anthropic: Claude 3 Sonnet", + organization: "Anthropic", + maxLength: 200000, + }, + "anthropic/claude-3-haiku": { + id: "anthropic/claude-3-haiku", + name: "Anthropic: Claude 3 Haiku", + organization: "Anthropic", + maxLength: 200000, + }, + "anthropic/claude-3-opus:beta": { + id: "anthropic/claude-3-opus:beta", + name: "Anthropic: Claude 3 Opus (self-moderated)", + organization: "Anthropic", + maxLength: 200000, + }, + "anthropic/claude-3-sonnet:beta": { + id: "anthropic/claude-3-sonnet:beta", + name: "Anthropic: Claude 3 Sonnet (self-moderated)", + organization: "Anthropic", + maxLength: 200000, + }, + "anthropic/claude-3-haiku:beta": { + id: "anthropic/claude-3-haiku:beta", + name: "Anthropic: Claude 3 Haiku (self-moderated)", + organization: "Anthropic", + maxLength: 200000, + }, "meta-llama/llama-2-70b-chat": { id: "meta-llama/llama-2-70b-chat", name: "Meta: Llama v2 70B Chat", organization: "Meta-llama", maxLength: 4096, }, + "nousresearch/nous-capybara-34b": { + id: "nousresearch/nous-capybara-34b", + name: "Nous: Capybara 34B", + organization: "Nousresearch", + maxLength: 32768, + }, "jondurbin/airoboros-l2-70b": { id: "jondurbin/airoboros-l2-70b", name: "Airoboros 70B", organization: "Jondurbin", maxLength: 4096, }, + "jondurbin/bagel-34b": { + id: "jondurbin/bagel-34b", + name: "Bagel 34B v0.2", + organization: "Jondurbin", + maxLength: 8000, + }, "austism/chronos-hermes-13b": { id: "austism/chronos-hermes-13b", name: "Chronos Hermes 13B v2", @@ -465,7 +513,13 @@ const MODELS = { id: "mistralai/mistral-7b-instruct", name: "Mistral 7B Instruct", organization: "Mistralai", - maxLength: 8192, + maxLength: 32768, + }, + "gryphe/mythomax-l2-13b": { + id: "gryphe/mythomax-l2-13b", + name: "MythoMax 13B", + organization: "Gryphe", + maxLength: 4096, }, "openchat/openchat-7b": { id: "openchat/openchat-7b", @@ -473,18 +527,42 @@ const MODELS = { organization: "Openchat", maxLength: 8192, }, + "undi95/toppy-m-7b": { + id: "undi95/toppy-m-7b", + name: "Toppy M 7B", + organization: "Undi95", + maxLength: 4096, + }, "lizpreciatior/lzlv-70b-fp16-hf": { id: "lizpreciatior/lzlv-70b-fp16-hf", name: "lzlv 70B", organization: "Lizpreciatior", maxLength: 4096, }, + "mistralai/mixtral-8x7b-instruct": { + id: "mistralai/mixtral-8x7b-instruct", + name: "Mixtral 8x7B Instruct", + organization: "Mistralai", + maxLength: 32768, + }, "cognitivecomputations/dolphin-mixtral-8x7b": { id: "cognitivecomputations/dolphin-mixtral-8x7b", name: "Dolphin 2.6 Mixtral 8x7B ðŸ¬", organization: "Cognitivecomputations", maxLength: 32000, }, + "neversleep/noromaid-mixtral-8x7b-instruct": { + id: "neversleep/noromaid-mixtral-8x7b-instruct", + name: "Noromaid Mixtral 8x7B Instruct", + organization: "Neversleep", + maxLength: 8000, + }, + "nousresearch/nous-hermes-2-mixtral-8x7b-dpo": { + id: "nousresearch/nous-hermes-2-mixtral-8x7b-dpo", + name: "Nous: Hermes 2 Mixtral 8x7B DPO", + organization: "Nousresearch", + maxLength: 32000, + }, "rwkv/rwkv-5-world-3b": { id: "rwkv/rwkv-5-world-3b", name: "RWKV v5 World 3B", @@ -507,7 +585,19 @@ const MODELS = { id: "google/gemma-7b-it", name: "Google: Gemma 7B", organization: "Google", - maxLength: 8000, + maxLength: 8192, + }, + "databricks/dbrx-instruct": { + id: "databricks/dbrx-instruct", + name: "Databricks: DBRX 132B Instruct", + organization: "Databricks", + maxLength: 32768, + }, + "huggingfaceh4/zephyr-orpo-141b-a35b": { + id: "huggingfaceh4/zephyr-orpo-141b-a35b", + name: "Zephyr 141B-A35B", + organization: "Huggingfaceh4", + maxLength: 65536, }, "anthropic/claude-2": { id: "anthropic/claude-2", @@ -565,58 +655,124 @@ const MODELS = { }, "anthropic/claude-2:beta": { id: "anthropic/claude-2:beta", - name: "Anthropic: Claude v2 (experimental)", + name: "Anthropic: Claude v2 (self-moderated)", organization: "Anthropic", maxLength: 200000, }, "anthropic/claude-2.1:beta": { id: "anthropic/claude-2.1:beta", - name: "Anthropic: Claude v2.1 (experimental)", + name: "Anthropic: Claude v2.1 (self-moderated)", organization: "Anthropic", maxLength: 200000, }, "anthropic/claude-2.0:beta": { id: "anthropic/claude-2.0:beta", - name: "Anthropic: Claude v2.0 (experimental)", + name: "Anthropic: Claude v2.0 (self-moderated)", organization: "Anthropic", maxLength: 100000, }, "anthropic/claude-instant-1:beta": { id: "anthropic/claude-instant-1:beta", - name: "Anthropic: Claude Instant v1 (experimental)", + name: "Anthropic: Claude Instant v1 (self-moderated)", organization: "Anthropic", maxLength: 100000, }, + "mistralai/mixtral-8x22b": { + id: "mistralai/mixtral-8x22b", + name: "Mistral: Mixtral 8x22B (base)", + organization: "Mistralai", + maxLength: 65536, + }, "huggingfaceh4/zephyr-7b-beta:free": { id: "huggingfaceh4/zephyr-7b-beta:free", name: "Hugging Face: Zephyr 7B (free)", organization: "Huggingfaceh4", maxLength: 4096, }, - "openchat/openchat-7b:free": { - id: "openchat/openchat-7b:free", - name: "OpenChat 3.5 (free)", - organization: "Openchat", + "mistralai/mixtral-8x7b-instruct:nitro": { + id: "mistralai/mixtral-8x7b-instruct:nitro", + name: "Mixtral 8x7B Instruct (nitro)", + organization: "Mistralai", + maxLength: 32768, + }, + "meta-llama/llama-2-70b-chat:nitro": { + id: "meta-llama/llama-2-70b-chat:nitro", + name: "Meta: Llama v2 70B Chat (nitro)", + organization: "Meta-llama", + maxLength: 4096, + }, + "gryphe/mythomax-l2-13b:nitro": { + id: "gryphe/mythomax-l2-13b:nitro", + name: "MythoMax 13B (nitro)", + organization: "Gryphe", + maxLength: 4096, + }, + "mistralai/mistral-7b-instruct:nitro": { + id: "mistralai/mistral-7b-instruct:nitro", + name: "Mistral 7B Instruct (nitro)", + organization: "Mistralai", + maxLength: 32768, + }, + "google/gemma-7b-it:nitro": { + id: "google/gemma-7b-it:nitro", + name: "Google: Gemma 7B (nitro)", + organization: "Google", + maxLength: 8192, + }, + "databricks/dbrx-instruct:nitro": { + id: "databricks/dbrx-instruct:nitro", + name: "Databricks: DBRX 132B Instruct (nitro)", + organization: "Databricks", + maxLength: 32768, + }, + "gryphe/mythomax-l2-13b:extended": { + id: "gryphe/mythomax-l2-13b:extended", + name: "MythoMax 13B (extended)", + organization: "Gryphe", maxLength: 8192, }, "mistralai/mistral-tiny": { id: "mistralai/mistral-tiny", - name: "Mistral: Tiny", + name: "Mistral Tiny", organization: "Mistralai", maxLength: 32000, }, "mistralai/mistral-small": { id: "mistralai/mistral-small", - name: "Mistral: Small", + name: "Mistral Small", organization: "Mistralai", maxLength: 32000, }, "mistralai/mistral-medium": { id: "mistralai/mistral-medium", - name: "Mistral: Medium", + name: "Mistral Medium", + organization: "Mistralai", + maxLength: 32000, + }, + "mistralai/mistral-large": { + id: "mistralai/mistral-large", + name: "Mistral Large", organization: "Mistralai", maxLength: 32000, }, + "cohere/command": { + id: "cohere/command", + name: "Cohere: Command", + organization: "Cohere", + maxLength: 4096, + }, + "cohere/command-r": { + id: "cohere/command-r", + name: "Cohere: Command R", + organization: "Cohere", + maxLength: 128000, + }, + "cohere/command-r-plus": { + id: "cohere/command-r-plus", + name: "Cohere: Command R+", + organization: "Cohere", + maxLength: 128000, + }, }; module.exports.MODELS = MODELS; diff --git a/server/utils/AiProviders/openRouter/scripts/parse.mjs b/server/utils/AiProviders/openRouter/scripts/parse.mjs index fb3b562b5cbbaf75507f4215d46767d1b8d841ad..11c67b22c6ead5c688fa9d7937b62776176b87ed 100644 --- a/server/utils/AiProviders/openRouter/scripts/parse.mjs +++ b/server/utils/AiProviders/openRouter/scripts/parse.mjs @@ -6,7 +6,7 @@ // copy outputs into the export in ../models.js // Update the date below if you run this again because OpenRouter added new models. -// Last Collected: Feb 23, 2024 +// Last Collected: Apr 14, 2024 import fs from "fs"; diff --git a/server/utils/AiProviders/perplexity/models.js b/server/utils/AiProviders/perplexity/models.js index 95cd8eac79a1a7f4daafb59e49c30c632e53dce5..8bed2a5a0958bb2368b556a4e8d2cb7a471d4055 100644 --- a/server/utils/AiProviders/perplexity/models.js +++ b/server/utils/AiProviders/perplexity/models.js @@ -19,21 +19,11 @@ const MODELS = { name: "sonar-medium-online", maxLength: 12000, }, - "codellama-34b-instruct": { - id: "codellama-34b-instruct", - name: "codellama-34b-instruct", - maxLength: 16384, - }, "codellama-70b-instruct": { id: "codellama-70b-instruct", name: "codellama-70b-instruct", maxLength: 16384, }, - "llama-2-70b-chat": { - id: "llama-2-70b-chat", - name: "llama-2-70b-chat", - maxLength: 4096, - }, "mistral-7b-instruct": { id: "mistral-7b-instruct", name: "mistral-7b-instruct", @@ -44,26 +34,6 @@ const MODELS = { name: "mixtral-8x7b-instruct", maxLength: 16384, }, - "pplx-7b-chat": { - id: "pplx-7b-chat", - name: "pplx-7b-chat", - maxLength: 16384, - }, - "pplx-7b-online": { - id: "pplx-7b-online", - name: "pplx-7b-online", - maxLength: 12000, - }, - "pplx-70b-chat": { - id: "pplx-70b-chat", - name: "pplx-70b-chat", - maxLength: 8192, - }, - "pplx-70b-online": { - id: "pplx-70b-online", - name: "pplx-70b-online", - maxLength: 4000, - }, }; module.exports.MODELS = MODELS; diff --git a/server/utils/AiProviders/perplexity/scripts/chat_models.txt b/server/utils/AiProviders/perplexity/scripts/chat_models.txt index 97ba9017a423535af8951699609ae2c3bf8aecfa..41fce0f016799b01b56e25c8240cd88eb43a6a26 100644 --- a/server/utils/AiProviders/perplexity/scripts/chat_models.txt +++ b/server/utils/AiProviders/perplexity/scripts/chat_models.txt @@ -1,15 +1,9 @@ -| Model | Parameter Count | Context Length | Model Type | -| :-------------------------- | :-------------- | :------------- | :-------------- | -| `sonar-small-chat` | 7B | 16384 | Chat Completion | -| `sonar-small-online` | 7B | 12000 | Chat Completion | -| `sonar-medium-chat` | 8x7B | 16384 | Chat Completion | -| `sonar-medium-online` | 8x7B | 12000 | Chat Completion | -| `codellama-34b-instruct`[3] | 34B | 16384 | Chat Completion | -| `codellama-70b-instruct` | 70B | 16384 | Chat Completion | -| `llama-2-70b-chat`[3] | 70B | 4096 | Chat Completion | -| `mistral-7b-instruct` [1] | 7B | 16384 | Chat Completion | -| `mixtral-8x7b-instruct` | 8x7B | 16384 | Chat Completion | -| `pplx-7b-chat`[2] [3] | 7B | 16384 | Chat Completion | -| `pplx-7b-online`[2] [3] | 7B | 12000 | Chat Completion | -| `pplx-70b-chat`[3] | 70B | 8192 | Chat Completion | -| `pplx-70b-online`[3] | 70B | 4000 | Chat Completion | \ No newline at end of file +| Model | Parameter Count | Context Length | Model Type | +| :-------------------- | :-------------- | :------------- | :-------------- | +| `sonar-small-chat` | 7B | 16384 | Chat Completion | +| `sonar-small-online` | 7B | 12000 | Chat Completion | +| `sonar-medium-chat` | 8x7B | 16384 | Chat Completion | +| `sonar-medium-online` | 8x7B | 12000 | Chat Completion | +| `codellama-70b-instruct` | 70B | 16384 | Chat Completion | +| `mistral-7b-instruct` [1] | 7B | 16384 | Chat Completion | +| `mixtral-8x7b-instruct` | 8x7B | 16384 | Chat Completion | \ No newline at end of file diff --git a/server/utils/AiProviders/perplexity/scripts/parse.mjs b/server/utils/AiProviders/perplexity/scripts/parse.mjs index d2064354afc3333a74053668a8be55eddb47e9ed..1858eafb879abdefd7b540c5c2a11ebf25b0f06d 100644 --- a/server/utils/AiProviders/perplexity/scripts/parse.mjs +++ b/server/utils/AiProviders/perplexity/scripts/parse.mjs @@ -8,7 +8,7 @@ // copy outputs into the export in ../models.js // Update the date below if you run this again because Perplexity added new models. -// Last Collected: Feb 23, 2024 +// Last Collected: Apr 14, 2024 import fs from "fs"; diff --git a/server/utils/AiProviders/togetherAi/models.js b/server/utils/AiProviders/togetherAi/models.js index ad940bc3976e9e736e327fd57cf62f5f72546e07..6fad3969bd1966939ac2b58ea000d7b68c74d045 100644 --- a/server/utils/AiProviders/togetherAi/models.js +++ b/server/utils/AiProviders/togetherAi/models.js @@ -1,8 +1,26 @@ const MODELS = { - "togethercomputer/alpaca-7b": { - id: "togethercomputer/alpaca-7b", - organization: "Stanford", - name: "Alpaca (7B)", + "zero-one-ai/Yi-34B-Chat": { + id: "zero-one-ai/Yi-34B-Chat", + organization: "01.AI", + name: "01-ai Yi Chat (34B)", + maxLength: 4096, + }, + "allenai/OLMo-7B-Instruct": { + id: "allenai/OLMo-7B-Instruct", + organization: "Allen AI", + name: "OLMo Instruct (7B)", + maxLength: 2048, + }, + "allenai/OLMo-7B-Twin-2T": { + id: "allenai/OLMo-7B-Twin-2T", + organization: "Allen AI", + name: "OLMo Twin-2T (7B)", + maxLength: 2048, + }, + "allenai/OLMo-7B": { + id: "allenai/OLMo-7B", + organization: "Allen AI", + name: "OLMo (7B)", maxLength: 2048, }, "Austism/chronos-hermes-13b": { @@ -11,95 +29,149 @@ const MODELS = { name: "Chronos Hermes (13B)", maxLength: 2048, }, - "togethercomputer/CodeLlama-13b-Instruct": { - id: "togethercomputer/CodeLlama-13b-Instruct", - organization: "Meta", - name: "Code Llama Instruct (13B)", - maxLength: 8192, + "cognitivecomputations/dolphin-2.5-mixtral-8x7b": { + id: "cognitivecomputations/dolphin-2.5-mixtral-8x7b", + organization: "cognitivecomputations", + name: "Dolphin 2.5 Mixtral 8x7b", + maxLength: 32768, }, - "togethercomputer/CodeLlama-34b-Instruct": { - id: "togethercomputer/CodeLlama-34b-Instruct", - organization: "Meta", - name: "Code Llama Instruct (34B)", + "databricks/dbrx-instruct": { + id: "databricks/dbrx-instruct", + organization: "databricks", + name: "DBRX Instruct", + maxLength: 32000, + }, + "deepseek-ai/deepseek-coder-33b-instruct": { + id: "deepseek-ai/deepseek-coder-33b-instruct", + organization: "DeepSeek", + name: "Deepseek Coder Instruct (33B)", + maxLength: 16384, + }, + "deepseek-ai/deepseek-llm-67b-chat": { + id: "deepseek-ai/deepseek-llm-67b-chat", + organization: "DeepSeek", + name: "DeepSeek LLM Chat (67B)", + maxLength: 4096, + }, + "garage-bAInd/Platypus2-70B-instruct": { + id: "garage-bAInd/Platypus2-70B-instruct", + organization: "garage-bAInd", + name: "Platypus2 Instruct (70B)", + maxLength: 4096, + }, + "google/gemma-2b-it": { + id: "google/gemma-2b-it", + organization: "Google", + name: "Gemma Instruct (2B)", maxLength: 8192, }, - "togethercomputer/CodeLlama-7b-Instruct": { - id: "togethercomputer/CodeLlama-7b-Instruct", - organization: "Meta", - name: "Code Llama Instruct (7B)", + "google/gemma-7b-it": { + id: "google/gemma-7b-it", + organization: "Google", + name: "Gemma Instruct (7B)", maxLength: 8192, }, - "DiscoResearch/DiscoLM-mixtral-8x7b-v2": { - id: "DiscoResearch/DiscoLM-mixtral-8x7b-v2", - organization: "DiscoResearch", - name: "DiscoLM Mixtral 8x7b", - maxLength: 32768, + "Gryphe/MythoMax-L2-13b": { + id: "Gryphe/MythoMax-L2-13b", + organization: "Gryphe", + name: "MythoMax-L2 (13B)", + maxLength: 4096, }, - "togethercomputer/falcon-40b-instruct": { - id: "togethercomputer/falcon-40b-instruct", - organization: "TII UAE", - name: "Falcon Instruct (40B)", - maxLength: 2048, + "lmsys/vicuna-13b-v1.5": { + id: "lmsys/vicuna-13b-v1.5", + organization: "LM Sys", + name: "Vicuna v1.5 (13B)", + maxLength: 4096, }, - "togethercomputer/falcon-7b-instruct": { - id: "togethercomputer/falcon-7b-instruct", - organization: "TII UAE", - name: "Falcon Instruct (7B)", - maxLength: 2048, + "lmsys/vicuna-7b-v1.5": { + id: "lmsys/vicuna-7b-v1.5", + organization: "LM Sys", + name: "Vicuna v1.5 (7B)", + maxLength: 4096, }, - "togethercomputer/GPT-NeoXT-Chat-Base-20B": { - id: "togethercomputer/GPT-NeoXT-Chat-Base-20B", - organization: "Together", - name: "GPT-NeoXT-Chat-Base (20B)", - maxLength: 2048, + "codellama/CodeLlama-13b-Instruct-hf": { + id: "codellama/CodeLlama-13b-Instruct-hf", + organization: "Meta", + name: "Code Llama Instruct (13B)", + maxLength: 16384, }, - "togethercomputer/llama-2-13b-chat": { - id: "togethercomputer/llama-2-13b-chat", + "codellama/CodeLlama-34b-Instruct-hf": { + id: "codellama/CodeLlama-34b-Instruct-hf", organization: "Meta", - name: "LLaMA-2 Chat (13B)", + name: "Code Llama Instruct (34B)", + maxLength: 16384, + }, + "codellama/CodeLlama-70b-Instruct-hf": { + id: "codellama/CodeLlama-70b-Instruct-hf", + organization: "Meta", + name: "Code Llama Instruct (70B)", maxLength: 4096, }, - "togethercomputer/llama-2-70b-chat": { - id: "togethercomputer/llama-2-70b-chat", + "codellama/CodeLlama-7b-Instruct-hf": { + id: "codellama/CodeLlama-7b-Instruct-hf", + organization: "Meta", + name: "Code Llama Instruct (7B)", + maxLength: 16384, + }, + "meta-llama/Llama-2-70b-chat-hf": { + id: "meta-llama/Llama-2-70b-chat-hf", organization: "Meta", name: "LLaMA-2 Chat (70B)", maxLength: 4096, }, - "togethercomputer/llama-2-7b-chat": { - id: "togethercomputer/llama-2-7b-chat", + "meta-llama/Llama-2-13b-chat-hf": { + id: "meta-llama/Llama-2-13b-chat-hf", organization: "Meta", - name: "LLaMA-2 Chat (7B)", + name: "LLaMA-2 Chat (13B)", maxLength: 4096, }, - "togethercomputer/Llama-2-7B-32K-Instruct": { - id: "togethercomputer/Llama-2-7B-32K-Instruct", - organization: "Together", - name: "LLaMA-2-7B-32K-Instruct (7B)", - maxLength: 32768, + "meta-llama/Llama-2-7b-chat-hf": { + id: "meta-llama/Llama-2-7b-chat-hf", + organization: "Meta", + name: "LLaMA-2 Chat (7B)", + maxLength: 4096, }, "mistralai/Mistral-7B-Instruct-v0.1": { id: "mistralai/Mistral-7B-Instruct-v0.1", - organization: "MistralAI", - name: "Mistral (7B) Instruct v0.1", - maxLength: 4096, + organization: "mistralai", + name: "Mistral (7B) Instruct", + maxLength: 8192, }, "mistralai/Mistral-7B-Instruct-v0.2": { id: "mistralai/Mistral-7B-Instruct-v0.2", - organization: "MistralAI", + organization: "mistralai", name: "Mistral (7B) Instruct v0.2", maxLength: 32768, }, "mistralai/Mixtral-8x7B-Instruct-v0.1": { id: "mistralai/Mixtral-8x7B-Instruct-v0.1", - organization: "MistralAI", - name: "Mixtral-8x7B Instruct", + organization: "mistralai", + name: "Mixtral-8x7B Instruct (46.7B)", maxLength: 32768, }, - "Gryphe/MythoMax-L2-13b": { - id: "Gryphe/MythoMax-L2-13b", - organization: "Gryphe", - name: "MythoMax-L2 (13B)", - maxLength: 4096, + "NousResearch/Nous-Capybara-7B-V1p9": { + id: "NousResearch/Nous-Capybara-7B-V1p9", + organization: "NousResearch", + name: "Nous Capybara v1.9 (7B)", + maxLength: 8192, + }, + "NousResearch/Nous-Hermes-2-Mistral-7B-DPO": { + id: "NousResearch/Nous-Hermes-2-Mistral-7B-DPO", + organization: "NousResearch", + name: "Nous Hermes 2 - Mistral DPO (7B)", + maxLength: 32768, + }, + "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": { + id: "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", + organization: "NousResearch", + name: "Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B)", + maxLength: 32768, + }, + "NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT": { + id: "NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT", + organization: "NousResearch", + name: "Nous Hermes 2 - Mixtral 8x7B-SFT (46.7B)", + maxLength: 32768, }, "NousResearch/Nous-Hermes-llama-2-7b": { id: "NousResearch/Nous-Hermes-llama-2-7b", @@ -113,66 +185,96 @@ const MODELS = { name: "Nous Hermes Llama-2 (13B)", maxLength: 4096, }, - "NousResearch/Nous-Hermes-Llama2-70b": { - id: "NousResearch/Nous-Hermes-Llama2-70b", - organization: "NousResearch", - name: "Nous Hermes Llama-2 (70B)", - maxLength: 4096, - }, "NousResearch/Nous-Hermes-2-Yi-34B": { id: "NousResearch/Nous-Hermes-2-Yi-34B", organization: "NousResearch", name: "Nous Hermes-2 Yi (34B)", maxLength: 4096, }, - "NousResearch/Nous-Capybara-7B-V1p9": { - id: "NousResearch/Nous-Capybara-7B-V1p9", - organization: "NousResearch", - name: "Nous Capybara v1.9 (7B)", - maxLength: 8192, - }, "openchat/openchat-3.5-1210": { id: "openchat/openchat-3.5-1210", organization: "OpenChat", - name: "OpenChat 3.5 1210 (7B)", + name: "OpenChat 3.5 (7B)", maxLength: 8192, }, - "teknium/OpenHermes-2-Mistral-7B": { - id: "teknium/OpenHermes-2-Mistral-7B", - organization: "teknium", - name: "OpenHermes-2-Mistral (7B)", - maxLength: 4096, - }, - "teknium/OpenHermes-2p5-Mistral-7B": { - id: "teknium/OpenHermes-2p5-Mistral-7B", - organization: "teknium", - name: "OpenHermes-2.5-Mistral (7B)", - maxLength: 4096, - }, "Open-Orca/Mistral-7B-OpenOrca": { id: "Open-Orca/Mistral-7B-OpenOrca", organization: "OpenOrca", name: "OpenOrca Mistral (7B) 8K", maxLength: 8192, }, - "garage-bAInd/Platypus2-70B-instruct": { - id: "garage-bAInd/Platypus2-70B-instruct", - organization: "garage-bAInd", - name: "Platypus2 Instruct (70B)", - maxLength: 4096, + "Qwen/Qwen1.5-0.5B-Chat": { + id: "Qwen/Qwen1.5-0.5B-Chat", + organization: "Qwen", + name: "Qwen 1.5 Chat (0.5B)", + maxLength: 32768, }, - "togethercomputer/Pythia-Chat-Base-7B-v0.16": { - id: "togethercomputer/Pythia-Chat-Base-7B-v0.16", - organization: "Together", - name: "Pythia-Chat-Base (7B)", - maxLength: 2048, + "Qwen/Qwen1.5-1.8B-Chat": { + id: "Qwen/Qwen1.5-1.8B-Chat", + organization: "Qwen", + name: "Qwen 1.5 Chat (1.8B)", + maxLength: 32768, + }, + "Qwen/Qwen1.5-4B-Chat": { + id: "Qwen/Qwen1.5-4B-Chat", + organization: "Qwen", + name: "Qwen 1.5 Chat (4B)", + maxLength: 32768, + }, + "Qwen/Qwen1.5-7B-Chat": { + id: "Qwen/Qwen1.5-7B-Chat", + organization: "Qwen", + name: "Qwen 1.5 Chat (7B)", + maxLength: 32768, + }, + "Qwen/Qwen1.5-14B-Chat": { + id: "Qwen/Qwen1.5-14B-Chat", + organization: "Qwen", + name: "Qwen 1.5 Chat (14B)", + maxLength: 32768, }, - "togethercomputer/Qwen-7B-Chat": { - id: "togethercomputer/Qwen-7B-Chat", + "Qwen/Qwen1.5-32B-Chat": { + id: "Qwen/Qwen1.5-32B-Chat", organization: "Qwen", - name: "Qwen-Chat (7B)", + name: "Qwen 1.5 Chat (32B)", + maxLength: 32768, + }, + "Qwen/Qwen1.5-72B-Chat": { + id: "Qwen/Qwen1.5-72B-Chat", + organization: "Qwen", + name: "Qwen 1.5 Chat (72B)", + maxLength: 32768, + }, + "snorkelai/Snorkel-Mistral-PairRM-DPO": { + id: "snorkelai/Snorkel-Mistral-PairRM-DPO", + organization: "Snorkel AI", + name: "Snorkel Mistral PairRM DPO (7B)", + maxLength: 32768, + }, + "togethercomputer/alpaca-7b": { + id: "togethercomputer/alpaca-7b", + organization: "Stanford", + name: "Alpaca (7B)", + maxLength: 2048, + }, + "teknium/OpenHermes-2-Mistral-7B": { + id: "teknium/OpenHermes-2-Mistral-7B", + organization: "Teknium", + name: "OpenHermes-2-Mistral (7B)", maxLength: 8192, }, + "teknium/OpenHermes-2p5-Mistral-7B": { + id: "teknium/OpenHermes-2p5-Mistral-7B", + organization: "Teknium", + name: "OpenHermes-2.5-Mistral (7B)", + maxLength: 8192, + }, + "togethercomputer/Llama-2-7B-32K-Instruct": { + id: "togethercomputer/Llama-2-7B-32K-Instruct", + organization: "Together", + name: "LLaMA-2-7B-32K-Instruct (7B)", + maxLength: 32768, + }, "togethercomputer/RedPajama-INCITE-Chat-3B-v1": { id: "togethercomputer/RedPajama-INCITE-Chat-3B-v1", organization: "Together", @@ -185,40 +287,34 @@ const MODELS = { name: "RedPajama-INCITE Chat (7B)", maxLength: 2048, }, - "upstage/SOLAR-0-70b-16bit": { - id: "upstage/SOLAR-0-70b-16bit", - organization: "Upstage", - name: "SOLAR v0 (70B)", - maxLength: 4096, - }, "togethercomputer/StripedHyena-Nous-7B": { id: "togethercomputer/StripedHyena-Nous-7B", organization: "Together", name: "StripedHyena Nous (7B)", maxLength: 32768, }, - "lmsys/vicuna-7b-v1.5": { - id: "lmsys/vicuna-7b-v1.5", - organization: "LM Sys", - name: "Vicuna v1.5 (7B)", + "Undi95/ReMM-SLERP-L2-13B": { + id: "Undi95/ReMM-SLERP-L2-13B", + organization: "Undi95", + name: "ReMM SLERP L2 (13B)", maxLength: 4096, }, - "lmsys/vicuna-13b-v1.5": { - id: "lmsys/vicuna-13b-v1.5", - organization: "LM Sys", - name: "Vicuna v1.5 (13B)", + "Undi95/Toppy-M-7B": { + id: "Undi95/Toppy-M-7B", + organization: "Undi95", + name: "Toppy M (7B)", maxLength: 4096, }, - "lmsys/vicuna-13b-v1.5-16k": { - id: "lmsys/vicuna-13b-v1.5-16k", - organization: "LM Sys", - name: "Vicuna v1.5 16K (13B)", - maxLength: 16384, + "WizardLM/WizardLM-13B-V1.2": { + id: "WizardLM/WizardLM-13B-V1.2", + organization: "WizardLM", + name: "WizardLM v1.2 (13B)", + maxLength: 4096, }, - "zero-one-ai/Yi-34B-Chat": { - id: "zero-one-ai/Yi-34B-Chat", - organization: "01.AI", - name: "01-ai Yi Chat (34B)", + "upstage/SOLAR-10.7B-Instruct-v1.0": { + id: "upstage/SOLAR-10.7B-Instruct-v1.0", + organization: "upstage", + name: "Upstage SOLAR Instruct v1 (11B)", maxLength: 4096, }, }; diff --git a/server/utils/AiProviders/togetherAi/scripts/chat_models.txt b/server/utils/AiProviders/togetherAi/scripts/chat_models.txt index 81c23bf4a06dc81fd36b3df538bce24d5539a28d..03f0414cb83a42ebbaa69fd84f75f681d4de46e5 100644 --- a/server/utils/AiProviders/togetherAi/scripts/chat_models.txt +++ b/server/utils/AiProviders/togetherAi/scripts/chat_models.txt @@ -1,39 +1,55 @@ -| Organization | Model Name | Model String for API | Max Seq Length | -| ------------- | ---------------------------- | -------------------------------------------- | -------------- | -| Stanford | Alpaca (7B) | togethercomputer/alpaca-7b | 2048 | -| Austism | Chronos Hermes (13B) | Austism/chronos-hermes-13b | 2048 | -| Meta | Code Llama Instruct (13B) | togethercomputer/CodeLlama-13b-Instruct | 8192 | -| Meta | Code Llama Instruct (34B) | togethercomputer/CodeLlama-34b-Instruct | 8192 | -| Meta | Code Llama Instruct (7B) | togethercomputer/CodeLlama-7b-Instruct | 8192 | -| DiscoResearch | DiscoLM Mixtral 8x7b | DiscoResearch/DiscoLM-mixtral-8x7b-v2 | 32768 | -| TII UAE | Falcon Instruct (40B) | togethercomputer/falcon-40b-instruct | 2048 | -| TII UAE | Falcon Instruct (7B) | togethercomputer/falcon-7b-instruct | 2048 | -| Together | GPT-NeoXT-Chat-Base (20B) | togethercomputer/GPT-NeoXT-Chat-Base-20B | 2048 | -| Meta | LLaMA-2 Chat (13B) | togethercomputer/llama-2-13b-chat | 4096 | -| Meta | LLaMA-2 Chat (70B) | togethercomputer/llama-2-70b-chat | 4096 | -| Meta | LLaMA-2 Chat (7B) | togethercomputer/llama-2-7b-chat | 4096 | -| Together | LLaMA-2-7B-32K-Instruct (7B) | togethercomputer/Llama-2-7B-32K-Instruct | 32768 | -| MistralAI | Mistral (7B) Instruct v0.1 | mistralai/Mistral-7B-Instruct-v0.1 | 4096 | -| MistralAI | Mistral (7B) Instruct v0.2 | mistralai/Mistral-7B-Instruct-v0.2 | 32768 | -| MistralAI | Mixtral-8x7B Instruct | mistralai/Mixtral-8x7B-Instruct-v0.1 | 32768 | -| Gryphe | MythoMax-L2 (13B) | Gryphe/MythoMax-L2-13b | 4096 | -| NousResearch | Nous Hermes LLaMA-2 (7B) | NousResearch/Nous-Hermes-llama-2-7b | 4096 | -| NousResearch | Nous Hermes Llama-2 (13B) | NousResearch/Nous-Hermes-Llama2-13b | 4096 | -| NousResearch | Nous Hermes Llama-2 (70B) | NousResearch/Nous-Hermes-Llama2-70b | 4096 | -| NousResearch | Nous Hermes-2 Yi (34B) | NousResearch/Nous-Hermes-2-Yi-34B | 4096 | -| NousResearch | Nous Capybara v1.9 (7B) | NousResearch/Nous-Capybara-7B-V1p9 | 8192 | -| OpenChat | OpenChat 3.5 1210 (7B) | openchat/openchat-3.5-1210 | 8192 | -| teknium | OpenHermes-2-Mistral (7B) | teknium/OpenHermes-2-Mistral-7B | 4096 | -| teknium | OpenHermes-2.5-Mistral (7B) | teknium/OpenHermes-2p5-Mistral-7B | 4096 | -| OpenOrca | OpenOrca Mistral (7B) 8K | Open-Orca/Mistral-7B-OpenOrca | 8192 | -| garage-bAInd | Platypus2 Instruct (70B) | garage-bAInd/Platypus2-70B-instruct | 4096 | -| Together | Pythia-Chat-Base (7B) | togethercomputer/Pythia-Chat-Base-7B-v0.16 | 2048 | -| Qwen | Qwen-Chat (7B) | togethercomputer/Qwen-7B-Chat | 8192 | -| Together | RedPajama-INCITE Chat (3B) | togethercomputer/RedPajama-INCITE-Chat-3B-v1 | 2048 | -| Together | RedPajama-INCITE Chat (7B) | togethercomputer/RedPajama-INCITE-7B-Chat | 2048 | -| Upstage | SOLAR v0 (70B) | upstage/SOLAR-0-70b-16bit | 4096 | -| Together | StripedHyena Nous (7B) | togethercomputer/StripedHyena-Nous-7B | 32768 | -| LM Sys | Vicuna v1.5 (7B) | lmsys/vicuna-7b-v1.5 | 4096 | -| LM Sys | Vicuna v1.5 (13B) | lmsys/vicuna-13b-v1.5 | 4096 | -| LM Sys | Vicuna v1.5 16K (13B) | lmsys/vicuna-13b-v1.5-16k | 16384 | -| 01.AI | 01-ai Yi Chat (34B) | zero-one-ai/Yi-34B-Chat | 4096 | \ No newline at end of file +| Organization | Model Name | Model String for API | Context length | +| --- | --- | --- | --- | +| 01.AI | 01-ai Yi Chat (34B) | zero-one-ai/Yi-34B-Chat | 4096 | +| Allen AI | OLMo Instruct (7B) | allenai/OLMo-7B-Instruct | 2048 | +| Allen AI | OLMo Twin-2T (7B) | allenai/OLMo-7B-Twin-2T | 2048 | +| Allen AI | OLMo (7B) | allenai/OLMo-7B | 2048 | +| Austism | Chronos Hermes (13B) | Austism/chronos-hermes-13b | 2048 | +| cognitivecomputations | Dolphin 2.5 Mixtral 8x7b | cognitivecomputations/dolphin-2.5-mixtral-8x7b | 32768 | +| databricks | DBRX Instruct | databricks/dbrx-instruct | 32000 | +| DeepSeek | Deepseek Coder Instruct (33B) | deepseek-ai/deepseek-coder-33b-instruct | 16384 | +| DeepSeek | DeepSeek LLM Chat (67B) | deepseek-ai/deepseek-llm-67b-chat | 4096 | +| garage-bAInd | Platypus2 Instruct (70B) | garage-bAInd/Platypus2-70B-instruct | 4096 | +| Google | Gemma Instruct (2B) | google/gemma-2b-it | 8192 | +| Google | Gemma Instruct (7B) | google/gemma-7b-it | 8192 | +| Gryphe | MythoMax-L2 (13B) | Gryphe/MythoMax-L2-13b | 4096 | +| LM Sys | Vicuna v1.5 (13B) | lmsys/vicuna-13b-v1.5 | 4096 | +| LM Sys | Vicuna v1.5 (7B) | lmsys/vicuna-7b-v1.5 | 4096 | +| Meta | Code Llama Instruct (13B) | codellama/CodeLlama-13b-Instruct-hf | 16384 | +| Meta | Code Llama Instruct (34B) | codellama/CodeLlama-34b-Instruct-hf | 16384 | +| Meta | Code Llama Instruct (70B) | codellama/CodeLlama-70b-Instruct-hf | 4096 | +| Meta | Code Llama Instruct (7B) | codellama/CodeLlama-7b-Instruct-hf | 16384 | +| Meta | LLaMA-2 Chat (70B) | meta-llama/Llama-2-70b-chat-hf | 4096 | +| Meta | LLaMA-2 Chat (13B) | meta-llama/Llama-2-13b-chat-hf | 4096 | +| Meta | LLaMA-2 Chat (7B) | meta-llama/Llama-2-7b-chat-hf | 4096 | +| mistralai | Mistral (7B) Instruct | mistralai/Mistral-7B-Instruct-v0.1 | 8192 | +| mistralai | Mistral (7B) Instruct v0.2 | mistralai/Mistral-7B-Instruct-v0.2 | 32768 | +| mistralai | Mixtral-8x7B Instruct (46.7B) | mistralai/Mixtral-8x7B-Instruct-v0.1 | 32768 | +| NousResearch | Nous Capybara v1.9 (7B) | NousResearch/Nous-Capybara-7B-V1p9 | 8192 | +| NousResearch | Nous Hermes 2 - Mistral DPO (7B) | NousResearch/Nous-Hermes-2-Mistral-7B-DPO | 32768 | +| NousResearch | Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) | NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO | 32768 | +| NousResearch | Nous Hermes 2 - Mixtral 8x7B-SFT (46.7B) | NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT | 32768 | +| NousResearch | Nous Hermes LLaMA-2 (7B) | NousResearch/Nous-Hermes-llama-2-7b | 4096 | +| NousResearch | Nous Hermes Llama-2 (13B) | NousResearch/Nous-Hermes-Llama2-13b | 4096 | +| NousResearch | Nous Hermes-2 Yi (34B) | NousResearch/Nous-Hermes-2-Yi-34B | 4096 | +| OpenChat | OpenChat 3.5 (7B) | openchat/openchat-3.5-1210 | 8192 | +| OpenOrca | OpenOrca Mistral (7B) 8K | Open-Orca/Mistral-7B-OpenOrca | 8192 | +| Qwen | Qwen 1.5 Chat (0.5B) | Qwen/Qwen1.5-0.5B-Chat | 32768 | +| Qwen | Qwen 1.5 Chat (1.8B) | Qwen/Qwen1.5-1.8B-Chat | 32768 | +| Qwen | Qwen 1.5 Chat (4B) | Qwen/Qwen1.5-4B-Chat | 32768 | +| Qwen | Qwen 1.5 Chat (7B) | Qwen/Qwen1.5-7B-Chat | 32768 | +| Qwen | Qwen 1.5 Chat (14B) | Qwen/Qwen1.5-14B-Chat | 32768 | +| Qwen | Qwen 1.5 Chat (32B) | Qwen/Qwen1.5-32B-Chat | 32768 | +| Qwen | Qwen 1.5 Chat (72B) | Qwen/Qwen1.5-72B-Chat | 32768 | +| Snorkel AI | Snorkel Mistral PairRM DPO (7B) | snorkelai/Snorkel-Mistral-PairRM-DPO | 32768 | +| Stanford | Alpaca (7B) | togethercomputer/alpaca-7b | 2048 | +| Teknium | OpenHermes-2-Mistral (7B) | teknium/OpenHermes-2-Mistral-7B | 8192 | +| Teknium | OpenHermes-2.5-Mistral (7B) | teknium/OpenHermes-2p5-Mistral-7B | 8192 | +| Together | LLaMA-2-7B-32K-Instruct (7B) | togethercomputer/Llama-2-7B-32K-Instruct | 32768 | +| Together | RedPajama-INCITE Chat (3B) | togethercomputer/RedPajama-INCITE-Chat-3B-v1 | 2048 | +| Together | RedPajama-INCITE Chat (7B) | togethercomputer/RedPajama-INCITE-7B-Chat | 2048 | +| Together | StripedHyena Nous (7B) | togethercomputer/StripedHyena-Nous-7B | 32768 | +| Undi95 | ReMM SLERP L2 (13B) | Undi95/ReMM-SLERP-L2-13B | 4096 | +| Undi95 | Toppy M (7B) | Undi95/Toppy-M-7B | 4096 | +| WizardLM | WizardLM v1.2 (13B) | WizardLM/WizardLM-13B-V1.2 | 4096 | +| upstage | Upstage SOLAR Instruct v1 (11B) | upstage/SOLAR-10.7B-Instruct-v1.0 | 4096 | \ No newline at end of file diff --git a/server/utils/AiProviders/togetherAi/scripts/parse.mjs b/server/utils/AiProviders/togetherAi/scripts/parse.mjs index b96d40ab1d0f6f8ce81dfeb1cbc7d332a404820f..b7840401225288abc43266e729cc4c5af035e40d 100644 --- a/server/utils/AiProviders/togetherAi/scripts/parse.mjs +++ b/server/utils/AiProviders/togetherAi/scripts/parse.mjs @@ -8,7 +8,9 @@ // copy outputs into the export in ../models.js // Update the date below if you run this again because TogetherAI added new models. -// Last Collected: Jan 10, 2023 +// Last Collected: Apr 14, 2024 +// Since last collection Together's docs are broken. I just copied the HTML table +// and had claude3 convert to markdown and it works well enough. import fs from "fs";