This project is mirrored from https://github.com/Mintplex-Labs/anything-llm.
Pull mirroring updated .
- May 11, 2024
-
-
Timothy Carambat authored
* prevent accidental lockout from restrict chars in single pass mode * update error message
-
- May 10, 2024
-
-
Sean Hatfield authored
* add max tokens field to generic openai llm connector * add max_tokens property to generic openai agent provider
-
- May 08, 2024
-
-
Sean Hatfield authored
* add text gen web ui LLM provider support * update README * README typo * update TextWebUI display name patch workspace<>model support for provider --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- May 02, 2024
-
-
Sean Hatfield authored
* koboldcpp LLM support * update .env.examples for koboldcpp support * update LLM preference order update koboldcpp comments --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
Sean Hatfield authored
* getChatCompletion working WIP streaming * WIP * working streaming WIP abort stream * implement cohere embedder support * remove inputType option from cohere embedder * fix cohere LLM from not aborting stream when canceled by user * Patch Cohere implemention * add cohere to onboarding --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Apr 23, 2024
-
-
Timothy Carambat authored
* Add generic OpenAI endpoint support * allow any input for model in case provider does not support models endpoint
-
- Apr 19, 2024
-
-
Timothy Carambat authored
* Add LMStudio embedding endpoint support * update alive path check for HEAD remove commented JSX * update comment
-
Timothy Carambat authored
* Add support for Gemini-1.5 Pro bump @google/generative-ai pkg Toggle apiVersion if beta model selected resolves #1109 * update API messages due to package change
-
- Apr 16, 2024
-
-
Timothy Carambat authored
V1 of agent support via built-in `@agent` that can be invoked alongside normal workspace RAG chat.
-
- Apr 10, 2024
-
-
Timothy Carambat authored
* expand support for non-US azure deployments * update conditional
-
- Apr 05, 2024
-
-
Timothy Carambat authored
* Enable per-workspace provider/model combination * cleanup * remove resetWorkspaceChatModels and wipeWorkspaceModelPreference to prevent workspace from resetting model * add space --------- Co-authored-by:
shatfield4 <seanhatfield5@gmail.com>
-
- Apr 02, 2024
-
-
Ivan Skodje authored
* [FEAT] Added port checks in updateENV.validDockerizedUrl to prevent docker from assuming it cannot access localhost URLs * [CHORE] Updated error message to include Linux URL * Patch port checking for general loopbacks * typo --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Mar 29, 2024
-
-
timothycarambat authored
-
- Mar 22, 2024
-
-
Timothy Carambat authored
-
- Mar 14, 2024
-
-
Timothy Carambat authored
-
Timothy Carambat authored
* Support External Transcription providers * patch files * update docs * fix return data
-
Sean Hatfield authored
add Haiku model support
-
- Mar 06, 2024
-
-
Sean Hatfield authored
* implement new version of anthropic sdk and support new models * remove handleAnthropicStream and move to handleStream inside anthropic provider * update useGetProvidersModels for new anthropic models
-
Sean Hatfield authored
* Groq LLM support complete * update useGetProvidersModels for groq models * Add definiations update comments and error log reports add example envs --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Feb 27, 2024
-
-
Timothy Carambat authored
* Add Ollama embedder model support calls * update docs
-
- Feb 24, 2024
-
-
Sean Hatfield authored
* WIP openrouter integration * add OpenRouter options to onboarding flow and data handling * add todo to fix headers for rankings * OpenRouter LLM support complete * Fix hanging response stream with OpenRouter update tagline update comment * update timeout comment * wait for first chunk to start timer * sort OpenRouter models by organization * uppercase first letter of organization * sort grouped models by org --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Feb 22, 2024
-
-
Sean Hatfield authored
* add LLM support for perplexity * update README & example env * fix ENV keys in example env files * slight changes for QA of perplexity support * Update Perplexity AI name --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Feb 06, 2024
-
-
Sean Hatfield authored
* WIP event logging - new table for events and new settings view for viewing * WIP add logging * UI for log rows * rename files to Logging to prevent getting gitignore * add metadata for all logging events and colored badges in logs page * remove unneeded comment * cleanup namespace for logging * clean up backend calls * update logging to show to => from settings changes * add logging for invitations, created, deleted, and accepted * add logging for user created, updated, suspended, or removed * add logging for workspace deleted * add logging for chat logs exported * add logging for API keys, LLM, embedder, vector db, embed chat, and reset button * modify event logs * update to event log types * simplify rendering of event badges --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
Timothy Carambat authored
-
- Jan 26, 2024
-
-
Hakeem Abbas authored
* feature: Integrate Astra as vectorDBProvider feature: Integrate Astra as vectorDBProvider * Update .env.example * Add env.example to docker example file Update spellcheck fo Astra Update Astra key for vector selection Update order of AstraDB options Resize Astra logo image to 330x330 Update methods of Astra to take in latest vectorDB params like TopN and more Update Astra interface to support default methods and avoid crash errors from 404 collections Update Astra interface to comply to max chunk insertion limitations Update Astra interface to dynamically set dimensionality from chunk 0 size on creation * reset workspaces --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Jan 23, 2024
-
-
Sean Hatfield authored
* migrate pinecone package to latest version and migrate pinecone vectordb provider class * remove pinecone environment name env variable and update docs to reflect removal & serverless support complete * migrate query for pinecone db * typo in log --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Jan 19, 2024
-
-
Timothy Carambat authored
* add support for exporting to json and csv in workspace chats * safety encode URL options * remove message about openai fine tuning on export success * all defaults to jsonl * Persist special env keys on updates --------- Co-authored-by:
shatfield4 <seanhatfield5@gmail.com>
-
- Jan 18, 2024
-
-
Timothy Carambat authored
* feat: Add support for Zilliz Cloud by Milvus * update placeholder text update data handling stmt * update zilliz descriptor
-
Sean Hatfield authored
Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Jan 17, 2024
-
-
Sean Hatfield authored
* add support for mistral api * update docs to show support for Mistral * add default temp to all providers, suggest different results per provider --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
Sean Hatfield authored
* WIP model selection per workspace (migrations and openai saves properly * revert OpenAiOption * add support for models per workspace for anthropic, localAi, ollama, openAi, and togetherAi * remove unneeded comments * update logic for when LLMProvider is reset, reset Ai provider files with master * remove frontend/api reset of workspace chat and move logic to updateENV add postUpdate callbacks to envs * set preferred model for chat on class instantiation * remove extra param * linting * remove unused var * refactor chat model selection on workspace * linting * add fallback for base path to localai models --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Jan 12, 2024
-
-
Shuyoou authored
* issue #543 support milvus vector db * migrate Milvus to use MilvusClient instead of ORM normalize env setup for docs/implementation feat: embedder model dimension added * update comments --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Jan 10, 2024
-
-
Sean Hatfield authored
* add Together AI LLM support * update readme to support together ai * Patch togetherAI implementation * add model sorting/option labels by organization for model selection * linting + add data handling for TogetherAI * change truthy statement patch validLLMSelection method --------- Co-authored-by:
timothycarambat <rambat1010@gmail.com>
-
- Jan 02, 2024
-
-
Timothy Carambat authored
-
- Dec 28, 2023
-
-
Timothy Carambat authored
* Prevent external service localhost question * add 0.0.0.0 to docker-invalid URL * clarify hint
-
Timothy Carambat authored
* Add support for Ollama as LLM provider resolves #493
-
Timothy Carambat authored
resolves #489
-
- Dec 20, 2023
-
-
timothycarambat authored
update file picker spacing for attributes
-
- Dec 08, 2023
-
-
Timothy Carambat authored
fix: cleanup code for embedding length clarify resolves #388
-
- Dec 07, 2023
-
-
Timothy Carambat authored
* Implement use of native embedder (all-Mini-L6-v2) stop showing prisma queries during dev * Add native embedder as an available embedder selection * wrap model loader in try/catch * print progress on download * add built-in LLM support (expiermental) * Update to progress output for embedder * move embedder selection options to component * saftey checks for modelfile * update ref * Hide selection when on hosted subdomain * update documentation hide localLlama when on hosted * saftey checks for storage of models * update dockerfile to pre-build Llama.cpp bindings * update lockfile * add langchain doc comment * remove extraneous --no-metal option * Show data handling for private LLM * persist model in memory for N+1 chats * update import update dev comment on token model size * update primary README * chore: more readme updates and remove screenshots - too much to maintain, just use the app! * remove screeshot link
-