Skip to content
Snippets Groups Projects
This project is mirrored from https://github.com/Mintplex-Labs/anything-llm. Pull mirroring updated .
  1. Mar 22, 2024
  2. Mar 19, 2024
  3. Mar 14, 2024
  4. Mar 09, 2024
  5. Mar 07, 2024
  6. Mar 06, 2024
  7. Feb 27, 2024
  8. Feb 24, 2024
    • Sean Hatfield's avatar
      [FEAT] OpenRouter integration (#784) · 633f4252
      Sean Hatfield authored
      
      * WIP openrouter integration
      
      * add OpenRouter options to onboarding flow and data handling
      
      * add todo to fix headers for rankings
      
      * OpenRouter LLM support complete
      
      * Fix hanging response stream with OpenRouter
      update tagline
      update comment
      
      * update timeout comment
      
      * wait for first chunk to start timer
      
      * sort OpenRouter models by organization
      
      * uppercase first letter of organization
      
      * sort grouped models by org
      
      ---------
      
      Co-authored-by: default avatartimothycarambat <rambat1010@gmail.com>
      Unverified
      633f4252
  9. Feb 22, 2024
  10. Feb 19, 2024
  11. Feb 12, 2024
  12. Feb 06, 2024
  13. Jan 29, 2024
  14. Jan 26, 2024
    • Hakeem Abbas's avatar
      feature: Integrate Astra as vectorDBProvider (#648) · 5614e2ed
      Hakeem Abbas authored
      
      * feature: Integrate Astra as vectorDBProvider
      
      feature: Integrate Astra as vectorDBProvider
      
      * Update .env.example
      
      * Add env.example to docker example file
      Update spellcheck fo Astra
      Update Astra key for vector selection
      Update order of AstraDB options
      Resize Astra logo image to 330x330
      Update methods of Astra to take in latest vectorDB params like TopN and more
      Update Astra interface to support default methods and avoid crash errors from 404 collections
      Update Astra interface to comply to max chunk insertion limitations
      Update Astra interface to dynamically set dimensionality from chunk 0 size on creation
      
      * reset workspaces
      
      ---------
      
      Co-authored-by: default avatartimothycarambat <rambat1010@gmail.com>
      Unverified
      5614e2ed
  15. Jan 23, 2024
  16. Jan 18, 2024
  17. Jan 17, 2024
  18. Jan 12, 2024
  19. Jan 10, 2024
  20. Jan 09, 2024
  21. Jan 06, 2024
  22. Jan 05, 2024
  23. Dec 28, 2023
  24. Dec 20, 2023
  25. Dec 19, 2023
  26. Dec 17, 2023
  27. Dec 16, 2023
  28. Dec 14, 2023
    • Timothy Carambat's avatar
      Document Processor v2 (#442) · 719521c3
      Timothy Carambat authored
      * wip: init refactor of document processor to JS
      
      * add NodeJs PDF support
      
      * wip: partity with python processor
      feat: add pptx support
      
      * fix: forgot files
      
      * Remove python scripts totally
      
      * wip:update docker to boot new collector
      
      * add package.json support
      
      * update dockerfile for new build
      
      * update gitignore and linting
      
      * add more protections on file lookup
      
      * update package.json
      
      * test build
      
      * update docker commands to use cap-add=SYS_ADMIN so web scraper can run
      update all scripts to reflect this
      remove docker build for branch
      Unverified
      719521c3
  29. Dec 13, 2023
  30. Dec 08, 2023
  31. Dec 07, 2023
    • timothycarambat's avatar
    • Timothy Carambat's avatar
      [Feature] AnythingLLM use locally hosted Llama.cpp and GGUF files for inferencing (#413) · 655ebd94
      Timothy Carambat authored
      * Implement use of native embedder (all-Mini-L6-v2)
      stop showing prisma queries during dev
      
      * Add native embedder as an available embedder selection
      
      * wrap model loader in try/catch
      
      * print progress on download
      
      * add built-in LLM support (expiermental)
      
      * Update to progress output for embedder
      
      * move embedder selection options to component
      
      * saftey checks for modelfile
      
      * update ref
      
      * Hide selection when on hosted subdomain
      
      * update documentation
      hide localLlama when on hosted
      
      * saftey checks for storage of models
      
      * update dockerfile to pre-build Llama.cpp bindings
      
      * update lockfile
      
      * add langchain doc comment
      
      * remove extraneous --no-metal option
      
      * Show data handling for private LLM
      
      * persist model in memory for N+1 chats
      
      * update import
      update dev comment on token model size
      
      * update primary README
      
      * chore: more readme updates and remove screenshots - too much to maintain, just use the app!
      
      * remove screeshot link
      Unverified
      655ebd94
    • timothycarambat's avatar
      chore: remove unused NO_DEBUG env · fecfb0fa
      timothycarambat authored
      fecfb0fa
  32. Dec 06, 2023
Loading