Skip to content
Snippets Groups Projects
  • Timothy Carambat's avatar
    655ebd94
    [Feature] AnythingLLM use locally hosted Llama.cpp and GGUF files for inferencing (#413) · 655ebd94
    Timothy Carambat authored
    * Implement use of native embedder (all-Mini-L6-v2)
    stop showing prisma queries during dev
    
    * Add native embedder as an available embedder selection
    
    * wrap model loader in try/catch
    
    * print progress on download
    
    * add built-in LLM support (expiermental)
    
    * Update to progress output for embedder
    
    * move embedder selection options to component
    
    * saftey checks for modelfile
    
    * update ref
    
    * Hide selection when on hosted subdomain
    
    * update documentation
    hide localLlama when on hosted
    
    * saftey checks for storage of models
    
    * update dockerfile to pre-build Llama.cpp bindings
    
    * update lockfile
    
    * add langchain doc comment
    
    * remove extraneous --no-metal option
    
    * Show data handling for private LLM
    
    * persist model in memory for N+1 chats
    
    * update import
    update dev comment on token model size
    
    * update primary README
    
    * chore: more readme updates and remove screenshots - too much to maintain, just use the app!
    
    * remove screeshot link
    [Feature] AnythingLLM use locally hosted Llama.cpp and GGUF files for inferencing (#413)
    Timothy Carambat authored
    * Implement use of native embedder (all-Mini-L6-v2)
    stop showing prisma queries during dev
    
    * Add native embedder as an available embedder selection
    
    * wrap model loader in try/catch
    
    * print progress on download
    
    * add built-in LLM support (expiermental)
    
    * Update to progress output for embedder
    
    * move embedder selection options to component
    
    * saftey checks for modelfile
    
    * update ref
    
    * Hide selection when on hosted subdomain
    
    * update documentation
    hide localLlama when on hosted
    
    * saftey checks for storage of models
    
    * update dockerfile to pre-build Llama.cpp bindings
    
    * update lockfile
    
    * add langchain doc comment
    
    * remove extraneous --no-metal option
    
    * Show data handling for private LLM
    
    * persist model in memory for N+1 chats
    
    * update import
    update dev comment on token model size
    
    * update primary README
    
    * chore: more readme updates and remove screenshots - too much to maintain, just use the app!
    
    * remove screeshot link
Code owners
Assign users and groups as approvers for specific file changes. Learn more.