Projects with this topic
-
Crush
🔧 🔗 https://github.com/charmbracelet/crush The glamourous AI coding agent for your favourite terminal💘 Updated -
🔧 🔗 https://github.com/All-Hands-AI/OpenHands🙌 OpenHands: Code Less, Make MoreUpdated -
huggingface.co/transformers https://github.com/huggingface/transformers🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.Updated -
🔧 🔗 https://github.com/langfuse/langfuse🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management, playground, datasets. Integrates with LlamaIndex, Langchain, OpenAI SDK, LUpdated -
🔧 🔗 https://github.com/BerriAI/litellmPython SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
🕸 ️🔗 docs.litellm.ai/docs/Updated -
🔧 🔗 https://github.com/langfuse/langfuse-docs🪢 Langfuse documentation -- Langfuse is the open source LLM Engineering Platform. Observability, evals, prompt management, playground and metrics to debug andUpdated -
🔧 🔗 https://github.com/IBM/mcp-context-forge A Model Context Protocol (MCP) Gateway & Registry. Serves as a central management point for tools, resources, and prompts that can be accessed by MCP-compatible LLM applications. Converts REST API endpoints to MCP, composes virtual MCP servers with added security and observability, and converts between protocols (stdio, SSE, Streamable HTTP).Updated -
🔧 🔗 https://github.com/openai/codex Lightweight coding agent that runs in your terminalUpdated -
🔧 🔗 https://github.com/modelscope/ms-swiftSWIFT (Scalable lightWeight Infrastructure for Fine-Tuning) Use PEFT or Full-parameter to finetune 400+ LLMs or 100+ MLLMs.
Updated -
🔧 🔗 https://github.com/vllm-project/vllm-ascend Community maintained hardware plugin for vLLM on AscendUpdated -
🔧 🔗 https://github.com/containers/ramalamaRamaLama is an open-source developer tool that simplifies the local serving of AI models from any source and facilitates their use for inference in production, all
Updated -
🔧 🔗 https://github.com/vllm-project/vllmA high-throughput and memory-efficient inference and serving engine for LLMs
Updated -
🔧 🔗 https://github.com/mendableai/firecrawl🔥 Turn entire websites into LLM-ready markdown or structured data. Scrape, crawl and extract with a single API.Updated -
https://github.com/joaomdmoura/crewAI Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
crewai.com
Updated -
https://github.com/InternLM/xtuner An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
Updated -
https://github.com/InternLM/lmdeploy LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
🔗 lmdeploy.readthedocs.io/en/latest/Updated -
🔧 🔗 https://github.com/gpustack/gpustackManage GPU clusters for running LLMs
Updated -
🔧 🔗 https://github.com/modelscope/evalscope A streamlined and customizable framework for efficient large model evaluation and performance benchmarkingUpdated -
🔧 🔗 https://github.com/flashinfer-ai/flashinferFlashInfer: Kernel Library for LLM Serving
Updated -
Catwalk
🔧 🔗 https://github.com/charmbracelet/catwalk🐈 A collection of LLM inference providers and modelsUpdated