Explore projects
-
-
https://github.com/InternLM/lmdeploy LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
๐ lmdeploy.readthedocs.io/en/latest/Updated -
๐ง ๐ https://github.com/pytorch/pytorchTensors and Dynamic neural networks in Python with strong GPU acceleration
Updated -
๐ง ๐ https://github.com/modelscope/dash-inferDashInfer is a native LLM inference engine aiming to deliver industry-leading performance atop various hardware architectures, including x86 and ARMv9.
Updated -
-
https://github.com/janhq/nitro.git now: https://github.com/janhq/cortex.git Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM, ONNX). Powers
๐ JanUpdated -
๐ง ๐ https://github.com/bytedance/Elkeid Elkeid is an open source solution that can meet the security requirements of various workloadsUpdated -
Port of OpenAI's Whisper model in C/C++
Updated -
-
๐ง ๐ https://github.com/AppFlowy-IO/AppFlowy Bring projects, wikis, and teams together with AI. AppFlowy is an AI collaborative workspace where you achieve more without losing control of your data. The bestUpdated -
๐ง ๐ https://github.com/microsoft/quicreachA 'ping' equivalent tool for QUIC.
Updated -
Updated
-
-
A high-throughput and memory-efficient inference and serving engine for LLMs
Updated -
๐ง ๐ https://github.com/ghostty-org/ghostty๐ป Ghostty is a fast, feature-rich, and cross-platform terminal emulator that uses platform-native UI and GPU acceleration.Updated -
User-Mode Driver for Tenstorrent hardware
Updated