I
inferrence
Projects with this topic
-
🔧 🔗 https://github.com/lmstudio-ai/mlx-engine.git Apple MLX engine for LM StudioUpdated -
Real-time inference for Stable Diffusion - 0.88s latency. Covers AITemplate, nvFuser, TensorRT, FlashAttention. Join our Discord communty: https://discord.com/invite/TgHXuSJEk6
Updated