A high-throughput and memory-efficient inference and serving engine for LLMs
The all-in-one Desktop & Docker AI application with built-in RAG, AI agents, No-code agent builder, MCP compatibility, and more.