Mooncake is the serving platform for Kimi
An Easy-to-Use and High-Performance AI Deployment Framework
Distribute and run LLMs with a single file
TT-NN operator library, and TT-Metalium low level kernel programming
Fast Multimodal LLM on Mobile Devices
A @ClickHouse fork that supports high-performance vector search
UCCL is an efficient communication library for GPUs
High-speed Large Language Model Serving for Local Deployment
Production ready toolkit to run AI locally
Locally run an Instruction-Tuned Chat-Style LLM
Implements a reference architecture for creating information systems