Stars
A light weight vLLM simulator, for mocking out replicas.
llm-d is a Kubernetes-native high-performance distributed LLM inference framework
Gateway API Inference Extension
A high-throughput and memory-efficient inference and serving engine for LLMs
A high-performance distributed file system designed to address the challenges of AI training and inference workloads.
vLLM’s reference system for K8S-native cluster-wide deployment with community-driven performance optimization
LangChain for Go, the easiest way to write LLM-based programs in Go
GUI tool for visualizing the result data of deBruijn sequence complexity distribution study
KubeStellar - a flexible solution for multi-cluster configuration management for edge, multi-cloud, and hybrid cloud
the main repository for the multicluster global hub