kubeai-project / kubeai Star 1.1k Code Issues Pull requests Discussions AI Inference Operator for Kubernetes. The easiest way to serve ML models in production. Supports VLMs, LLMs, embeddings, and speech-to-text. kubernetes ai k8s whisper autoscaler openai-api llm vllm faster-whisper ollama vllm-operator ollama-operator inference-operator Updated Dec 15, 2025 Go
BudEcosystem / Awesome-vLLM-plugins Star 5 Code Issues Pull requests A curated list of plugins built on top of vLLM plugins vllm vllm-operator vllm-serve vllm-integration vllm-plugins Updated Dec 12, 2025
brokedba / vllm-lab Star 2 Code Issues Pull requests This Repository contains terraform configuration for vllm production-stack in the cloud managed K8s gke aks civo eks oke vllm llmcache vllm-operator vllm-serve vllm-production-stack Updated Nov 10, 2025 HCL
pakolee867 / vllm-production-stack-terraform Star 0 Code Issues Pull requests ai gke aws-eks aks civo eks oke azure-aks gcp-gke lmops vllm llm-inference llmcache vllm-operator nebius vllm-serve vllm-production-stack oci-oke Updated Dec 24, 2025 HCL