The Jozu Blog
Serving LLMs at Scale with KitOps, Kubeflow, and KServe
Learn how to deploy and serve large language models at scale using KitOps for packaging, Kubeflow for orchestration, and KServe for production-grade inference on Kubernetes.
Stop Rebuilding Docker Images: Deploy ML Models at Scale with Argo and KitOps
Learn how to run scalable ML inference with Argo Workflows and KitOps ModelKits. Deploy models without rebuilding Docker images using Jozu Hub governance.
Top Open Source Tools for Kubernetes ML: From Development to Production
From Development to Production Running machine learning on Kubernetes has evolved from experimental curiosity to production necessity. But with hundreds of tools claiming to solve ML (machine learning) deployment, which ones should you consider?
Jozu Launches Enterprise Support for CNCF-Backed ModelPack and KitOps Standards
How KitOps and Weights & Biases Work Together for Reliable Model Versioning
Learn how to combine Weights & Biases experiment tracking with KitOps ModelKits for reproducible ML workflows. This tutorial shows you how to train, package, and deploy models to production with full lineage tracking, automatic SBOM generation, and security scanning—eliminating the 'works on my machine' problem for ML deployments.
How to Turn ML Training Notebook into Deployable ModelKits with KitOps and Marimo
Learn how to transform your ML training notebooks into deployable ModelKits using KitOps and Marimo. This comprehensive tutorial covers packaging your machine learning models with all dependencies, datasets, and code into a single, shareable artifact for seamless deployment.