Deploy production-grade Milvus vector database clusters in minutes. Standalone and distributed topologies, billion-scale ANN search, and full Day-2 operations.
Vectors at Scale
ANN Search Latency
Deploy Milvus in 4 steps
Install KubeBlocks
# Add Helm repo helm repo add kubeblocks https://apecloud.github.io/helm-charts helm repo update # Install KubeBlocks helm install kubeblocks kubeblocks/kubeblocks \ --namespace kb-system --create-namespace
Install Milvus Addon
helm upgrade -i kb-addon-milvus kubeblocks/milvus \ -n kb-system
Create a Milvus Cluster
apiVersion: apps.kubeblocks.io/v1
kind: Cluster
metadata:
name: milvus-cluster
namespace: demo
spec:
terminationPolicy: Delete
topology: standalone
componentSpecs:
- name: milvus
serviceVersion: "2.5.13"
replicas: 1Cluster is Ready
$ kubectl get cluster milvus-cluster -n demo NAME CLUSTER-DEF STATUS AGE milvus-cluster Running 2m
Trusted by Engineering Teams at Scale
A single Milvus pod runs all coordinator and worker roles as goroutines within one process. etcd and MinIO are deployed as companion pods. Ideal for development, CI pipelines, and single-node deployments.
Single pod runs all roles: Proxy, RootCoord, QueryCoord, DataCoord, IndexCoord, QueryNode, DataNode, IndexNode
etcd and MinIO deployed as companion pods for metadata and object storage
PVC-backed local segment cache for fast cold-start
Minimal resource footprint — runs on a single Kubernetes node
Same CRD API as distributed mode — switch topologies without application changes
{cluster}-milvus:19530milvus processKubeBlocks automates the hardest parts of running Milvus on Kubernetes — so your team doesn't have to.
No SSH into pods, no shell scripts. Submit an OpsRequest and KubeBlocks handles the rest.
Scaling & Availability
Horizontal Scaling
Scale querynode, datanode, indexnode, and proxy independently. All workers are stateless — scale out without data migration.
Vertical Scaling
Resize CPU and memory for any Milvus component via OpsRequest with a rolling restart that keeps the cluster available.
Volume Expansion
Expand PVCs for components that use persistent storage without pod restarts on supported storage classes.
Rolling Restart
Controlled rolling restarts with health checks between each pod to maintain service availability.
Stop / Start
Suspend the Milvus cluster to eliminate compute cost; resume with full state from MinIO and etcd.
Configuration & Data
Dynamic Configuration
Tune Milvus parameters via OpsRequest — dataCoord settings, queryNode cache sizes, etcd endpoints — without full restarts.
Version Upgrade
Rolling upgrades from 2.3.x to 2.5.x with component-ordered sequencing: workers first, proxy, then mixcoord.
Prometheus Metrics
Per-component metrics exposed on port 9091; compatible with Prometheus and Grafana dashboards.
Backup & Restore
Collection-level backup via milvus-backup (v0.5.9+); restore to a new cluster from any stored snapshot.
Expose via LoadBalancer
Expose the gRPC proxy endpoint on port 19530 via a Kubernetes LoadBalancer service for external client access.
Also on KubeBlocks
Milvus 2.5+ supports Apache Kafka as its write-ahead log (WAL) for higher throughput streaming. KubeBlocks manages both Milvus and Kafka clusters — deploy the full AI data stack on a single operator.
Open source and production-ready. Enterprise customers get dedicated onboarding, migration support, and SLA guarantees.