Deploy production-grade Qdrant vector database clusters on Kubernetes in minutes. Automate distributed HA with per-shard Raft replication and rolling upgrades.
Supported Versions
Open Source
Deploy Qdrant in 2 steps
Install KubeBlocks & Qdrant Addon
# Add Helm repo helm repo add kubeblocks https://apecloud.github.io/helm-charts helm repo update # Install KubeBlocks helm install kubeblocks kubeblocks/kubeblocks \ --namespace kb-system --create-namespace # Install Qdrant addon helm upgrade -i kb-addon-qdrant kubeblocks/qdrant \ --namespace kb-system
Create a Qdrant Cluster
apiVersion: apps.kubeblocks.io/v1
kind: Cluster
metadata:
name: qdrant-cluster
namespace: demo
spec:
clusterDef: qdrant
topology: cluster
terminationPolicy: Delete
componentSpecs:
- name: qdrant
replicas: 3
serviceVersion: "1.16.3"Trusted by Engineering Teams at Scale
KubeBlocks deploys Qdrant as a distributed cluster where each collection shard elects its own Raft leader. Data is replicated across nodes; any surviving node continues serving vector search queries when a peer fails.
KubeBlocks deploys Qdrant in distributed mode with 3 or more nodes. Each collection shard is replicated across nodes using Raft consensus — when a node fails, the surviving replicas elect a new shard leader and continue serving queries. KubeBlocks restarts the failed pod; it rejoins the cluster and syncs shard data automatically.
Per-shard Raft consensus — every collection write is replicated to a quorum before acknowledged
Automatic shard leader re-election when a node fails — replicated shards on surviving nodes continue serving queries
All nodes accept REST and gRPC search requests — clients connect to any surviving node
Peer discovery via Kubernetes headless service DNS (no external etcd or ZooKeeper)
HNSW indexing with configurable m and ef_construct parameters per collection
REST API on :6333 (includes GET /metrics), gRPC on :6334, Raft P2P on :6335
{cluster}-qdrant-qdrant:6333{cluster}-qdrant-qdrant:6334No SSH into pods, no shell scripts. Submit an OpsRequest and KubeBlocks handles the rest.
Availability & Scaling
Horizontal Scaling
Add or remove nodes online. KubeBlocks joins new pods to the Qdrant cluster; shards are redistributed automatically to utilize new capacity.
Vertical Scaling
Resize CPU and memory on running nodes with a rolling pod restart strategy — the cluster continues serving during the operation.
Volume Expansion
Expand PVC storage for vector data without pod restarts on supported storage classes.
Rolling Restart
Controlled pod restarts that maintain shard availability throughout — the cluster keeps serving queries.
Stop / Start
Suspend the cluster to eliminate compute cost; resume with full vector data and shard state intact.
Backup & Restore
Back up vector data to S3-compatible storage via the datafile method. Restore to a new cluster from any backup snapshot.
Configuration & Observability
Multi-Engine Consistency
Manage Qdrant alongside PostgreSQL, MySQL, Redis, and 35+ other engines using the same OpsRequest API and kubectl tooling.
Version Upgrade
Rolling upgrades across supported versions (e.g. 1.13.4 → 1.16.3) with health checks between each pod restart.
Prometheus Metrics
Per-node metrics via GET /metrics on :6333 (built-in, no sidecar needed). Compatible with Grafana dashboards.
Expose via LoadBalancer
Expose REST or gRPC APIs externally via a LoadBalancer or NodePort service for external client access.
KubeBlocks automates the hardest parts of running Qdrant on Kubernetes — so your team doesn't have to.
Deploy a production-grade Qdrant vector database cluster in minutes with distributed HA, per-shard Raft replication, and full Day-2 operations — all open source.