mirror of
https://github.com/github/awesome-copilot.git
synced 2026-04-18 06:05:55 +00:00
3.4 KiB
3.4 KiB
name, description
| name | description |
|---|---|
| qdrant-monitoring-debugging | Diagnoses Qdrant production issues using metrics and observability tools. Use when someone reports 'optimizer stuck', 'indexing too slow', 'memory too high', 'OOM crash', 'queries are slow', 'latency spike', or 'search was fast now it's slow'. Also use when performance degrades without obvious config changes. |
How to Debug Qdrant with Metrics
First check optimizer status. Most production issues trace back to active optimizations competing for resources. If optimizer is clean, check memory, then request metrics.
Optimizer Stuck or Too Slow
Use when: optimizer running for hours, not finishing, or showing errors.
- Use
/collections/{collection_name}/optimizationsendpoint (v1.17+) to check status Optimization monitoring - Query with optional detail flags:
?with=queued,completed,idle_segments - Returns: queued optimizations count, active optimizer type, involved segments, progress tracking
- Web UI has an Optimizations tab with timeline view and per-task duration metrics Web UI
- If
optimizer_statusshows an error in collection info, check logs for disk full or corrupted segments - Large merges and HNSW rebuilds legitimately take hours on big datasets. Check progress before assuming it's stuck.
Memory Seems Too High
Use when: memory exceeds expectations, node crashes with OOM, or memory keeps growing.
- Process memory metrics available via
/metrics(RSS, allocated bytes, page faults) - Qdrant uses two types of RAM: resident memory (data structures, quantized vectors) and OS page cache (cached disk reads). Page cache filling available RAM is normal. Memory article
- If resident memory (RSSAnon) exceeds 80% of total RAM, investigate
- Check
/telemetryfor per-collection breakdown of point counts and vector configurations - Estimate expected memory:
num_vectors * dimensions * 4 bytes * 1.5for vectors, plus payload and index overhead Capacity planning - Common causes of unexpected growth: quantized vectors with
always_ram=true, too many payload indexes, largemax_segment_sizeduring optimization
Queries Are Slow
Use when: queries slower than expected and you need to identify the cause.
- Track
rest_responses_avg_duration_secondsandrest_responses_max_duration_secondsper endpoint - Use histogram metric
rest_responses_duration_seconds(v1.8+) for percentile analysis in Grafana - Equivalent gRPC metrics with
grpc_responses_prefix - Check optimizer status first. Active optimizations compete for CPU and I/O, degrading search latency.
- Check segment count via collection info. Too many unmerged segments after bulk upload causes slower search.
- Compare filtered vs unfiltered query times. Large gap means missing payload index. Payload index
What NOT to Do
- Ignore optimizer status when debugging slow queries (most common root cause)
- Assume memory leak when page cache fills RAM (normal OS behavior)
- Make config changes while optimizer is running (causes cascading re-optimizations)
- Blame Qdrant before checking if bulk upload just finished (unmerged segments)