docs/CaseStudies.md: updated ARNES numbers

This commit is contained in:
Aliaksandr Valialkin 2020-04-06 16:20:11 +03:00
parent 65ba430632
commit f5121d1e5f

View File

@ -246,15 +246,14 @@ SNMPCollector to collect metrics from network devices and switching from InfluxD
Numbers: Numbers:
- 2 single node instances - 2 single node instances per DC (one for prometheus and one for influxdb metrics)
- Active time series per VictoriaMetrics instance: 1,7M - Active time series per VictoriaMetrics instance: ~500k (prometheus) + ~320k (influxdb)
- Ingestion rate per VictoriaMetrics instance: 75K points/second - Ingestion rate per VictoriaMetrics instance: 45k/s (prometheus) / 30k/s (influxdb)
- Query duration: median is ~5ms, 99th percentile is ~45ms - Query duration: median is ~5ms, 99th percentile is ~45ms
- Total number of datapoints: 1+ Trillion - Total number of datapoints per instance: 390B (prometheus), 110B (influxdb)
- Average time series churn rate: ~350k/day
- Average datapoint size on drive: 0.4 bytes - Average datapoint size on drive: 0.4 bytes
- Disk usage: 380GB for ~4 months of data - Disk usage per VictoriaMetrics instance: 125GB (prometheus), 185GB (influxdb)
- Index size: 2,6GB - Index size per VictoriaMetrics instance: 1.6GB (prometheus), 1.2GB (influcdb)
We are running 1 Prometheus, 1 VictoriaMetrics and 1 Grafana server in each datacenter on baremetal servers, scraping 350+ targets We are running 1 Prometheus, 1 VictoriaMetrics and 1 Grafana server in each datacenter on baremetal servers, scraping 350+ targets
(and 3k+ devices collected via SNMPCollector sending metrics directly to VM). Each Prometheus is scraping all targets, (and 3k+ devices collected via SNMPCollector sending metrics directly to VM). Each Prometheus is scraping all targets,