Shard Autoscaling
- Owners:
- Status:
Implemented
- Related Tickets:
- Other docs:
- n/a
This document aims to address existing issues preventing users from leveraging the Horizontal Pod Autoscaler (HPA) in conjunction with Prometheus sharding.
- Why
- Pitfalls of the current solution
- Goals
- Non-Goals
- Audience
- How
- Scale-down alternatives
- Action Plan
Why
Managing Prometheus instances with highly variable workloads can be quite challenging. When the number of scraped targets can scale up and down between a few items to thousands over a short period of time (e.g. within a day), operators are forced to opt for the safest scenario, ending up with over-provisioned instances. Additionally, the WAL replay often becomes a problem after a restart as it takes a long time to be replayed into memory. Without available replicas, this leads to unnecessary downtime.
To address these issues, this document outlines a solution to enhance the Prometheus Operator by enabling the usage of Horizontal Pod Autoscalers (HPAs) to scale shards up and down. This improvement aims to simplify the process of scaling Prometheus instances, making it more accessible and efficient for operators.
Pitfalls of the current solution
Prometheus can scale vertically pretty well, but instances that grow up to hundreds of GiB/TiB in WAL size start face several challenges:
- Eventual cardinality explosions have a big blast radius if a single Prometheus instance is responsible for scraping the majority of monitored applications.
- Recovering from a crash takes several minutes due to the WAL replay.
While the Prometheus and PrometheusAgent CRDs implement sharding for scrape configurations, there are still pending issues and missing features that prevent adoption.
- Environments where the load is undistributed throughout the day/week, operators need to constantly adjust number of shards or shard size.
- Scaling down Prometheus Servers causes data loss, as there’s no guarantee that stored data is persisted in centralized storage before the container is deleted.
Goals
- Enable scaling of the Prometheus shards up and down via Horizontal Pod Autoscaler objects.
Non-Goals
- Implement the autoscaling logic into the Prometheus operator.
- Automatic balancing of targets to ensure homogeneous load across all Prometheus pods.
Audience
- Users looking for a shard autoscaling mechanism with a good cost-reliability ratio.
How
Today, there are a few strategies to measure the load of Prometheus instances.
-
Perhaps you can use Kubernetes Metrics API to measure load in terms of CPU and/or Memory usage.
-
Another strategy might be to measure the amount of samples ingested per second
- Prometheus Agents:
rate(prometheus_agent_samples_appended_total[5m])
- Prometheus Servers:
rate(prometheus_tsdb_head_samples_appended_total[5m])
- Or even measuring samples queried per second for Prometheus instances under heavy query load.
- Prometheus Servers:
rate(prometheus_engine_query_samples_total[5m])
These are all data that could be configured as input for different Horizontal Pod Autoscalers, but how do HPAs know if the amount of shard is bigger/lower than desired?
Scale subresource
When working with any resource, including CRDs, HPAs depend on a subresource called Scale. This proposal suggests to implement the scale subresource for the Prometheus and PrometheusAgent CRDs. Instead of working on the “replicas” count, it will operate on the “shards” count because the purpose of scaling up (resp. down) is to distribute the same number of targets across more (resp. less) Prometheus instances.
With only this change, Prometheus Agents can already be horizontally scaled without problems, but for Prometheus Servers it gets a little more complicated.
Graceful scale-down of Prometheus Servers
Prometheus Servers don’t behave like the usual stateless applications. A few cases require Prometheus pods to still be available even after a scale-down event is triggered:
- Thanos proxying queries back to the sidecars to retrieve data that hasn’t been uploaded to object storage yet.
- Prometheus servers with long retention periods still hold data that users might want to query.
If the operator deletes Prometheus pods in excess during a scale-down operation, data loss will be a common problem, so a more complex strategy needs to be used here.
We propose that Prometheus-Operator introduces a new shardRetentionPolicy
field to the Prometheus CRD:
apiVersion: monitoring.coreos.com
kind: Prometheus
metadata:
name: example
spec:
shardRetentionPolicy:
# Default: Delete
whenScaled: Retain|Delete
# if not specified, the operator picks up the retention period if it's defined,
# otherwise (e.g. size-based retention) the statefulset lives forever.
retain:
retentionPeriod: 3d
Inspired by the StatefulSet API, when setting .shardRetentionPolicy.whenScaled
is set to Delete
the operator would simply delete the underlying StatefulSet on the spot, not caring about the data stored in that particular shard. Although not the safest approach, it is exactly the current behavior, and changing it might surprise a lot of users.
If .shardRetentionPolicy.whenScaled
is set to Retain
Prometheus-Operator will keep the shard running for a customizable period of time. By default, the Operator will keep the Shard for the whole retention period, or forever if only size-based retention is defined. Alternatively, the user can define a custom retention period.
If a retention time is defined, the deletion logic is controlled by an annotation present on the StatefulSet:
operator.prometheus.io/deletion-timestamp: X
On scale downs, the configuration of all shards will be re-arranged to make sure that the “scaled-down” Prometheus pods don’t scrape targets anymore but they will still be available for queries.
When the deletion timestamp set in the annotation is reached, the Operator will trigger the deletion of the statefulset associated with the shard.
Scaling up after scaling down
If a scale-up event happens while the Prometheus pods marked for deletion are still running, the annotation will be removed and the scrape targets will be redistributed to all active shards.
We intentionally don’t want to spin up new instances while others that are marked for deletion are still running as shard names and containers’ environment variables that configure each shard’s target group will be too complex to coordinate.
Graceful scale-down of Prometheus Agents
Prometheus Agents are different than servers since queries are not available in this mode. Their only responsibility is scraping metrics and pushing them via remote-write to a long-term storage backend, making the scale-down experience much easier to handle.
When receiving the SIGTERM signal, the Prometheus Agent should gracefully handle the signal by finishing all remote-write queues before ending the process. Prometheus-Operator, by default, adjusts the Graceful Termination Period of Prometheus/PrometheusAgent pods to 600s. Ten minutes should be enough for them to flush the remote-write queue, but, if needed, users can redefine Graceful Termination Period using Strategic Merge Patch.
Since there’s no use case for retaining Prometheus Agents, its CRD will not be extended with the RetentionPolicy
mentioned in Graceful scale-down of Prometheus Servers
Scale down Alternatives
Dump & Backfill
During scale-down, the Prometheus-Operator could read TSDB Blocks from the Prometheus instance being deleted and backfill it into another instance.
Advantages:
- Prometheus can be shut down immediately without data loss.
Disadvantages:
- Loading TSDB Blocks into memory is expensive, requiring Prometheus-Operator to run with big Memory/CPU requests.
- Complex and hard to coordinate workflow. (Require restarts to reload TSDB blocks, hard to identify possible corruptions)
Snapshot & Upload on shutdown
During scale-down, Prometheus-Operator could send an HTTP request to Prometheus’ Snapshot endpoint. Thanos sidecar could be extended to watch the snapshot directory and automatically upload snapshots to Object storage.
There is a related issue open already: https://github.com/thanos-io/thanos/issues/6263
Advantages:
- Prometheus can be shut down immediately without data loss.
Disadvantages:
- Prometheus running without Thanos sidecars won’t benefit from this strategy.
Action Plan
-
Re-implement Pull Request #4735, but this time focusing on Shards and making sure we add the selectorpath to the scale subresource. Enabling horizontal autoscaling for Prometheus Agents.
-
Implement the graceful shutdown strategy of Prometheus servers.