I don't recall that anybody mentioned something similar but taking a
memory profile of the running server would useful.
You can use promtool for this:
promtool debug pprof <Prometheus URL>
On Sun, Sep 15, 2019 at 10:51 PM Kevin Black <kevinbla...@gmail.com> wrote:
>
> I have some namespaces with just a few thousand metrics, but I have a couple gigs higher memory usage over comparable prometheus instances. The only difference I can find is that the instance with high memory usage is in a k8s environment with significantly more pods/services (~57,000). We use k8s pod/service discovery, but can't filter to namespaces since the targets will live in different namespaces in a dynamic fashion. So we rely on pod/service discovery, and then use relabel_configs with a keep action on a specific label which keeps a handful of targets. Is it known/expected that a large discovery even without actually scraping the metrics would have a couple GB memory impact? Is there something else we could be doing to mitigate this usage?
>
> --
> You received this message because you are subscribed to the Google Groups "Prometheus Users" group.
> To unsubscribe from this group and stop receiving emails from it, send an email to promethe...@googlegroups.com.