--
You received this message because you are subscribed to the Google Groups "rabbitmq-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to rabbitmq-user...@googlegroups.com.
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/946c2296-e920-4412-8d06-25167ed780c4n%40googlegroups.com.
You can see top 10 processes for 3.9 (above) and 3.7 (below) here https://p.defau.lt/?8x4Tu9xL75yCXYuhkXNV4g . Unfortunatelly it doesn‘t tell me much as I don‘t know RabbitMQ that well.
I see that version 3.9 runs queue_metrics_metrics_collector and queue_coarse_metrics_metrics_collector constantly. Is this because I have few thousand classic queues mirrored? I have the same amount on 3.7. Both installations are idle for the past couple of days with no producers or consumers attached to queues.
--
Vilius
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/CAA81d0untKMyVEbdenppd9NtUPK7XjB3svdauSQO%2Bb%2BW8v_Nsg%40mail.gmail.com.
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/DBBPR01MB7705E402327EA849392CAFCE92EA9%40DBBPR01MB7705.eurprd01.prod.exchangelabs.com.
Hi,
you are on to something here. I have found that indead 3.9 on Google Cloud is shipped with prometheus plugin enabled. Disabling it didn‘t give any results however.
Then I found that it also ships with prometheus scraping sidecar. Disabled that too, still no results.
The remaining plugins are:
rabbitmq_management
rabbitmq_management_agent
rabbitmq_peer_discovery_common
rabbitmq_peer_discovery_k8s
rabbitmq_web_dispatch
Essentially they are the same as with 3.7 with the exception of peer_discovery_k8s which is enabled because of the cluster setup.
Your suggestion to disable statistics collection helped to save ~20% of CPU usage. However I have struggled with collect_statistics setting because it seems that it cannot be changed unless I disable management plugin. Is this a known bug? How do I disable statistics without disabling management plugin?
And finally I found that GKE is using „rabbitmqctl status“ every 10 seconds as a liveness and a readiness check. I have increased that to 120 seconds and CPU usage dropped 3 times! I tried to change “rabbitmqctl status” to simple “rabbitmq-diagnostics ping” but that did help much. As soon as I lowered readiness check to 10 or 20 seconds CPU jumped up considerably again even with simple ping.
I started reading through https://www.rabbitmq.com/monitoring.html#readiness-probes but it got me confused. The documentation talks about readiness probes but at the same time it talks about Kubernetes and node restarts. Readiness probe doesn’t restart applications so is that paragraph named incorrectly or I’m misunderstanding it?
What would be your suggestions for liveness and readiness probe? I’m thinking about “rabbitmq-diagnostics ping” for liveness and “rabbitmq-diagnostics check_running” for readiness, but I’m still wondering why these commands produce essentially the same CPU load as “rabbitmqctl status”?
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/CAA81d0t4qFkd_2UXnciUaL6V%3DZAXuCgOE4R3G9z6ZHfUGNaSqA%40mail.gmail.com.
It is the same hardware, the docker image is built by different providers though. All feature flags are enabled as far as I can see.
Plugin list is pretty basic:
[E*] rabbitmq_management 3.9.13
[e*] rabbitmq_management_agent 3.9.13
[e*] rabbitmq_peer_discovery_common 3.9.13
[E*] rabbitmq_peer_discovery_k8s 3.9.13
[E*] rabbitmq_prometheus 3.9.13
[e*] rabbitmq_web_dispatch 3.9.13
Disabling prometheus didn‘t produce any noticeable difference.
--
Vilius
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/47fc03e4-335d-4e0b-8958-75b3334cb6f6n%40googlegroups.com.
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/AS8PR01MB77045E326833E261AED3CD4592EA9%40AS8PR01MB7704.eurprd01.prod.exchangelabs.com.
We are using Google Cloud, so my initial idea was to use what is provided by Google (it‘s their click-to-deploy configuration). I assumed they know what they are doing, apparently not :)
Since we already fully tested our application with the current cluster setup and failover behavior, I‘m afraid it‘s too late to migrate, but I will keep operator in mind during our next big dev cycle.
For the moment I will migrate to TCP probes then. Before I try to read operator’s source code, could you just tell me which port is used in its probes? Is it the port used for inter-node communication or something else? Also is it the same for liveness probe?
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/CAA81d0tzPNYnw-WHXXfZfCo9c9%3DAS3VrNF1khKRP%2BcEmhNDV5w%40mail.gmail.com.
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/AS8PR01MB7704898B49D2B14191C7AB8492ED9%40AS8PR01MB7704.eurprd01.prod.exchangelabs.com.
Thank you for your help. I have changed the probes to TCP and now CPU usage more or les sis workable. I still don‘t like it, because it‘s ~0,09 of core for every node used constantly, but now it‘s a least bearable.
One last question regarding collect_statistics parameter. Is it possible disable statistics without disabling management plugin? At the moment it seems to be always set to ‘fine’ statistics if management plugin is enabled.
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/CAA81d0sLYaqL-TMF1q9Z4o7D24%2BNMvb_qvoRbXgkkKfT-5gtsQ%40mail.gmail.com.
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/AS8PR01MB7704A719C1B74548E504E5F592EC9%40AS8PR01MB7704.eurprd01.prod.exchangelabs.com.
Hi,
I have 3 nodes at the moment, but I‘m worried that this “additional constant” usage will double or triple on top of the “actual” usage under heavier loads. Or maybe I’m just too paranoid, I don’t know.
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/CAA81d0uJttjBcwJ%2BBqYHjt4XALAc5A-zmPB5aGigNpg-J4qC7A%40mail.gmail.com.