Hi Team,We have 2 different openshift cluster site A is hosted on ibm cloud ( ROKS ) and site B hosted on AWS cloud ( ROSA ). site A is where we generate the token and link at site B. We have exposed several services from site A to site B and vice versa, in site B we have several namespaces where the skupper is installed and exposed the services. As these namespaces are connected each other using skupper link all the N number of services are visible in all the namespaces in site A and site B where the skupper is configured or skupper is linked.Question 1: site A is the one where we generate token which is basically a main site for us so how much traffic that skupper-router can handle ? is there limitation of number services that to be exposed ?
Question 2: Most of the times we have noticed below error in few of the namespaces which are there in site B and skupper router breaks the connectivity between the respective namespace from site B with the site A namespace.2025-03-26 08:27:08.006312 +0000 SERVER (error) [C45] Connection to skupper-inter-router-namespacesiteA.siteAcluster.xyzdomain:443 failed: amqp:resource-limit-exceeded local-idle-timeout expired
So, is there any specific limit set for this queueing mechanism ( amqp ) so it could handle only so n so number of requests and after that it gives this above error. if so can we increase this limit to handle more requests.Thanks,Vishavjeet S
--
You received this message because you are subscribed to the Google Groups "Skupper" group.
To unsubscribe from this group and stop receiving emails from it, send an email to skupper+u...@googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/skupper/29a68cd1-b9b7-4dc7-806b-3c4a8a970addn%40googlegroups.com.
Thanks Ganesh for the information !We have several microservices running in both sites and there are several API calls executed. Network in our both sites is stable. By any chance can we increase this heartbeat time from 16 sec to some other number ?
The error which I mentioned in the previous thread, what proactive action needs to be taken to avoid such issues ? OR what action needs to be taken when such an error occurs?
In our case it happens very frequently in lower environments like development where the usage of microservices is high.We have not explicitly set up any resource quota for skupper-router it is the default while installing through the skupper-site configmap. However we have noticed the higher memory utilisation of skupper-router in Site A, below is the screenshot for your reference.This is the memory utilisation of skupper-router from site B namespace where we have high traffic.As our skupper configuration is done using YAML method by skupper-site configmap, is there any way we can explicitly set the resource quota or is that okay if we just simply scale up the skupper-router replica's from 1 to 2 or the desired number where the load is high ?
client version 1.4.2
transport version quay.io/skupper/skupper-router:2.4.2 (sha256:c0dfccf2c269)
controller version quay.io/skupper/service-controller:1.4.2 (sha256:06f3ef0047c6)
config-sync version quay.io/skupper/config-sync:1.4.2 (sha256:2a350062450a)
flow-collector version quay.io/skupper/flow-collector:1.4.2 (sha256:c3b1e43cce4a)
You received this message because you are subscribed to a topic in the Google Groups "Skupper" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/skupper/odSfUA6Qz9k/unsubscribe.
To unsubscribe from this group and all its topics, send an email to skupper+u...@googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/skupper/CA%2BYqnOpTC%2BL4-t3cZFCWSGOVK_1-83Vk2JQvoM-LOX3GjLKvxA%40mail.gmail.com.
Hi Ganesh,Today i noticed that skupper-router pod was consuming memory more than 19GB so to address this I scaled up pods to 2 replicas, it's been more than 5 hours and still memory utilisation is more than 19GB for old pod and around 2GB for the newly scaled up pod.1. How the traffic will be handled/routed in case of more than 1 skupper-router replica ?
2. Is there anything that I need to do explicitly to release the memory consumption or balance the traffic flow ?
3. As you mentioned, recently "discovered an issue where the router processes the vanflow records slowly which is leading to memory growth and your team is working on fix", so this fix is for a which specific skupper version as we are currently running on skupper 1.4.2 version.
Hi Ganesh,
On Point 1: Yes, we noticed some bad errors which broke the connectivity in application from Site A to Site B, so I scaled down pods to 1 replica and rolling restart worked fine.
On Point 2: Might be in debug logs there are chances of exposing the DNS names or cluster details to everybody so is there any option where I can share the logs with you directly .. like 1:1 over your email or any box link or something that is authenticated ?
On Point 3: Currently we are running on 1.42. skupper version so when we plan for an upgrade from 1.x to 2.x is it straight forward or there are some changes ? does it involve the downtime or deleting the existing skupper links and recreate it ?
Hi Ganesh,
I've attached the skstat-m.txt file for your reference from one of our specific namespaces where we noticed the high memory consumption.