I was submitting jobs fine on dataProc clusters until I changed VMs to get better CPU performances. Now I get this error and I am stuck. I can't stop the cluster because it has a local disk, stopping and reseting the VMs manually does resolve the issue. I creating a new cluster is not a good option
ERROR: (gcloud.dataproc.jobs.submit.pyspark) HttpError accessing .... response: <{'x-debug-tracking-id': '9633867706168606723;o=0', 'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'content-encoding': 'gzip', 'date': 'Wed, 04 Oct 2023 18:06:34 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': 429}>, content <{
"error": {
"code": 429,
"message": "No agent on master node(s) found to be active in the past 300 seconds.\nLast reported times: [gca-dev-cluster-m seconds: 1696241723\nnanos: 125792000\n]. This may indicate high memory usage in Dataproc master or an unhealthy Dataproc master node",
"status": "RESOURCE_EXHAUSTED"
}
}
>
This may be due to network connectivity issues. Please check your network settings, and the status of the service you are trying to reach.