--
You received this message because you are subscribed to the Google Groups "Isilon Technical User Group" group.
To unsubscribe from this group and stop receiving emails from it, send an email to isilon-user-group+unsubscribe@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
What does "isi job history" show you?
--
You received this message because you are subscribed to the Google Groups "Isilon Technical User Group" group.
To unsubscribe from this group and stop receiving emails from it, send an email to isilon-user-gr...@googlegroups.com.
To unsubscribe from this group and stop receiving emails from it, send an email to isilon-user-group+unsubscribe@googlegroups.com.
--
You received this message because you are subscribed to the Google Groups "Isilon Technical User Group" group.
To unsubscribe from this group and stop receiving emails from it, send an email to isilon-user-group+unsubscribe@googlegroups.com.
What is the general situation on the cluster
in terms of CPU load, disk IOPS/latencies
and disk stalls or other events causing group changes?
Do you have SSDs for metadata acceleration?
To unsubscribe from this group and stop receiving emails from it, send an email to isilon-user-gr...@googlegroups.com.
--
You received this message because you are subscribed to the Google Groups "Isilon Technical User Group" group.
To unsubscribe from this group and stop receiving emails from it, send an email to isilon-user-gr...@googlegroups.com.
Check your file counts per pool and SSD utilization. If you have enough capacity you should ask your sales team for a GNA threshold variance so you can enable it. We are using a lot more SSD capacity since
This is a known issue with the code version you are on along with FSA failing, an explanation is listed below.
FSAnalyze fails with 'database is locked' due to isi_job_d restarting and briefly having 2 locks on the database simultaneously. What usually causes isi_job_d to restart is the worker manager queue gets filled up with messages sent from the coordinator to the manager every minute when checking cluster load to throttle the job engine. The size of the message queue is 100. After 100 minutes this queue gets full and causes isi_job_d to stop and restart.
We can do a workaround to avoid the issue in the future. What we need to do is to increase the sysctl timeout rate. The default rate for the below command is 60 seconds, we want to increase that timeout to 240 seconds. If you are available for a Webex we can make that change.
Command that identifies the timeout rate:
isi_gconfig -t job-config core.load_balance_interval_sec=
--
You received this message because you are subscribed to the Google Groups "Isilon Technical User Group" group.
To unsubscribe from this group and stop receiving emails from it, send an email to isilon-user-group+unsubscribe@googlegroups.com.