"Cannot open dets table",schema,[{file,"/var/lib/rabbitmq/mnesia/rabbit@ornery-clownfish-crmq-1/"Cannot open dets table",schema,[{file,"/var/lib/rabbitmq/mnesia/rabbit@ornery-clownfish-crmq-1/schema.DAT"
I also see on logs after restart :
dets: file "/var/lib/rabbitmq/mnesia/rabbit@ornery-clownfish-crmq-1/schema.DAT" not properly closed, repairing ...
I also check files in replicas that are not working ( just before they restart ) and it seems that there are missing file.
I did the same test with an other pvc ( still in glusterfs ) on :
- name: log
mountPath: /var/log/rabbitmq/
And i don't get any issue ...
And if i use an other StorageClass there are no issue at all.
Thanks a lot.Regards,Léo
--
You received this message because you are subscribed to the Google Groups "rabbitmq-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to rabbitmq-user...@googlegroups.com.
To view this discussion on the web, visit https://groups.google.com/d/msgid/rabbitmq-users/1a64f720-a909-496a-b463-efcd8099510en%40googlegroups.com.
Hello,
We still looking for the issue on our side and we found that it's finally not a race/locking issue.
It can't be because, yes we use glusterfs so we might think that the data is shared between the 3 pod but in our case we are declaring a pvc for each pod.
Then each pod have is own data.
The issue still coming from the file schema.DAT because the file is corrupted.
This file is "Schema Definition Export and Import" for more information check : https://www.rabbitmq.com/definitions.html#overview
the thing is that his file is storing information "metadata or topology. Users, vhosts, queues, exchanges, bindings, runtime parameters ..."
And the interesting part is : "Definitions are stored in an internal database and replicated across all cluster nodes. Every node in a cluster has its own replica of all definitions. When a part of definitions changes, the update is performed on all nodes in a single transaction. This means that in practice, definitions can be exported from any cluster node with the same result."
Then in our case because our file is corrupted when crmq-1 try to access this file to update it's own it fail then restart.
The unknown part is now why this file is corrupted when we use glusterfs.