ganeti-watcher pid=42043 ERROR Can't acquire lock on state file /var/lib/ganeti/watcher.cc1ca697-6ef8-478e-97ef-e1ed25ad281b.data: File already locked
gnt-cluster modify --ipolicy-spindle-ratio=128# gnt-job info 70184
- Job ID: 70184
Status: success
Received: 2015-06-25 15:30:04.390484
Processing start: 2015-06-25 15:30:05.368574 (delta 0.978090s)
Processing end: 2015-06-25 15:35:22.434716 (delta 317.066142s)
Total processing time: 318.044232 seconds
Opcodes:
- Opcode: OP_GROUP_VERIFY_DISKS
Status: success
Processing start: 2015-06-25 15:30:05.368574
Execution start: 2015-06-25 15:30:05.897476
Processing end: 2015-06-25 15:35:22.434697
Input fields:
OP_ID: OP_GROUP_VERIFY_DISKS
comment: None
debug_level: 0
depends: None
group_name: cc1ca697-6ef8-478e-97ef-e1ed25ad281b
priority: 10
reason:
- - gnt:watcher
- Verifying disks of group cc1ca697-6ef8-478e-97ef-e1ed25ad281b
- 1435239004291613184
- - gnt:opcode:group_verify_disks
- job=70184;index=0
- 1435239004390484000
- - gnt:daemon:masterd:pickup:op_group_verify_disks
- job=70184;index=0
- 1435239005265840896
Result:
-
-
-
Execution log:
# gnt-cluster version
Software version: 2.12.4
Internode protocol: 2120000
Configuration format: 2120000
OS api version: 20
Export interface: 0
VCS version: (ganeti) version v2.12.4
# gnt-cluster --version
gnt-cluster (ganeti v2.12.4) 2.12.4
# gnt-cluster verify
Submitted jobs 70187, 70188
Waiting for job 70187 ...
Thu Jun 25 15:47:59 2015 * Verifying cluster config
Thu Jun 25 15:47:59 2015 * Verifying cluster certificate files
Thu Jun 25 15:47:59 2015 * Verifying hypervisor parameters
Thu Jun 25 15:47:59 2015 * Verifying all nodes belong to an existing group
Waiting for job 70188 ...
Thu Jun 25 15:48:01 2015 * Verifying group 'default'
Thu Jun 25 15:48:01 2015 * Gathering data (3 nodes)
Thu Jun 25 15:48:01 2015 * Gathering information about nodes (3 nodes)
Thu Jun 25 15:48:10 2015 * Gathering disk information (3 nodes)
Thu Jun 25 15:52:00 2015 * Verifying configuration file consistency
Thu Jun 25 15:52:00 2015 * Verifying node status
Thu Jun 25 15:52:00 2015 * Verifying instance status
Thu Jun 25 15:52:00 2015 * Verifying orphan volumes
Thu Jun 25 15:52:00 2015 * Verifying N+1 Memory redundancy
Thu Jun 25 15:52:00 2015 * Other Notes
Thu Jun 25 15:52:00 2015 - NOTICE: 1 non-redundant instance(s) found.
Thu Jun 25 15:52:00 2015 * Hooks Results
lvs /dev/ganeti/52c6d3df-1e11-4802-a597-cb254de78b30.disk1_datadevices {
dir = "/dev"
scan = [ "/dev" ]
obtain_device_list_from_udev = 1
preferred_names = [ ]
filter = ["r|/dev/cdrom|", "r|/dev/drbd[0-9]+|" ]
cache_dir = "/run/lvm"
cache_file_prefix = ""
write_cache_state = 1
sysfs_scan = 1
multipath_component_detection = 1
md_component_detection = 1
md_chunk_alignment = 1
data_alignment_detection = 1
data_alignment = 0
data_alignment_offset_detection = 1
ignore_suspended_devices = 0
disable_after_error_count = 0
require_restorefile_with_uuid = 1
pv_min_size = 2048
issue_discards = 0
}
log {
verbose = 0
syslog = 1
overwrite = 0
level = 0
indent = 1
command_names = 0
prefix = " "
}
backup {
backup = 1
backup_dir = "/etc/lvm/backup"
archive = 1
archive_dir = "/etc/lvm/archive"
retain_min = 10
retain_days = 30
}
shell {
history_size = 100
}
global {
umask = 077
test = 0
units = "h"
si_unit_consistency = 1
activation = 1
proc = "/proc"
locking_type = 1
wait_for_locks = 1
fallback_to_clustered_locking = 1
fallback_to_local_locking = 1
locking_dir = "/run/lock/lvm"
prioritise_write_locks = 1
abort_on_internal_errors = 0
detect_internal_vg_cache_corruption = 0
metadata_read_only = 0
mirror_segtype_default = "mirror"
use_lvmetad = 0
}
activation {
checks = 0
udev_sync = 1
udev_rules = 1
verify_udev_operations = 0
retry_deactivation = 1
missing_stripe_filler = "error"
use_linear_target = 1
reserved_stack = 64
reserved_memory = 8192
process_priority = -18
mirror_region_size = 512
readahead = "auto"
raid_fault_policy = "warn"
mirror_log_fault_policy = "allocate"
mirror_image_fault_policy = "remove"
snapshot_autoextend_threshold = 100
snapshot_autoextend_percent = 20
thin_pool_autoextend_threshold = 100
thin_pool_autoextend_percent = 20
thin_check_executable = "/sbin/thin_check -q"
use_mlockall = 0
monitoring = 0
polling_interval = 15
}
dmeventd {
mirror_library = "libdevmapper-event-lvm2mirror.so"
snapshot_library = "libdevmapper-event-lvm2snapshot.so"
thin_library = "libdevmapper-event-lvm2thin.so"
}
...
which hypervisor,
which disk templates are you using
Also, is there anything unusal about your cluster, like an unusual high number of disks
gnt-cluster modify --ipolicy-spindle-ratio=128when gnt-instance info takes a lot of time, is that lock-related (the INSTANCE_QUERY_DATA job would be in status waiting, and gnt-debug locks would tell you why)
# gnt-debug locks
Name Mode Owner Pending
cluster/BGL shared 70832,70834 -
cluster/[lockset] - - -
instance/[lockset] - - -
instance/xxxxx shared 70832 -
...
instance/xxxxx shared 70832 -
instance/new_instance exclusive 70834 -
instance/xxxxx shared 70832 -
node-alloc/[lockset] shared 70832 exclusive:70834
node/16accc41-71bc-479f-b425-a4ce48f57d6d shared 70832 -
node/3406be72-5628-42d5-a0af-92bcf0861033 shared 70832 -
node/1421562c-6e0a-4194-8827-9c2605c71bb3 shared 70832 -
node/[lockset] - - -
nodegroup/[lockset] - - -
nodegroup/cc1ca697-6ef8-478e-97ef-e1ed25ad281b shared 70832 -
# gnt-job list
70832 running GROUP_VERIFY_DISKS(cc1ca697-6ef8-478e-97ef-e1ed25ad281b)
70833 success OS_DIAGNOSE
70834 waiting INSTANCE_CREATE(new_instance)
Anything unusual in noded's log?
filter = ["a|^/dev/sd*|", "r|.*|"]
...