However I have encountered another challenge while running a burn-in test after initializing the ganeti cluster. The cluster details and error are shown below:
root@node1:~# /usr/lib/ganeti/tools/burnin -o debootstrap+default -H kvm:kernel_path=/vmlinuz,initrd_path=/initrd.img --disk-size 1024m --no-name-check --no-ip-check
burn-in.example.com- Testing global parameters
- Creating instances
- Setting instance runtime memory
Set memory to 128 MB
- Replacing disks on the same nodes
run replace_on_secondary
run replace_on_primary
- Changing the secondary node
- Growing disks
increase disk/0 by 128 MB
- Failing over instances
- Migrating instances
migration and migration cleanup
- Exporting and re-importing instances
remove instance
remove export
- Non-idempotent opcode failed, aborting
- Error detected: opcode buffer follows:
Thu Apr 6 14:14:06 2017 * creating instance disks...
Thu Apr 6 14:14:10 2017 - INFO: - device disk/0: 0.80% done, 3m 57s remaining (estimated)
Thu Apr 6 14:15:10 2017 * running the instance OS create scripts...
Thu Apr 6 14:25:33 2017 * starting instance...
Thu Apr 6 14:25:36 2017 STEP 1/6 Check device existence
Thu Apr 6 14:25:37 2017 - INFO: Checking volume groups
Thu Apr 6 14:25:37 2017 STEP 2/6 Check peer consistency
Thu Apr 6 14:25:37 2017 - INFO: Checking disk/0 consistency on node
node1.example.comThu Apr 6 14:25:37 2017 STEP 3/6 Allocate new storage
Thu Apr 6 14:25:38 2017 STEP 4/6 Changing drbd configuration
Thu Apr 6 14:25:38 2017 - INFO: Detaching disk/0 drbd from local storage
Thu Apr 6 14:25:38 2017 - INFO: Renaming the old LVs on the target node
Thu Apr 6 14:25:38 2017 - INFO: Renaming the new LVs on the target node
Thu Apr 6 14:25:39 2017 STEP 5/6 Sync devices
Thu Apr 6 14:25:39 2017 - INFO: - device disk/0: 0.80% done, 43m 41s remaining (estimated)
Thu Apr 6 14:26:40 2017 - INFO: - device disk/0: 82.20% done, 8s remaining (estimated)
Thu Apr 6 14:26:48 2017 STEP 6/6 Removing old storage
Thu Apr 6 14:26:48 2017 - INFO: Remove logical volumes for disk/0
Thu Apr 6 14:26:49 2017 STEP 1/6 Check device existence
Thu Apr 6 14:26:50 2017 - INFO: Checking volume groups
Thu Apr 6 14:26:50 2017 STEP 2/6 Check peer consistency
Thu Apr 6 14:26:50 2017 - INFO: Checking disk/0 consistency on node
node2.example.comThu Apr 6 14:26:50 2017 STEP 3/6 Allocate new storage
Thu Apr 6 14:26:51 2017 STEP 4/6 Changing drbd configuration
Thu Apr 6 14:26:51 2017 - INFO: Detaching disk/0 drbd from local storage
Thu Apr 6 14:26:51 2017 - INFO: Renaming the old LVs on the target node
Thu Apr 6 14:26:51 2017 - INFO: Renaming the new LVs on the target node
Thu Apr 6 14:26:52 2017 STEP 5/6 Sync devices
Thu Apr 6 14:26:53 2017 - INFO: - device disk/0: 0.80% done, 43m 40s remaining (estimated)
Thu Apr 6 14:27:53 2017 STEP 6/6 Removing old storage
Thu Apr 6 14:27:53 2017 - INFO: Remove logical volumes for disk/0
Thu Apr 6 14:27:54 2017 STEP 1/6 Check device existence
Thu Apr 6 14:27:54 2017 - INFO: Checking volume groups
Thu Apr 6 14:27:54 2017 STEP 2/6 Check peer consistency
Thu Apr 6 14:27:54 2017 - INFO: Checking disk/0 consistency on node
node1.example.comThu Apr 6 14:27:55 2017 STEP 3/6 Allocate new storage
Thu Apr 6 14:27:55 2017 - INFO: Adding new local storage on
node3.example.com for disk/0
Thu Apr 6 14:27:55 2017 STEP 4/6 Changing drbd configuration
Thu Apr 6 14:27:55 2017 - INFO: activating a new drbd on
node3.example.com for disk/0
Thu Apr 6 14:27:56 2017 - INFO: Shutting down drbd for disk/0 on old node
Thu Apr 6 14:27:56 2017 - INFO: Detaching primary drbds from the network (=> standalone)
Thu Apr 6 14:27:56 2017 - INFO: Updating instance configuration
Thu Apr 6 14:27:56 2017 - INFO: Attaching primary drbds to new secondary (standalone => connected)
Thu Apr 6 14:27:57 2017 STEP 5/6 Sync devices
Thu Apr 6 14:27:57 2017 - INFO: - device disk/0: 1.20% done, 2m 3s remaining (estimated)
Thu Apr 6 14:28:58 2017 STEP 6/6 Removing old storage
Thu Apr 6 14:28:58 2017 - INFO: Remove logical volumes for 0
Thu Apr 6 14:29:01 2017 - INFO: - device disk/0: 9.10% done, 14s remaining (estimated)
Thu Apr 6 14:29:16 2017 * checking disk consistency between source and target
Thu Apr 6 14:29:16 2017 * shutting down instance on source node
Thu Apr 6 14:31:18 2017 * deactivating the instance's disks on source node
Thu Apr 6 14:31:19 2017 * activating the instance's disks on target node
node3.example.comThu Apr 6 14:31:23 2017 * checking disk consistency between source and target
Thu Apr 6 14:31:24 2017 * changing into standalone mode
Thu Apr 6 14:31:24 2017 * changing disks into dual-master mode
Thu Apr 6 14:31:25 2017 * wait until resync is done
Thu Apr 6 14:31:26 2017 * starting memory transfer
Thu Apr 6 14:31:30 2017 * memory transfer complete
Thu Apr 6 14:31:31 2017 * wait until resync is done
Thu Apr 6 14:31:31 2017 * changing into standalone mode
Thu Apr 6 14:31:31 2017 * changing disks into single-master mode
Thu Apr 6 14:31:32 2017 * wait until resync is done
Thu Apr 6 14:31:32 2017 * done
Thu Apr 6 14:31:33 2017 - INFO: Not checking memory on the secondary node as instance will not be started
Thu Apr 6 14:31:33 2017 * checking where the instance actually runs (if this hangs, the hypervisor might be in a bad state)
Thu Apr 6 14:31:33 2017 * instance confirmed to be running on its primary node (
node1.example.com)
Thu Apr 6 14:31:33 2017 * wait until resync is done
Thu Apr 6 14:31:33 2017 * changing into standalone mode
Thu Apr 6 14:31:34 2017 * changing disks into single-master mode
Thu Apr 6 14:31:34 2017 * wait until resync is done
Thu Apr 6 14:31:35 2017 * done
Thu Apr 6 14:33:43 2017 snapshot/0 is now listening, starting export
Thu Apr 6 14:33:51 2017 snapshot/0 sent 0M, 0.0 MiB/s
Thu Apr 6 14:33:56 2017 - WARNING: export 'export-disk0-2017-04-06_14_33_44-Zntkkx' on
node1.example.com failed: Exited with status 1
Thu Apr 6 14:33:56 2017 snapshot/0 failed to send data: Exited with status 1 (recent output: DUMP: Date of this level 0 dump: Thu Apr 6 14:33:45 2017
DUMP: Dumping /dev/mapper/ganeti-6346d65f--15a9--468a--a7d2--c4b7f7555896.disk0_data.snap-1 (an unlisted file system) to standard output
DUMP: Label: none
DUMP: Writing 10 Kilobyte records
DUMP: mapping (Pass I) [regular files]
DUMP: mapping (Pass II) [directories]
DUMP: estimated 192396 blocks.
DUMP: Volume 1 started with block 1 at: Thu Apr 6 14:33:45 2017
DUMP: dumping (Pass III) [directories]
socat: E SSL_connect(): error:14082174:SSL routines:SSL3_CHECK_CERT_AND_ALGORITHM:dh key too small
dd: dd: error writing 'standard output': Broken pipe
DUMP: Broken pipe
DUMP: The ENTIRE dump is aborted.)
Thu Apr 6 14:33:56 2017 - WARNING: Aborting import 'import-disk0-2017-04-06_14_33_39-lKCYBL' on 6c87e66e-dbf5-4cd1-9345-9fb62c753014
Thu Apr 6 14:33:57 2017 - WARNING: import 'import-disk0-2017-04-06_14_33_39-lKCYBL' on
node3.example.com failed: Exited due to signal 15
Thu Apr 6 14:33:57 2017 snapshot/0 failed to receive data: Exited due to signal 15 (recent output: socat: W exiting on signal 15)
Removing instances
Traceback (most recent call last):
File "/usr/lib/ganeti/tools/burnin", line 21, in <module>
sys.exit(main.Main())
File "/usr/share/ganeti/2.10/ganeti/tools/burnin.py", line 1174, in Main
return Burner().BurninCluster()
File "/usr/share/ganeti/2.10/ganeti/tools/burnin.py", line 1113, in BurninCluster
self.BurnImportExport()
File "/usr/share/ganeti/2.10/ganeti/tools/burnin.py", line 283, in wrapper
val = fn(self, *args, **kwargs)
File "/usr/share/ganeti/2.10/ganeti/tools/burnin.py", line 303, in batched
val = fn(self, *args, **kwargs)
File "/usr/share/ganeti/2.10/ganeti/tools/burnin.py", line 815, in BurnImportExport
self.ExecOrQueue(instance, [exp_op, rem_op, imp_op, erem_op])
File "/usr/share/ganeti/2.10/ganeti/tools/burnin.py", line 411, in ExecOrQueue
val = self.ExecOp(self.queue_retry, *ops) # pylint: disable=W0142
File "/usr/share/ganeti/2.10/ganeti/tools/burnin.py", line 403, in ExecOp
return self.MaybeRetry(rval, "opcode", self._ExecOp, *ops)
File "/usr/share/ganeti/2.10/ganeti/tools/burnin.py", line 360, in MaybeRetry
val = fn(*args)
File "/usr/share/ganeti/2.10/ganeti/tools/burnin.py", line 385, in _ExecOp
results = cli.PollJob(job_id, cl=
self.cl, feedback_fn=self.Feedback)
File "/usr/share/ganeti/2.10/ganeti/cli.py", line 2291, in PollJob
return GenericPollJob(job_id, _LuxiJobPollCb(cl), reporter)
File "/usr/share/ganeti/2.10/ganeti/cli.py", line 2113, in GenericPollJob
errors.MaybeRaise(msg)
File "/usr/share/ganeti/2.10/ganeti/errors.py", line 510, in MaybeRaise
raise errcls(*args)
ganeti.errors.OpExecError: Export failed, errors in disk export: disk(s) 0