"the long numeric ID"
- What does that look like? /dev/disk/by-id are _not_ numeric
automatically replaced?
- What automatic? Are you using /etc/zfs/zfs_pool_alert and if you disable it do things work?
- if things have been replaced, have you waited for resilver to complete?
Now I can't
get a mirror to attach to newpool
- what is the command you are using for this to happen? What is the outputIn general, be as specific as you can be. I _know_ you have a problem but I'm sure you want to tell me more than just that :)
root@storage:/var/lib/zfs# zpool detach canaanpool mirror-0 /dev/sdd1
I'll be testing this but from the top of my head this your problem
I had read that partitions are recommended mentioned on some forum post from a couple of years ago - couldn't tell if that was still the case. From my zpool status, you can tell that the original pool was using partitions. All these drives use GUID Partition Tables created under either OpenSolaris or Nexenta - is that relevant? When I tried to examine the partition structure with gparted, it wanted to repair the partitions, perhaps because of endianness? Could this be part of the trouble with the device names?
d. After I had upgraded to 0.6.9, pulled the failing disk, and rebooted, another unused disk automatically replaced the failed disk (as if I had physically replaced the disk). I'm sure this occurred b/ c the old disk had been sdc and the new disk was given the sdc label (even though I had already switched to using /dev/disk/by-id).
I thought this was weird behavior, but maybe it was b/c I had switched to /dev/disk/by-id before upgrading version, and the old version was still using "sdc" internally?
/var/lib/zfs? Didn't notice. I was going to rm it, but it disappeared when I exported all pools. Re-importing in any manner hasn't helped.
8372648596947368235 - in other words, just a string of digits, I believe 19 digits long. Using that, I was able to detach the new drive from the mirror, but not able to reattach it. I've never seen such a disk identifier except in some forum post about zfs. The post was old, on Solaris, and off-point (somebody was trying to detach a drive from a raidz).
Does that make sense?
Now.... if only we could understand why the b***rd won't detach the
other vdev ...
This would actually give me hope that you should be able to similarly
detach the other vdev by physically removing it first...?
You might, of course, zap the zpool.cache, create a /dev/disk/subset
folder containing just symlinks to the devices you want zpool to notice
and then
zpool import -d /dev/disk/subset poolname
That should have the same effect as yanking the disk but without the
physical exercise and risk of damaging your harddrive in the process :)
> But I'm not sure why you brought that up.
Well, like I promised I was going to _test_ my first idea on how to
change the syntax on detach. I was simply showing [1] that I was wrong :)
[1] as a footnote, as well
I meant duplicate device names _simultaneously visible_. This is not
technically feasible and therefore not what happened. Besides, we have
found a perfectly reasonable explanation for the long numeric IDs
appearing in my other response, based on the actual output of zpool that
you had saved (bravo!).
I was fantasizing along the lines of having two identically-named pools
visible at one given time. Never mind
That was a very fast resilver - approximately 60MBps. zpool iostat reported write speed to be very very low. I believe the resilver retained the data on the disk and basically verified checksums. Interesting.
The problem is not fixed at all. But gnu parted no longer chokes on the partition table.
The problem is not fixed at all. But gnu parted no longer chokes on the partition table. I guess I should try just using msdos partition table. I would think gpt would pose no problem, support is compiled into the kernel.