[OpenIndiana-discuss] Zfs import fails (OpenIndiana-discuss Digest, Vol 31, Issue 28)
Ong Yu-Phing
ong.yu.phing at group.ong-ong.com
Fri Feb 8 02:09:35 UTC 2013
I had a similar situation where after export, my disk paths were wrongly
picked up by ZFS, even though they were recognised by the system.
===
The zpool import fails with this sort of message:
pool1 UNAVAIL insufficient replicas
mirror-0 UNAVAIL corrupted data
c3t0d0 ONLINE
c3t1d0 ONLINE
When I check the disk labels on the disks:
zdb -l /dev/dsk/c3t0d0s0
--------------------------------------------
LABEL 0
--------------------------------------------
version: 28
name: 'pool1'
state: 1
txg: 7173951
pool_guid: 8370873525947507187
hostid: 13162267
hostname: 'openindiana'
top_guid: 15064987019855796782
guid: 4751459059166773513
vdev_children: 3
vdev_tree:
type: 'mirror'
id: 0
guid: 15064987019855796782
metaslab_array: 30
metaslab_shift: 34
ashift: 9
asize: 1998985625600
is_log: 0
create_txg: 4
children[0]:
type: 'disk'
id: 0
guid: 4751459059166773513
path: '/dev/dsk/c2t0d0s0'
devid: 'id1,sd at n600605b002e26410183d783b0e56515a/a'
phys_path: '/pci at 0,0/pci8086,3410 at 9/pci1014,3c7 at 0/sd at 0,0:a'
whole_disk: 1
DTL: 4119
create_txg: 4
children[1]:
type: 'disk'
id: 1
guid: 7277976899319815787
path: '/dev/dsk/c2t1d0s0'
devid: 'id1,sd at n600605b002e264101839b39f123c1210/a'
phys_path: '/pci at 0,0/pci8086,3410 at 9/pci1014,3c7 at 0/sd at 1,0:a'
whole_disk: 1
DTL: 4118
create_txg: 4
So I noticed that even though the disk is c3t0d0, the path shows in the
disk label as c2t0d0s0 (and similarly for the "child"/partner mirrored
disk, c3t1d0, yet path shows as c2t1d0), is this what is causing the
really strange behaviour, when even though the devices are considered
ONLINE, the pool itself is UAVAIL?
===
Maybe you can check if you are hitting the same situation?
Unfortunately, I never was able to fix this (caused by a live upgrade of
OI from 148 to 151a3); instead, I booted from a live USB, then did a ZFS
send to transfer the filesystems to another server.
On 07/02/2013 21:47, Ram Chander wrote:
> "format" detects all 108 disks . I suspect it could be zfs issue. I ran out
> of options now.
>
> root at host1:~# format
> Searching for disks...done
> AVAILABLE DISK SELECTIONS:
> 0. c1t0d0 <DELL-PERCH700-2.10 cyl 54665 alt 2 hd 255 sec 252>
> /pci at 0,0/pci8086,340b at 4/pci1028,1f17 at 0/sd at 0,0
> 1. c5t1d0 <Coraid-EtherDrive SRX42-V0.4-2.73TB>
> /ethdrv/sd at 1,0
> 2. c5t1d1 <Coraid-EtherDrive SRX42-V0.4-2.73TB>
> /ethdrv/sd at 1,1
> 3. c5t1d2 <Coraid-EtherDrive SRX42-V0.4-2.73TB>
> /ethdrv/sd at 1,2
> 4. c5t1d3 <Coraid-EtherDrive SRX42-V0.4-2.73TB>
> /ethdrv/sd at 1,3
> 5. c5t1d4 <Coraid-EtherDrive SRX42-V0.4-2.73TB>
> /ethdrv/sd at 1,4
> ................ etc
>
> root at host1~# devfsadm -Cv
> root at host1:~#
>
>
>
> On Thu, Feb 7, 2013 at 6:44 PM, Sa?o Kiselkov <skiselkov.ml at gmail.com>wrote:
>
>> You have an issue with conectivity to your drives on the Coraid HBA
>> card. I suggest querying your HBA via its management tools to make sure
>> you can discover all the drives on your network. Chances are, they're
>> not all visible, which is why your pool is having trouble.
>>
>> --
>> Saso
>>
>> On 02/07/2013 01:49 PM, Ram Chander wrote:
>>> The drives are in Coraid connected to the server via Coraid 10G HBA
>> card.
>>> Its exported and imported on the same host but after OS upgrade ( format
>> OS
>>> disk and install again ). Before OS upgrade, zpool export was issued and
>>> when tried to import it faults as below. The pool is not functional in
>> any
>>> system now. Tried -d /dev/dsk option but no luck.
>>>
>>>
>>> root at storage1:~# zpool import
>>> pool: pool1
>>> id: 10136140719439709374
>>> state: FAULTED
>>> status: The pool was last accessed by another system.
>>> action: The pool cannot be imported due to damaged devices or data.
>>> The pool may be active on another system, but can be imported
>> using
>>> the '-f' flag.
>>> see: http://illumos.org/msg/ZFS-8000-EY
>>> config:
>>>
>>> pool1 FAULTED corrupted data
>>> raidz1-0 ONLINE
>>> c5t1d0 UNAVAIL corrupted data
>>> c5t1d1 UNAVAIL corrupted data
>>> c5t1d2 UNAVAIL corrupted data
>>> c5t1d3 UNAVAIL corrupted data
>>> c5t1d4 UNAVAIL corrupted data
>>> raidz1-1 ONLINE
>>> c5t1d5 UNAVAIL corrupted data
>>> c5t1d6 UNAVAIL corrupted data
>>> c5t1d7 UNAVAIL corrupted data
>>> c5t1d8 UNAVAIL corrupted data
>>> c5t1d9 UNAVAIL corrupted data
>>> raidz1-2 ONLINE
>>> c5t1d10 UNAVAIL corrupted data
>>> c5t1d11 UNAVAIL corrupted data
>>> c5t1d12 UNAVAIL corrupted data
>>> c5t1d13 UNAVAIL corrupted data
>>> c5t1d14 UNAVAIL corrupted data
>>> raidz1-3 ONLINE
>>> c5t1d15 UNAVAIL corrupted data
>>> c5t1d16 UNAVAIL corrupted data
>>> c5t1d17 UNAVAIL corrupted data
>>> c5t1d18 UNAVAIL corrupted data
>>> c5t1d19 UNAVAIL corrupted data
>>> raidz1-4 ONLINE
>>> c5t1d20 UNAVAIL corrupted data
>>> c5t1d21 UNAVAIL corrupted data
>>> c5t1d22 UNAVAIL corrupted data
>>> c5t1d23 UNAVAIL corrupted data
>>> c5t1d24 UNAVAIL corrupted data
>>> raidz1-5 ONLINE
>>> c5t1d25 UNAVAIL corrupted data
>>> c5t1d26 UNAVAIL corrupted data
>>> c5t1d27 UNAVAIL corrupted data
>>> c5t1d28 UNAVAIL corrupted data
>>> c5t1d29 UNAVAIL corrupted data
>>> raidz1-6 ONLINE
>>> c5t1d30 UNAVAIL corrupted data
>>> c5t1d31 UNAVAIL corrupted data
>>> c5t1d32 UNAVAIL corrupted data
>>> c5t1d33 UNAVAIL corrupted data
>>> c5t1d34 UNAVAIL corrupted data
>>> raidz1-7 ONLINE
>>> c5t2d0 UNAVAIL corrupted data
>>> c5t2d1 UNAVAIL corrupted data
>>> c5t2d2 UNAVAIL corrupted data
>>> c5t2d3 UNAVAIL corrupted data
>>> c5t2d4 UNAVAIL corrupted data
>>> raidz1-8 ONLINE
>>> c5t2d5 UNAVAIL corrupted data
>>> c5t2d6 UNAVAIL corrupted data
>>> c5t2d7 UNAVAIL corrupted data
>>> c5t2d8 UNAVAIL corrupted data
>>> c5t2d9 UNAVAIL corrupted data
>>> raidz1-9 ONLINE
>>> c5t2d10 UNAVAIL corrupted data
>>> c5t2d11 UNAVAIL corrupted data
>>> c5t2d12 UNAVAIL corrupted data
>>> c5t2d13 UNAVAIL corrupted data
>>> c5t2d14 UNAVAIL corrupted data
>>> raidz1-10 ONLINE
>>> c5t2d15 UNAVAIL corrupted data
>>> c5t2d16 UNAVAIL corrupted data
>>> c5t2d17 UNAVAIL corrupted data
>>> c5t2d18 UNAVAIL corrupted data
>>> c5t2d19 UNAVAIL corrupted data
>>> raidz1-11 ONLINE
>>> c5t2d20 UNAVAIL corrupted data
>>> c5t2d21 UNAVAIL corrupted data
>>> c5t2d22 UNAVAIL corrupted data
>>> c5t2d23 UNAVAIL corrupted data
>>> c5t2d24 UNAVAIL corrupted data
>>> raidz1-12 ONLINE
>>> c5t2d25 UNAVAIL corrupted data
>>> c5t2d26 UNAVAIL corrupted data
>>> c5t2d27 UNAVAIL corrupted data
>>> c5t2d28 UNAVAIL corrupted data
>>> c5t2d29 UNAVAIL corrupted data
>>> raidz1-13 ONLINE
>>> c5t2d30 UNAVAIL corrupted data
>>> c5t2d31 UNAVAIL corrupted data
>>> c5t2d32 UNAVAIL corrupted data
>>> c5t2d33 UNAVAIL corrupted data
>>> c5t2d34 UNAVAIL corrupted data
>>> raidz1-14 ONLINE
>>> c5t3d0 ONLINE
>>> c5t3d1 ONLINE
>>> c5t3d2 ONLINE
>>> c5t3d3 ONLINE
>>> c5t3d4 ONLINE
>>> raidz1-15 ONLINE
>>> c5t3d5 ONLINE
>>> c5t3d6 ONLINE
>>> c5t3d7 ONLINE
>>> c5t3d8 ONLINE
>>> c5t3d9 ONLINE
>>> raidz1-16 ONLINE
>>> c5t3d10 ONLINE
>>> c5t3d11 ONLINE
>>> c5t3d12 ONLINE
>>> c5t3d13 ONLINE
>>> c5t3d14 ONLINE
>>> raidz1-17 ONLINE
>>> c5t3d15 ONLINE
>>> c5t3d16 ONLINE
>>> c5t3d17 ONLINE
>>> c5t3d18 ONLINE
>>> c5t3d19 ONLINE
>>> raidz1-18 ONLINE
>>> c5t3d20 ONLINE
>>> c5t3d21 ONLINE
>>> c5t3d22 ONLINE
>>> c5t3d23 ONLINE
>>> c5t3d24 ONLINE
>>> raidz1-19 ONLINE
>>> c5t3d25 ONLINE
>>> c5t3d26 ONLINE
>>> c5t3d27 ONLINE
>>> c5t3d28 ONLINE
>>> c5t3d29 ONLINE
>>> raidz1-20 ONLINE
>>> c5t3d30 UNAVAIL corrupted data
>>> c5t3d31 UNAVAIL corrupted data
>>> c5t3d32 UNAVAIL corrupted data
>>> c5t3d33 UNAVAIL corrupted data
>>> c5t1d35 UNAVAIL corrupted data
>>>
>>> root at host:~# zpool import -FfX pool1
>>> cannot import 'pool1': one or more devices is currently unavailable
>>>
>>> root at host:~# zpool import -f pool1
>>> cannot import 'pool1': I/O error
>>> Destroy and re-create the pool from
>>> a backup source.
>>>
>>>
>>>
>>> On Wed, Feb 6, 2013 at 5:58 PM, Edward Ned Harvey (openindiana) <
>>> openindiana at nedharvey.com> wrote:
>>>
>>>>> From: Ram Chander [mailto:ramquick at gmail.com]
>>>>>
>>>>> I had a zpool thats exported on another system and when i try to
>> import,
>>>>> it fails. Any idea how to recover ?
>>>> Start by proving there isn't some other problem. Import the pool again
>> on
>>>> the same system that did the export. Assuming you can successfully
>> import,
>>>> capture a "zpool status" and then export again and get back to your new
>>>> system...
>>>>
>>>> Show us the zpool status for the pool while it's functional in the old
>>>> system.
>>>>
>>>> Your error message said missing device. ("one or more devices currently
>>>> unavailable"). Make sure you "devfsadm -Cv" on the new system, and make
>>>> sure the new disks are all appearing.
>>>>
>>>> Your pool isn't based on partitions or slices, is it? If so, you'll
>> have
>>>> to specify the devices manually. (I think it's zpool import -d)
>>>>
>>>> What type of disk controllers do you have in the new & old systems?
>> Many
>>>> HBA's will occupy some space for their config & meta data on the drives,
>>>> transparently to the OS. This makes the drives incompatible with other
>>>> systems, other than similar compatible HBA's.
>>>>
>>>> Ideally, you'll have simple braindead SATA/SAS controllers on both the
>>>> source and destination machines. Because they won't add any custom
>> data to
>>>> the drives; they just present the drive to the OS, simple as that.
>>>>
>>>>
>>>> _______________________________________________
>>>> OpenIndiana-discuss mailing list
>>>> OpenIndiana-discuss at openindiana.org
>>>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>>>>
>>> _______________________________________________
>>> OpenIndiana-discuss mailing list
>>> OpenIndiana-discuss at openindiana.org
>>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>>>
>>
>> _______________________________________________
>> OpenIndiana-discuss mailing list
>> OpenIndiana-discuss at openindiana.org
>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>>
>
More information about the OpenIndiana-discuss
mailing list