[OpenIndiana-discuss] [zfs] problem on my zpool
Clement BRIZARD
clement at brizou.fr
Wed Oct 23 07:30:29 UTC 2013
I'm still a student so when I'll get a job I will change all that.
apparently mixing vdevs with different number of disk is not the best
idea either
Le 23/10/2013 09:21, Mark a écrit :
>
> One suggestion for the future for better reliability.
>
> Avoid using WDC WD20EARS, they aren't great, and hopeless on SAS
> controllers.
>
> Avoid mixing disk brands or models in a zpool.
> The different behaviour can trigger issues.
>
> On 23/10/2013 7:46 p.m., Clement BRIZARD wrote:
>> I cleared the "degraded" disk. we will see what happens in 131hours
>>
>> pool: nas
>> state: ONLINE
>> status: One or more devices is currently being resilvered. The pool
>> will
>> continue to function, possibly in a degraded state.
>> action: Wait for the resilver to complete.
>> scan: resilver in progress since Wed Oct 23 08:25:56 2013
>> 2.23G scanned out of 22.2T at 48.6M/s, 133h22m to go
>> 6.10M resilvered, 0.01% done
>> config:
>>
>> NAME STATE READ WRITE CKSUM
>> CAP Product
>> nas ONLINE 0 0 0
>> raidz1-0 ONLINE 0 0 0
>> c8t50024E9004993E6Ed0p0 ONLINE 0 0 0 2
>> TB SAMSUNG HD204UI
>> c8t50024E92062E7524d0 ONLINE 0 0 0 2
>> TB SAMSUNG HD204UI
>> c8t50024E900495BE84d0p0 ONLINE 0 0 0 2
>> TB SAMSUNG HD204UI
>> c8t50014EE25A5EEC23d0p0 ONLINE 0 0 0 2
>> TB WDC WD20EARS-00M
>> c8t50024E9003F03980d0p0 ONLINE 0 0 0 2
>> TB SAMSUNG HD204UI
>> c8t50014EE2B0D3EFC8d0 ONLINE 0 0 0 2
>> TB WDC WD20EARX-00P
>> c8t50014EE6561DDB4Cd0p0 ONLINE 0 0 0 2
>> TB WDC WD20EARS-00M
>> c8t50024E9003F03A09d0p0 ONLINE 0 0 0 2
>> TB SAMSUNG HD204UI
>> raidz1-1 ONLINE 0 0 0
>> c50t8d0 ONLINE 0 0 0
>> (resilvering) 2 TB ST2000DL004 HD20
>> c2d0 ONLINE 0 0 0
>> (resilvering) 2 TB
>> c1d0 ONLINE 0 0 0
>> (resilvering) 2 TB
>> c50t11d0 ONLINE 0 0 0 2
>> TB SAMSUNG HD204UI
>> c50t10d0 ONLINE 0 0 0
>> (resilvering) 2 TB SAMSUNG HD204UI
>>
>>
>>
>>
>> Le 23/10/2013 08:43, Clement BRIZARD a écrit :
>>> I woke up this morning and so you're messages, unfortunately I had to
>>> reboot, the server completely froze.
>>> Now I have that :
>>>
>>> pool: nas
>>> state: DEGRADED
>>> status: One or more devices is currently being resilvered. The pool
>>> will
>>> continue to function, possibly in a degraded state.
>>> action: Wait for the resilver to complete.
>>> scan: resilver in progress since Wed Oct 23 08:19:42 2013
>>> 5.81G scanned out of 22.2T at 49.2M/s, 131h43m to go
>>> 15.6M resilvered, 0.03% done
>>> config:
>>>
>>> NAME STATE READ WRITE CKSUM
>>> nas DEGRADED 0 0 0
>>> raidz1-0 DEGRADED 0 0 0
>>> c8t50024E9004993E6Ed0p0 ONLINE 0 0 0
>>> c8t50024E92062E7524d0 ONLINE 0 0 0
>>> c8t50024E900495BE84d0p0 ONLINE 0 0 0
>>> c8t50014EE25A5EEC23d0p0 ONLINE 0 0 0
>>> c8t50024E9003F03980d0p0 ONLINE 0 0 0
>>> c8t50014EE2B0D3EFC8d0 ONLINE 0 0 0
>>> c8t50014EE6561DDB4Cd0p0 DEGRADED 0 0 0 too many
>>> errors
>>> c8t50024E9003F03A09d0p0 ONLINE 0 0 0
>>> raidz1-1 ONLINE 0 0 0
>>> c50t8d0 ONLINE 0 0 0
>>> (resilvering)
>>> c2d0 ONLINE 0 0 0
>>> (resilvering)
>>> c1d0 ONLINE 0 0 0
>>> (resilvering)
>>> c50t11d0 ONLINE 0 0 0
>>> c50t10d0 ONLINE 0 0 0
>>> (resilvering)
>>>
>>>
>>>
>>>
>>>
>>> Le 23/10/2013 08:00, Jason Matthews a écrit :
>>>>
>>>> first, dont reboot. if you do you might not be able remount the pool.
>>>> the data you see is from the disks that are functioning. listing the
>>>> files and copying complete files are two different things. if you
>>>> dont have a backup you may need to copy whatever partial data you can
>>>> from the broken pool.
>>>>
>>>> now let's start by getting the disks back in good shape.
>>>>
>>>> clear the degraded disk
>>>> zpool clear c8t50014EE6561DDB4Cd0p0
>>>>
>>>> reseat the missing disks in the hopes they come back then clear them
>>>>
>>>> check cfgadm -al and make sure they are connected and configured
>>>>
>>>> when you reseat them check the messages (or dmesg) to see if the
>>>> system notices the re-insertion. if it does see the disk installed
>>>> clear the disks in the pool in effort to bring the pool back to an
>>>> operational state.
>>>>
>>>> Sent from Jasons' hand held
>>>>
>>>> On Oct 22, 2013, at 5:04 PM, Clement BRIZARD <clement at brizou.fr>
>>>> wrote:
>>>>
>>>>> Hello everybody,
>>>>> I have a problem with my pool, I had some slowdowns lately on my nfs
>>>>> share of my zfs pool. A weekly scrub began and is still running but
>>>>> it worries me, it currently returne that
>>>>>
>>>>> pool: nas
>>>>> state: UNAVAIL
>>>>> status: One or more devices are faulted in response to IO failures.
>>>>> action: Make sure the affected devices are connected, then run
>>>>> 'zpool clear'.
>>>>> see: http://illumos.org/msg/ZFS-8000-HC
>>>>> scan: scrub in progress since Sun Oct 20 19:29:23 2013
>>>>> 15.2T scanned out of 22.2T at 84.0M/s, 24h5m to go
>>>>> 1.29G repaired, 68.67% done
>>>>> config:
>>>>>
>>>>> NAME STATE READ WRITE CKSUM
>>>>> nas UNAVAIL 63 2 0
>>>>> insufficient replicas
>>>>> raidz1-0 DEGRADED 0 0 0
>>>>> c8t50024E9004993E6Ed0p0 ONLINE 0 0 0
>>>>> c8t50024E92062E7524d0 ONLINE 0 0 0
>>>>> c8t50024E900495BE84d0p0 ONLINE 0 0 0
>>>>> c8t50014EE25A5EEC23d0p0 ONLINE 0 0 0
>>>>> c8t50024E9003F03980d0p0 ONLINE 0 0 1 (repairing)
>>>>> c8t50014EE2B0D3EFC8d0 ONLINE 0 0 0
>>>>> c8t50014EE6561DDB4Cd0p0 DEGRADED 0 0 211 too many
>>>>> errors (repairing)
>>>>> c8t50024E9003F03A09d0p0 ONLINE 0 0 18 (repairing)
>>>>> raidz1-1 UNAVAIL 131 9 0
>>>>> insufficient replicas
>>>>> c50t8d0 REMOVED 0 0 0 (repairing)
>>>>> c2d0 ONLINE 0 0 0 (repairing)
>>>>> c1d0 ONLINE 0 0 0 (repairing)
>>>>> c50t11d0 ONLINE 0 0 0 (repairing)
>>>>> c50t10d0 REMOVED 0 0 0
>>>>>
>>>>> errors: 10972861 data errors, use '-v' for a list
>>>>>
>>>>>
>>>>> really weird, I haven't disconnected any disk. For several hours
>>>>> even if it said that the pool was unavailable I was browsing on it
>>>>> via nfs. I can't anymore.
>>>>>
>>>>>
>>>>> What do you think I should do ?
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> OpenIndiana-discuss mailing list
>>>>> OpenIndiana-discuss at openindiana.org
>>>>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>>>> _______________________________________________
>>>> OpenIndiana-discuss mailing list
>>>> OpenIndiana-discuss at openindiana.org
>>>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>>>
>>>
>>> _______________________________________________
>>> OpenIndiana-discuss mailing list
>>> OpenIndiana-discuss at openindiana.org
>>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>>
>>
>> _______________________________________________
>> OpenIndiana-discuss mailing list
>> OpenIndiana-discuss at openindiana.org
>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>
>
> _______________________________________________
> OpenIndiana-discuss mailing list
> OpenIndiana-discuss at openindiana.org
> http://openindiana.org/mailman/listinfo/openindiana-discuss
More information about the OpenIndiana-discuss
mailing list