[OpenIndiana-discuss] [zfs] problem on my zpool

Clement BRIZARD clement at brizou.fr
Wed Oct 23 07:30:29 UTC 2013


I'm still a student so when I'll get a job I will change all that. 
apparently mixing vdevs with different number of disk is not the best 
idea either

Le 23/10/2013 09:21, Mark a écrit :
>
> One suggestion for the future for better reliability.
>
> Avoid using WDC WD20EARS, they aren't great, and hopeless on SAS 
> controllers.
>
> Avoid mixing disk brands or models in a zpool.
> The different behaviour can trigger issues.
>
> On 23/10/2013 7:46 p.m., Clement BRIZARD wrote:
>> I cleared the "degraded" disk. we will see what happens  in 131hours
>>
>>    pool: nas
>>   state: ONLINE
>> status: One or more devices is currently being resilvered.  The pool 
>> will
>>      continue to function, possibly in a degraded state.
>> action: Wait for the resilver to complete.
>>    scan: resilver in progress since Wed Oct 23 08:25:56 2013
>>      2.23G scanned out of 22.2T at 48.6M/s, 133h22m to go
>>      6.10M resilvered, 0.01% done
>> config:
>>
>>      NAME                         STATE     READ WRITE CKSUM
>> CAP            Product
>>      nas                          ONLINE       0     0     0
>>        raidz1-0                   ONLINE       0     0     0
>>          c8t50024E9004993E6Ed0p0  ONLINE       0     0     0 2
>> TB           SAMSUNG HD204UI
>>          c8t50024E92062E7524d0    ONLINE       0     0     0 2
>> TB           SAMSUNG HD204UI
>>          c8t50024E900495BE84d0p0  ONLINE       0     0     0 2
>> TB           SAMSUNG HD204UI
>>          c8t50014EE25A5EEC23d0p0  ONLINE       0     0     0 2
>> TB           WDC WD20EARS-00M
>>          c8t50024E9003F03980d0p0  ONLINE       0     0     0 2
>> TB           SAMSUNG HD204UI
>>          c8t50014EE2B0D3EFC8d0    ONLINE       0     0     0 2
>> TB           WDC WD20EARX-00P
>>          c8t50014EE6561DDB4Cd0p0  ONLINE       0     0     0 2
>> TB           WDC WD20EARS-00M
>>          c8t50024E9003F03A09d0p0  ONLINE       0     0     0 2
>> TB           SAMSUNG HD204UI
>>        raidz1-1                   ONLINE       0     0     0
>>          c50t8d0                  ONLINE       0     0     0
>> (resilvering)     2 TB           ST2000DL004 HD20
>>          c2d0                     ONLINE       0     0     0
>> (resilvering)     2 TB
>>          c1d0                     ONLINE       0     0     0
>> (resilvering)     2 TB
>>          c50t11d0                 ONLINE       0     0     0 2
>> TB           SAMSUNG HD204UI
>>          c50t10d0                 ONLINE       0     0     0
>> (resilvering)     2 TB           SAMSUNG HD204UI
>>
>>
>>
>>
>> Le 23/10/2013 08:43, Clement BRIZARD a écrit :
>>> I woke up this morning and so you're messages, unfortunately I had to
>>> reboot, the server completely froze.
>>> Now I have that :
>>>
>>>   pool: nas
>>>  state: DEGRADED
>>> status: One or more devices is currently being resilvered. The pool 
>>> will
>>>     continue to function, possibly in a degraded state.
>>> action: Wait for the resilver to complete.
>>>   scan: resilver in progress since Wed Oct 23 08:19:42 2013
>>>     5.81G scanned out of 22.2T at 49.2M/s, 131h43m to go
>>>     15.6M resilvered, 0.03% done
>>> config:
>>>
>>>     NAME                         STATE     READ WRITE CKSUM
>>>     nas                          DEGRADED     0     0     0
>>>       raidz1-0                   DEGRADED     0     0     0
>>>         c8t50024E9004993E6Ed0p0  ONLINE       0     0     0
>>>         c8t50024E92062E7524d0    ONLINE       0     0     0
>>>         c8t50024E900495BE84d0p0  ONLINE       0     0     0
>>>         c8t50014EE25A5EEC23d0p0  ONLINE       0     0     0
>>>         c8t50024E9003F03980d0p0  ONLINE       0     0     0
>>>         c8t50014EE2B0D3EFC8d0    ONLINE       0     0     0
>>>         c8t50014EE6561DDB4Cd0p0  DEGRADED     0     0     0 too many
>>> errors
>>>         c8t50024E9003F03A09d0p0  ONLINE       0     0     0
>>>       raidz1-1                   ONLINE       0     0     0
>>>         c50t8d0                  ONLINE       0     0     0 
>>> (resilvering)
>>>         c2d0                     ONLINE       0     0     0 
>>> (resilvering)
>>>         c1d0                     ONLINE       0     0     0 
>>> (resilvering)
>>>         c50t11d0                 ONLINE       0     0     0
>>>         c50t10d0                 ONLINE       0     0     0 
>>> (resilvering)
>>>
>>>
>>>
>>>
>>>
>>> Le 23/10/2013 08:00, Jason Matthews a écrit :
>>>>
>>>> first, dont reboot. if you do you might not be able remount the pool.
>>>> the data you see is from the disks that are functioning. listing the
>>>> files and copying complete files are two different things. if you
>>>> dont have a backup you may need to copy whatever partial data you can
>>>> from the broken pool.
>>>>
>>>> now let's start by getting the disks back in good shape.
>>>>
>>>> clear the degraded disk
>>>> zpool clear c8t50014EE6561DDB4Cd0p0
>>>>
>>>> reseat the missing disks in the hopes they come back then clear them
>>>>
>>>> check cfgadm -al and make sure they are connected and configured
>>>>
>>>> when you reseat them check the messages (or dmesg) to see if the
>>>> system notices the re-insertion. if it does see the disk installed
>>>> clear the disks in the pool in effort to bring the pool back to an
>>>> operational state.
>>>>
>>>> Sent from Jasons' hand held
>>>>
>>>> On Oct 22, 2013, at 5:04 PM, Clement BRIZARD <clement at brizou.fr> 
>>>> wrote:
>>>>
>>>>> Hello everybody,
>>>>> I have a problem with my pool, I had some slowdowns lately on my nfs
>>>>> share of my zfs pool. A weekly scrub began and is still running but
>>>>> it worries me, it currently returne that
>>>>>
>>>>>   pool: nas
>>>>> state: UNAVAIL
>>>>> status: One or more devices are faulted in response to IO failures.
>>>>> action: Make sure the affected devices are connected, then run
>>>>> 'zpool clear'.
>>>>>    see: http://illumos.org/msg/ZFS-8000-HC
>>>>>   scan: scrub in progress since Sun Oct 20 19:29:23 2013
>>>>>     15.2T scanned out of 22.2T at 84.0M/s, 24h5m to go
>>>>>     1.29G repaired, 68.67% done
>>>>> config:
>>>>>
>>>>>     NAME                         STATE     READ WRITE CKSUM
>>>>>     nas                          UNAVAIL     63     2 0
>>>>> insufficient replicas
>>>>>       raidz1-0                   DEGRADED     0     0 0
>>>>>         c8t50024E9004993E6Ed0p0  ONLINE       0     0 0
>>>>>         c8t50024E92062E7524d0    ONLINE       0     0 0
>>>>>         c8t50024E900495BE84d0p0  ONLINE       0     0 0
>>>>>         c8t50014EE25A5EEC23d0p0  ONLINE       0     0 0
>>>>>         c8t50024E9003F03980d0p0  ONLINE       0     0 1 (repairing)
>>>>>         c8t50014EE2B0D3EFC8d0    ONLINE       0     0 0
>>>>>         c8t50014EE6561DDB4Cd0p0  DEGRADED     0     0 211 too many
>>>>> errors  (repairing)
>>>>>         c8t50024E9003F03A09d0p0  ONLINE       0     0 18 (repairing)
>>>>>       raidz1-1                   UNAVAIL    131     9 0
>>>>> insufficient replicas
>>>>>         c50t8d0                  REMOVED      0     0 0 (repairing)
>>>>>         c2d0                     ONLINE       0     0 0 (repairing)
>>>>>         c1d0                     ONLINE       0     0 0 (repairing)
>>>>>         c50t11d0                 ONLINE       0     0 0 (repairing)
>>>>>         c50t10d0                 REMOVED      0     0 0
>>>>>
>>>>> errors: 10972861 data errors, use '-v' for a list
>>>>>
>>>>>
>>>>> really weird, I haven't disconnected any disk. For several hours
>>>>> even if it said that the pool was unavailable I was browsing on it
>>>>> via nfs. I can't anymore.
>>>>>
>>>>>
>>>>> What do you think I should do ?
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> OpenIndiana-discuss mailing list
>>>>> OpenIndiana-discuss at openindiana.org
>>>>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>>>> _______________________________________________
>>>> OpenIndiana-discuss mailing list
>>>> OpenIndiana-discuss at openindiana.org
>>>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>>>
>>>
>>> _______________________________________________
>>> OpenIndiana-discuss mailing list
>>> OpenIndiana-discuss at openindiana.org
>>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>>
>>
>> _______________________________________________
>> OpenIndiana-discuss mailing list
>> OpenIndiana-discuss at openindiana.org
>> http://openindiana.org/mailman/listinfo/openindiana-discuss
>
>
> _______________________________________________
> OpenIndiana-discuss mailing list
> OpenIndiana-discuss at openindiana.org
> http://openindiana.org/mailman/listinfo/openindiana-discuss




More information about the OpenIndiana-discuss mailing list